From: Piotr Karbowski (grsec-scrape) Date: Fri, 16 Mar 2012 02:02:09 +0000 (+0100) Subject: grsec-scrape autocommit. 2 new patch(es). X-Git-Url: http://git.ipfire.org/cgi-bin/gitweb.cgi?a=commitdiff_plain;h=6ecae204a0c5acca39d9eccc89d7700a08ef7c52;p=thirdparty%2Fgrsecurity-scrape.git grsec-scrape autocommit. 2 new patch(es). New patch: grsecurity-2.9-3.2.11-201203152127.patch New patch: grsecurity-2.9-2.6.32.58-201203152126.patch --- diff --git a/test/changelog-test.txt b/test/changelog-test.txt index e49144a..e8a9269 100644 --- a/test/changelog-test.txt +++ b/test/changelog-test.txt @@ -1,3 +1,29 @@ +commit e4b05b65c645c412eceb9c950ee7b4771627e6b1 +Merge: e55aa68 258c015 +Author: Brad Spengler +Date: Thu Mar 15 20:59:19 2012 -0400 + + Merge branch 'pax-test' into grsec-test + +commit 258c0159fa6dd5044ca984eeaad57bb6e21bacea +Author: Brad Spengler +Date: Thu Mar 15 20:59:05 2012 -0400 + + fix ARM compilation + +commit e55aa68f4bb20e75cd7423123aa612c2a69590c0 +Merge: 8f95ea9 55b7573 +Author: Brad Spengler +Date: Wed Mar 14 19:33:41 2012 -0400 + + Merge branch 'pax-test' into grsec-test + +commit 55b7573f6c2f3be26fb39c7bd6a9d742d02811ca +Author: Brad Spengler +Date: Wed Mar 14 19:33:15 2012 -0400 + + Update to pax-linux-3.2.10-test28.patch + commit 8f95ea9f718c293794a1f6bdd2a5f5f336f7bd64 Merge: c8786a2 886ac5e Author: Brad Spengler diff --git a/test/grsecurity-2.9-2.6.32.58-201203152126.patch b/test/grsecurity-2.9-2.6.32.58-201203152126.patch new file mode 100644 index 0000000..61f4945 --- /dev/null +++ b/test/grsecurity-2.9-2.6.32.58-201203152126.patch @@ -0,0 +1,106916 @@ +diff --git a/Documentation/dontdiff b/Documentation/dontdiff +index e1efc40..76e689e 100644 +--- a/Documentation/dontdiff ++++ b/Documentation/dontdiff +@@ -1,15 +1,19 @@ + *.a + *.aux + *.bin ++*.cis + *.cpio + *.csp ++*.dbg + *.dsp + *.dvi + *.elf + *.eps + *.fw ++*.gcno + *.gen.S + *.gif ++*.gmo + *.grep + *.grp + *.gz +@@ -38,8 +42,10 @@ + *.tab.h + *.tex + *.ver ++*.vim + *.xml + *_MODULES ++*_reg_safe.h + *_vga16.c + *~ + *.9 +@@ -49,11 +55,16 @@ + 53c700_d.h + CVS + ChangeSet ++GPATH ++GRTAGS ++GSYMS ++GTAGS + Image + Kerntypes + Module.markers + Module.symvers + PENDING ++PERF* + SCCS + System.map* + TAGS +@@ -76,7 +87,11 @@ btfixupprep + build + bvmlinux + bzImage* ++capability_names.h ++capflags.c + classlist.h* ++clut_vga16.c ++common-cmds.h + comp*.log + compile.h* + conf +@@ -84,6 +99,8 @@ config + config-* + config_data.h* + config_data.gz* ++config.c ++config.tmp + conmakehash + consolemap_deftbl.c* + cpustr.h +@@ -97,19 +114,23 @@ elfconfig.h* + fixdep + fore200e_mkfirm + fore200e_pca_fw.c* ++gate.lds + gconf + gen-devlist + gen_crc32table + gen_init_cpio + genksyms + *_gray256.c ++hash ++hid-example + ihex2fw + ikconfig.h* + initramfs_data.cpio ++initramfs_data.cpio.bz2 + initramfs_data.cpio.gz + initramfs_list + kallsyms +-kconfig ++kern_constants.h + keywords.c + ksym.c* + ksym.h* +@@ -117,6 +138,7 @@ kxgettext + lkc_defs.h + lex.c + lex.*.c ++lib1funcs.S + logo_*.c + logo_*_clut224.c + logo_*_mono.c +@@ -127,13 +149,16 @@ machtypes.h + map + maui_boot.h + mconf ++mdp + miboot* + mk_elfconfig + mkboot + mkbugboot + mkcpustr + mkdep ++mkpiggy + mkprep ++mkregtable + mktables + mktree + modpost +@@ -149,6 +174,7 @@ patches* + pca200e.bin + pca200e_ecd.bin2 + piggy.gz ++piggy.S + piggyback + pnmtologo + ppc_defs.h* +@@ -157,12 +183,15 @@ qconf + raid6altivec*.c + raid6int*.c + raid6tables.c ++regdb.c + relocs ++rlim_names.h + series + setup + setup.bin + setup.elf + sImage ++slabinfo + sm_tbl* + split-include + syscalltab.h +@@ -171,6 +200,7 @@ tftpboot.img + timeconst.h + times.h* + trix_boot.h ++user_constants.h + utsrelease.h* + vdso-syms.lds + vdso.lds +@@ -186,14 +216,20 @@ version.h* + vmlinux + vmlinux-* + vmlinux.aout ++vmlinux.bin.all ++vmlinux.bin.bz2 + vmlinux.lds ++vmlinux.relocs ++voffset.h + vsyscall.lds + vsyscall_32.lds + wanxlfw.inc + uImage + unifdef ++utsrelease.h + wakeup.bin + wakeup.elf + wakeup.lds + zImage* + zconf.hash.c ++zoffset.h +diff --git a/Documentation/kernel-parameters.txt b/Documentation/kernel-parameters.txt +index c840e7d..f4c451c 100644 +--- a/Documentation/kernel-parameters.txt ++++ b/Documentation/kernel-parameters.txt +@@ -1837,6 +1837,13 @@ and is between 256 and 4096 characters. It is defined in the file + the specified number of seconds. This is to be used if + your oopses keep scrolling off the screen. + ++ pax_nouderef [X86] disables UDEREF. Most likely needed under certain ++ virtualization environments that don't cope well with the ++ expand down segment used by UDEREF on X86-32 or the frequent ++ page table updates on X86-64. ++ ++ pax_softmode= 0/1 to disable/enable PaX softmode on boot already. ++ + pcbit= [HW,ISDN] + + pcd. [PARIDE] +diff --git a/MAINTAINERS b/MAINTAINERS +index 613da5d..4fe3eda 100644 +--- a/MAINTAINERS ++++ b/MAINTAINERS +@@ -5725,6 +5725,14 @@ L: netdev@vger.kernel.org + S: Maintained + F: drivers/net/vmxnet3/ + ++VMware PVSCSI driver ++M: Alok Kataria ++M: VMware PV-Drivers ++L: linux-scsi@vger.kernel.org ++S: Maintained ++F: drivers/scsi/vmw_pvscsi.c ++F: drivers/scsi/vmw_pvscsi.h ++ + VOLTAGE AND CURRENT REGULATOR FRAMEWORK + M: Liam Girdwood + M: Mark Brown +diff --git a/Makefile b/Makefile +index ed78982..cb8fa66 100644 +--- a/Makefile ++++ b/Makefile +@@ -221,8 +221,9 @@ CONFIG_SHELL := $(shell if [ -x "$$BASH" ]; then echo $$BASH; \ + + HOSTCC = gcc + HOSTCXX = g++ +-HOSTCFLAGS = -Wall -Wmissing-prototypes -Wstrict-prototypes -O2 -fomit-frame-pointer +-HOSTCXXFLAGS = -O2 ++HOSTCFLAGS = -Wall -W -Wmissing-prototypes -Wstrict-prototypes -Wno-unused-parameter -Wno-missing-field-initializers -O2 -fomit-frame-pointer -fno-delete-null-pointer-checks ++HOSTCFLAGS += $(call cc-option, -Wno-empty-body) ++HOSTCXXFLAGS = -O2 -Wall -W -fno-delete-null-pointer-checks + + # Decide whether to build built-in, modular, or both. + # Normally, just do built-in. +@@ -376,8 +377,8 @@ export RCS_TAR_IGNORE := --exclude SCCS --exclude BitKeeper --exclude .svn --exc + # Rules shared between *config targets and build targets + + # Basic helpers built in scripts/ +-PHONY += scripts_basic +-scripts_basic: ++PHONY += scripts_basic gcc-plugins ++scripts_basic: gcc-plugins + $(Q)$(MAKE) $(build)=scripts/basic + + # To avoid any implicit rule to kick in, define an empty command. +@@ -403,7 +404,7 @@ endif + # of make so .config is not included in this case either (for *config). + + no-dot-config-targets := clean mrproper distclean \ +- cscope TAGS tags help %docs check% \ ++ cscope gtags TAGS tags help %docs check% \ + include/linux/version.h headers_% \ + kernelrelease kernelversion + +@@ -526,6 +527,50 @@ else + KBUILD_CFLAGS += -O2 + endif + ++ifndef DISABLE_PAX_PLUGINS ++ifeq ($(shell $(CONFIG_SHELL) $(srctree)/scripts/gcc-plugin.sh "$(HOSTCC)" "$(CC)"), y) ++ifndef DISABLE_PAX_CONSTIFY_PLUGIN ++CONSTIFY_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/constify_plugin.so -DCONSTIFY_PLUGIN ++endif ++ifdef CONFIG_PAX_MEMORY_STACKLEAK ++STACKLEAK_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/stackleak_plugin.so -DSTACKLEAK_PLUGIN ++STACKLEAK_PLUGIN_CFLAGS += -fplugin-arg-stackleak_plugin-track-lowest-sp=100 ++endif ++ifdef CONFIG_KALLOCSTAT_PLUGIN ++KALLOCSTAT_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/kallocstat_plugin.so ++endif ++ifdef CONFIG_PAX_KERNEXEC_PLUGIN ++KERNEXEC_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/kernexec_plugin.so ++KERNEXEC_PLUGIN_CFLAGS += -fplugin-arg-kernexec_plugin-method=$(CONFIG_PAX_KERNEXEC_PLUGIN_METHOD) -DKERNEXEC_PLUGIN ++KERNEXEC_PLUGIN_AFLAGS := -DKERNEXEC_PLUGIN ++endif ++ifdef CONFIG_CHECKER_PLUGIN ++ifeq ($(call cc-ifversion, -ge, 0406, y), y) ++CHECKER_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/checker_plugin.so -DCHECKER_PLUGIN ++endif ++endif ++COLORIZE_PLUGIN_CFLAGS := -fplugin=$(objtree)/tools/gcc/colorize_plugin.so ++GCC_PLUGINS_CFLAGS := $(CONSTIFY_PLUGIN_CFLAGS) $(STACKLEAK_PLUGIN_CFLAGS) $(KALLOCSTAT_PLUGIN_CFLAGS) ++GCC_PLUGINS_CFLAGS += $(KERNEXEC_PLUGIN_CFLAGS) $(CHECKER_PLUGIN_CFLAGS) $(COLORIZE_PLUGIN_CFLAGS) ++GCC_PLUGINS_AFLAGS := $(KERNEXEC_PLUGIN_AFLAGS) ++export CONSTIFY_PLUGIN STACKLEAK_PLUGIN KERNEXEC_PLUGIN CHECKER_PLUGIN ++ifeq ($(KBUILD_EXTMOD),) ++gcc-plugins: ++ $(Q)$(MAKE) $(build)=tools/gcc ++else ++gcc-plugins: ; ++endif ++else ++gcc-plugins: ++ifeq ($(call cc-ifversion, -ge, 0405, y), y) ++ $(error Your gcc installation does not support plugins. If the necessary headers for plugin support are missing, they should be installed. On Debian, apt-get install gcc--plugin-dev. If you choose to ignore this error and lessen the improvements provided by this patch, re-run make with the DISABLE_PAX_PLUGINS=y argument.)) ++else ++ $(Q)echo "warning, your gcc version does not support plugins, you should upgrade it to gcc 4.5 at least" ++endif ++ $(Q)echo "PAX_MEMORY_STACKLEAK and constification will be less secure" ++endif ++endif ++ + include $(srctree)/arch/$(SRCARCH)/Makefile + + ifneq ($(CONFIG_FRAME_WARN),0) +@@ -647,7 +692,7 @@ export mod_strip_cmd + + + ifeq ($(KBUILD_EXTMOD),) +-core-y += kernel/ mm/ fs/ ipc/ security/ crypto/ block/ ++core-y += kernel/ mm/ fs/ ipc/ security/ crypto/ block/ grsecurity/ + + vmlinux-dirs := $(patsubst %/,%,$(filter %/, $(init-y) $(init-m) \ + $(core-y) $(core-m) $(drivers-y) $(drivers-m) \ +@@ -868,6 +913,8 @@ vmlinux.o: $(modpost-init) $(vmlinux-main) FORCE + + # The actual objects are generated when descending, + # make sure no implicit rule kicks in ++$(sort $(vmlinux-init) $(vmlinux-main)) $(vmlinux-lds): KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++$(sort $(vmlinux-init) $(vmlinux-main)) $(vmlinux-lds): KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) + $(sort $(vmlinux-init) $(vmlinux-main)) $(vmlinux-lds): $(vmlinux-dirs) ; + + # Handle descending into subdirectories listed in $(vmlinux-dirs) +@@ -877,7 +924,7 @@ $(sort $(vmlinux-init) $(vmlinux-main)) $(vmlinux-lds): $(vmlinux-dirs) ; + # Error messages still appears in the original language + + PHONY += $(vmlinux-dirs) +-$(vmlinux-dirs): prepare scripts ++$(vmlinux-dirs): gcc-plugins prepare scripts + $(Q)$(MAKE) $(build)=$@ + + # Build the kernel release string +@@ -986,6 +1033,7 @@ prepare0: archprepare FORCE + $(Q)$(MAKE) $(build)=. missing-syscalls + + # All the preparing.. ++prepare: KBUILD_CFLAGS := $(filter-out $(GCC_PLUGINS_CFLAGS),$(KBUILD_CFLAGS)) + prepare: prepare0 + + # The asm symlink changes when $(ARCH) changes. +@@ -1127,6 +1175,8 @@ all: modules + # using awk while concatenating to the final file. + + PHONY += modules ++modules: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++modules: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) + modules: $(vmlinux-dirs) $(if $(KBUILD_BUILTIN),vmlinux) + $(Q)$(AWK) '!x[$$0]++' $(vmlinux-dirs:%=$(objtree)/%/modules.order) > $(objtree)/modules.order + @$(kecho) ' Building modules, stage 2.'; +@@ -1136,7 +1186,7 @@ modules: $(vmlinux-dirs) $(if $(KBUILD_BUILTIN),vmlinux) + + # Target to prepare building external modules + PHONY += modules_prepare +-modules_prepare: prepare scripts ++modules_prepare: gcc-plugins prepare scripts + + # Target to install modules + PHONY += modules_install +@@ -1201,7 +1251,7 @@ MRPROPER_FILES += .config .config.old include/asm .version .old_version \ + include/linux/autoconf.h include/linux/version.h \ + include/linux/utsrelease.h \ + include/linux/bounds.h include/asm*/asm-offsets.h \ +- Module.symvers Module.markers tags TAGS cscope* ++ Module.symvers Module.markers tags TAGS cscope* GPATH GTAGS GRTAGS GSYMS + + # clean - Delete most, but leave enough to build external modules + # +@@ -1245,7 +1295,7 @@ distclean: mrproper + @find $(srctree) $(RCS_FIND_IGNORE) \ + \( -name '*.orig' -o -name '*.rej' -o -name '*~' \ + -o -name '*.bak' -o -name '#*#' -o -name '.*.orig' \ +- -o -name '.*.rej' -o -size 0 \ ++ -o -name '.*.rej' -o -name '*.so' -o -size 0 \ + -o -name '*%' -o -name '.*.cmd' -o -name 'core' \) \ + -type f -print | xargs rm -f + +@@ -1292,6 +1342,7 @@ help: + @echo ' modules_prepare - Set up for building external modules' + @echo ' tags/TAGS - Generate tags file for editors' + @echo ' cscope - Generate cscope index' ++ @echo ' gtags - Generate GNU GLOBAL index' + @echo ' kernelrelease - Output the release version string' + @echo ' kernelversion - Output the version stored in Makefile' + @echo ' headers_install - Install sanitised kernel headers to INSTALL_HDR_PATH'; \ +@@ -1393,6 +1444,8 @@ PHONY += $(module-dirs) modules + $(module-dirs): crmodverdir $(objtree)/Module.symvers + $(Q)$(MAKE) $(build)=$(patsubst _module_%,%,$@) + ++modules: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++modules: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) + modules: $(module-dirs) + @$(kecho) ' Building modules, stage 2.'; + $(Q)$(MAKE) -f $(srctree)/scripts/Makefile.modpost +@@ -1448,7 +1501,7 @@ endif # KBUILD_EXTMOD + quiet_cmd_tags = GEN $@ + cmd_tags = $(CONFIG_SHELL) $(srctree)/scripts/tags.sh $@ + +-tags TAGS cscope: FORCE ++tags TAGS cscope gtags: FORCE + $(call cmd,tags) + + # Scripts to check various things for consistency +@@ -1513,17 +1566,21 @@ else + target-dir = $(if $(KBUILD_EXTMOD),$(dir $<),$(dir $@)) + endif + +-%.s: %.c prepare scripts FORCE ++%.s: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++%.s: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) ++%.s: %.c gcc-plugins prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) + %.i: %.c prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) +-%.o: %.c prepare scripts FORCE ++%.o: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++%.o: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) ++%.o: %.c gcc-plugins prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) + %.lst: %.c prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) +-%.s: %.S prepare scripts FORCE ++%.s: %.S gcc-plugins prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) +-%.o: %.S prepare scripts FORCE ++%.o: %.S gcc-plugins prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) + %.symtypes: %.c prepare scripts FORCE + $(Q)$(MAKE) $(build)=$(build-dir) $(target-dir)$(notdir $@) +@@ -1533,11 +1590,15 @@ endif + $(cmd_crmodverdir) + $(Q)$(MAKE) KBUILD_MODULES=$(if $(CONFIG_MODULES),1) \ + $(build)=$(build-dir) +-%/: prepare scripts FORCE ++%/: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++%/: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) ++%/: gcc-plugins prepare scripts FORCE + $(cmd_crmodverdir) + $(Q)$(MAKE) KBUILD_MODULES=$(if $(CONFIG_MODULES),1) \ + $(build)=$(build-dir) +-%.ko: prepare scripts FORCE ++%.ko: KBUILD_CFLAGS += $(GCC_PLUGINS_CFLAGS) ++%.ko: KBUILD_AFLAGS += $(GCC_PLUGINS_AFLAGS) ++%.ko: gcc-plugins prepare scripts FORCE + $(cmd_crmodverdir) + $(Q)$(MAKE) KBUILD_MODULES=$(if $(CONFIG_MODULES),1) \ + $(build)=$(build-dir) $(@:.ko=.o) +diff --git a/arch/alpha/include/asm/atomic.h b/arch/alpha/include/asm/atomic.h +index 610dff4..f396854 100644 +--- a/arch/alpha/include/asm/atomic.h ++++ b/arch/alpha/include/asm/atomic.h +@@ -251,6 +251,16 @@ static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) + #define atomic_dec(v) atomic_sub(1,(v)) + #define atomic64_dec(v) atomic64_sub(1,(v)) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + #define smp_mb__before_atomic_dec() smp_mb() + #define smp_mb__after_atomic_dec() smp_mb() + #define smp_mb__before_atomic_inc() smp_mb() +diff --git a/arch/alpha/include/asm/cache.h b/arch/alpha/include/asm/cache.h +index f199e69..af005f5 100644 +--- a/arch/alpha/include/asm/cache.h ++++ b/arch/alpha/include/asm/cache.h +@@ -4,19 +4,20 @@ + #ifndef __ARCH_ALPHA_CACHE_H + #define __ARCH_ALPHA_CACHE_H + ++#include + + /* Bytes per L1 (data) cache line. */ + #if defined(CONFIG_ALPHA_GENERIC) || defined(CONFIG_ALPHA_EV6) +-# define L1_CACHE_BYTES 64 + # define L1_CACHE_SHIFT 6 + #else + /* Both EV4 and EV5 are write-through, read-allocate, + direct-mapped, physical. + */ +-# define L1_CACHE_BYTES 32 + # define L1_CACHE_SHIFT 5 + #endif + ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) ++ + #define L1_CACHE_ALIGN(x) (((x)+(L1_CACHE_BYTES-1))&~(L1_CACHE_BYTES-1)) + #define SMP_CACHE_BYTES L1_CACHE_BYTES + +diff --git a/arch/alpha/include/asm/elf.h b/arch/alpha/include/asm/elf.h +index 5c75c1b..c82f878 100644 +--- a/arch/alpha/include/asm/elf.h ++++ b/arch/alpha/include/asm/elf.h +@@ -91,6 +91,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_NFPREG]; + + #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x1000000) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (current->personality & ADDR_LIMIT_32BIT ? 0x10000 : 0x120000000UL) ++ ++#define PAX_DELTA_MMAP_LEN (current->personality & ADDR_LIMIT_32BIT ? 14 : 28) ++#define PAX_DELTA_STACK_LEN (current->personality & ADDR_LIMIT_32BIT ? 14 : 19) ++#endif ++ + /* $0 is set by ld.so to a pointer to a function which might be + registered using atexit. This provides a mean for the dynamic + linker to call DT_FINI functions for shared libraries that have +diff --git a/arch/alpha/include/asm/pgtable.h b/arch/alpha/include/asm/pgtable.h +index 3f0c59f..cf1e100 100644 +--- a/arch/alpha/include/asm/pgtable.h ++++ b/arch/alpha/include/asm/pgtable.h +@@ -101,6 +101,17 @@ struct vm_area_struct; + #define PAGE_SHARED __pgprot(_PAGE_VALID | __ACCESS_BITS) + #define PAGE_COPY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW) + #define PAGE_READONLY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW) ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++# define PAGE_SHARED_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOE) ++# define PAGE_COPY_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW | _PAGE_FOE) ++# define PAGE_READONLY_NOEXEC __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW | _PAGE_FOE) ++#else ++# define PAGE_SHARED_NOEXEC PAGE_SHARED ++# define PAGE_COPY_NOEXEC PAGE_COPY ++# define PAGE_READONLY_NOEXEC PAGE_READONLY ++#endif ++ + #define PAGE_KERNEL __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE | _PAGE_KWE) + + #define _PAGE_NORMAL(x) __pgprot(_PAGE_VALID | __ACCESS_BITS | (x)) +diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c +index ebc3c89..20cfa63 100644 +--- a/arch/alpha/kernel/module.c ++++ b/arch/alpha/kernel/module.c +@@ -182,7 +182,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab, + + /* The small sections were sorted to the end of the segment. + The following should definitely cover them. */ +- gp = (u64)me->module_core + me->core_size - 0x8000; ++ gp = (u64)me->module_core_rw + me->core_size_rw - 0x8000; + got = sechdrs[me->arch.gotsecindex].sh_addr; + + for (i = 0; i < n; i++) { +diff --git a/arch/alpha/kernel/osf_sys.c b/arch/alpha/kernel/osf_sys.c +index a94e49c..d71dd44 100644 +--- a/arch/alpha/kernel/osf_sys.c ++++ b/arch/alpha/kernel/osf_sys.c +@@ -1172,7 +1172,7 @@ arch_get_unmapped_area_1(unsigned long addr, unsigned long len, + /* At this point: (!vma || addr < vma->vm_end). */ + if (limit - len < addr) + return -ENOMEM; +- if (!vma || addr + len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + return addr; + addr = vma->vm_end; + vma = vma->vm_next; +@@ -1208,6 +1208,10 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, + merely specific addresses, but regions of memory -- perhaps + this feature should be incorporated into all ports? */ + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(current->mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + if (addr) { + addr = arch_get_unmapped_area_1 (PAGE_ALIGN(addr), len, limit); + if (addr != (unsigned long) -ENOMEM) +@@ -1215,8 +1219,8 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, + } + + /* Next, try allocating at TASK_UNMAPPED_BASE. */ +- addr = arch_get_unmapped_area_1 (PAGE_ALIGN(TASK_UNMAPPED_BASE), +- len, limit); ++ addr = arch_get_unmapped_area_1 (PAGE_ALIGN(current->mm->mmap_base), len, limit); ++ + if (addr != (unsigned long) -ENOMEM) + return addr; + +diff --git a/arch/alpha/mm/fault.c b/arch/alpha/mm/fault.c +index 00a31de..2ded0f2 100644 +--- a/arch/alpha/mm/fault.c ++++ b/arch/alpha/mm/fault.c +@@ -54,6 +54,124 @@ __load_new_mm_context(struct mm_struct *next_mm) + __reload_thread(pcb); + } + ++#ifdef CONFIG_PAX_PAGEEXEC ++/* ++ * PaX: decide what to do with offenders (regs->pc = fault address) ++ * ++ * returns 1 when task should be killed ++ * 2 when patched PLT trampoline was detected ++ * 3 when unpatched PLT trampoline was detected ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ ++#ifdef CONFIG_PAX_EMUPLT ++ int err; ++ ++ do { /* PaX: patched PLT emulation #1 */ ++ unsigned int ldah, ldq, jmp; ++ ++ err = get_user(ldah, (unsigned int *)regs->pc); ++ err |= get_user(ldq, (unsigned int *)(regs->pc+4)); ++ err |= get_user(jmp, (unsigned int *)(regs->pc+8)); ++ ++ if (err) ++ break; ++ ++ if ((ldah & 0xFFFF0000U) == 0x277B0000U && ++ (ldq & 0xFFFF0000U) == 0xA77B0000U && ++ jmp == 0x6BFB0000U) ++ { ++ unsigned long r27, addr; ++ unsigned long addrh = (ldah | 0xFFFFFFFFFFFF0000UL) << 16; ++ unsigned long addrl = ldq | 0xFFFFFFFFFFFF0000UL; ++ ++ addr = regs->r27 + ((addrh ^ 0x80000000UL) + 0x80000000UL) + ((addrl ^ 0x8000UL) + 0x8000UL); ++ err = get_user(r27, (unsigned long *)addr); ++ if (err) ++ break; ++ ++ regs->r27 = r27; ++ regs->pc = r27; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: patched PLT emulation #2 */ ++ unsigned int ldah, lda, br; ++ ++ err = get_user(ldah, (unsigned int *)regs->pc); ++ err |= get_user(lda, (unsigned int *)(regs->pc+4)); ++ err |= get_user(br, (unsigned int *)(regs->pc+8)); ++ ++ if (err) ++ break; ++ ++ if ((ldah & 0xFFFF0000U) == 0x277B0000U && ++ (lda & 0xFFFF0000U) == 0xA77B0000U && ++ (br & 0xFFE00000U) == 0xC3E00000U) ++ { ++ unsigned long addr = br | 0xFFFFFFFFFFE00000UL; ++ unsigned long addrh = (ldah | 0xFFFFFFFFFFFF0000UL) << 16; ++ unsigned long addrl = lda | 0xFFFFFFFFFFFF0000UL; ++ ++ regs->r27 += ((addrh ^ 0x80000000UL) + 0x80000000UL) + ((addrl ^ 0x8000UL) + 0x8000UL); ++ regs->pc += 12 + (((addr ^ 0x00100000UL) + 0x00100000UL) << 2); ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: unpatched PLT emulation */ ++ unsigned int br; ++ ++ err = get_user(br, (unsigned int *)regs->pc); ++ ++ if (!err && (br & 0xFFE00000U) == 0xC3800000U) { ++ unsigned int br2, ldq, nop, jmp; ++ unsigned long addr = br | 0xFFFFFFFFFFE00000UL, resolver; ++ ++ addr = regs->pc + 4 + (((addr ^ 0x00100000UL) + 0x00100000UL) << 2); ++ err = get_user(br2, (unsigned int *)addr); ++ err |= get_user(ldq, (unsigned int *)(addr+4)); ++ err |= get_user(nop, (unsigned int *)(addr+8)); ++ err |= get_user(jmp, (unsigned int *)(addr+12)); ++ err |= get_user(resolver, (unsigned long *)(addr+16)); ++ ++ if (err) ++ break; ++ ++ if (br2 == 0xC3600000U && ++ ldq == 0xA77B000CU && ++ nop == 0x47FF041FU && ++ jmp == 0x6B7B0000U) ++ { ++ regs->r28 = regs->pc+4; ++ regs->r27 = addr+16; ++ regs->pc = resolver; ++ return 3; ++ } ++ } ++ } while (0); ++#endif ++ ++ return 1; ++} ++ ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 5; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif + + /* + * This routine handles page faults. It determines the address, +@@ -131,8 +249,29 @@ do_page_fault(unsigned long address, unsigned long mmcsr, + good_area: + si_code = SEGV_ACCERR; + if (cause < 0) { +- if (!(vma->vm_flags & VM_EXEC)) ++ if (!(vma->vm_flags & VM_EXEC)) { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (!(mm->pax_flags & MF_PAX_PAGEEXEC) || address != regs->pc) ++ goto bad_area; ++ ++ up_read(&mm->mmap_sem); ++ switch (pax_handle_fetch_fault(regs)) { ++ ++#ifdef CONFIG_PAX_EMUPLT ++ case 2: ++ case 3: ++ return; ++#endif ++ ++ } ++ pax_report_fault(regs, (void *)regs->pc, (void *)rdusp()); ++ do_group_exit(SIGKILL); ++#else + goto bad_area; ++#endif ++ ++ } + } else if (!cause) { + /* Allow reads even for write-only mappings */ + if (!(vma->vm_flags & (VM_READ | VM_WRITE))) +diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig +index b68faef..6dd1496 100644 +--- a/arch/arm/Kconfig ++++ b/arch/arm/Kconfig +@@ -14,6 +14,7 @@ config ARM + select SYS_SUPPORTS_APM_EMULATION + select HAVE_OPROFILE + select HAVE_ARCH_KGDB ++ select GENERIC_ATOMIC64 + select HAVE_KPROBES if (!XIP_KERNEL) + select HAVE_KRETPROBES if (HAVE_KPROBES) + select HAVE_FUNCTION_TRACER if (!XIP_KERNEL) +diff --git a/arch/arm/include/asm/atomic.h b/arch/arm/include/asm/atomic.h +index d0daeab..ca7e10e 100644 +--- a/arch/arm/include/asm/atomic.h ++++ b/arch/arm/include/asm/atomic.h +@@ -15,6 +15,10 @@ + #include + #include + ++#ifdef CONFIG_GENERIC_ATOMIC64 ++#include ++#endif ++ + #define ATOMIC_INIT(i) { (i) } + + #ifdef __KERNEL__ +@@ -24,8 +28,16 @@ + * strex/ldrex monitor on some implementations. The reason we can use it for + * atomic_set() is the clrex or dummy strex done on every exception return. + */ +-#define atomic_read(v) ((v)->counter) ++#define atomic_read(v) (*(volatile int *)&(v)->counter) ++static inline int atomic_read_unchecked(const atomic_unchecked_t *v) ++{ ++ return v->counter; ++} + #define atomic_set(v,i) (((v)->counter) = (i)) ++static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i) ++{ ++ v->counter = i; ++} + + #if __LINUX_ARM_ARCH__ >= 6 + +@@ -40,6 +52,35 @@ static inline void atomic_add(int i, atomic_t *v) + int result; + + __asm__ __volatile__("@ atomic_add\n" ++"1: ldrex %1, [%2]\n" ++" add %0, %1, %3\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" bvc 3f\n" ++"2: bkpt 0xf103\n" ++"3:\n" ++#endif ++ ++" strex %1, %0, [%2]\n" ++" teq %1, #0\n" ++" bne 1b" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++"\n4:\n" ++ _ASM_EXTABLE(2b, 4b) ++#endif ++ ++ : "=&r" (result), "=&r" (tmp) ++ : "r" (&v->counter), "Ir" (i) ++ : "cc"); ++} ++ ++static inline void atomic_add_unchecked(int i, atomic_unchecked_t *v) ++{ ++ unsigned long tmp; ++ int result; ++ ++ __asm__ __volatile__("@ atomic_add_unchecked\n" + "1: ldrex %0, [%2]\n" + " add %0, %0, %3\n" + " strex %1, %0, [%2]\n" +@@ -58,6 +99,42 @@ static inline int atomic_add_return(int i, atomic_t *v) + smp_mb(); + + __asm__ __volatile__("@ atomic_add_return\n" ++"1: ldrex %1, [%2]\n" ++" add %0, %1, %3\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" bvc 3f\n" ++" mov %0, %1\n" ++"2: bkpt 0xf103\n" ++"3:\n" ++#endif ++ ++" strex %1, %0, [%2]\n" ++" teq %1, #0\n" ++" bne 1b" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++"\n4:\n" ++ _ASM_EXTABLE(2b, 4b) ++#endif ++ ++ : "=&r" (result), "=&r" (tmp) ++ : "r" (&v->counter), "Ir" (i) ++ : "cc"); ++ ++ smp_mb(); ++ ++ return result; ++} ++ ++static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v) ++{ ++ unsigned long tmp; ++ int result; ++ ++ smp_mb(); ++ ++ __asm__ __volatile__("@ atomic_add_return_unchecked\n" + "1: ldrex %0, [%2]\n" + " add %0, %0, %3\n" + " strex %1, %0, [%2]\n" +@@ -78,6 +155,35 @@ static inline void atomic_sub(int i, atomic_t *v) + int result; + + __asm__ __volatile__("@ atomic_sub\n" ++"1: ldrex %1, [%2]\n" ++" sub %0, %1, %3\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" bvc 3f\n" ++"2: bkpt 0xf103\n" ++"3:\n" ++#endif ++ ++" strex %1, %0, [%2]\n" ++" teq %1, #0\n" ++" bne 1b" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++"\n4:\n" ++ _ASM_EXTABLE(2b, 4b) ++#endif ++ ++ : "=&r" (result), "=&r" (tmp) ++ : "r" (&v->counter), "Ir" (i) ++ : "cc"); ++} ++ ++static inline void atomic_sub_unchecked(int i, atomic_unchecked_t *v) ++{ ++ unsigned long tmp; ++ int result; ++ ++ __asm__ __volatile__("@ atomic_sub_unchecked\n" + "1: ldrex %0, [%2]\n" + " sub %0, %0, %3\n" + " strex %1, %0, [%2]\n" +@@ -96,11 +202,25 @@ static inline int atomic_sub_return(int i, atomic_t *v) + smp_mb(); + + __asm__ __volatile__("@ atomic_sub_return\n" +-"1: ldrex %0, [%2]\n" +-" sub %0, %0, %3\n" ++"1: ldrex %1, [%2]\n" ++" sub %0, %1, %3\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" bvc 3f\n" ++" mov %0, %1\n" ++"2: bkpt 0xf103\n" ++"3:\n" ++#endif ++ + " strex %1, %0, [%2]\n" + " teq %1, #0\n" + " bne 1b" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++"\n4:\n" ++ _ASM_EXTABLE(2b, 4b) ++#endif ++ + : "=&r" (result), "=&r" (tmp) + : "r" (&v->counter), "Ir" (i) + : "cc"); +@@ -132,6 +252,28 @@ static inline int atomic_cmpxchg(atomic_t *ptr, int old, int new) + return oldval; + } + ++static inline int atomic_cmpxchg_unchecked(atomic_unchecked_t *ptr, int old, int new) ++{ ++ unsigned long oldval, res; ++ ++ smp_mb(); ++ ++ do { ++ __asm__ __volatile__("@ atomic_cmpxchg_unchecked\n" ++ "ldrex %1, [%2]\n" ++ "mov %0, #0\n" ++ "teq %1, %3\n" ++ "strexeq %0, %4, [%2]\n" ++ : "=&r" (res), "=&r" (oldval) ++ : "r" (&ptr->counter), "Ir" (old), "r" (new) ++ : "cc"); ++ } while (res); ++ ++ smp_mb(); ++ ++ return oldval; ++} ++ + static inline void atomic_clear_mask(unsigned long mask, unsigned long *addr) + { + unsigned long tmp, tmp2; +@@ -207,6 +349,10 @@ static inline void atomic_clear_mask(unsigned long mask, unsigned long *addr) + #endif /* __LINUX_ARM_ARCH__ */ + + #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) ++static inline int atomic_xchg_unchecked(atomic_unchecked_t *v, int new) ++{ ++ return xchg(&v->counter, new); ++} + + static inline int atomic_add_unless(atomic_t *v, int a, int u) + { +@@ -220,11 +366,27 @@ static inline int atomic_add_unless(atomic_t *v, int a, int u) + #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) + + #define atomic_inc(v) atomic_add(1, v) ++static inline void atomic_inc_unchecked(atomic_unchecked_t *v) ++{ ++ atomic_add_unchecked(1, v); ++} + #define atomic_dec(v) atomic_sub(1, v) ++static inline void atomic_dec_unchecked(atomic_unchecked_t *v) ++{ ++ atomic_sub_unchecked(1, v); ++} + + #define atomic_inc_and_test(v) (atomic_add_return(1, v) == 0) ++static inline int atomic_inc_and_test_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_add_return_unchecked(1, v) == 0; ++} + #define atomic_dec_and_test(v) (atomic_sub_return(1, v) == 0) + #define atomic_inc_return(v) (atomic_add_return(1, v)) ++static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_add_return_unchecked(1, v); ++} + #define atomic_dec_return(v) (atomic_sub_return(1, v)) + #define atomic_sub_and_test(i, v) (atomic_sub_return(i, v) == 0) + +diff --git a/arch/arm/include/asm/cache.h b/arch/arm/include/asm/cache.h +index 66c160b..bca1449 100644 +--- a/arch/arm/include/asm/cache.h ++++ b/arch/arm/include/asm/cache.h +@@ -4,8 +4,10 @@ + #ifndef __ASMARM_CACHE_H + #define __ASMARM_CACHE_H + ++#include ++ + #define L1_CACHE_SHIFT CONFIG_ARM_L1_CACHE_SHIFT +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + /* + * Memory returned by kmalloc() may be used for DMA, so we must make +diff --git a/arch/arm/include/asm/cacheflush.h b/arch/arm/include/asm/cacheflush.h +index 3d0cdd2..19957c5 100644 +--- a/arch/arm/include/asm/cacheflush.h ++++ b/arch/arm/include/asm/cacheflush.h +@@ -216,13 +216,13 @@ struct cpu_cache_fns { + void (*dma_inv_range)(const void *, const void *); + void (*dma_clean_range)(const void *, const void *); + void (*dma_flush_range)(const void *, const void *); +-}; ++} __no_const; + + struct outer_cache_fns { + void (*inv_range)(unsigned long, unsigned long); + void (*clean_range)(unsigned long, unsigned long); + void (*flush_range)(unsigned long, unsigned long); +-}; ++} __no_const; + + /* + * Select the calling method +diff --git a/arch/arm/include/asm/elf.h b/arch/arm/include/asm/elf.h +index 6aac3f5..265536b 100644 +--- a/arch/arm/include/asm/elf.h ++++ b/arch/arm/include/asm/elf.h +@@ -109,7 +109,14 @@ int dump_task_regs(struct task_struct *t, elf_gregset_t *elfregs); + the loader. We need to make sure that it is out of the way of the program + that it will "exec", and that there is sufficient room for the brk. */ + +-#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3) ++#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2) ++ ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE 0x00008000UL ++ ++#define PAX_DELTA_MMAP_LEN ((current->personality == PER_LINUX_32BIT) ? 16 : 10) ++#define PAX_DELTA_STACK_LEN ((current->personality == PER_LINUX_32BIT) ? 16 : 10) ++#endif + + /* When the program starts, a1 contains a pointer to a function to be + registered with atexit, as per the SVR4 ABI. A value of 0 means we +diff --git a/arch/arm/include/asm/kmap_types.h b/arch/arm/include/asm/kmap_types.h +index c019949..388fdd1 100644 +--- a/arch/arm/include/asm/kmap_types.h ++++ b/arch/arm/include/asm/kmap_types.h +@@ -19,6 +19,7 @@ enum km_type { + KM_SOFTIRQ0, + KM_SOFTIRQ1, + KM_L2_CACHE, ++ KM_CLEARPAGE, + KM_TYPE_NR + }; + +diff --git a/arch/arm/include/asm/page.h b/arch/arm/include/asm/page.h +index 3a32af4..c8def8a 100644 +--- a/arch/arm/include/asm/page.h ++++ b/arch/arm/include/asm/page.h +@@ -122,7 +122,7 @@ struct cpu_user_fns { + void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr); + void (*cpu_copy_user_highpage)(struct page *to, struct page *from, + unsigned long vaddr); +-}; ++} __no_const; + + #ifdef MULTI_USER + extern struct cpu_user_fns cpu_user; +diff --git a/arch/arm/include/asm/system.h b/arch/arm/include/asm/system.h +index d65b2f5..9d87555 100644 +--- a/arch/arm/include/asm/system.h ++++ b/arch/arm/include/asm/system.h +@@ -86,6 +86,8 @@ void hook_fault_code(int nr, int (*fn)(unsigned long, unsigned int, + + #define xchg(ptr,x) \ + ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr)))) ++#define xchg_unchecked(ptr,x) \ ++ ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr)))) + + extern asmlinkage void __backtrace(void); + extern asmlinkage void c_backtrace(unsigned long fp, int pmode); +@@ -98,7 +100,7 @@ extern int cpu_architecture(void); + extern void cpu_init(void); + + void arm_machine_restart(char mode, const char *cmd); +-extern void (*arm_pm_restart)(char str, const char *cmd); ++extern void (*arm_pm_restart)(char str, const char *cmd) __noreturn; + + #define UDBG_UNDEFINED (1 << 0) + #define UDBG_SYSCALL (1 << 1) +@@ -505,6 +507,13 @@ static inline unsigned long long __cmpxchg64_mb(volatile void *ptr, + + #endif /* __LINUX_ARM_ARCH__ >= 6 */ + ++#define _ASM_EXTABLE(from, to) \ ++" .pushsection __ex_table,\"a\"\n"\ ++" .align 3\n" \ ++" .long " #from ", " #to"\n" \ ++" .popsection" ++ ++ + #endif /* __ASSEMBLY__ */ + + #define arch_align_stack(x) (x) +diff --git a/arch/arm/include/asm/uaccess.h b/arch/arm/include/asm/uaccess.h +index 1d6bd40..fba0cb9 100644 +--- a/arch/arm/include/asm/uaccess.h ++++ b/arch/arm/include/asm/uaccess.h +@@ -22,6 +22,8 @@ + #define VERIFY_READ 0 + #define VERIFY_WRITE 1 + ++extern void check_object_size(const void *ptr, unsigned long n, bool to); ++ + /* + * The exception table consists of pairs of addresses: the first is the + * address of an instruction that is allowed to fault, and the second is +@@ -387,8 +389,23 @@ do { \ + + + #ifdef CONFIG_MMU +-extern unsigned long __must_check __copy_from_user(void *to, const void __user *from, unsigned long n); +-extern unsigned long __must_check __copy_to_user(void __user *to, const void *from, unsigned long n); ++extern unsigned long __must_check ___copy_from_user(void *to, const void __user *from, unsigned long n); ++extern unsigned long __must_check ___copy_to_user(void __user *to, const void *from, unsigned long n); ++ ++static inline unsigned long __must_check __copy_from_user(void *to, const void __user *from, unsigned long n) ++{ ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); ++ return ___copy_from_user(to, from, n); ++} ++ ++static inline unsigned long __must_check __copy_to_user(void __user *to, const void *from, unsigned long n) ++{ ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); ++ return ___copy_to_user(to, from, n); ++} ++ + extern unsigned long __must_check __copy_to_user_std(void __user *to, const void *from, unsigned long n); + extern unsigned long __must_check __clear_user(void __user *addr, unsigned long n); + extern unsigned long __must_check __clear_user_std(void __user *addr, unsigned long n); +@@ -403,6 +420,9 @@ extern unsigned long __must_check __strnlen_user(const char __user *s, long n); + + static inline unsigned long __must_check copy_from_user(void *to, const void __user *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + if (access_ok(VERIFY_READ, from, n)) + n = __copy_from_user(to, from, n); + else /* security hole - plug it */ +@@ -412,6 +432,9 @@ static inline unsigned long __must_check copy_from_user(void *to, const void __u + + static inline unsigned long __must_check copy_to_user(void __user *to, const void *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + if (access_ok(VERIFY_WRITE, to, n)) + n = __copy_to_user(to, from, n); + return n; +diff --git a/arch/arm/kernel/armksyms.c b/arch/arm/kernel/armksyms.c +index 0e62770..e2c2cd6 100644 +--- a/arch/arm/kernel/armksyms.c ++++ b/arch/arm/kernel/armksyms.c +@@ -118,8 +118,8 @@ EXPORT_SYMBOL(__strncpy_from_user); + #ifdef CONFIG_MMU + EXPORT_SYMBOL(copy_page); + +-EXPORT_SYMBOL(__copy_from_user); +-EXPORT_SYMBOL(__copy_to_user); ++EXPORT_SYMBOL(___copy_from_user); ++EXPORT_SYMBOL(___copy_to_user); + EXPORT_SYMBOL(__clear_user); + + EXPORT_SYMBOL(__get_user_1); +diff --git a/arch/arm/kernel/kgdb.c b/arch/arm/kernel/kgdb.c +index ba8ccfe..2dc34dc 100644 +--- a/arch/arm/kernel/kgdb.c ++++ b/arch/arm/kernel/kgdb.c +@@ -190,7 +190,7 @@ void kgdb_arch_exit(void) + * and we handle the normal undef case within the do_undefinstr + * handler. + */ +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + #ifndef __ARMEB__ + .gdb_bpt_instr = {0xfe, 0xde, 0xff, 0xe7} + #else /* ! __ARMEB__ */ +diff --git a/arch/arm/kernel/process.c b/arch/arm/kernel/process.c +index 61f90d3..771ab27 100644 +--- a/arch/arm/kernel/process.c ++++ b/arch/arm/kernel/process.c +@@ -83,7 +83,7 @@ static int __init hlt_setup(char *__unused) + __setup("nohlt", nohlt_setup); + __setup("hlt", hlt_setup); + +-void arm_machine_restart(char mode, const char *cmd) ++__noreturn void arm_machine_restart(char mode, const char *cmd) + { + /* + * Clean and disable cache, and turn off interrupts +@@ -117,7 +117,7 @@ void arm_machine_restart(char mode, const char *cmd) + void (*pm_power_off)(void); + EXPORT_SYMBOL(pm_power_off); + +-void (*arm_pm_restart)(char str, const char *cmd) = arm_machine_restart; ++void (*arm_pm_restart)(char str, const char *cmd) __noreturn = arm_machine_restart; + EXPORT_SYMBOL_GPL(arm_pm_restart); + + +@@ -195,6 +195,7 @@ __setup("reboot=", reboot_setup); + + void machine_halt(void) + { ++ BUG(); + } + + +@@ -202,6 +203,7 @@ void machine_power_off(void) + { + if (pm_power_off) + pm_power_off(); ++ BUG(); + } + + void machine_restart(char *cmd) +diff --git a/arch/arm/kernel/setup.c b/arch/arm/kernel/setup.c +index c6c57b6..0c3b29e 100644 +--- a/arch/arm/kernel/setup.c ++++ b/arch/arm/kernel/setup.c +@@ -92,16 +92,16 @@ EXPORT_SYMBOL(elf_hwcap); + struct processor processor; + #endif + #ifdef MULTI_TLB +-struct cpu_tlb_fns cpu_tlb; ++struct cpu_tlb_fns cpu_tlb __read_only; + #endif + #ifdef MULTI_USER +-struct cpu_user_fns cpu_user; ++struct cpu_user_fns cpu_user __read_only; + #endif + #ifdef MULTI_CACHE +-struct cpu_cache_fns cpu_cache; ++struct cpu_cache_fns cpu_cache __read_only; + #endif + #ifdef CONFIG_OUTER_CACHE +-struct outer_cache_fns outer_cache; ++struct outer_cache_fns outer_cache __read_only; + #endif + + struct stack { +diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c +index 3f361a7..6e806e1 100644 +--- a/arch/arm/kernel/traps.c ++++ b/arch/arm/kernel/traps.c +@@ -247,6 +247,8 @@ static void __die(const char *str, int err, struct thread_info *thread, struct p + + DEFINE_SPINLOCK(die_lock); + ++extern void gr_handle_kernel_exploit(void); ++ + /* + * This function is protected against re-entrancy. + */ +@@ -271,6 +273,8 @@ NORET_TYPE void die(const char *str, struct pt_regs *regs, int err) + if (panic_on_oops) + panic("Fatal exception"); + ++ gr_handle_kernel_exploit(); ++ + do_exit(SIGSEGV); + } + +diff --git a/arch/arm/kernel/vmlinux.lds.S b/arch/arm/kernel/vmlinux.lds.S +index aecf87df..bed731b 100644 +--- a/arch/arm/kernel/vmlinux.lds.S ++++ b/arch/arm/kernel/vmlinux.lds.S +@@ -74,14 +74,18 @@ SECTIONS + #ifndef CONFIG_XIP_KERNEL + __init_begin = _stext; + INIT_DATA ++ EXIT_TEXT ++ EXIT_DATA + . = ALIGN(PAGE_SIZE); + __init_end = .; + #endif + } + + /DISCARD/ : { /* Exit code and data */ ++#ifdef CONFIG_XIP_KERNEL + EXIT_TEXT + EXIT_DATA ++#endif + *(.exitcall.exit) + *(.discard) + *(.ARM.exidx.exit.text) +diff --git a/arch/arm/lib/copy_from_user.S b/arch/arm/lib/copy_from_user.S +index e4fe124..0fc246b 100644 +--- a/arch/arm/lib/copy_from_user.S ++++ b/arch/arm/lib/copy_from_user.S +@@ -16,7 +16,7 @@ + /* + * Prototype: + * +- * size_t __copy_from_user(void *to, const void *from, size_t n) ++ * size_t ___copy_from_user(void *to, const void *from, size_t n) + * + * Purpose: + * +@@ -84,11 +84,11 @@ + + .text + +-ENTRY(__copy_from_user) ++ENTRY(___copy_from_user) + + #include "copy_template.S" + +-ENDPROC(__copy_from_user) ++ENDPROC(___copy_from_user) + + .section .fixup,"ax" + .align 0 +diff --git a/arch/arm/lib/copy_page.S b/arch/arm/lib/copy_page.S +index 6ee2f67..d1cce76 100644 +--- a/arch/arm/lib/copy_page.S ++++ b/arch/arm/lib/copy_page.S +@@ -10,6 +10,7 @@ + * ASM optimised string functions + */ + #include ++#include + #include + #include + #include +diff --git a/arch/arm/lib/copy_to_user.S b/arch/arm/lib/copy_to_user.S +index 1a71e15..ac7b258 100644 +--- a/arch/arm/lib/copy_to_user.S ++++ b/arch/arm/lib/copy_to_user.S +@@ -16,7 +16,7 @@ + /* + * Prototype: + * +- * size_t __copy_to_user(void *to, const void *from, size_t n) ++ * size_t ___copy_to_user(void *to, const void *from, size_t n) + * + * Purpose: + * +@@ -88,11 +88,11 @@ + .text + + ENTRY(__copy_to_user_std) +-WEAK(__copy_to_user) ++WEAK(___copy_to_user) + + #include "copy_template.S" + +-ENDPROC(__copy_to_user) ++ENDPROC(___copy_to_user) + + .section .fixup,"ax" + .align 0 +diff --git a/arch/arm/lib/uaccess.S b/arch/arm/lib/uaccess.S +index ffdd274..91017b6 100644 +--- a/arch/arm/lib/uaccess.S ++++ b/arch/arm/lib/uaccess.S +@@ -19,7 +19,7 @@ + + #define PAGE_SHIFT 12 + +-/* Prototype: int __copy_to_user(void *to, const char *from, size_t n) ++/* Prototype: int ___copy_to_user(void *to, const char *from, size_t n) + * Purpose : copy a block to user memory from kernel memory + * Params : to - user memory + * : from - kernel memory +@@ -39,7 +39,7 @@ USER( strgtbt r3, [r0], #1) @ May fault + sub r2, r2, ip + b .Lc2u_dest_aligned + +-ENTRY(__copy_to_user) ++ENTRY(___copy_to_user) + stmfd sp!, {r2, r4 - r7, lr} + cmp r2, #4 + blt .Lc2u_not_enough +@@ -277,14 +277,14 @@ USER( strgebt r3, [r0], #1) @ May fault + ldrgtb r3, [r1], #0 + USER( strgtbt r3, [r0], #1) @ May fault + b .Lc2u_finished +-ENDPROC(__copy_to_user) ++ENDPROC(___copy_to_user) + + .section .fixup,"ax" + .align 0 + 9001: ldmfd sp!, {r0, r4 - r7, pc} + .previous + +-/* Prototype: unsigned long __copy_from_user(void *to,const void *from,unsigned long n); ++/* Prototype: unsigned long ___copy_from_user(void *to,const void *from,unsigned long n); + * Purpose : copy a block from user memory to kernel memory + * Params : to - kernel memory + * : from - user memory +@@ -303,7 +303,7 @@ USER( ldrgtbt r3, [r1], #1) @ May fault + sub r2, r2, ip + b .Lcfu_dest_aligned + +-ENTRY(__copy_from_user) ++ENTRY(___copy_from_user) + stmfd sp!, {r0, r2, r4 - r7, lr} + cmp r2, #4 + blt .Lcfu_not_enough +@@ -543,7 +543,7 @@ USER( ldrgebt r3, [r1], #1) @ May fault + USER( ldrgtbt r3, [r1], #1) @ May fault + strgtb r3, [r0], #1 + b .Lcfu_finished +-ENDPROC(__copy_from_user) ++ENDPROC(___copy_from_user) + + .section .fixup,"ax" + .align 0 +diff --git a/arch/arm/lib/uaccess_with_memcpy.c b/arch/arm/lib/uaccess_with_memcpy.c +index 6b967ff..67d5b2b 100644 +--- a/arch/arm/lib/uaccess_with_memcpy.c ++++ b/arch/arm/lib/uaccess_with_memcpy.c +@@ -97,7 +97,7 @@ out: + } + + unsigned long +-__copy_to_user(void __user *to, const void *from, unsigned long n) ++___copy_to_user(void __user *to, const void *from, unsigned long n) + { + /* + * This test is stubbed out of the main function above to keep +diff --git a/arch/arm/mach-at91/pm.c b/arch/arm/mach-at91/pm.c +index 4028724..beec230 100644 +--- a/arch/arm/mach-at91/pm.c ++++ b/arch/arm/mach-at91/pm.c +@@ -348,7 +348,7 @@ static void at91_pm_end(void) + } + + +-static struct platform_suspend_ops at91_pm_ops ={ ++static const struct platform_suspend_ops at91_pm_ops ={ + .valid = at91_pm_valid_state, + .begin = at91_pm_begin, + .enter = at91_pm_enter, +diff --git a/arch/arm/mach-omap1/pm.c b/arch/arm/mach-omap1/pm.c +index 5218943..0a34552 100644 +--- a/arch/arm/mach-omap1/pm.c ++++ b/arch/arm/mach-omap1/pm.c +@@ -647,7 +647,7 @@ static struct irqaction omap_wakeup_irq = { + + + +-static struct platform_suspend_ops omap_pm_ops ={ ++static const struct platform_suspend_ops omap_pm_ops ={ + .prepare = omap_pm_prepare, + .enter = omap_pm_enter, + .finish = omap_pm_finish, +diff --git a/arch/arm/mach-omap2/pm24xx.c b/arch/arm/mach-omap2/pm24xx.c +index bff5c4e..d4c649b 100644 +--- a/arch/arm/mach-omap2/pm24xx.c ++++ b/arch/arm/mach-omap2/pm24xx.c +@@ -326,7 +326,7 @@ static void omap2_pm_finish(void) + enable_hlt(); + } + +-static struct platform_suspend_ops omap_pm_ops = { ++static const struct platform_suspend_ops omap_pm_ops = { + .prepare = omap2_pm_prepare, + .enter = omap2_pm_enter, + .finish = omap2_pm_finish, +diff --git a/arch/arm/mach-omap2/pm34xx.c b/arch/arm/mach-omap2/pm34xx.c +index 8946319..7d3e661 100644 +--- a/arch/arm/mach-omap2/pm34xx.c ++++ b/arch/arm/mach-omap2/pm34xx.c +@@ -401,7 +401,7 @@ static void omap3_pm_end(void) + return; + } + +-static struct platform_suspend_ops omap_pm_ops = { ++static const struct platform_suspend_ops omap_pm_ops = { + .begin = omap3_pm_begin, + .end = omap3_pm_end, + .prepare = omap3_pm_prepare, +diff --git a/arch/arm/mach-pnx4008/pm.c b/arch/arm/mach-pnx4008/pm.c +index b3d8d53..6e68ebc 100644 +--- a/arch/arm/mach-pnx4008/pm.c ++++ b/arch/arm/mach-pnx4008/pm.c +@@ -116,7 +116,7 @@ static int pnx4008_pm_valid(suspend_state_t state) + (state == PM_SUSPEND_MEM); + } + +-static struct platform_suspend_ops pnx4008_pm_ops = { ++static const struct platform_suspend_ops pnx4008_pm_ops = { + .enter = pnx4008_pm_enter, + .valid = pnx4008_pm_valid, + }; +diff --git a/arch/arm/mach-pxa/pm.c b/arch/arm/mach-pxa/pm.c +index 7693355..9beb00a 100644 +--- a/arch/arm/mach-pxa/pm.c ++++ b/arch/arm/mach-pxa/pm.c +@@ -95,7 +95,7 @@ void pxa_pm_finish(void) + pxa_cpu_pm_fns->finish(); + } + +-static struct platform_suspend_ops pxa_pm_ops = { ++static const struct platform_suspend_ops pxa_pm_ops = { + .valid = pxa_pm_valid, + .enter = pxa_pm_enter, + .prepare = pxa_pm_prepare, +diff --git a/arch/arm/mach-pxa/sharpsl_pm.c b/arch/arm/mach-pxa/sharpsl_pm.c +index 629e05d..06be589 100644 +--- a/arch/arm/mach-pxa/sharpsl_pm.c ++++ b/arch/arm/mach-pxa/sharpsl_pm.c +@@ -891,7 +891,7 @@ static void sharpsl_apm_get_power_status(struct apm_power_info *info) + } + + #ifdef CONFIG_PM +-static struct platform_suspend_ops sharpsl_pm_ops = { ++static const struct platform_suspend_ops sharpsl_pm_ops = { + .prepare = pxa_pm_prepare, + .finish = pxa_pm_finish, + .enter = corgi_pxa_pm_enter, +diff --git a/arch/arm/mach-sa1100/pm.c b/arch/arm/mach-sa1100/pm.c +index c83fdc8..ab9fc44 100644 +--- a/arch/arm/mach-sa1100/pm.c ++++ b/arch/arm/mach-sa1100/pm.c +@@ -120,7 +120,7 @@ unsigned long sleep_phys_sp(void *sp) + return virt_to_phys(sp); + } + +-static struct platform_suspend_ops sa11x0_pm_ops = { ++static const struct platform_suspend_ops sa11x0_pm_ops = { + .enter = sa11x0_pm_enter, + .valid = suspend_valid_only_mem, + }; +diff --git a/arch/arm/mm/fault.c b/arch/arm/mm/fault.c +index 3191cd6..c322981 100644 +--- a/arch/arm/mm/fault.c ++++ b/arch/arm/mm/fault.c +@@ -166,6 +166,13 @@ __do_user_fault(struct task_struct *tsk, unsigned long addr, + } + #endif + ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (fsr & FSR_LNX_PF) { ++ pax_report_fault(regs, (void *)regs->ARM_pc, (void *)regs->ARM_sp); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + tsk->thread.address = addr; + tsk->thread.error_code = fsr; + tsk->thread.trap_no = 14; +@@ -357,6 +364,33 @@ do_page_fault(unsigned long addr, unsigned int fsr, struct pt_regs *regs) + } + #endif /* CONFIG_MMU */ + ++#ifdef CONFIG_PAX_PAGEEXEC ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 20; i++) { ++ unsigned char c; ++ if (get_user(c, (__force unsigned char __user *)pc+i)) ++ printk(KERN_CONT "?? "); ++ else ++ printk(KERN_CONT "%02x ", c); ++ } ++ printk("\n"); ++ ++ printk(KERN_ERR "PAX: bytes at SP-4: "); ++ for (i = -1; i < 20; i++) { ++ unsigned long c; ++ if (get_user(c, (__force unsigned long __user *)sp+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08lx ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + /* + * First Level Translation Fault Handler + * +@@ -569,6 +603,20 @@ do_PrefetchAbort(unsigned long addr, unsigned int ifsr, struct pt_regs *regs) + const struct fsr_info *inf = ifsr_info + fsr_fs(ifsr); + struct siginfo info; + ++#ifdef CONFIG_PAX_REFCOUNT ++ if (fsr_fs(ifsr) == 2) { ++ unsigned int bkpt; ++ ++ if (!probe_kernel_address((unsigned int *)addr, bkpt) && bkpt == 0xe12f1073) { ++ current->thread.error_code = ifsr; ++ current->thread.trap_no = 0; ++ pax_report_refcount_overflow(regs); ++ fixup_exception(regs); ++ return; ++ } ++ } ++#endif ++ + if (!inf->fn(addr, ifsr | FSR_LNX_PF, regs)) + return; + +diff --git a/arch/arm/mm/mmap.c b/arch/arm/mm/mmap.c +index f5abc51..7ec524c 100644 +--- a/arch/arm/mm/mmap.c ++++ b/arch/arm/mm/mmap.c +@@ -63,6 +63,10 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, + if (len > TASK_SIZE) + return -ENOMEM; + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + if (addr) { + if (do_align) + addr = COLOUR_ALIGN(addr, pgoff); +@@ -70,15 +74,14 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, + addr = PAGE_ALIGN(addr); + + vma = find_vma(mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + if (len > mm->cached_hole_size) { +- start_addr = addr = mm->free_area_cache; ++ start_addr = addr = mm->free_area_cache; + } else { +- start_addr = addr = TASK_UNMAPPED_BASE; +- mm->cached_hole_size = 0; ++ start_addr = addr = mm->mmap_base; ++ mm->cached_hole_size = 0; + } + + full_search: +@@ -94,14 +97,14 @@ full_search: + * Start a new search - just in case we missed + * some holes. + */ +- if (start_addr != TASK_UNMAPPED_BASE) { +- start_addr = addr = TASK_UNMAPPED_BASE; ++ if (start_addr != mm->mmap_base) { ++ start_addr = addr = mm->mmap_base; + mm->cached_hole_size = 0; + goto full_search; + } + return -ENOMEM; + } +- if (!vma || addr + len <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* + * Remember the place where we stopped the search: + */ +diff --git a/arch/arm/plat-s3c/pm.c b/arch/arm/plat-s3c/pm.c +index 8d97db2..b66cfa5 100644 +--- a/arch/arm/plat-s3c/pm.c ++++ b/arch/arm/plat-s3c/pm.c +@@ -355,7 +355,7 @@ static void s3c_pm_finish(void) + s3c_pm_check_cleanup(); + } + +-static struct platform_suspend_ops s3c_pm_ops = { ++static const struct platform_suspend_ops s3c_pm_ops = { + .enter = s3c_pm_enter, + .prepare = s3c_pm_prepare, + .finish = s3c_pm_finish, +diff --git a/arch/avr32/include/asm/cache.h b/arch/avr32/include/asm/cache.h +index d3cf35a..0ba6053 100644 +--- a/arch/avr32/include/asm/cache.h ++++ b/arch/avr32/include/asm/cache.h +@@ -1,8 +1,10 @@ + #ifndef __ASM_AVR32_CACHE_H + #define __ASM_AVR32_CACHE_H + ++#include ++ + #define L1_CACHE_SHIFT 5 +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + /* + * Memory returned by kmalloc() may be used for DMA, so we must make +diff --git a/arch/avr32/include/asm/elf.h b/arch/avr32/include/asm/elf.h +index d5d1d41..856e2ed 100644 +--- a/arch/avr32/include/asm/elf.h ++++ b/arch/avr32/include/asm/elf.h +@@ -85,8 +85,14 @@ typedef struct user_fpu_struct elf_fpregset_t; + the loader. We need to make sure that it is out of the way of the program + that it will "exec", and that there is sufficient room for the brk. */ + +-#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3) ++#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE 0x00001000UL ++ ++#define PAX_DELTA_MMAP_LEN 15 ++#define PAX_DELTA_STACK_LEN 15 ++#endif + + /* This yields a mask that user programs can use to figure out what + instruction set this CPU supports. This could be done in user space, +diff --git a/arch/avr32/include/asm/kmap_types.h b/arch/avr32/include/asm/kmap_types.h +index b7f5c68..556135c 100644 +--- a/arch/avr32/include/asm/kmap_types.h ++++ b/arch/avr32/include/asm/kmap_types.h +@@ -22,7 +22,8 @@ D(10) KM_IRQ0, + D(11) KM_IRQ1, + D(12) KM_SOFTIRQ0, + D(13) KM_SOFTIRQ1, +-D(14) KM_TYPE_NR ++D(14) KM_CLEARPAGE, ++D(15) KM_TYPE_NR + }; + + #undef D +diff --git a/arch/avr32/mach-at32ap/pm.c b/arch/avr32/mach-at32ap/pm.c +index f021edf..32d680e 100644 +--- a/arch/avr32/mach-at32ap/pm.c ++++ b/arch/avr32/mach-at32ap/pm.c +@@ -176,7 +176,7 @@ out: + return 0; + } + +-static struct platform_suspend_ops avr32_pm_ops = { ++static const struct platform_suspend_ops avr32_pm_ops = { + .valid = avr32_pm_valid_state, + .enter = avr32_pm_enter, + }; +diff --git a/arch/avr32/mm/fault.c b/arch/avr32/mm/fault.c +index b61d86d..e292c7f 100644 +--- a/arch/avr32/mm/fault.c ++++ b/arch/avr32/mm/fault.c +@@ -41,6 +41,23 @@ static inline int notify_page_fault(struct pt_regs *regs, int trap) + + int exception_trace = 1; + ++#ifdef CONFIG_PAX_PAGEEXEC ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 20; i++) { ++ unsigned char c; ++ if (get_user(c, (unsigned char *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%02x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + /* + * This routine handles page faults. It determines the address and the + * problem, and then passes it off to one of the appropriate routines. +@@ -157,6 +174,16 @@ bad_area: + up_read(&mm->mmap_sem); + + if (user_mode(regs)) { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (mm->pax_flags & MF_PAX_PAGEEXEC) { ++ if (ecr == ECR_PROTECTION_X || ecr == ECR_TLB_MISS_X) { ++ pax_report_fault(regs, (void *)regs->pc, (void *)regs->sp); ++ do_group_exit(SIGKILL); ++ } ++ } ++#endif ++ + if (exception_trace && printk_ratelimit()) + printk("%s%s[%d]: segfault at %08lx pc %08lx " + "sp %08lx ecr %lu\n", +diff --git a/arch/blackfin/include/asm/cache.h b/arch/blackfin/include/asm/cache.h +index 93f6c63..d144953 100644 +--- a/arch/blackfin/include/asm/cache.h ++++ b/arch/blackfin/include/asm/cache.h +@@ -7,12 +7,14 @@ + #ifndef __ARCH_BLACKFIN_CACHE_H + #define __ARCH_BLACKFIN_CACHE_H + ++#include ++ + /* + * Bytes per L1 cache line + * Blackfin loads 32 bytes for cache + */ + #define L1_CACHE_SHIFT 5 +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + #define SMP_CACHE_BYTES L1_CACHE_BYTES + + #define ARCH_KMALLOC_MINALIGN L1_CACHE_BYTES +diff --git a/arch/blackfin/kernel/kgdb.c b/arch/blackfin/kernel/kgdb.c +index cce79d0..c406c85 100644 +--- a/arch/blackfin/kernel/kgdb.c ++++ b/arch/blackfin/kernel/kgdb.c +@@ -428,7 +428,7 @@ int kgdb_arch_handle_exception(int vector, int signo, + return -1; /* this means that we do not want to exit from the handler */ + } + +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + .gdb_bpt_instr = {0xa1}, + #ifdef CONFIG_SMP + .flags = KGDB_HW_BREAKPOINT|KGDB_THR_PROC_SWAP, +diff --git a/arch/blackfin/mach-common/pm.c b/arch/blackfin/mach-common/pm.c +index 8837be4..b2fb413 100644 +--- a/arch/blackfin/mach-common/pm.c ++++ b/arch/blackfin/mach-common/pm.c +@@ -255,7 +255,7 @@ static int bfin_pm_enter(suspend_state_t state) + return 0; + } + +-struct platform_suspend_ops bfin_pm_ops = { ++const struct platform_suspend_ops bfin_pm_ops = { + .enter = bfin_pm_enter, + .valid = bfin_pm_valid, + }; +diff --git a/arch/cris/include/arch-v10/arch/cache.h b/arch/cris/include/arch-v10/arch/cache.h +index aea2718..3639a60 100644 +--- a/arch/cris/include/arch-v10/arch/cache.h ++++ b/arch/cris/include/arch-v10/arch/cache.h +@@ -1,8 +1,9 @@ + #ifndef _ASM_ARCH_CACHE_H + #define _ASM_ARCH_CACHE_H + ++#include + /* Etrax 100LX have 32-byte cache-lines. */ +-#define L1_CACHE_BYTES 32 + #define L1_CACHE_SHIFT 5 ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #endif /* _ASM_ARCH_CACHE_H */ +diff --git a/arch/cris/include/arch-v32/arch/cache.h b/arch/cris/include/arch-v32/arch/cache.h +index dfc7305..417f5b3 100644 +--- a/arch/cris/include/arch-v32/arch/cache.h ++++ b/arch/cris/include/arch-v32/arch/cache.h +@@ -1,11 +1,12 @@ + #ifndef _ASM_CRIS_ARCH_CACHE_H + #define _ASM_CRIS_ARCH_CACHE_H + ++#include + #include + + /* A cache-line is 32 bytes. */ +-#define L1_CACHE_BYTES 32 + #define L1_CACHE_SHIFT 5 ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + void flush_dma_list(dma_descr_data *descr); + void flush_dma_descr(dma_descr_data *descr, int flush_buf); +diff --git a/arch/frv/include/asm/atomic.h b/arch/frv/include/asm/atomic.h +index 00a57af..c3ef0cd 100644 +--- a/arch/frv/include/asm/atomic.h ++++ b/arch/frv/include/asm/atomic.h +@@ -241,6 +241,16 @@ extern uint32_t __xchg_32(uint32_t i, volatile void *v); + #define atomic64_cmpxchg(v, old, new) (__cmpxchg_64(old, new, &(v)->counter)) + #define atomic64_xchg(v, new) (__xchg_64(new, &(v)->counter)) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + static __inline__ int atomic_add_unless(atomic_t *v, int a, int u) + { + int c, old; +diff --git a/arch/frv/include/asm/cache.h b/arch/frv/include/asm/cache.h +index 7dc0f0f..1e6a620 100644 +--- a/arch/frv/include/asm/cache.h ++++ b/arch/frv/include/asm/cache.h +@@ -12,10 +12,11 @@ + #ifndef __ASM_CACHE_H + #define __ASM_CACHE_H + ++#include + + /* bytes per L1 cache line */ + #define L1_CACHE_SHIFT (CONFIG_FRV_L1_CACHE_SHIFT) +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define ARCH_KMALLOC_MINALIGN L1_CACHE_BYTES + +diff --git a/arch/frv/include/asm/kmap_types.h b/arch/frv/include/asm/kmap_types.h +index f8e16b2..c73ff79 100644 +--- a/arch/frv/include/asm/kmap_types.h ++++ b/arch/frv/include/asm/kmap_types.h +@@ -23,6 +23,7 @@ enum km_type { + KM_IRQ1, + KM_SOFTIRQ0, + KM_SOFTIRQ1, ++ KM_CLEARPAGE, + KM_TYPE_NR + }; + +diff --git a/arch/frv/mm/elf-fdpic.c b/arch/frv/mm/elf-fdpic.c +index 385fd30..6c3d97e 100644 +--- a/arch/frv/mm/elf-fdpic.c ++++ b/arch/frv/mm/elf-fdpic.c +@@ -73,8 +73,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + if (addr) { + addr = PAGE_ALIGN(addr); + vma = find_vma(current->mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len)) + goto success; + } + +@@ -89,7 +88,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + for (; vma; vma = vma->vm_next) { + if (addr > limit) + break; +- if (addr + len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + goto success; + addr = vma->vm_end; + } +@@ -104,7 +103,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + for (; vma; vma = vma->vm_next) { + if (addr > limit) + break; +- if (addr + len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + goto success; + addr = vma->vm_end; + } +diff --git a/arch/h8300/include/asm/cache.h b/arch/h8300/include/asm/cache.h +index c635028..6d9445a 100644 +--- a/arch/h8300/include/asm/cache.h ++++ b/arch/h8300/include/asm/cache.h +@@ -1,8 +1,10 @@ + #ifndef __ARCH_H8300_CACHE_H + #define __ARCH_H8300_CACHE_H + ++#include ++ + /* bytes per L1 cache line */ +-#define L1_CACHE_BYTES 4 ++#define L1_CACHE_BYTES _AC(4,UL) + + /* m68k-elf-gcc 2.95.2 doesn't like these */ + +diff --git a/arch/ia64/hp/common/hwsw_iommu.c b/arch/ia64/hp/common/hwsw_iommu.c +index e4a80d8..11a7ea1 100644 +--- a/arch/ia64/hp/common/hwsw_iommu.c ++++ b/arch/ia64/hp/common/hwsw_iommu.c +@@ -17,7 +17,7 @@ + #include + #include + +-extern struct dma_map_ops sba_dma_ops, swiotlb_dma_ops; ++extern const struct dma_map_ops sba_dma_ops, swiotlb_dma_ops; + + /* swiotlb declarations & definitions: */ + extern int swiotlb_late_init_with_default_size (size_t size); +@@ -33,7 +33,7 @@ static inline int use_swiotlb(struct device *dev) + !sba_dma_ops.dma_supported(dev, *dev->dma_mask); + } + +-struct dma_map_ops *hwsw_dma_get_ops(struct device *dev) ++const struct dma_map_ops *hwsw_dma_get_ops(struct device *dev) + { + if (use_swiotlb(dev)) + return &swiotlb_dma_ops; +diff --git a/arch/ia64/hp/common/sba_iommu.c b/arch/ia64/hp/common/sba_iommu.c +index 01ae69b..35752fd 100644 +--- a/arch/ia64/hp/common/sba_iommu.c ++++ b/arch/ia64/hp/common/sba_iommu.c +@@ -2097,7 +2097,7 @@ static struct acpi_driver acpi_sba_ioc_driver = { + }, + }; + +-extern struct dma_map_ops swiotlb_dma_ops; ++extern const struct dma_map_ops swiotlb_dma_ops; + + static int __init + sba_init(void) +@@ -2211,7 +2211,7 @@ sba_page_override(char *str) + + __setup("sbapagesize=",sba_page_override); + +-struct dma_map_ops sba_dma_ops = { ++const struct dma_map_ops sba_dma_ops = { + .alloc_coherent = sba_alloc_coherent, + .free_coherent = sba_free_coherent, + .map_page = sba_map_page, +diff --git a/arch/ia64/ia32/binfmt_elf32.c b/arch/ia64/ia32/binfmt_elf32.c +index c69552b..c7122f4 100644 +--- a/arch/ia64/ia32/binfmt_elf32.c ++++ b/arch/ia64/ia32/binfmt_elf32.c +@@ -45,6 +45,13 @@ randomize_stack_top(unsigned long stack_top); + + #define elf_read_implies_exec(ex, have_pt_gnu_stack) (!(have_pt_gnu_stack)) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (current->personality == PER_LINUX32 ? 0x08048000UL : 0x4000000000000000UL) ++ ++#define PAX_DELTA_MMAP_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13) ++#define PAX_DELTA_STACK_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13) ++#endif ++ + /* Ugly but avoids duplication */ + #include "../../../fs/binfmt_elf.c" + +diff --git a/arch/ia64/ia32/ia32priv.h b/arch/ia64/ia32/ia32priv.h +index 0f15349..26b3429 100644 +--- a/arch/ia64/ia32/ia32priv.h ++++ b/arch/ia64/ia32/ia32priv.h +@@ -296,7 +296,14 @@ typedef struct compat_siginfo { + #define ELF_DATA ELFDATA2LSB + #define ELF_ARCH EM_386 + +-#define IA32_STACK_TOP IA32_PAGE_OFFSET ++#ifdef CONFIG_PAX_RANDUSTACK ++#define __IA32_DELTA_STACK (current->mm->delta_stack) ++#else ++#define __IA32_DELTA_STACK 0UL ++#endif ++ ++#define IA32_STACK_TOP (IA32_PAGE_OFFSET - __IA32_DELTA_STACK) ++ + #define IA32_GATE_OFFSET IA32_PAGE_OFFSET + #define IA32_GATE_END IA32_PAGE_OFFSET + PAGE_SIZE + +diff --git a/arch/ia64/include/asm/atomic.h b/arch/ia64/include/asm/atomic.h +index 88405cb..de5ca5d 100644 +--- a/arch/ia64/include/asm/atomic.h ++++ b/arch/ia64/include/asm/atomic.h +@@ -210,6 +210,16 @@ atomic64_add_negative (__s64 i, atomic64_t *v) + #define atomic64_inc(v) atomic64_add(1, (v)) + #define atomic64_dec(v) atomic64_sub(1, (v)) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + /* Atomic operations are already serializing */ + #define smp_mb__before_atomic_dec() barrier() + #define smp_mb__after_atomic_dec() barrier() +diff --git a/arch/ia64/include/asm/cache.h b/arch/ia64/include/asm/cache.h +index e7482bd..d1c9b8e 100644 +--- a/arch/ia64/include/asm/cache.h ++++ b/arch/ia64/include/asm/cache.h +@@ -1,6 +1,7 @@ + #ifndef _ASM_IA64_CACHE_H + #define _ASM_IA64_CACHE_H + ++#include + + /* + * Copyright (C) 1998-2000 Hewlett-Packard Co +@@ -9,7 +10,7 @@ + + /* Bytes per L1 (data) cache line. */ + #define L1_CACHE_SHIFT CONFIG_IA64_L1_CACHE_SHIFT +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #ifdef CONFIG_SMP + # define SMP_CACHE_SHIFT L1_CACHE_SHIFT +diff --git a/arch/ia64/include/asm/dma-mapping.h b/arch/ia64/include/asm/dma-mapping.h +index 8d3c79c..71b3af6 100644 +--- a/arch/ia64/include/asm/dma-mapping.h ++++ b/arch/ia64/include/asm/dma-mapping.h +@@ -12,7 +12,7 @@ + + #define ARCH_HAS_DMA_GET_REQUIRED_MASK + +-extern struct dma_map_ops *dma_ops; ++extern const struct dma_map_ops *dma_ops; + extern struct ia64_machine_vector ia64_mv; + extern void set_iommu_machvec(void); + +@@ -24,7 +24,7 @@ extern void machvec_dma_sync_sg(struct device *, struct scatterlist *, int, + static inline void *dma_alloc_coherent(struct device *dev, size_t size, + dma_addr_t *daddr, gfp_t gfp) + { +- struct dma_map_ops *ops = platform_dma_get_ops(dev); ++ const struct dma_map_ops *ops = platform_dma_get_ops(dev); + void *caddr; + + caddr = ops->alloc_coherent(dev, size, daddr, gfp); +@@ -35,7 +35,7 @@ static inline void *dma_alloc_coherent(struct device *dev, size_t size, + static inline void dma_free_coherent(struct device *dev, size_t size, + void *caddr, dma_addr_t daddr) + { +- struct dma_map_ops *ops = platform_dma_get_ops(dev); ++ const struct dma_map_ops *ops = platform_dma_get_ops(dev); + debug_dma_free_coherent(dev, size, caddr, daddr); + ops->free_coherent(dev, size, caddr, daddr); + } +@@ -49,13 +49,13 @@ static inline void dma_free_coherent(struct device *dev, size_t size, + + static inline int dma_mapping_error(struct device *dev, dma_addr_t daddr) + { +- struct dma_map_ops *ops = platform_dma_get_ops(dev); ++ const struct dma_map_ops *ops = platform_dma_get_ops(dev); + return ops->mapping_error(dev, daddr); + } + + static inline int dma_supported(struct device *dev, u64 mask) + { +- struct dma_map_ops *ops = platform_dma_get_ops(dev); ++ const struct dma_map_ops *ops = platform_dma_get_ops(dev); + return ops->dma_supported(dev, mask); + } + +diff --git a/arch/ia64/include/asm/elf.h b/arch/ia64/include/asm/elf.h +index 86eddee..b116bb4 100644 +--- a/arch/ia64/include/asm/elf.h ++++ b/arch/ia64/include/asm/elf.h +@@ -43,6 +43,13 @@ + */ + #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x800000000UL) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (current->personality == PER_LINUX32 ? 0x08048000UL : 0x4000000000000000UL) ++ ++#define PAX_DELTA_MMAP_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13) ++#define PAX_DELTA_STACK_LEN (current->personality == PER_LINUX32 ? 16 : 3*PAGE_SHIFT - 13) ++#endif ++ + #define PT_IA_64_UNWIND 0x70000001 + + /* IA-64 relocations: */ +diff --git a/arch/ia64/include/asm/machvec.h b/arch/ia64/include/asm/machvec.h +index 367d299..9ad4279 100644 +--- a/arch/ia64/include/asm/machvec.h ++++ b/arch/ia64/include/asm/machvec.h +@@ -45,7 +45,7 @@ typedef void ia64_mv_kernel_launch_event_t(void); + /* DMA-mapping interface: */ + typedef void ia64_mv_dma_init (void); + typedef u64 ia64_mv_dma_get_required_mask (struct device *); +-typedef struct dma_map_ops *ia64_mv_dma_get_ops(struct device *); ++typedef const struct dma_map_ops *ia64_mv_dma_get_ops(struct device *); + + /* + * WARNING: The legacy I/O space is _architected_. Platforms are +@@ -251,7 +251,7 @@ extern void machvec_init_from_cmdline(const char *cmdline); + # endif /* CONFIG_IA64_GENERIC */ + + extern void swiotlb_dma_init(void); +-extern struct dma_map_ops *dma_get_ops(struct device *); ++extern const struct dma_map_ops *dma_get_ops(struct device *); + + /* + * Define default versions so we can extend machvec for new platforms without having +diff --git a/arch/ia64/include/asm/pgtable.h b/arch/ia64/include/asm/pgtable.h +index 8840a69..cdb63d9 100644 +--- a/arch/ia64/include/asm/pgtable.h ++++ b/arch/ia64/include/asm/pgtable.h +@@ -12,7 +12,7 @@ + * David Mosberger-Tang + */ + +- ++#include + #include + #include + #include +@@ -143,6 +143,17 @@ + #define PAGE_READONLY __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R) + #define PAGE_COPY __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R) + #define PAGE_COPY_EXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_RX) ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++# define PAGE_SHARED_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_RW) ++# define PAGE_READONLY_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R) ++# define PAGE_COPY_NOEXEC __pgprot(__ACCESS_BITS | _PAGE_PL_3 | _PAGE_AR_R) ++#else ++# define PAGE_SHARED_NOEXEC PAGE_SHARED ++# define PAGE_READONLY_NOEXEC PAGE_READONLY ++# define PAGE_COPY_NOEXEC PAGE_COPY ++#endif ++ + #define PAGE_GATE __pgprot(__ACCESS_BITS | _PAGE_PL_0 | _PAGE_AR_X_RX) + #define PAGE_KERNEL __pgprot(__DIRTY_BITS | _PAGE_PL_0 | _PAGE_AR_RWX) + #define PAGE_KERNELRX __pgprot(__ACCESS_BITS | _PAGE_PL_0 | _PAGE_AR_RX) +diff --git a/arch/ia64/include/asm/spinlock.h b/arch/ia64/include/asm/spinlock.h +index 239ecdc..f94170e 100644 +--- a/arch/ia64/include/asm/spinlock.h ++++ b/arch/ia64/include/asm/spinlock.h +@@ -72,7 +72,7 @@ static __always_inline void __ticket_spin_unlock(raw_spinlock_t *lock) + unsigned short *p = (unsigned short *)&lock->lock + 1, tmp; + + asm volatile ("ld2.bias %0=[%1]" : "=r"(tmp) : "r"(p)); +- ACCESS_ONCE(*p) = (tmp + 2) & ~1; ++ ACCESS_ONCE_RW(*p) = (tmp + 2) & ~1; + } + + static __always_inline void __ticket_spin_unlock_wait(raw_spinlock_t *lock) +diff --git a/arch/ia64/include/asm/uaccess.h b/arch/ia64/include/asm/uaccess.h +index 449c8c0..432a3d2 100644 +--- a/arch/ia64/include/asm/uaccess.h ++++ b/arch/ia64/include/asm/uaccess.h +@@ -257,7 +257,7 @@ __copy_from_user (void *to, const void __user *from, unsigned long count) + const void *__cu_from = (from); \ + long __cu_len = (n); \ + \ +- if (__access_ok(__cu_to, __cu_len, get_fs())) \ ++ if (__cu_len > 0 && __cu_len <= INT_MAX && __access_ok(__cu_to, __cu_len, get_fs())) \ + __cu_len = __copy_user(__cu_to, (__force void __user *) __cu_from, __cu_len); \ + __cu_len; \ + }) +@@ -269,7 +269,7 @@ __copy_from_user (void *to, const void __user *from, unsigned long count) + long __cu_len = (n); \ + \ + __chk_user_ptr(__cu_from); \ +- if (__access_ok(__cu_from, __cu_len, get_fs())) \ ++ if (__cu_len > 0 && __cu_len <= INT_MAX && __access_ok(__cu_from, __cu_len, get_fs())) \ + __cu_len = __copy_user((__force void __user *) __cu_to, __cu_from, __cu_len); \ + __cu_len; \ + }) +diff --git a/arch/ia64/kernel/dma-mapping.c b/arch/ia64/kernel/dma-mapping.c +index f2c1600..969398a 100644 +--- a/arch/ia64/kernel/dma-mapping.c ++++ b/arch/ia64/kernel/dma-mapping.c +@@ -3,7 +3,7 @@ + /* Set this to 1 if there is a HW IOMMU in the system */ + int iommu_detected __read_mostly; + +-struct dma_map_ops *dma_ops; ++const struct dma_map_ops *dma_ops; + EXPORT_SYMBOL(dma_ops); + + #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16) +@@ -16,7 +16,7 @@ static int __init dma_init(void) + } + fs_initcall(dma_init); + +-struct dma_map_ops *dma_get_ops(struct device *dev) ++const struct dma_map_ops *dma_get_ops(struct device *dev) + { + return dma_ops; + } +diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c +index 1481b0a..e7d38ff 100644 +--- a/arch/ia64/kernel/module.c ++++ b/arch/ia64/kernel/module.c +@@ -315,8 +315,7 @@ module_alloc (unsigned long size) + void + module_free (struct module *mod, void *module_region) + { +- if (mod && mod->arch.init_unw_table && +- module_region == mod->module_init) { ++ if (mod && mod->arch.init_unw_table && module_region == mod->module_init_rx) { + unw_remove_unwind_table(mod->arch.init_unw_table); + mod->arch.init_unw_table = NULL; + } +@@ -502,15 +501,39 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings, + } + + static inline int ++in_init_rx (const struct module *mod, uint64_t addr) ++{ ++ return addr - (uint64_t) mod->module_init_rx < mod->init_size_rx; ++} ++ ++static inline int ++in_init_rw (const struct module *mod, uint64_t addr) ++{ ++ return addr - (uint64_t) mod->module_init_rw < mod->init_size_rw; ++} ++ ++static inline int + in_init (const struct module *mod, uint64_t addr) + { +- return addr - (uint64_t) mod->module_init < mod->init_size; ++ return in_init_rx(mod, addr) || in_init_rw(mod, addr); ++} ++ ++static inline int ++in_core_rx (const struct module *mod, uint64_t addr) ++{ ++ return addr - (uint64_t) mod->module_core_rx < mod->core_size_rx; ++} ++ ++static inline int ++in_core_rw (const struct module *mod, uint64_t addr) ++{ ++ return addr - (uint64_t) mod->module_core_rw < mod->core_size_rw; + } + + static inline int + in_core (const struct module *mod, uint64_t addr) + { +- return addr - (uint64_t) mod->module_core < mod->core_size; ++ return in_core_rx(mod, addr) || in_core_rw(mod, addr); + } + + static inline int +@@ -693,7 +716,14 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend, + break; + + case RV_BDREL: +- val -= (uint64_t) (in_init(mod, val) ? mod->module_init : mod->module_core); ++ if (in_init_rx(mod, val)) ++ val -= (uint64_t) mod->module_init_rx; ++ else if (in_init_rw(mod, val)) ++ val -= (uint64_t) mod->module_init_rw; ++ else if (in_core_rx(mod, val)) ++ val -= (uint64_t) mod->module_core_rx; ++ else if (in_core_rw(mod, val)) ++ val -= (uint64_t) mod->module_core_rw; + break; + + case RV_LTV: +@@ -828,15 +858,15 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind + * addresses have been selected... + */ + uint64_t gp; +- if (mod->core_size > MAX_LTOFF) ++ if (mod->core_size_rx + mod->core_size_rw > MAX_LTOFF) + /* + * This takes advantage of fact that SHF_ARCH_SMALL gets allocated + * at the end of the module. + */ +- gp = mod->core_size - MAX_LTOFF / 2; ++ gp = mod->core_size_rx + mod->core_size_rw - MAX_LTOFF / 2; + else +- gp = mod->core_size / 2; +- gp = (uint64_t) mod->module_core + ((gp + 7) & -8); ++ gp = (mod->core_size_rx + mod->core_size_rw) / 2; ++ gp = (uint64_t) mod->module_core_rx + ((gp + 7) & -8); + mod->arch.gp = gp; + DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp); + } +diff --git a/arch/ia64/kernel/pci-dma.c b/arch/ia64/kernel/pci-dma.c +index f6b1ff0..de773fb 100644 +--- a/arch/ia64/kernel/pci-dma.c ++++ b/arch/ia64/kernel/pci-dma.c +@@ -43,7 +43,7 @@ struct device fallback_dev = { + .dma_mask = &fallback_dev.coherent_dma_mask, + }; + +-extern struct dma_map_ops intel_dma_ops; ++extern const struct dma_map_ops intel_dma_ops; + + static int __init pci_iommu_init(void) + { +@@ -96,15 +96,34 @@ int iommu_dma_supported(struct device *dev, u64 mask) + } + EXPORT_SYMBOL(iommu_dma_supported); + ++extern void *intel_alloc_coherent(struct device *hwdev, size_t size, dma_addr_t *dma_handle, gfp_t flags); ++extern void intel_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle); ++extern int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems, enum dma_data_direction dir, struct dma_attrs *attrs); ++extern void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist, int nelems, enum dma_data_direction dir, struct dma_attrs *attrs); ++extern dma_addr_t intel_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, struct dma_attrs *attrs); ++extern void intel_unmap_page(struct device *dev, dma_addr_t dev_addr, size_t size, enum dma_data_direction dir, struct dma_attrs *attrs); ++extern int intel_mapping_error(struct device *dev, dma_addr_t dma_addr); ++ ++static const struct dma_map_ops intel_iommu_dma_ops = { ++ /* from drivers/pci/intel-iommu.c:intel_dma_ops */ ++ .alloc_coherent = intel_alloc_coherent, ++ .free_coherent = intel_free_coherent, ++ .map_sg = intel_map_sg, ++ .unmap_sg = intel_unmap_sg, ++ .map_page = intel_map_page, ++ .unmap_page = intel_unmap_page, ++ .mapping_error = intel_mapping_error, ++ ++ .sync_single_for_cpu = machvec_dma_sync_single, ++ .sync_sg_for_cpu = machvec_dma_sync_sg, ++ .sync_single_for_device = machvec_dma_sync_single, ++ .sync_sg_for_device = machvec_dma_sync_sg, ++ .dma_supported = iommu_dma_supported, ++}; ++ + void __init pci_iommu_alloc(void) + { +- dma_ops = &intel_dma_ops; +- +- dma_ops->sync_single_for_cpu = machvec_dma_sync_single; +- dma_ops->sync_sg_for_cpu = machvec_dma_sync_sg; +- dma_ops->sync_single_for_device = machvec_dma_sync_single; +- dma_ops->sync_sg_for_device = machvec_dma_sync_sg; +- dma_ops->dma_supported = iommu_dma_supported; ++ dma_ops = &intel_iommu_dma_ops; + + /* + * The order of these functions is important for +diff --git a/arch/ia64/kernel/pci-swiotlb.c b/arch/ia64/kernel/pci-swiotlb.c +index 285aae8..61dbab6 100644 +--- a/arch/ia64/kernel/pci-swiotlb.c ++++ b/arch/ia64/kernel/pci-swiotlb.c +@@ -21,7 +21,7 @@ static void *ia64_swiotlb_alloc_coherent(struct device *dev, size_t size, + return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); + } + +-struct dma_map_ops swiotlb_dma_ops = { ++const struct dma_map_ops swiotlb_dma_ops = { + .alloc_coherent = ia64_swiotlb_alloc_coherent, + .free_coherent = swiotlb_free_coherent, + .map_page = swiotlb_map_page, +diff --git a/arch/ia64/kernel/sys_ia64.c b/arch/ia64/kernel/sys_ia64.c +index 609d500..7dde2a8 100644 +--- a/arch/ia64/kernel/sys_ia64.c ++++ b/arch/ia64/kernel/sys_ia64.c +@@ -43,6 +43,13 @@ arch_get_unmapped_area (struct file *filp, unsigned long addr, unsigned long len + if (REGION_NUMBER(addr) == RGN_HPAGE) + addr = 0; + #endif ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ addr = mm->free_area_cache; ++ else ++#endif ++ + if (!addr) + addr = mm->free_area_cache; + +@@ -61,14 +68,14 @@ arch_get_unmapped_area (struct file *filp, unsigned long addr, unsigned long len + for (vma = find_vma(mm, addr); ; vma = vma->vm_next) { + /* At this point: (!vma || addr < vma->vm_end). */ + if (TASK_SIZE - len < addr || RGN_MAP_LIMIT - len < REGION_OFFSET(addr)) { +- if (start_addr != TASK_UNMAPPED_BASE) { ++ if (start_addr != mm->mmap_base) { + /* Start a new search --- just in case we missed some holes. */ +- addr = TASK_UNMAPPED_BASE; ++ addr = mm->mmap_base; + goto full_search; + } + return -ENOMEM; + } +- if (!vma || addr + len <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* Remember the address where we stopped this search: */ + mm->free_area_cache = addr + len; + return addr; +diff --git a/arch/ia64/kernel/topology.c b/arch/ia64/kernel/topology.c +index 8f06035..b3a5818 100644 +--- a/arch/ia64/kernel/topology.c ++++ b/arch/ia64/kernel/topology.c +@@ -282,7 +282,7 @@ static ssize_t cache_show(struct kobject * kobj, struct attribute * attr, char * + return ret; + } + +-static struct sysfs_ops cache_sysfs_ops = { ++static const struct sysfs_ops cache_sysfs_ops = { + .show = cache_show + }; + +diff --git a/arch/ia64/kernel/vmlinux.lds.S b/arch/ia64/kernel/vmlinux.lds.S +index 0a0c77b..8e55a81 100644 +--- a/arch/ia64/kernel/vmlinux.lds.S ++++ b/arch/ia64/kernel/vmlinux.lds.S +@@ -190,7 +190,7 @@ SECTIONS + /* Per-cpu data: */ + . = ALIGN(PERCPU_PAGE_SIZE); + PERCPU_VADDR(PERCPU_ADDR, :percpu) +- __phys_per_cpu_start = __per_cpu_load; ++ __phys_per_cpu_start = per_cpu_load; + . = __phys_per_cpu_start + PERCPU_PAGE_SIZE; /* ensure percpu data fits + * into percpu page size + */ +diff --git a/arch/ia64/mm/fault.c b/arch/ia64/mm/fault.c +index 19261a9..1611b7a 100644 +--- a/arch/ia64/mm/fault.c ++++ b/arch/ia64/mm/fault.c +@@ -72,6 +72,23 @@ mapped_kernel_page_is_present (unsigned long address) + return pte_present(pte); + } + ++#ifdef CONFIG_PAX_PAGEEXEC ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 8; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + void __kprobes + ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *regs) + { +@@ -145,9 +162,23 @@ ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *re + mask = ( (((isr >> IA64_ISR_X_BIT) & 1UL) << VM_EXEC_BIT) + | (((isr >> IA64_ISR_W_BIT) & 1UL) << VM_WRITE_BIT)); + +- if ((vma->vm_flags & mask) != mask) ++ if ((vma->vm_flags & mask) != mask) { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (!(vma->vm_flags & VM_EXEC) && (mask & VM_EXEC)) { ++ if (!(mm->pax_flags & MF_PAX_PAGEEXEC) || address != regs->cr_iip) ++ goto bad_area; ++ ++ up_read(&mm->mmap_sem); ++ pax_report_fault(regs, (void *)regs->cr_iip, (void *)regs->r12); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + goto bad_area; + ++ } ++ + survive: + /* + * If for any reason at all we couldn't handle the fault, make +diff --git a/arch/ia64/mm/hugetlbpage.c b/arch/ia64/mm/hugetlbpage.c +index b0f6157..a082bbc 100644 +--- a/arch/ia64/mm/hugetlbpage.c ++++ b/arch/ia64/mm/hugetlbpage.c +@@ -172,7 +172,7 @@ unsigned long hugetlb_get_unmapped_area(struct file *file, unsigned long addr, u + /* At this point: (!vmm || addr < vmm->vm_end). */ + if (REGION_OFFSET(addr) + len > RGN_MAP_LIMIT) + return -ENOMEM; +- if (!vmm || (addr + len) <= vmm->vm_start) ++ if (check_heap_stack_gap(vmm, addr, len)) + return addr; + addr = ALIGN(vmm->vm_end, HPAGE_SIZE); + } +diff --git a/arch/ia64/mm/init.c b/arch/ia64/mm/init.c +index 1857766..05cc6a3 100644 +--- a/arch/ia64/mm/init.c ++++ b/arch/ia64/mm/init.c +@@ -122,6 +122,19 @@ ia64_init_addr_space (void) + vma->vm_start = current->thread.rbs_bot & PAGE_MASK; + vma->vm_end = vma->vm_start + PAGE_SIZE; + vma->vm_flags = VM_DATA_DEFAULT_FLAGS|VM_GROWSUP|VM_ACCOUNT; ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (current->mm->pax_flags & MF_PAX_PAGEEXEC) { ++ vma->vm_flags &= ~VM_EXEC; ++ ++#ifdef CONFIG_PAX_MPROTECT ++ if (current->mm->pax_flags & MF_PAX_MPROTECT) ++ vma->vm_flags &= ~VM_MAYEXEC; ++#endif ++ ++ } ++#endif ++ + vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); + down_write(¤t->mm->mmap_sem); + if (insert_vm_struct(current->mm, vma)) { +diff --git a/arch/ia64/sn/pci/pci_dma.c b/arch/ia64/sn/pci/pci_dma.c +index 98b6849..8046766 100644 +--- a/arch/ia64/sn/pci/pci_dma.c ++++ b/arch/ia64/sn/pci/pci_dma.c +@@ -464,7 +464,7 @@ int sn_pci_legacy_write(struct pci_bus *bus, u16 port, u32 val, u8 size) + return ret; + } + +-static struct dma_map_ops sn_dma_ops = { ++static const struct dma_map_ops sn_dma_ops = { + .alloc_coherent = sn_dma_alloc_coherent, + .free_coherent = sn_dma_free_coherent, + .map_page = sn_dma_map_page, +diff --git a/arch/m32r/include/asm/cache.h b/arch/m32r/include/asm/cache.h +index 40b3ee9..8c2c112 100644 +--- a/arch/m32r/include/asm/cache.h ++++ b/arch/m32r/include/asm/cache.h +@@ -1,8 +1,10 @@ + #ifndef _ASM_M32R_CACHE_H + #define _ASM_M32R_CACHE_H + ++#include ++ + /* L1 cache line size */ + #define L1_CACHE_SHIFT 4 +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #endif /* _ASM_M32R_CACHE_H */ +diff --git a/arch/m32r/lib/usercopy.c b/arch/m32r/lib/usercopy.c +index 82abd15..d95ae5d 100644 +--- a/arch/m32r/lib/usercopy.c ++++ b/arch/m32r/lib/usercopy.c +@@ -14,6 +14,9 @@ + unsigned long + __generic_copy_to_user(void __user *to, const void *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + prefetch(from); + if (access_ok(VERIFY_WRITE, to, n)) + __copy_user(to,from,n); +@@ -23,6 +26,9 @@ __generic_copy_to_user(void __user *to, const void *from, unsigned long n) + unsigned long + __generic_copy_from_user(void *to, const void __user *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + prefetchw(to); + if (access_ok(VERIFY_READ, from, n)) + __copy_user_zeroing(to,from,n); +diff --git a/arch/m68k/include/asm/cache.h b/arch/m68k/include/asm/cache.h +index ecafbe1..432c3e4 100644 +--- a/arch/m68k/include/asm/cache.h ++++ b/arch/m68k/include/asm/cache.h +@@ -4,9 +4,11 @@ + #ifndef __ARCH_M68K_CACHE_H + #define __ARCH_M68K_CACHE_H + ++#include ++ + /* bytes per L1 cache line */ + #define L1_CACHE_SHIFT 4 +-#define L1_CACHE_BYTES (1<< L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define ARCH_KMALLOC_MINALIGN L1_CACHE_BYTES + +diff --git a/arch/microblaze/include/asm/cache.h b/arch/microblaze/include/asm/cache.h +index c209c47..2ba96e2 100644 +--- a/arch/microblaze/include/asm/cache.h ++++ b/arch/microblaze/include/asm/cache.h +@@ -13,11 +13,12 @@ + #ifndef _ASM_MICROBLAZE_CACHE_H + #define _ASM_MICROBLAZE_CACHE_H + ++#include + #include + + #define L1_CACHE_SHIFT 2 + /* word-granular cache in microblaze */ +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define SMP_CACHE_BYTES L1_CACHE_BYTES + +diff --git a/arch/mips/Kconfig b/arch/mips/Kconfig +index fd7620f..63d73a6 100644 +--- a/arch/mips/Kconfig ++++ b/arch/mips/Kconfig +@@ -5,6 +5,7 @@ config MIPS + select HAVE_IDE + select HAVE_OPROFILE + select HAVE_ARCH_KGDB ++ select GENERIC_ATOMIC64 if !64BIT + # Horrible source of confusion. Die, die, die ... + select EMBEDDED + select RTC_LIB if !LEMOTE_FULOONG2E +diff --git a/arch/mips/Makefile b/arch/mips/Makefile +index 77f5021..2b1db8a 100644 +--- a/arch/mips/Makefile ++++ b/arch/mips/Makefile +@@ -51,6 +51,8 @@ endif + cflags-y := -ffunction-sections + cflags-y += $(call cc-option, -mno-check-zero-division) + ++cflags-y += -Wno-sign-compare -Wno-extra ++ + ifdef CONFIG_32BIT + ld-emul = $(32bit-emul) + vmlinux-32 = vmlinux +diff --git a/arch/mips/alchemy/devboards/pm.c b/arch/mips/alchemy/devboards/pm.c +index 632f986..fd0378d 100644 +--- a/arch/mips/alchemy/devboards/pm.c ++++ b/arch/mips/alchemy/devboards/pm.c +@@ -78,7 +78,7 @@ static void db1x_pm_end(void) + + } + +-static struct platform_suspend_ops db1x_pm_ops = { ++static const struct platform_suspend_ops db1x_pm_ops = { + .valid = suspend_valid_only_mem, + .begin = db1x_pm_begin, + .enter = db1x_pm_enter, +diff --git a/arch/mips/include/asm/atomic.h b/arch/mips/include/asm/atomic.h +index 09e7128..111035b 100644 +--- a/arch/mips/include/asm/atomic.h ++++ b/arch/mips/include/asm/atomic.h +@@ -21,6 +21,10 @@ + #include + #include + ++#ifdef CONFIG_GENERIC_ATOMIC64 ++#include ++#endif ++ + #define ATOMIC_INIT(i) { (i) } + + /* +@@ -782,6 +786,16 @@ static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) + */ + #define atomic64_add_negative(i, v) (atomic64_add_return(i, (v)) < 0) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + #endif /* CONFIG_64BIT */ + + /* +diff --git a/arch/mips/include/asm/cache.h b/arch/mips/include/asm/cache.h +index 37f175c..c7a3065 100644 +--- a/arch/mips/include/asm/cache.h ++++ b/arch/mips/include/asm/cache.h +@@ -9,10 +9,11 @@ + #ifndef _ASM_CACHE_H + #define _ASM_CACHE_H + ++#include + #include + + #define L1_CACHE_SHIFT CONFIG_MIPS_L1_CACHE_SHIFT +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define SMP_CACHE_SHIFT L1_CACHE_SHIFT + #define SMP_CACHE_BYTES L1_CACHE_BYTES +diff --git a/arch/mips/include/asm/elf.h b/arch/mips/include/asm/elf.h +index 7990694..4e93acf 100644 +--- a/arch/mips/include/asm/elf.h ++++ b/arch/mips/include/asm/elf.h +@@ -368,4 +368,11 @@ extern int dump_task_fpu(struct task_struct *, elf_fpregset_t *); + #define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2) + #endif + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL) ++ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#endif ++ + #endif /* _ASM_ELF_H */ +diff --git a/arch/mips/include/asm/page.h b/arch/mips/include/asm/page.h +index f266295..627cfff 100644 +--- a/arch/mips/include/asm/page.h ++++ b/arch/mips/include/asm/page.h +@@ -93,7 +93,7 @@ extern void copy_user_highpage(struct page *to, struct page *from, + #ifdef CONFIG_CPU_MIPS32 + typedef struct { unsigned long pte_low, pte_high; } pte_t; + #define pte_val(x) ((x).pte_low | ((unsigned long long)(x).pte_high << 32)) +- #define __pte(x) ({ pte_t __pte = {(x), ((unsigned long long)(x)) >> 32}; __pte; }) ++ #define __pte(x) ({ pte_t __pte = {(x), (x) >> 32}; __pte; }) + #else + typedef struct { unsigned long long pte; } pte_t; + #define pte_val(x) ((x).pte) +diff --git a/arch/mips/include/asm/reboot.h b/arch/mips/include/asm/reboot.h +index e48c0bf..f3acf65 100644 +--- a/arch/mips/include/asm/reboot.h ++++ b/arch/mips/include/asm/reboot.h +@@ -9,7 +9,7 @@ + #ifndef _ASM_REBOOT_H + #define _ASM_REBOOT_H + +-extern void (*_machine_restart)(char *command); +-extern void (*_machine_halt)(void); ++extern void (*__noreturn _machine_restart)(char *command); ++extern void (*__noreturn _machine_halt)(void); + + #endif /* _ASM_REBOOT_H */ +diff --git a/arch/mips/include/asm/system.h b/arch/mips/include/asm/system.h +index 83b5509..9fa24a23 100644 +--- a/arch/mips/include/asm/system.h ++++ b/arch/mips/include/asm/system.h +@@ -230,6 +230,6 @@ extern void per_cpu_trap_init(void); + */ + #define __ARCH_WANT_UNLOCKED_CTXSW + +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) ((x) & ~0xfUL) + + #endif /* _ASM_SYSTEM_H */ +diff --git a/arch/mips/kernel/binfmt_elfn32.c b/arch/mips/kernel/binfmt_elfn32.c +index 9fdd8bc..fcf9d68 100644 +--- a/arch/mips/kernel/binfmt_elfn32.c ++++ b/arch/mips/kernel/binfmt_elfn32.c +@@ -50,6 +50,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_NFPREG]; + #undef ELF_ET_DYN_BASE + #define ELF_ET_DYN_BASE (TASK32_SIZE / 3 * 2) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL) ++ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#endif ++ + #include + #include + #include +diff --git a/arch/mips/kernel/binfmt_elfo32.c b/arch/mips/kernel/binfmt_elfo32.c +index ff44823..cf0b48a 100644 +--- a/arch/mips/kernel/binfmt_elfo32.c ++++ b/arch/mips/kernel/binfmt_elfo32.c +@@ -52,6 +52,13 @@ typedef elf_fpreg_t elf_fpregset_t[ELF_NFPREG]; + #undef ELF_ET_DYN_BASE + #define ELF_ET_DYN_BASE (TASK32_SIZE / 3 * 2) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT_ADDR) ? 0x00400000UL : 0x00400000UL) ++ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT_ADDR) ? 27-PAGE_SHIFT : 36-PAGE_SHIFT) ++#endif ++ + #include + + /* +diff --git a/arch/mips/kernel/kgdb.c b/arch/mips/kernel/kgdb.c +index 50c9bb8..efdd5f8 100644 +--- a/arch/mips/kernel/kgdb.c ++++ b/arch/mips/kernel/kgdb.c +@@ -245,6 +245,7 @@ int kgdb_arch_handle_exception(int vector, int signo, int err_code, + return -1; + } + ++/* cannot be const */ + struct kgdb_arch arch_kgdb_ops; + + /* +diff --git a/arch/mips/kernel/process.c b/arch/mips/kernel/process.c +index f3d73e1..bb3f57a 100644 +--- a/arch/mips/kernel/process.c ++++ b/arch/mips/kernel/process.c +@@ -470,15 +470,3 @@ unsigned long get_wchan(struct task_struct *task) + out: + return pc; + } +- +-/* +- * Don't forget that the stack pointer must be aligned on a 8 bytes +- * boundary for 32-bits ABI and 16 bytes for 64-bits ABI. +- */ +-unsigned long arch_align_stack(unsigned long sp) +-{ +- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space) +- sp -= get_random_int() & ~PAGE_MASK; +- +- return sp & ALMASK; +-} +diff --git a/arch/mips/kernel/reset.c b/arch/mips/kernel/reset.c +index 060563a..7fbf310 100644 +--- a/arch/mips/kernel/reset.c ++++ b/arch/mips/kernel/reset.c +@@ -19,8 +19,8 @@ + * So handle all using function pointers to machine specific + * functions. + */ +-void (*_machine_restart)(char *command); +-void (*_machine_halt)(void); ++void (*__noreturn _machine_restart)(char *command); ++void (*__noreturn _machine_halt)(void); + void (*pm_power_off)(void); + + EXPORT_SYMBOL(pm_power_off); +@@ -29,16 +29,19 @@ void machine_restart(char *command) + { + if (_machine_restart) + _machine_restart(command); ++ BUG(); + } + + void machine_halt(void) + { + if (_machine_halt) + _machine_halt(); ++ BUG(); + } + + void machine_power_off(void) + { + if (pm_power_off) + pm_power_off(); ++ BUG(); + } +diff --git a/arch/mips/kernel/syscall.c b/arch/mips/kernel/syscall.c +index 3f7f466..3abe0b5 100644 +--- a/arch/mips/kernel/syscall.c ++++ b/arch/mips/kernel/syscall.c +@@ -102,17 +102,21 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, + do_color_align = 0; + if (filp || (flags & MAP_SHARED)) + do_color_align = 1; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(current->mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + if (addr) { + if (do_color_align) + addr = COLOUR_ALIGN(addr, pgoff); + else + addr = PAGE_ALIGN(addr); + vmm = find_vma(current->mm, addr); +- if (task_size - len >= addr && +- (!vmm || addr + len <= vmm->vm_start)) ++ if (task_size - len >= addr && check_heap_stack_gap(vmm, addr, len)) + return addr; + } +- addr = TASK_UNMAPPED_BASE; ++ addr = current->mm->mmap_base; + if (do_color_align) + addr = COLOUR_ALIGN(addr, pgoff); + else +@@ -122,7 +126,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, + /* At this point: (!vmm || addr < vmm->vm_end). */ + if (task_size - len < addr) + return -ENOMEM; +- if (!vmm || addr + len <= vmm->vm_start) ++ if (check_heap_stack_gap(vmm, addr, len)) + return addr; + addr = vmm->vm_end; + if (do_color_align) +diff --git a/arch/mips/mm/fault.c b/arch/mips/mm/fault.c +index e97a7a2..f18f5b0 100644 +--- a/arch/mips/mm/fault.c ++++ b/arch/mips/mm/fault.c +@@ -26,6 +26,23 @@ + #include + #include /* For VMALLOC_END */ + ++#ifdef CONFIG_PAX_PAGEEXEC ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 5; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + /* + * This routine handles page faults. It determines the address, + * and the problem, and then passes it off to one of the appropriate +diff --git a/arch/mn10300/proc-mn103e010/include/proc/cache.h b/arch/mn10300/proc-mn103e010/include/proc/cache.h +index bdc1f9a..e8de5c5 100644 +--- a/arch/mn10300/proc-mn103e010/include/proc/cache.h ++++ b/arch/mn10300/proc-mn103e010/include/proc/cache.h +@@ -11,12 +11,14 @@ + #ifndef _ASM_PROC_CACHE_H + #define _ASM_PROC_CACHE_H + ++#include ++ + /* L1 cache */ + + #define L1_CACHE_NWAYS 4 /* number of ways in caches */ + #define L1_CACHE_NENTRIES 256 /* number of entries in each way */ +-#define L1_CACHE_BYTES 16 /* bytes per entry */ + #define L1_CACHE_SHIFT 4 /* shift for bytes per entry */ ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) /* bytes per entry */ + #define L1_CACHE_WAYDISP 0x1000 /* displacement of one way from the next */ + + #define L1_CACHE_TAG_VALID 0x00000001 /* cache tag valid bit */ +diff --git a/arch/parisc/include/asm/atomic.h b/arch/parisc/include/asm/atomic.h +index 8bc9e96..26554f8 100644 +--- a/arch/parisc/include/asm/atomic.h ++++ b/arch/parisc/include/asm/atomic.h +@@ -336,6 +336,16 @@ static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) + + #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + #else /* CONFIG_64BIT */ + + #include +diff --git a/arch/parisc/include/asm/cache.h b/arch/parisc/include/asm/cache.h +index 32c2cca..a7b3a64 100644 +--- a/arch/parisc/include/asm/cache.h ++++ b/arch/parisc/include/asm/cache.h +@@ -5,6 +5,7 @@ + #ifndef __ARCH_PARISC_CACHE_H + #define __ARCH_PARISC_CACHE_H + ++#include + + /* + * PA 2.0 processors have 64-byte cachelines; PA 1.1 processors have +@@ -15,13 +16,13 @@ + * just ruin performance. + */ + #ifdef CONFIG_PA20 +-#define L1_CACHE_BYTES 64 + #define L1_CACHE_SHIFT 6 + #else +-#define L1_CACHE_BYTES 32 + #define L1_CACHE_SHIFT 5 + #endif + ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) ++ + #ifndef __ASSEMBLY__ + + #define L1_CACHE_ALIGN(x) (((x)+(L1_CACHE_BYTES-1))&~(L1_CACHE_BYTES-1)) +diff --git a/arch/parisc/include/asm/elf.h b/arch/parisc/include/asm/elf.h +index 9c802eb..0592e41 100644 +--- a/arch/parisc/include/asm/elf.h ++++ b/arch/parisc/include/asm/elf.h +@@ -343,6 +343,13 @@ struct pt_regs; /* forward declaration... */ + + #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x01000000) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE 0x10000UL ++ ++#define PAX_DELTA_MMAP_LEN 16 ++#define PAX_DELTA_STACK_LEN 16 ++#endif ++ + /* This yields a mask that user programs can use to figure out what + instruction set this CPU supports. This could be done in user space, + but it's not easy, and we've already done it here. */ +diff --git a/arch/parisc/include/asm/pgtable.h b/arch/parisc/include/asm/pgtable.h +index a27d2e2..18fd845 100644 +--- a/arch/parisc/include/asm/pgtable.h ++++ b/arch/parisc/include/asm/pgtable.h +@@ -207,6 +207,17 @@ + #define PAGE_EXECREAD __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_EXEC |_PAGE_ACCESSED) + #define PAGE_COPY PAGE_EXECREAD + #define PAGE_RWX __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_WRITE | _PAGE_EXEC |_PAGE_ACCESSED) ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++# define PAGE_SHARED_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_WRITE | _PAGE_ACCESSED) ++# define PAGE_COPY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_ACCESSED) ++# define PAGE_READONLY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_READ | _PAGE_ACCESSED) ++#else ++# define PAGE_SHARED_NOEXEC PAGE_SHARED ++# define PAGE_COPY_NOEXEC PAGE_COPY ++# define PAGE_READONLY_NOEXEC PAGE_READONLY ++#endif ++ + #define PAGE_KERNEL __pgprot(_PAGE_KERNEL) + #define PAGE_KERNEL_RO __pgprot(_PAGE_KERNEL & ~_PAGE_WRITE) + #define PAGE_KERNEL_UNC __pgprot(_PAGE_KERNEL | _PAGE_NO_CACHE) +diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c +index 2120746..8d70a5e 100644 +--- a/arch/parisc/kernel/module.c ++++ b/arch/parisc/kernel/module.c +@@ -95,16 +95,38 @@ + + /* three functions to determine where in the module core + * or init pieces the location is */ ++static inline int in_init_rx(struct module *me, void *loc) ++{ ++ return (loc >= me->module_init_rx && ++ loc < (me->module_init_rx + me->init_size_rx)); ++} ++ ++static inline int in_init_rw(struct module *me, void *loc) ++{ ++ return (loc >= me->module_init_rw && ++ loc < (me->module_init_rw + me->init_size_rw)); ++} ++ + static inline int in_init(struct module *me, void *loc) + { +- return (loc >= me->module_init && +- loc <= (me->module_init + me->init_size)); ++ return in_init_rx(me, loc) || in_init_rw(me, loc); ++} ++ ++static inline int in_core_rx(struct module *me, void *loc) ++{ ++ return (loc >= me->module_core_rx && ++ loc < (me->module_core_rx + me->core_size_rx)); ++} ++ ++static inline int in_core_rw(struct module *me, void *loc) ++{ ++ return (loc >= me->module_core_rw && ++ loc < (me->module_core_rw + me->core_size_rw)); + } + + static inline int in_core(struct module *me, void *loc) + { +- return (loc >= me->module_core && +- loc <= (me->module_core + me->core_size)); ++ return in_core_rx(me, loc) || in_core_rw(me, loc); + } + + static inline int in_local(struct module *me, void *loc) +@@ -364,13 +386,13 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr, + } + + /* align things a bit */ +- me->core_size = ALIGN(me->core_size, 16); +- me->arch.got_offset = me->core_size; +- me->core_size += gots * sizeof(struct got_entry); ++ me->core_size_rw = ALIGN(me->core_size_rw, 16); ++ me->arch.got_offset = me->core_size_rw; ++ me->core_size_rw += gots * sizeof(struct got_entry); + +- me->core_size = ALIGN(me->core_size, 16); +- me->arch.fdesc_offset = me->core_size; +- me->core_size += fdescs * sizeof(Elf_Fdesc); ++ me->core_size_rw = ALIGN(me->core_size_rw, 16); ++ me->arch.fdesc_offset = me->core_size_rw; ++ me->core_size_rw += fdescs * sizeof(Elf_Fdesc); + + me->arch.got_max = gots; + me->arch.fdesc_max = fdescs; +@@ -388,7 +410,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend) + + BUG_ON(value == 0); + +- got = me->module_core + me->arch.got_offset; ++ got = me->module_core_rw + me->arch.got_offset; + for (i = 0; got[i].addr; i++) + if (got[i].addr == value) + goto out; +@@ -406,7 +428,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend) + #ifdef CONFIG_64BIT + static Elf_Addr get_fdesc(struct module *me, unsigned long value) + { +- Elf_Fdesc *fdesc = me->module_core + me->arch.fdesc_offset; ++ Elf_Fdesc *fdesc = me->module_core_rw + me->arch.fdesc_offset; + + if (!value) { + printk(KERN_ERR "%s: zero OPD requested!\n", me->name); +@@ -424,7 +446,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value) + + /* Create new one */ + fdesc->addr = value; +- fdesc->gp = (Elf_Addr)me->module_core + me->arch.got_offset; ++ fdesc->gp = (Elf_Addr)me->module_core_rw + me->arch.got_offset; + return (Elf_Addr)fdesc; + } + #endif /* CONFIG_64BIT */ +@@ -848,7 +870,7 @@ register_unwind_table(struct module *me, + + table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr; + end = table + sechdrs[me->arch.unwind_section].sh_size; +- gp = (Elf_Addr)me->module_core + me->arch.got_offset; ++ gp = (Elf_Addr)me->module_core_rw + me->arch.got_offset; + + DEBUGP("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n", + me->arch.unwind_section, table, end, gp); +diff --git a/arch/parisc/kernel/sys_parisc.c b/arch/parisc/kernel/sys_parisc.c +index 9147391..f3d949a 100644 +--- a/arch/parisc/kernel/sys_parisc.c ++++ b/arch/parisc/kernel/sys_parisc.c +@@ -43,7 +43,7 @@ static unsigned long get_unshared_area(unsigned long addr, unsigned long len) + /* At this point: (!vma || addr < vma->vm_end). */ + if (TASK_SIZE - len < addr) + return -ENOMEM; +- if (!vma || addr + len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + return addr; + addr = vma->vm_end; + } +@@ -79,7 +79,7 @@ static unsigned long get_shared_area(struct address_space *mapping, + /* At this point: (!vma || addr < vma->vm_end). */ + if (TASK_SIZE - len < addr) + return -ENOMEM; +- if (!vma || addr + len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + return addr; + addr = DCACHE_ALIGN(vma->vm_end - offset) + offset; + if (addr < vma->vm_end) /* handle wraparound */ +@@ -98,7 +98,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, + if (flags & MAP_FIXED) + return addr; + if (!addr) +- addr = TASK_UNMAPPED_BASE; ++ addr = current->mm->mmap_base; + + if (filp) { + addr = get_shared_area(filp->f_mapping, addr, len, pgoff); +diff --git a/arch/parisc/kernel/traps.c b/arch/parisc/kernel/traps.c +index 8b58bf0..7afff03 100644 +--- a/arch/parisc/kernel/traps.c ++++ b/arch/parisc/kernel/traps.c +@@ -733,9 +733,7 @@ void notrace handle_interruption(int code, struct pt_regs *regs) + + down_read(¤t->mm->mmap_sem); + vma = find_vma(current->mm,regs->iaoq[0]); +- if (vma && (regs->iaoq[0] >= vma->vm_start) +- && (vma->vm_flags & VM_EXEC)) { +- ++ if (vma && (regs->iaoq[0] >= vma->vm_start)) { + fault_address = regs->iaoq[0]; + fault_space = regs->iasq[0]; + +diff --git a/arch/parisc/mm/fault.c b/arch/parisc/mm/fault.c +index c6afbfc..c5839f6 100644 +--- a/arch/parisc/mm/fault.c ++++ b/arch/parisc/mm/fault.c +@@ -15,6 +15,7 @@ + #include + #include + #include ++#include + + #include + #include +@@ -52,7 +53,7 @@ DEFINE_PER_CPU(struct exception_data, exception_data); + static unsigned long + parisc_acctyp(unsigned long code, unsigned int inst) + { +- if (code == 6 || code == 16) ++ if (code == 6 || code == 7 || code == 16) + return VM_EXEC; + + switch (inst & 0xf0000000) { +@@ -138,6 +139,116 @@ parisc_acctyp(unsigned long code, unsigned int inst) + } + #endif + ++#ifdef CONFIG_PAX_PAGEEXEC ++/* ++ * PaX: decide what to do with offenders (instruction_pointer(regs) = fault address) ++ * ++ * returns 1 when task should be killed ++ * 2 when rt_sigreturn trampoline was detected ++ * 3 when unpatched PLT trampoline was detected ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ ++#ifdef CONFIG_PAX_EMUPLT ++ int err; ++ ++ do { /* PaX: unpatched PLT emulation */ ++ unsigned int bl, depwi; ++ ++ err = get_user(bl, (unsigned int *)instruction_pointer(regs)); ++ err |= get_user(depwi, (unsigned int *)(instruction_pointer(regs)+4)); ++ ++ if (err) ++ break; ++ ++ if (bl == 0xEA9F1FDDU && depwi == 0xD6801C1EU) { ++ unsigned int ldw, bv, ldw2, addr = instruction_pointer(regs)-12; ++ ++ err = get_user(ldw, (unsigned int *)addr); ++ err |= get_user(bv, (unsigned int *)(addr+4)); ++ err |= get_user(ldw2, (unsigned int *)(addr+8)); ++ ++ if (err) ++ break; ++ ++ if (ldw == 0x0E801096U && ++ bv == 0xEAC0C000U && ++ ldw2 == 0x0E881095U) ++ { ++ unsigned int resolver, map; ++ ++ err = get_user(resolver, (unsigned int *)(instruction_pointer(regs)+8)); ++ err |= get_user(map, (unsigned int *)(instruction_pointer(regs)+12)); ++ if (err) ++ break; ++ ++ regs->gr[20] = instruction_pointer(regs)+8; ++ regs->gr[21] = map; ++ regs->gr[22] = resolver; ++ regs->iaoq[0] = resolver | 3UL; ++ regs->iaoq[1] = regs->iaoq[0] + 4; ++ return 3; ++ } ++ } ++ } while (0); ++#endif ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++ ++#ifndef CONFIG_PAX_EMUSIGRT ++ if (!(current->mm->pax_flags & MF_PAX_EMUTRAMP)) ++ return 1; ++#endif ++ ++ do { /* PaX: rt_sigreturn emulation */ ++ unsigned int ldi1, ldi2, bel, nop; ++ ++ err = get_user(ldi1, (unsigned int *)instruction_pointer(regs)); ++ err |= get_user(ldi2, (unsigned int *)(instruction_pointer(regs)+4)); ++ err |= get_user(bel, (unsigned int *)(instruction_pointer(regs)+8)); ++ err |= get_user(nop, (unsigned int *)(instruction_pointer(regs)+12)); ++ ++ if (err) ++ break; ++ ++ if ((ldi1 == 0x34190000U || ldi1 == 0x34190002U) && ++ ldi2 == 0x3414015AU && ++ bel == 0xE4008200U && ++ nop == 0x08000240U) ++ { ++ regs->gr[25] = (ldi1 & 2) >> 1; ++ regs->gr[20] = __NR_rt_sigreturn; ++ regs->gr[31] = regs->iaoq[1] + 16; ++ regs->sr[0] = regs->iasq[1]; ++ regs->iaoq[0] = 0x100UL; ++ regs->iaoq[1] = regs->iaoq[0] + 4; ++ regs->iasq[0] = regs->sr[2]; ++ regs->iasq[1] = regs->sr[2]; ++ return 2; ++ } ++ } while (0); ++#endif ++ ++ return 1; ++} ++ ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 5; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + int fixup_exception(struct pt_regs *regs) + { + const struct exception_table_entry *fix; +@@ -192,8 +303,33 @@ good_area: + + acc_type = parisc_acctyp(code,regs->iir); + +- if ((vma->vm_flags & acc_type) != acc_type) ++ if ((vma->vm_flags & acc_type) != acc_type) { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && (acc_type & VM_EXEC) && ++ (address & ~3UL) == instruction_pointer(regs)) ++ { ++ up_read(&mm->mmap_sem); ++ switch (pax_handle_fetch_fault(regs)) { ++ ++#ifdef CONFIG_PAX_EMUPLT ++ case 3: ++ return; ++#endif ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++ case 2: ++ return; ++#endif ++ ++ } ++ pax_report_fault(regs, (void *)instruction_pointer(regs), (void *)regs->gr[30]); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + goto bad_area; ++ } + + /* + * If for any reason at all we couldn't handle the fault, make +diff --git a/arch/powerpc/Makefile b/arch/powerpc/Makefile +index c107b74..409dc0f 100644 +--- a/arch/powerpc/Makefile ++++ b/arch/powerpc/Makefile +@@ -74,6 +74,8 @@ KBUILD_AFLAGS += -Iarch/$(ARCH) + KBUILD_CFLAGS += -msoft-float -pipe -Iarch/$(ARCH) $(CFLAGS-y) + CPP = $(CC) -E $(KBUILD_CFLAGS) + ++cflags-y += -Wno-sign-compare -Wno-extra ++ + CHECKFLAGS += -m$(CONFIG_WORD_SIZE) -D__powerpc__ -D__powerpc$(CONFIG_WORD_SIZE)__ + + ifeq ($(CONFIG_PPC64),y) +diff --git a/arch/powerpc/include/asm/cache.h b/arch/powerpc/include/asm/cache.h +index 81de6eb..d5d0e24 100644 +--- a/arch/powerpc/include/asm/cache.h ++++ b/arch/powerpc/include/asm/cache.h +@@ -3,6 +3,7 @@ + + #ifdef __KERNEL__ + ++#include + + /* bytes per L1 cache line */ + #if defined(CONFIG_8xx) || defined(CONFIG_403GCX) +@@ -18,7 +19,7 @@ + #define L1_CACHE_SHIFT 7 + #endif + +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define SMP_CACHE_BYTES L1_CACHE_BYTES + +diff --git a/arch/powerpc/include/asm/device.h b/arch/powerpc/include/asm/device.h +index 6d94d27..50d4cad 100644 +--- a/arch/powerpc/include/asm/device.h ++++ b/arch/powerpc/include/asm/device.h +@@ -14,7 +14,7 @@ struct dev_archdata { + struct device_node *of_node; + + /* DMA operations on that device */ +- struct dma_map_ops *dma_ops; ++ const struct dma_map_ops *dma_ops; + + /* + * When an iommu is in use, dma_data is used as a ptr to the base of the +diff --git a/arch/powerpc/include/asm/dma-mapping.h b/arch/powerpc/include/asm/dma-mapping.h +index e281dae..2b8a784 100644 +--- a/arch/powerpc/include/asm/dma-mapping.h ++++ b/arch/powerpc/include/asm/dma-mapping.h +@@ -69,9 +69,9 @@ static inline unsigned long device_to_mask(struct device *dev) + #ifdef CONFIG_PPC64 + extern struct dma_map_ops dma_iommu_ops; + #endif +-extern struct dma_map_ops dma_direct_ops; ++extern const struct dma_map_ops dma_direct_ops; + +-static inline struct dma_map_ops *get_dma_ops(struct device *dev) ++static inline const struct dma_map_ops *get_dma_ops(struct device *dev) + { + /* We don't handle the NULL dev case for ISA for now. We could + * do it via an out of line call but it is not needed for now. The +@@ -84,7 +84,7 @@ static inline struct dma_map_ops *get_dma_ops(struct device *dev) + return dev->archdata.dma_ops; + } + +-static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops) ++static inline void set_dma_ops(struct device *dev, const struct dma_map_ops *ops) + { + dev->archdata.dma_ops = ops; + } +@@ -118,7 +118,7 @@ static inline void set_dma_offset(struct device *dev, dma_addr_t off) + + static inline int dma_supported(struct device *dev, u64 mask) + { +- struct dma_map_ops *dma_ops = get_dma_ops(dev); ++ const struct dma_map_ops *dma_ops = get_dma_ops(dev); + + if (unlikely(dma_ops == NULL)) + return 0; +@@ -132,7 +132,7 @@ static inline int dma_supported(struct device *dev, u64 mask) + + static inline int dma_set_mask(struct device *dev, u64 dma_mask) + { +- struct dma_map_ops *dma_ops = get_dma_ops(dev); ++ const struct dma_map_ops *dma_ops = get_dma_ops(dev); + + if (unlikely(dma_ops == NULL)) + return -EIO; +@@ -147,7 +147,7 @@ static inline int dma_set_mask(struct device *dev, u64 dma_mask) + static inline void *dma_alloc_coherent(struct device *dev, size_t size, + dma_addr_t *dma_handle, gfp_t flag) + { +- struct dma_map_ops *dma_ops = get_dma_ops(dev); ++ const struct dma_map_ops *dma_ops = get_dma_ops(dev); + void *cpu_addr; + + BUG_ON(!dma_ops); +@@ -162,7 +162,7 @@ static inline void *dma_alloc_coherent(struct device *dev, size_t size, + static inline void dma_free_coherent(struct device *dev, size_t size, + void *cpu_addr, dma_addr_t dma_handle) + { +- struct dma_map_ops *dma_ops = get_dma_ops(dev); ++ const struct dma_map_ops *dma_ops = get_dma_ops(dev); + + BUG_ON(!dma_ops); + +@@ -173,7 +173,7 @@ static inline void dma_free_coherent(struct device *dev, size_t size, + + static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) + { +- struct dma_map_ops *dma_ops = get_dma_ops(dev); ++ const struct dma_map_ops *dma_ops = get_dma_ops(dev); + + if (dma_ops->mapping_error) + return dma_ops->mapping_error(dev, dma_addr); +diff --git a/arch/powerpc/include/asm/elf.h b/arch/powerpc/include/asm/elf.h +index 5698502..5db093c 100644 +--- a/arch/powerpc/include/asm/elf.h ++++ b/arch/powerpc/include/asm/elf.h +@@ -179,8 +179,19 @@ typedef elf_fpreg_t elf_vsrreghalf_t32[ELF_NVSRHALFREG]; + the loader. We need to make sure that it is out of the way of the program + that it will "exec", and that there is sufficient room for the brk. */ + +-extern unsigned long randomize_et_dyn(unsigned long base); +-#define ELF_ET_DYN_BASE (randomize_et_dyn(0x20000000)) ++#define ELF_ET_DYN_BASE (0x20000000) ++ ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (0x10000000UL) ++ ++#ifdef __powerpc64__ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT) ? 16 : 28) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT) ? 16 : 28) ++#else ++#define PAX_DELTA_MMAP_LEN 15 ++#define PAX_DELTA_STACK_LEN 15 ++#endif ++#endif + + /* + * Our registers are always unsigned longs, whether we're a 32 bit +@@ -275,9 +286,6 @@ extern int arch_setup_additional_pages(struct linux_binprm *bprm, + (0x7ff >> (PAGE_SHIFT - 12)) : \ + (0x3ffff >> (PAGE_SHIFT - 12))) + +-extern unsigned long arch_randomize_brk(struct mm_struct *mm); +-#define arch_randomize_brk arch_randomize_brk +- + #endif /* __KERNEL__ */ + + /* +diff --git a/arch/powerpc/include/asm/iommu.h b/arch/powerpc/include/asm/iommu.h +index edfc980..1766f59 100644 +--- a/arch/powerpc/include/asm/iommu.h ++++ b/arch/powerpc/include/asm/iommu.h +@@ -116,6 +116,9 @@ extern void iommu_init_early_iSeries(void); + extern void iommu_init_early_dart(void); + extern void iommu_init_early_pasemi(void); + ++/* dma-iommu.c */ ++extern int dma_iommu_dma_supported(struct device *dev, u64 mask); ++ + #ifdef CONFIG_PCI + extern void pci_iommu_init(void); + extern void pci_direct_iommu_init(void); +diff --git a/arch/powerpc/include/asm/kmap_types.h b/arch/powerpc/include/asm/kmap_types.h +index 9163695..5a00112 100644 +--- a/arch/powerpc/include/asm/kmap_types.h ++++ b/arch/powerpc/include/asm/kmap_types.h +@@ -26,6 +26,7 @@ enum km_type { + KM_SOFTIRQ1, + KM_PPC_SYNC_PAGE, + KM_PPC_SYNC_ICACHE, ++ KM_CLEARPAGE, + KM_TYPE_NR + }; + +diff --git a/arch/powerpc/include/asm/page.h b/arch/powerpc/include/asm/page.h +index ff24254..fe45b21 100644 +--- a/arch/powerpc/include/asm/page.h ++++ b/arch/powerpc/include/asm/page.h +@@ -116,8 +116,9 @@ extern phys_addr_t kernstart_addr; + * and needs to be executable. This means the whole heap ends + * up being executable. + */ +-#define VM_DATA_DEFAULT_FLAGS32 (VM_READ | VM_WRITE | VM_EXEC | \ +- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) ++#define VM_DATA_DEFAULT_FLAGS32 \ ++ (((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0) | \ ++ VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) + + #define VM_DATA_DEFAULT_FLAGS64 (VM_READ | VM_WRITE | \ + VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) +@@ -145,6 +146,9 @@ extern phys_addr_t kernstart_addr; + #define is_kernel_addr(x) ((x) >= PAGE_OFFSET) + #endif + ++#define ktla_ktva(addr) (addr) ++#define ktva_ktla(addr) (addr) ++ + #ifndef __ASSEMBLY__ + + #undef STRICT_MM_TYPECHECKS +diff --git a/arch/powerpc/include/asm/page_64.h b/arch/powerpc/include/asm/page_64.h +index 3f17b83..1f9e766 100644 +--- a/arch/powerpc/include/asm/page_64.h ++++ b/arch/powerpc/include/asm/page_64.h +@@ -180,15 +180,18 @@ do { \ + * stack by default, so in the absense of a PT_GNU_STACK program header + * we turn execute permission off. + */ +-#define VM_STACK_DEFAULT_FLAGS32 (VM_READ | VM_WRITE | VM_EXEC | \ +- VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) ++#define VM_STACK_DEFAULT_FLAGS32 \ ++ (((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0) | \ ++ VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) + + #define VM_STACK_DEFAULT_FLAGS64 (VM_READ | VM_WRITE | \ + VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC) + ++#ifndef CONFIG_PAX_PAGEEXEC + #define VM_STACK_DEFAULT_FLAGS \ + (test_thread_flag(TIF_32BIT) ? \ + VM_STACK_DEFAULT_FLAGS32 : VM_STACK_DEFAULT_FLAGS64) ++#endif + + #include + +diff --git a/arch/powerpc/include/asm/pci.h b/arch/powerpc/include/asm/pci.h +index b5ea626..40308222 100644 +--- a/arch/powerpc/include/asm/pci.h ++++ b/arch/powerpc/include/asm/pci.h +@@ -65,8 +65,8 @@ static inline int pci_get_legacy_ide_irq(struct pci_dev *dev, int channel) + } + + #ifdef CONFIG_PCI +-extern void set_pci_dma_ops(struct dma_map_ops *dma_ops); +-extern struct dma_map_ops *get_pci_dma_ops(void); ++extern void set_pci_dma_ops(const struct dma_map_ops *dma_ops); ++extern const struct dma_map_ops *get_pci_dma_ops(void); + #else /* CONFIG_PCI */ + #define set_pci_dma_ops(d) + #define get_pci_dma_ops() NULL +diff --git a/arch/powerpc/include/asm/pgtable.h b/arch/powerpc/include/asm/pgtable.h +index 2a5da06..d65bea2 100644 +--- a/arch/powerpc/include/asm/pgtable.h ++++ b/arch/powerpc/include/asm/pgtable.h +@@ -2,6 +2,7 @@ + #define _ASM_POWERPC_PGTABLE_H + #ifdef __KERNEL__ + ++#include + #ifndef __ASSEMBLY__ + #include /* For TASK_SIZE */ + #include +diff --git a/arch/powerpc/include/asm/pte-hash32.h b/arch/powerpc/include/asm/pte-hash32.h +index 4aad413..85d86bf 100644 +--- a/arch/powerpc/include/asm/pte-hash32.h ++++ b/arch/powerpc/include/asm/pte-hash32.h +@@ -21,6 +21,7 @@ + #define _PAGE_FILE 0x004 /* when !present: nonlinear file mapping */ + #define _PAGE_USER 0x004 /* usermode access allowed */ + #define _PAGE_GUARDED 0x008 /* G: prohibit speculative access */ ++#define _PAGE_EXEC _PAGE_GUARDED + #define _PAGE_COHERENT 0x010 /* M: enforce memory coherence (SMP systems) */ + #define _PAGE_NO_CACHE 0x020 /* I: cache inhibit */ + #define _PAGE_WRITETHRU 0x040 /* W: cache write-through */ +diff --git a/arch/powerpc/include/asm/ptrace.h b/arch/powerpc/include/asm/ptrace.h +index 8c34149..78f425a 100644 +--- a/arch/powerpc/include/asm/ptrace.h ++++ b/arch/powerpc/include/asm/ptrace.h +@@ -103,7 +103,7 @@ extern unsigned long profile_pc(struct pt_regs *regs); + } while(0) + + struct task_struct; +-extern unsigned long ptrace_get_reg(struct task_struct *task, int regno); ++extern unsigned long ptrace_get_reg(struct task_struct *task, unsigned int regno); + extern int ptrace_put_reg(struct task_struct *task, int regno, + unsigned long data); + +diff --git a/arch/powerpc/include/asm/reg.h b/arch/powerpc/include/asm/reg.h +index 32a7c30..be3a8bb 100644 +--- a/arch/powerpc/include/asm/reg.h ++++ b/arch/powerpc/include/asm/reg.h +@@ -191,6 +191,7 @@ + #define SPRN_DBCR 0x136 /* e300 Data Breakpoint Control Reg */ + #define SPRN_DSISR 0x012 /* Data Storage Interrupt Status Register */ + #define DSISR_NOHPTE 0x40000000 /* no translation found */ ++#define DSISR_GUARDED 0x10000000 /* fetch from guarded storage */ + #define DSISR_PROTFAULT 0x08000000 /* protection fault */ + #define DSISR_ISSTORE 0x02000000 /* access was a store */ + #define DSISR_DABRMATCH 0x00400000 /* hit data breakpoint */ +diff --git a/arch/powerpc/include/asm/swiotlb.h b/arch/powerpc/include/asm/swiotlb.h +index 8979d4c..d2fd0d3 100644 +--- a/arch/powerpc/include/asm/swiotlb.h ++++ b/arch/powerpc/include/asm/swiotlb.h +@@ -13,7 +13,7 @@ + + #include + +-extern struct dma_map_ops swiotlb_dma_ops; ++extern const struct dma_map_ops swiotlb_dma_ops; + + static inline void dma_mark_clean(void *addr, size_t size) {} + +diff --git a/arch/powerpc/include/asm/system.h b/arch/powerpc/include/asm/system.h +index 094a12a..877a60a 100644 +--- a/arch/powerpc/include/asm/system.h ++++ b/arch/powerpc/include/asm/system.h +@@ -531,7 +531,7 @@ __cmpxchg_local(volatile void *ptr, unsigned long old, unsigned long new, + #define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n)) + #endif + +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) ((x) & ~0xfUL) + + /* Used in very early kernel initialization. */ + extern unsigned long reloc_offset(void); +diff --git a/arch/powerpc/include/asm/uaccess.h b/arch/powerpc/include/asm/uaccess.h +index bd0fb84..a42a14b 100644 +--- a/arch/powerpc/include/asm/uaccess.h ++++ b/arch/powerpc/include/asm/uaccess.h +@@ -13,6 +13,8 @@ + #define VERIFY_READ 0 + #define VERIFY_WRITE 1 + ++extern void check_object_size(const void *ptr, unsigned long n, bool to); ++ + /* + * The fs value determines whether argument validity checking should be + * performed or not. If get_fs() == USER_DS, checking is performed, with +@@ -327,52 +329,6 @@ do { \ + extern unsigned long __copy_tofrom_user(void __user *to, + const void __user *from, unsigned long size); + +-#ifndef __powerpc64__ +- +-static inline unsigned long copy_from_user(void *to, +- const void __user *from, unsigned long n) +-{ +- unsigned long over; +- +- if (access_ok(VERIFY_READ, from, n)) +- return __copy_tofrom_user((__force void __user *)to, from, n); +- if ((unsigned long)from < TASK_SIZE) { +- over = (unsigned long)from + n - TASK_SIZE; +- return __copy_tofrom_user((__force void __user *)to, from, +- n - over) + over; +- } +- return n; +-} +- +-static inline unsigned long copy_to_user(void __user *to, +- const void *from, unsigned long n) +-{ +- unsigned long over; +- +- if (access_ok(VERIFY_WRITE, to, n)) +- return __copy_tofrom_user(to, (__force void __user *)from, n); +- if ((unsigned long)to < TASK_SIZE) { +- over = (unsigned long)to + n - TASK_SIZE; +- return __copy_tofrom_user(to, (__force void __user *)from, +- n - over) + over; +- } +- return n; +-} +- +-#else /* __powerpc64__ */ +- +-#define __copy_in_user(to, from, size) \ +- __copy_tofrom_user((to), (from), (size)) +- +-extern unsigned long copy_from_user(void *to, const void __user *from, +- unsigned long n); +-extern unsigned long copy_to_user(void __user *to, const void *from, +- unsigned long n); +-extern unsigned long copy_in_user(void __user *to, const void __user *from, +- unsigned long n); +- +-#endif /* __powerpc64__ */ +- + static inline unsigned long __copy_from_user_inatomic(void *to, + const void __user *from, unsigned long n) + { +@@ -396,6 +352,10 @@ static inline unsigned long __copy_from_user_inatomic(void *to, + if (ret == 0) + return 0; + } ++ ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); ++ + return __copy_tofrom_user((__force void __user *)to, from, n); + } + +@@ -422,6 +382,10 @@ static inline unsigned long __copy_to_user_inatomic(void __user *to, + if (ret == 0) + return 0; + } ++ ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); ++ + return __copy_tofrom_user(to, (__force const void __user *)from, n); + } + +@@ -439,6 +403,92 @@ static inline unsigned long __copy_to_user(void __user *to, + return __copy_to_user_inatomic(to, from, size); + } + ++#ifndef __powerpc64__ ++ ++static inline unsigned long __must_check copy_from_user(void *to, ++ const void __user *from, unsigned long n) ++{ ++ unsigned long over; ++ ++ if ((long)n < 0) ++ return n; ++ ++ if (access_ok(VERIFY_READ, from, n)) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); ++ return __copy_tofrom_user((__force void __user *)to, from, n); ++ } ++ if ((unsigned long)from < TASK_SIZE) { ++ over = (unsigned long)from + n - TASK_SIZE; ++ if (!__builtin_constant_p(n - over)) ++ check_object_size(to, n - over, false); ++ return __copy_tofrom_user((__force void __user *)to, from, ++ n - over) + over; ++ } ++ return n; ++} ++ ++static inline unsigned long __must_check copy_to_user(void __user *to, ++ const void *from, unsigned long n) ++{ ++ unsigned long over; ++ ++ if ((long)n < 0) ++ return n; ++ ++ if (access_ok(VERIFY_WRITE, to, n)) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); ++ return __copy_tofrom_user(to, (__force void __user *)from, n); ++ } ++ if ((unsigned long)to < TASK_SIZE) { ++ over = (unsigned long)to + n - TASK_SIZE; ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n - over, true); ++ return __copy_tofrom_user(to, (__force void __user *)from, ++ n - over) + over; ++ } ++ return n; ++} ++ ++#else /* __powerpc64__ */ ++ ++#define __copy_in_user(to, from, size) \ ++ __copy_tofrom_user((to), (from), (size)) ++ ++static inline unsigned long __must_check copy_from_user(void *to, const void __user *from, unsigned long n) ++{ ++ if ((long)n < 0 || n > INT_MAX) ++ return n; ++ ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); ++ ++ if (likely(access_ok(VERIFY_READ, from, n))) ++ n = __copy_from_user(to, from, n); ++ else ++ memset(to, 0, n); ++ return n; ++} ++ ++static inline unsigned long __must_check copy_to_user(void __user *to, const void *from, unsigned long n) ++{ ++ if ((long)n < 0 || n > INT_MAX) ++ return n; ++ ++ if (likely(access_ok(VERIFY_WRITE, to, n))) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); ++ n = __copy_to_user(to, from, n); ++ } ++ return n; ++} ++ ++extern unsigned long copy_in_user(void __user *to, const void __user *from, ++ unsigned long n); ++ ++#endif /* __powerpc64__ */ ++ + extern unsigned long __clear_user(void __user *addr, unsigned long size); + + static inline unsigned long clear_user(void __user *addr, unsigned long size) +diff --git a/arch/powerpc/kernel/cacheinfo.c b/arch/powerpc/kernel/cacheinfo.c +index bb37b1d..01fe9ce 100644 +--- a/arch/powerpc/kernel/cacheinfo.c ++++ b/arch/powerpc/kernel/cacheinfo.c +@@ -642,7 +642,7 @@ static struct kobj_attribute *cache_index_opt_attrs[] = { + &cache_assoc_attr, + }; + +-static struct sysfs_ops cache_index_ops = { ++static const struct sysfs_ops cache_index_ops = { + .show = cache_index_show, + }; + +diff --git a/arch/powerpc/kernel/dma-iommu.c b/arch/powerpc/kernel/dma-iommu.c +index 37771a5..648530c 100644 +--- a/arch/powerpc/kernel/dma-iommu.c ++++ b/arch/powerpc/kernel/dma-iommu.c +@@ -70,7 +70,7 @@ static void dma_iommu_unmap_sg(struct device *dev, struct scatterlist *sglist, + } + + /* We support DMA to/from any memory page via the iommu */ +-static int dma_iommu_dma_supported(struct device *dev, u64 mask) ++int dma_iommu_dma_supported(struct device *dev, u64 mask) + { + struct iommu_table *tbl = get_iommu_table_base(dev); + +diff --git a/arch/powerpc/kernel/dma-swiotlb.c b/arch/powerpc/kernel/dma-swiotlb.c +index e96cbbd..bdd6d41 100644 +--- a/arch/powerpc/kernel/dma-swiotlb.c ++++ b/arch/powerpc/kernel/dma-swiotlb.c +@@ -31,7 +31,7 @@ unsigned int ppc_swiotlb_enable; + * map_page, and unmap_page on highmem, use normal dma_ops + * for everything else. + */ +-struct dma_map_ops swiotlb_dma_ops = { ++const struct dma_map_ops swiotlb_dma_ops = { + .alloc_coherent = dma_direct_alloc_coherent, + .free_coherent = dma_direct_free_coherent, + .map_sg = swiotlb_map_sg_attrs, +diff --git a/arch/powerpc/kernel/dma.c b/arch/powerpc/kernel/dma.c +index 6215062..ebea59c 100644 +--- a/arch/powerpc/kernel/dma.c ++++ b/arch/powerpc/kernel/dma.c +@@ -134,7 +134,7 @@ static inline void dma_direct_sync_single_range(struct device *dev, + } + #endif + +-struct dma_map_ops dma_direct_ops = { ++const struct dma_map_ops dma_direct_ops = { + .alloc_coherent = dma_direct_alloc_coherent, + .free_coherent = dma_direct_free_coherent, + .map_sg = dma_direct_map_sg, +diff --git a/arch/powerpc/kernel/exceptions-64e.S b/arch/powerpc/kernel/exceptions-64e.S +index 24dcc0e..a300455 100644 +--- a/arch/powerpc/kernel/exceptions-64e.S ++++ b/arch/powerpc/kernel/exceptions-64e.S +@@ -455,6 +455,7 @@ storage_fault_common: + std r14,_DAR(r1) + std r15,_DSISR(r1) + addi r3,r1,STACK_FRAME_OVERHEAD ++ bl .save_nvgprs + mr r4,r14 + mr r5,r15 + ld r14,PACA_EXGEN+EX_R14(r13) +@@ -464,8 +465,7 @@ storage_fault_common: + cmpdi r3,0 + bne- 1f + b .ret_from_except_lite +-1: bl .save_nvgprs +- mr r5,r3 ++1: mr r5,r3 + addi r3,r1,STACK_FRAME_OVERHEAD + ld r4,_DAR(r1) + bl .bad_page_fault +diff --git a/arch/powerpc/kernel/exceptions-64s.S b/arch/powerpc/kernel/exceptions-64s.S +index 1808876..9fd206a 100644 +--- a/arch/powerpc/kernel/exceptions-64s.S ++++ b/arch/powerpc/kernel/exceptions-64s.S +@@ -818,10 +818,10 @@ handle_page_fault: + 11: ld r4,_DAR(r1) + ld r5,_DSISR(r1) + addi r3,r1,STACK_FRAME_OVERHEAD ++ bl .save_nvgprs + bl .do_page_fault + cmpdi r3,0 + beq+ 13f +- bl .save_nvgprs + mr r5,r3 + addi r3,r1,STACK_FRAME_OVERHEAD + lwz r4,_DAR(r1) +diff --git a/arch/powerpc/kernel/ibmebus.c b/arch/powerpc/kernel/ibmebus.c +index a4c8b38..1b09ad9 100644 +--- a/arch/powerpc/kernel/ibmebus.c ++++ b/arch/powerpc/kernel/ibmebus.c +@@ -127,7 +127,7 @@ static int ibmebus_dma_supported(struct device *dev, u64 mask) + return 1; + } + +-static struct dma_map_ops ibmebus_dma_ops = { ++static const struct dma_map_ops ibmebus_dma_ops = { + .alloc_coherent = ibmebus_alloc_coherent, + .free_coherent = ibmebus_free_coherent, + .map_sg = ibmebus_map_sg, +diff --git a/arch/powerpc/kernel/irq.c b/arch/powerpc/kernel/irq.c +index 8564a41..67f3471 100644 +--- a/arch/powerpc/kernel/irq.c ++++ b/arch/powerpc/kernel/irq.c +@@ -490,9 +490,6 @@ struct irq_host *irq_alloc_host(struct device_node *of_node, + host->ops = ops; + host->of_node = of_node_get(of_node); + +- if (host->ops->match == NULL) +- host->ops->match = default_irq_host_match; +- + spin_lock_irqsave(&irq_big_lock, flags); + + /* If it's a legacy controller, check for duplicates and +@@ -567,7 +564,12 @@ struct irq_host *irq_find_host(struct device_node *node) + */ + spin_lock_irqsave(&irq_big_lock, flags); + list_for_each_entry(h, &irq_hosts, link) +- if (h->ops->match(h, node)) { ++ if (h->ops->match) { ++ if (h->ops->match(h, node)) { ++ found = h; ++ break; ++ } ++ } else if (default_irq_host_match(h, node)) { + found = h; + break; + } +diff --git a/arch/powerpc/kernel/kgdb.c b/arch/powerpc/kernel/kgdb.c +index 641c74b..8339ad7 100644 +--- a/arch/powerpc/kernel/kgdb.c ++++ b/arch/powerpc/kernel/kgdb.c +@@ -126,7 +126,7 @@ static int kgdb_handle_breakpoint(struct pt_regs *regs) + if (kgdb_handle_exception(0, SIGTRAP, 0, regs) != 0) + return 0; + +- if (*(u32 *) (regs->nip) == *(u32 *) (&arch_kgdb_ops.gdb_bpt_instr)) ++ if (*(u32 *) (regs->nip) == *(const u32 *) (&arch_kgdb_ops.gdb_bpt_instr)) + regs->nip += 4; + + return 1; +@@ -353,7 +353,7 @@ int kgdb_arch_handle_exception(int vector, int signo, int err_code, + /* + * Global data + */ +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + .gdb_bpt_instr = {0x7d, 0x82, 0x10, 0x08}, + }; + +diff --git a/arch/powerpc/kernel/module.c b/arch/powerpc/kernel/module.c +index 477c663..4f50234 100644 +--- a/arch/powerpc/kernel/module.c ++++ b/arch/powerpc/kernel/module.c +@@ -31,11 +31,24 @@ + + LIST_HEAD(module_bug_list); + ++#ifdef CONFIG_PAX_KERNEXEC + void *module_alloc(unsigned long size) + { + if (size == 0) + return NULL; + ++ return vmalloc(size); ++} ++ ++void *module_alloc_exec(unsigned long size) ++#else ++void *module_alloc(unsigned long size) ++#endif ++ ++{ ++ if (size == 0) ++ return NULL; ++ + return vmalloc_exec(size); + } + +@@ -45,6 +58,13 @@ void module_free(struct module *mod, void *module_region) + vfree(module_region); + } + ++#ifdef CONFIG_PAX_KERNEXEC ++void module_free_exec(struct module *mod, void *module_region) ++{ ++ module_free(mod, module_region); ++} ++#endif ++ + static const Elf_Shdr *find_section(const Elf_Ehdr *hdr, + const Elf_Shdr *sechdrs, + const char *name) +diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c +index f832773..0507238 100644 +--- a/arch/powerpc/kernel/module_32.c ++++ b/arch/powerpc/kernel/module_32.c +@@ -162,7 +162,7 @@ int module_frob_arch_sections(Elf32_Ehdr *hdr, + me->arch.core_plt_section = i; + } + if (!me->arch.core_plt_section || !me->arch.init_plt_section) { +- printk("Module doesn't contain .plt or .init.plt sections.\n"); ++ printk("Module %s doesn't contain .plt or .init.plt sections.\n", me->name); + return -ENOEXEC; + } + +@@ -203,11 +203,16 @@ static uint32_t do_plt_call(void *location, + + DEBUGP("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location); + /* Init, or core PLT? */ +- if (location >= mod->module_core +- && location < mod->module_core + mod->core_size) ++ if ((location >= mod->module_core_rx && location < mod->module_core_rx + mod->core_size_rx) || ++ (location >= mod->module_core_rw && location < mod->module_core_rw + mod->core_size_rw)) + entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr; +- else ++ else if ((location >= mod->module_init_rx && location < mod->module_init_rx + mod->init_size_rx) || ++ (location >= mod->module_init_rw && location < mod->module_init_rw + mod->init_size_rw)) + entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr; ++ else { ++ printk(KERN_ERR "%s: invalid R_PPC_REL24 entry found\n", mod->name); ++ return ~0UL; ++ } + + /* Find this entry, or if that fails, the next avail. entry */ + while (entry->jump[0]) { +diff --git a/arch/powerpc/kernel/pci-common.c b/arch/powerpc/kernel/pci-common.c +index cadbed6..b9bbb00 100644 +--- a/arch/powerpc/kernel/pci-common.c ++++ b/arch/powerpc/kernel/pci-common.c +@@ -50,14 +50,14 @@ resource_size_t isa_mem_base; + unsigned int ppc_pci_flags = 0; + + +-static struct dma_map_ops *pci_dma_ops = &dma_direct_ops; ++static const struct dma_map_ops *pci_dma_ops = &dma_direct_ops; + +-void set_pci_dma_ops(struct dma_map_ops *dma_ops) ++void set_pci_dma_ops(const struct dma_map_ops *dma_ops) + { + pci_dma_ops = dma_ops; + } + +-struct dma_map_ops *get_pci_dma_ops(void) ++const struct dma_map_ops *get_pci_dma_ops(void) + { + return pci_dma_ops; + } +diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c +index 7b816da..8d5c277 100644 +--- a/arch/powerpc/kernel/process.c ++++ b/arch/powerpc/kernel/process.c +@@ -539,8 +539,8 @@ void show_regs(struct pt_regs * regs) + * Lookup NIP late so we have the best change of getting the + * above info out without failing + */ +- printk("NIP ["REG"] %pS\n", regs->nip, (void *)regs->nip); +- printk("LR ["REG"] %pS\n", regs->link, (void *)regs->link); ++ printk("NIP ["REG"] %pA\n", regs->nip, (void *)regs->nip); ++ printk("LR ["REG"] %pA\n", regs->link, (void *)regs->link); + #endif + show_stack(current, (unsigned long *) regs->gpr[1]); + if (!user_mode(regs)) +@@ -1034,10 +1034,10 @@ void show_stack(struct task_struct *tsk, unsigned long *stack) + newsp = stack[0]; + ip = stack[STACK_FRAME_LR_SAVE]; + if (!firstframe || ip != lr) { +- printk("["REG"] ["REG"] %pS", sp, ip, (void *)ip); ++ printk("["REG"] ["REG"] %pA", sp, ip, (void *)ip); + #ifdef CONFIG_FUNCTION_GRAPH_TRACER + if ((ip == rth || ip == mrth) && curr_frame >= 0) { +- printk(" (%pS)", ++ printk(" (%pA)", + (void *)current->ret_stack[curr_frame].ret); + curr_frame--; + } +@@ -1057,7 +1057,7 @@ void show_stack(struct task_struct *tsk, unsigned long *stack) + struct pt_regs *regs = (struct pt_regs *) + (sp + STACK_FRAME_OVERHEAD); + lr = regs->link; +- printk("--- Exception: %lx at %pS\n LR = %pS\n", ++ printk("--- Exception: %lx at %pA\n LR = %pA\n", + regs->trap, (void *)regs->nip, (void *)lr); + firstframe = 1; + } +@@ -1134,58 +1134,3 @@ void thread_info_cache_init(void) + } + + #endif /* THREAD_SHIFT < PAGE_SHIFT */ +- +-unsigned long arch_align_stack(unsigned long sp) +-{ +- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space) +- sp -= get_random_int() & ~PAGE_MASK; +- return sp & ~0xf; +-} +- +-static inline unsigned long brk_rnd(void) +-{ +- unsigned long rnd = 0; +- +- /* 8MB for 32bit, 1GB for 64bit */ +- if (is_32bit_task()) +- rnd = (long)(get_random_int() % (1<<(23-PAGE_SHIFT))); +- else +- rnd = (long)(get_random_int() % (1<<(30-PAGE_SHIFT))); +- +- return rnd << PAGE_SHIFT; +-} +- +-unsigned long arch_randomize_brk(struct mm_struct *mm) +-{ +- unsigned long base = mm->brk; +- unsigned long ret; +- +-#ifdef CONFIG_PPC_STD_MMU_64 +- /* +- * If we are using 1TB segments and we are allowed to randomise +- * the heap, we can put it above 1TB so it is backed by a 1TB +- * segment. Otherwise the heap will be in the bottom 1TB +- * which always uses 256MB segments and this may result in a +- * performance penalty. +- */ +- if (!is_32bit_task() && (mmu_highuser_ssize == MMU_SEGSIZE_1T)) +- base = max_t(unsigned long, mm->brk, 1UL << SID_SHIFT_1T); +-#endif +- +- ret = PAGE_ALIGN(base + brk_rnd()); +- +- if (ret < mm->brk) +- return mm->brk; +- +- return ret; +-} +- +-unsigned long randomize_et_dyn(unsigned long base) +-{ +- unsigned long ret = PAGE_ALIGN(base + brk_rnd()); +- +- if (ret < base) +- return base; +- +- return ret; +-} +diff --git a/arch/powerpc/kernel/ptrace.c b/arch/powerpc/kernel/ptrace.c +index ef14988..856c4bc 100644 +--- a/arch/powerpc/kernel/ptrace.c ++++ b/arch/powerpc/kernel/ptrace.c +@@ -86,7 +86,7 @@ static int set_user_trap(struct task_struct *task, unsigned long trap) + /* + * Get contents of register REGNO in task TASK. + */ +-unsigned long ptrace_get_reg(struct task_struct *task, int regno) ++unsigned long ptrace_get_reg(struct task_struct *task, unsigned int regno) + { + if (task->thread.regs == NULL) + return -EIO; +@@ -894,7 +894,7 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) + + CHECK_FULL_REGS(child->thread.regs); + if (index < PT_FPR0) { +- tmp = ptrace_get_reg(child, (int) index); ++ tmp = ptrace_get_reg(child, index); + } else { + flush_fp_to_thread(child); + tmp = ((unsigned long *)child->thread.fpr) +diff --git a/arch/powerpc/kernel/signal_32.c b/arch/powerpc/kernel/signal_32.c +index d670429..2bc59b2 100644 +--- a/arch/powerpc/kernel/signal_32.c ++++ b/arch/powerpc/kernel/signal_32.c +@@ -857,7 +857,7 @@ int handle_rt_signal32(unsigned long sig, struct k_sigaction *ka, + /* Save user registers on the stack */ + frame = &rt_sf->uc.uc_mcontext; + addr = frame; +- if (vdso32_rt_sigtramp && current->mm->context.vdso_base) { ++ if (vdso32_rt_sigtramp && current->mm->context.vdso_base != ~0UL) { + if (save_user_regs(regs, frame, 0, 1)) + goto badframe; + regs->link = current->mm->context.vdso_base + vdso32_rt_sigtramp; +diff --git a/arch/powerpc/kernel/signal_64.c b/arch/powerpc/kernel/signal_64.c +index 2fe6fc6..ada0d96 100644 +--- a/arch/powerpc/kernel/signal_64.c ++++ b/arch/powerpc/kernel/signal_64.c +@@ -429,7 +429,7 @@ int handle_rt_signal64(int signr, struct k_sigaction *ka, siginfo_t *info, + current->thread.fpscr.val = 0; + + /* Set up to return from userspace. */ +- if (vdso64_rt_sigtramp && current->mm->context.vdso_base) { ++ if (vdso64_rt_sigtramp && current->mm->context.vdso_base != ~0UL) { + regs->link = current->mm->context.vdso_base + vdso64_rt_sigtramp; + } else { + err |= setup_trampoline(__NR_rt_sigreturn, &frame->tramp[0]); +diff --git a/arch/powerpc/kernel/sys_ppc32.c b/arch/powerpc/kernel/sys_ppc32.c +index b97c2d6..dd01a6a 100644 +--- a/arch/powerpc/kernel/sys_ppc32.c ++++ b/arch/powerpc/kernel/sys_ppc32.c +@@ -563,10 +563,10 @@ asmlinkage long compat_sys_sysctl(struct __sysctl_args32 __user *args) + if (oldlenp) { + if (!error) { + if (get_user(oldlen, oldlenp) || +- put_user(oldlen, (compat_size_t __user *)compat_ptr(tmp.oldlenp))) ++ put_user(oldlen, (compat_size_t __user *)compat_ptr(tmp.oldlenp)) || ++ copy_to_user(args->__unused, tmp.__unused, sizeof(tmp.__unused))) + error = -EFAULT; + } +- copy_to_user(args->__unused, tmp.__unused, sizeof(tmp.__unused)); + } + return error; + } +diff --git a/arch/powerpc/kernel/traps.c b/arch/powerpc/kernel/traps.c +index 6f0ae1a..e4b6a56 100644 +--- a/arch/powerpc/kernel/traps.c ++++ b/arch/powerpc/kernel/traps.c +@@ -99,6 +99,8 @@ static void pmac_backlight_unblank(void) + static inline void pmac_backlight_unblank(void) { } + #endif + ++extern void gr_handle_kernel_exploit(void); ++ + int die(const char *str, struct pt_regs *regs, long err) + { + static struct { +@@ -168,6 +170,8 @@ int die(const char *str, struct pt_regs *regs, long err) + if (panic_on_oops) + panic("Fatal exception"); + ++ gr_handle_kernel_exploit(); ++ + oops_exit(); + do_exit(err); + +diff --git a/arch/powerpc/kernel/vdso.c b/arch/powerpc/kernel/vdso.c +index 137dc22..fe57a79 100644 +--- a/arch/powerpc/kernel/vdso.c ++++ b/arch/powerpc/kernel/vdso.c +@@ -36,6 +36,7 @@ + #include + #include + #include ++#include + + #include "setup.h" + +@@ -220,7 +221,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + vdso_base = VDSO32_MBASE; + #endif + +- current->mm->context.vdso_base = 0; ++ current->mm->context.vdso_base = ~0UL; + + /* vDSO has a problem and was disabled, just don't "enable" it for the + * process +@@ -240,7 +241,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + vdso_base = get_unmapped_area(NULL, vdso_base, + (vdso_pages << PAGE_SHIFT) + + ((VDSO_ALIGNMENT - 1) & PAGE_MASK), +- 0, 0); ++ 0, MAP_PRIVATE | MAP_EXECUTABLE); + if (IS_ERR_VALUE(vdso_base)) { + rc = vdso_base; + goto fail_mmapsem; +diff --git a/arch/powerpc/kernel/vio.c b/arch/powerpc/kernel/vio.c +index 77f6421..829564a 100644 +--- a/arch/powerpc/kernel/vio.c ++++ b/arch/powerpc/kernel/vio.c +@@ -601,11 +601,12 @@ static void vio_dma_iommu_unmap_sg(struct device *dev, + vio_cmo_dealloc(viodev, alloc_size); + } + +-struct dma_map_ops vio_dma_mapping_ops = { ++static const struct dma_map_ops vio_dma_mapping_ops = { + .alloc_coherent = vio_dma_iommu_alloc_coherent, + .free_coherent = vio_dma_iommu_free_coherent, + .map_sg = vio_dma_iommu_map_sg, + .unmap_sg = vio_dma_iommu_unmap_sg, ++ .dma_supported = dma_iommu_dma_supported, + .map_page = vio_dma_iommu_map_page, + .unmap_page = vio_dma_iommu_unmap_page, + +@@ -857,7 +858,6 @@ static void vio_cmo_bus_remove(struct vio_dev *viodev) + + static void vio_cmo_set_dma_ops(struct vio_dev *viodev) + { +- vio_dma_mapping_ops.dma_supported = dma_iommu_ops.dma_supported; + viodev->dev.archdata.dma_ops = &vio_dma_mapping_ops; + } + +diff --git a/arch/powerpc/lib/usercopy_64.c b/arch/powerpc/lib/usercopy_64.c +index 5eea6f3..5d10396 100644 +--- a/arch/powerpc/lib/usercopy_64.c ++++ b/arch/powerpc/lib/usercopy_64.c +@@ -9,22 +9,6 @@ + #include + #include + +-unsigned long copy_from_user(void *to, const void __user *from, unsigned long n) +-{ +- if (likely(access_ok(VERIFY_READ, from, n))) +- n = __copy_from_user(to, from, n); +- else +- memset(to, 0, n); +- return n; +-} +- +-unsigned long copy_to_user(void __user *to, const void *from, unsigned long n) +-{ +- if (likely(access_ok(VERIFY_WRITE, to, n))) +- n = __copy_to_user(to, from, n); +- return n; +-} +- + unsigned long copy_in_user(void __user *to, const void __user *from, + unsigned long n) + { +@@ -35,7 +19,5 @@ unsigned long copy_in_user(void __user *to, const void __user *from, + return n; + } + +-EXPORT_SYMBOL(copy_from_user); +-EXPORT_SYMBOL(copy_to_user); + EXPORT_SYMBOL(copy_in_user); + +diff --git a/arch/powerpc/mm/fault.c b/arch/powerpc/mm/fault.c +index e7dae82..877ce0d 100644 +--- a/arch/powerpc/mm/fault.c ++++ b/arch/powerpc/mm/fault.c +@@ -30,6 +30,10 @@ + #include + #include + #include ++#include ++#include ++#include ++#include + + #include + #include +@@ -40,6 +44,7 @@ + #include + #include + #include ++#include + + + #ifdef CONFIG_KPROBES +@@ -64,6 +69,33 @@ static inline int notify_page_fault(struct pt_regs *regs) + } + #endif + ++#ifdef CONFIG_PAX_PAGEEXEC ++/* ++ * PaX: decide what to do with offenders (regs->nip = fault address) ++ * ++ * returns 1 when task should be killed ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ return 1; ++} ++ ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 5; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int __user *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + /* + * Check whether the instruction at regs->nip is a store using + * an update addressing form which will update r1. +@@ -134,7 +166,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address, + * indicate errors in DSISR but can validly be set in SRR1. + */ + if (trap == 0x400) +- error_code &= 0x48200000; ++ error_code &= 0x58200000; + else + is_write = error_code & DSISR_ISSTORE; + #else +@@ -250,7 +282,7 @@ good_area: + * "undefined". Of those that can be set, this is the only + * one which seems bad. + */ +- if (error_code & 0x10000000) ++ if (error_code & DSISR_GUARDED) + /* Guarded storage error. */ + goto bad_area; + #endif /* CONFIG_8xx */ +@@ -265,7 +297,7 @@ good_area: + * processors use the same I/D cache coherency mechanism + * as embedded. + */ +- if (error_code & DSISR_PROTFAULT) ++ if (error_code & (DSISR_PROTFAULT | DSISR_GUARDED)) + goto bad_area; + #endif /* CONFIG_PPC_STD_MMU */ + +@@ -335,6 +367,23 @@ bad_area: + bad_area_nosemaphore: + /* User mode accesses cause a SIGSEGV */ + if (user_mode(regs)) { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (mm->pax_flags & MF_PAX_PAGEEXEC) { ++#ifdef CONFIG_PPC_STD_MMU ++ if (is_exec && (error_code & (DSISR_PROTFAULT | DSISR_GUARDED))) { ++#else ++ if (is_exec && regs->nip == address) { ++#endif ++ switch (pax_handle_fetch_fault(regs)) { ++ } ++ ++ pax_report_fault(regs, (void *)regs->nip, (void *)regs->gpr[PT_R1]); ++ do_group_exit(SIGKILL); ++ } ++ } ++#endif ++ + _exception(SIGSEGV, regs, code, address); + return 0; + } +diff --git a/arch/powerpc/mm/mem.c b/arch/powerpc/mm/mem.c +index 5973631..ad617af 100644 +--- a/arch/powerpc/mm/mem.c ++++ b/arch/powerpc/mm/mem.c +@@ -250,7 +250,7 @@ static int __init mark_nonram_nosave(void) + { + unsigned long lmb_next_region_start_pfn, + lmb_region_max_pfn; +- int i; ++ unsigned int i; + + for (i = 0; i < lmb.memory.cnt - 1; i++) { + lmb_region_max_pfn = +diff --git a/arch/powerpc/mm/mmap_64.c b/arch/powerpc/mm/mmap_64.c +index 0d957a4..26d968f 100644 +--- a/arch/powerpc/mm/mmap_64.c ++++ b/arch/powerpc/mm/mmap_64.c +@@ -99,10 +99,22 @@ void arch_pick_mmap_layout(struct mm_struct *mm) + */ + if (mmap_is_legacy()) { + mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area; + mm->unmap_area = arch_unmap_area; + } else { + mm->mmap_base = mmap_base(); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base -= mm->delta_mmap + mm->delta_stack; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area_topdown; + mm->unmap_area = arch_unmap_area_topdown; + } +diff --git a/arch/powerpc/mm/slice.c b/arch/powerpc/mm/slice.c +index ba51948..23009d9 100644 +--- a/arch/powerpc/mm/slice.c ++++ b/arch/powerpc/mm/slice.c +@@ -98,7 +98,7 @@ static int slice_area_is_free(struct mm_struct *mm, unsigned long addr, + if ((mm->task_size - len) < addr) + return 0; + vma = find_vma(mm, addr); +- return (!vma || (addr + len) <= vma->vm_start); ++ return check_heap_stack_gap(vma, addr, len); + } + + static int slice_low_has_vma(struct mm_struct *mm, unsigned long slice) +@@ -256,7 +256,7 @@ full_search: + addr = _ALIGN_UP(addr + 1, 1ul << SLICE_HIGH_SHIFT); + continue; + } +- if (!vma || addr + len <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* + * Remember the place where we stopped the search: + */ +@@ -313,10 +313,14 @@ static unsigned long slice_find_area_topdown(struct mm_struct *mm, + } + } + +- addr = mm->mmap_base; +- while (addr > len) { ++ if (mm->mmap_base < len) ++ addr = -ENOMEM; ++ else ++ addr = mm->mmap_base - len; ++ ++ while (!IS_ERR_VALUE(addr)) { + /* Go down by chunk size */ +- addr = _ALIGN_DOWN(addr - len, 1ul << pshift); ++ addr = _ALIGN_DOWN(addr, 1ul << pshift); + + /* Check for hit with different page size */ + mask = slice_range_to_mask(addr, len); +@@ -336,7 +340,7 @@ static unsigned long slice_find_area_topdown(struct mm_struct *mm, + * return with success: + */ + vma = find_vma(mm, addr); +- if (!vma || (addr + len) <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* remember the address as a hint for next time */ + if (use_cache) + mm->free_area_cache = addr; +@@ -348,7 +352,7 @@ static unsigned long slice_find_area_topdown(struct mm_struct *mm, + mm->cached_hole_size = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = vma->vm_start; ++ addr = skip_heap_stack_gap(vma, len); + } + + /* +@@ -426,6 +430,11 @@ unsigned long slice_get_unmapped_area(unsigned long addr, unsigned long len, + if (fixed && addr > (mm->task_size - len)) + return -EINVAL; + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!fixed && (mm->pax_flags & MF_PAX_RANDMMAP)) ++ addr = 0; ++#endif ++ + /* If hint, make sure it matches our alignment restrictions */ + if (!fixed && addr) { + addr = _ALIGN_UP(addr, 1ul << pshift); +diff --git a/arch/powerpc/platforms/52xx/lite5200_pm.c b/arch/powerpc/platforms/52xx/lite5200_pm.c +index b5c753d..8f01abe 100644 +--- a/arch/powerpc/platforms/52xx/lite5200_pm.c ++++ b/arch/powerpc/platforms/52xx/lite5200_pm.c +@@ -235,7 +235,7 @@ static void lite5200_pm_end(void) + lite5200_pm_target_state = PM_SUSPEND_ON; + } + +-static struct platform_suspend_ops lite5200_pm_ops = { ++static const struct platform_suspend_ops lite5200_pm_ops = { + .valid = lite5200_pm_valid, + .begin = lite5200_pm_begin, + .prepare = lite5200_pm_prepare, +diff --git a/arch/powerpc/platforms/52xx/mpc52xx_pm.c b/arch/powerpc/platforms/52xx/mpc52xx_pm.c +index a55b0b6..478c18e 100644 +--- a/arch/powerpc/platforms/52xx/mpc52xx_pm.c ++++ b/arch/powerpc/platforms/52xx/mpc52xx_pm.c +@@ -180,7 +180,7 @@ void mpc52xx_pm_finish(void) + iounmap(mbar); + } + +-static struct platform_suspend_ops mpc52xx_pm_ops = { ++static const struct platform_suspend_ops mpc52xx_pm_ops = { + .valid = mpc52xx_pm_valid, + .prepare = mpc52xx_pm_prepare, + .enter = mpc52xx_pm_enter, +diff --git a/arch/powerpc/platforms/83xx/suspend.c b/arch/powerpc/platforms/83xx/suspend.c +index 08e65fc..643d3ac 100644 +--- a/arch/powerpc/platforms/83xx/suspend.c ++++ b/arch/powerpc/platforms/83xx/suspend.c +@@ -273,7 +273,7 @@ static int mpc83xx_is_pci_agent(void) + return ret; + } + +-static struct platform_suspend_ops mpc83xx_suspend_ops = { ++static const struct platform_suspend_ops mpc83xx_suspend_ops = { + .valid = mpc83xx_suspend_valid, + .begin = mpc83xx_suspend_begin, + .enter = mpc83xx_suspend_enter, +diff --git a/arch/powerpc/platforms/cell/iommu.c b/arch/powerpc/platforms/cell/iommu.c +index ca5bfdf..1602e09 100644 +--- a/arch/powerpc/platforms/cell/iommu.c ++++ b/arch/powerpc/platforms/cell/iommu.c +@@ -642,7 +642,7 @@ static int dma_fixed_dma_supported(struct device *dev, u64 mask) + + static int dma_set_mask_and_switch(struct device *dev, u64 dma_mask); + +-struct dma_map_ops dma_iommu_fixed_ops = { ++const struct dma_map_ops dma_iommu_fixed_ops = { + .alloc_coherent = dma_fixed_alloc_coherent, + .free_coherent = dma_fixed_free_coherent, + .map_sg = dma_fixed_map_sg, +diff --git a/arch/powerpc/platforms/ps3/system-bus.c b/arch/powerpc/platforms/ps3/system-bus.c +index e34b305..20e48ec 100644 +--- a/arch/powerpc/platforms/ps3/system-bus.c ++++ b/arch/powerpc/platforms/ps3/system-bus.c +@@ -694,7 +694,7 @@ static int ps3_dma_supported(struct device *_dev, u64 mask) + return mask >= DMA_BIT_MASK(32); + } + +-static struct dma_map_ops ps3_sb_dma_ops = { ++static const struct dma_map_ops ps3_sb_dma_ops = { + .alloc_coherent = ps3_alloc_coherent, + .free_coherent = ps3_free_coherent, + .map_sg = ps3_sb_map_sg, +@@ -704,7 +704,7 @@ static struct dma_map_ops ps3_sb_dma_ops = { + .unmap_page = ps3_unmap_page, + }; + +-static struct dma_map_ops ps3_ioc0_dma_ops = { ++static const struct dma_map_ops ps3_ioc0_dma_ops = { + .alloc_coherent = ps3_alloc_coherent, + .free_coherent = ps3_free_coherent, + .map_sg = ps3_ioc0_map_sg, +diff --git a/arch/powerpc/platforms/pseries/Kconfig b/arch/powerpc/platforms/pseries/Kconfig +index f0e6f28..60d53ed 100644 +--- a/arch/powerpc/platforms/pseries/Kconfig ++++ b/arch/powerpc/platforms/pseries/Kconfig +@@ -2,6 +2,8 @@ config PPC_PSERIES + depends on PPC64 && PPC_BOOK3S + bool "IBM pSeries & new (POWER5-based) iSeries" + select MPIC ++ select PCI_MSI ++ select XICS + select PPC_I8259 + select PPC_RTAS + select RTAS_ERROR_LOGGING +diff --git a/arch/s390/Kconfig b/arch/s390/Kconfig +index 43c0aca..42c045b 100644 +--- a/arch/s390/Kconfig ++++ b/arch/s390/Kconfig +@@ -194,28 +194,26 @@ config AUDIT_ARCH + + config S390_SWITCH_AMODE + bool "Switch kernel/user addressing modes" ++ default y + help + This option allows to switch the addressing modes of kernel and user +- space. The kernel parameter switch_amode=on will enable this feature, +- default is disabled. Enabling this (via kernel parameter) on machines +- earlier than IBM System z9-109 EC/BC will reduce system performance. ++ space. Enabling this on machines earlier than IBM System z9-109 EC/BC ++ will reduce system performance. + + Note that this option will also be selected by selecting the execute +- protection option below. Enabling the execute protection via the +- noexec kernel parameter will also switch the addressing modes, +- independent of the switch_amode kernel parameter. ++ protection option below. Enabling the execute protection will also ++ switch the addressing modes, independent of this option. + + + config S390_EXEC_PROTECT + bool "Data execute protection" ++ default y + select S390_SWITCH_AMODE + help + This option allows to enable a buffer overflow protection for user + space programs and it also selects the addressing mode option above. +- The kernel parameter noexec=on will enable this feature and also +- switch the addressing modes, default is disabled. Enabling this (via +- kernel parameter) on machines earlier than IBM System z9-109 EC/BC +- will reduce system performance. ++ Enabling this on machines earlier than IBM System z9-109 EC/BC will ++ reduce system performance. + + comment "Code generation options" + +diff --git a/arch/s390/include/asm/atomic.h b/arch/s390/include/asm/atomic.h +index ae7c8f9..3f01a0c 100644 +--- a/arch/s390/include/asm/atomic.h ++++ b/arch/s390/include/asm/atomic.h +@@ -362,6 +362,16 @@ static inline int atomic64_add_unless(atomic64_t *v, long long a, long long u) + #define atomic64_dec_and_test(_v) (atomic64_sub_return(1, _v) == 0) + #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) + ++#define atomic64_read_unchecked(v) atomic64_read(v) ++#define atomic64_set_unchecked(v, i) atomic64_set((v), (i)) ++#define atomic64_add_unchecked(a, v) atomic64_add((a), (v)) ++#define atomic64_add_return_unchecked(a, v) atomic64_add_return((a), (v)) ++#define atomic64_sub_unchecked(a, v) atomic64_sub((a), (v)) ++#define atomic64_inc_unchecked(v) atomic64_inc(v) ++#define atomic64_inc_return_unchecked(v) atomic64_inc_return(v) ++#define atomic64_dec_unchecked(v) atomic64_dec(v) ++#define atomic64_cmpxchg_unchecked(v, o, n) atomic64_cmpxchg((v), (o), (n)) ++ + #define smp_mb__before_atomic_dec() smp_mb() + #define smp_mb__after_atomic_dec() smp_mb() + #define smp_mb__before_atomic_inc() smp_mb() +diff --git a/arch/s390/include/asm/cache.h b/arch/s390/include/asm/cache.h +index 9b86681..c5140db 100644 +--- a/arch/s390/include/asm/cache.h ++++ b/arch/s390/include/asm/cache.h +@@ -11,8 +11,10 @@ + #ifndef __ARCH_S390_CACHE_H + #define __ARCH_S390_CACHE_H + +-#define L1_CACHE_BYTES 256 ++#include ++ + #define L1_CACHE_SHIFT 8 ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define __read_mostly __attribute__((__section__(".data.read_mostly"))) + +diff --git a/arch/s390/include/asm/elf.h b/arch/s390/include/asm/elf.h +index e885442..e3a2817 100644 +--- a/arch/s390/include/asm/elf.h ++++ b/arch/s390/include/asm/elf.h +@@ -164,6 +164,13 @@ extern unsigned int vdso_enabled; + that it will "exec", and that there is sufficient room for the brk. */ + #define ELF_ET_DYN_BASE (STACK_TOP / 3 * 2) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_31BIT) ? 0x10000UL : 0x80000000UL) ++ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_31BIT) ? 15 : 26) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_31BIT) ? 15 : 26) ++#endif ++ + /* This yields a mask that user programs can use to figure out what + instruction set this CPU supports. */ + +diff --git a/arch/s390/include/asm/setup.h b/arch/s390/include/asm/setup.h +index e37478e..9ce0e9f 100644 +--- a/arch/s390/include/asm/setup.h ++++ b/arch/s390/include/asm/setup.h +@@ -50,13 +50,13 @@ extern unsigned long memory_end; + void detect_memory_layout(struct mem_chunk chunk[]); + + #ifdef CONFIG_S390_SWITCH_AMODE +-extern unsigned int switch_amode; ++#define switch_amode (1) + #else + #define switch_amode (0) + #endif + + #ifdef CONFIG_S390_EXEC_PROTECT +-extern unsigned int s390_noexec; ++#define s390_noexec (1) + #else + #define s390_noexec (0) + #endif +diff --git a/arch/s390/include/asm/uaccess.h b/arch/s390/include/asm/uaccess.h +index 8377e91..e28e6f1 100644 +--- a/arch/s390/include/asm/uaccess.h ++++ b/arch/s390/include/asm/uaccess.h +@@ -232,6 +232,10 @@ static inline unsigned long __must_check + copy_to_user(void __user *to, const void *from, unsigned long n) + { + might_fault(); ++ ++ if ((long)n < 0) ++ return n; ++ + if (access_ok(VERIFY_WRITE, to, n)) + n = __copy_to_user(to, from, n); + return n; +@@ -257,6 +261,9 @@ copy_to_user(void __user *to, const void *from, unsigned long n) + static inline unsigned long __must_check + __copy_from_user(void *to, const void __user *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + if (__builtin_constant_p(n) && (n <= 256)) + return uaccess.copy_from_user_small(n, from, to); + else +@@ -283,6 +290,10 @@ static inline unsigned long __must_check + copy_from_user(void *to, const void __user *from, unsigned long n) + { + might_fault(); ++ ++ if ((long)n < 0) ++ return n; ++ + if (access_ok(VERIFY_READ, from, n)) + n = __copy_from_user(to, from, n); + else +diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c +index 639380a..72e3c02 100644 +--- a/arch/s390/kernel/module.c ++++ b/arch/s390/kernel/module.c +@@ -166,11 +166,11 @@ module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs, + + /* Increase core size by size of got & plt and set start + offsets for got and plt. */ +- me->core_size = ALIGN(me->core_size, 4); +- me->arch.got_offset = me->core_size; +- me->core_size += me->arch.got_size; +- me->arch.plt_offset = me->core_size; +- me->core_size += me->arch.plt_size; ++ me->core_size_rw = ALIGN(me->core_size_rw, 4); ++ me->arch.got_offset = me->core_size_rw; ++ me->core_size_rw += me->arch.got_size; ++ me->arch.plt_offset = me->core_size_rx; ++ me->core_size_rx += me->arch.plt_size; + return 0; + } + +@@ -256,7 +256,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + if (info->got_initialized == 0) { + Elf_Addr *gotent; + +- gotent = me->module_core + me->arch.got_offset + ++ gotent = me->module_core_rw + me->arch.got_offset + + info->got_offset; + *gotent = val; + info->got_initialized = 1; +@@ -280,7 +280,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + else if (r_type == R_390_GOTENT || + r_type == R_390_GOTPLTENT) + *(unsigned int *) loc = +- (val + (Elf_Addr) me->module_core - loc) >> 1; ++ (val + (Elf_Addr) me->module_core_rw - loc) >> 1; + else if (r_type == R_390_GOT64 || + r_type == R_390_GOTPLT64) + *(unsigned long *) loc = val; +@@ -294,7 +294,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + case R_390_PLTOFF64: /* 16 bit offset from GOT to PLT. */ + if (info->plt_initialized == 0) { + unsigned int *ip; +- ip = me->module_core + me->arch.plt_offset + ++ ip = me->module_core_rx + me->arch.plt_offset + + info->plt_offset; + #ifndef CONFIG_64BIT + ip[0] = 0x0d105810; /* basr 1,0; l 1,6(1); br 1 */ +@@ -319,7 +319,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + val - loc + 0xffffUL < 0x1ffffeUL) || + (r_type == R_390_PLT32DBL && + val - loc + 0xffffffffULL < 0x1fffffffeULL))) +- val = (Elf_Addr) me->module_core + ++ val = (Elf_Addr) me->module_core_rx + + me->arch.plt_offset + + info->plt_offset; + val += rela->r_addend - loc; +@@ -341,7 +341,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + case R_390_GOTOFF32: /* 32 bit offset to GOT. */ + case R_390_GOTOFF64: /* 64 bit offset to GOT. */ + val = val + rela->r_addend - +- ((Elf_Addr) me->module_core + me->arch.got_offset); ++ ((Elf_Addr) me->module_core_rw + me->arch.got_offset); + if (r_type == R_390_GOTOFF16) + *(unsigned short *) loc = val; + else if (r_type == R_390_GOTOFF32) +@@ -351,7 +351,7 @@ apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab, + break; + case R_390_GOTPC: /* 32 bit PC relative offset to GOT. */ + case R_390_GOTPCDBL: /* 32 bit PC rel. off. to GOT shifted by 1. */ +- val = (Elf_Addr) me->module_core + me->arch.got_offset + ++ val = (Elf_Addr) me->module_core_rw + me->arch.got_offset + + rela->r_addend - loc; + if (r_type == R_390_GOTPC) + *(unsigned int *) loc = val; +diff --git a/arch/s390/kernel/setup.c b/arch/s390/kernel/setup.c +index 0b2573a..71a22ec 100644 +--- a/arch/s390/kernel/setup.c ++++ b/arch/s390/kernel/setup.c +@@ -306,9 +306,6 @@ static int __init early_parse_mem(char *p) + early_param("mem", early_parse_mem); + + #ifdef CONFIG_S390_SWITCH_AMODE +-unsigned int switch_amode = 0; +-EXPORT_SYMBOL_GPL(switch_amode); +- + static int set_amode_and_uaccess(unsigned long user_amode, + unsigned long user32_amode) + { +@@ -334,17 +331,6 @@ static int set_amode_and_uaccess(unsigned long user_amode, + return 0; + } + } +- +-/* +- * Switch kernel/user addressing modes? +- */ +-static int __init early_parse_switch_amode(char *p) +-{ +- switch_amode = 1; +- return 0; +-} +-early_param("switch_amode", early_parse_switch_amode); +- + #else /* CONFIG_S390_SWITCH_AMODE */ + static inline int set_amode_and_uaccess(unsigned long user_amode, + unsigned long user32_amode) +@@ -353,24 +339,6 @@ static inline int set_amode_and_uaccess(unsigned long user_amode, + } + #endif /* CONFIG_S390_SWITCH_AMODE */ + +-#ifdef CONFIG_S390_EXEC_PROTECT +-unsigned int s390_noexec = 0; +-EXPORT_SYMBOL_GPL(s390_noexec); +- +-/* +- * Enable execute protection? +- */ +-static int __init early_parse_noexec(char *p) +-{ +- if (!strncmp(p, "off", 3)) +- return 0; +- switch_amode = 1; +- s390_noexec = 1; +- return 0; +-} +-early_param("noexec", early_parse_noexec); +-#endif /* CONFIG_S390_EXEC_PROTECT */ +- + static void setup_addressing_mode(void) + { + if (s390_noexec) { +diff --git a/arch/s390/mm/mmap.c b/arch/s390/mm/mmap.c +index 0ab74ae..c8b68f9 100644 +--- a/arch/s390/mm/mmap.c ++++ b/arch/s390/mm/mmap.c +@@ -78,10 +78,22 @@ void arch_pick_mmap_layout(struct mm_struct *mm) + */ + if (mmap_is_legacy()) { + mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area; + mm->unmap_area = arch_unmap_area; + } else { + mm->mmap_base = mmap_base(); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base -= mm->delta_mmap + mm->delta_stack; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area_topdown; + mm->unmap_area = arch_unmap_area_topdown; + } +@@ -153,10 +165,22 @@ void arch_pick_mmap_layout(struct mm_struct *mm) + */ + if (mmap_is_legacy()) { + mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ + mm->get_unmapped_area = s390_get_unmapped_area; + mm->unmap_area = arch_unmap_area; + } else { + mm->mmap_base = mmap_base(); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base -= mm->delta_mmap + mm->delta_stack; ++#endif ++ + mm->get_unmapped_area = s390_get_unmapped_area_topdown; + mm->unmap_area = arch_unmap_area_topdown; + } +diff --git a/arch/score/include/asm/cache.h b/arch/score/include/asm/cache.h +index ae3d59f..f65f075 100644 +--- a/arch/score/include/asm/cache.h ++++ b/arch/score/include/asm/cache.h +@@ -1,7 +1,9 @@ + #ifndef _ASM_SCORE_CACHE_H + #define _ASM_SCORE_CACHE_H + ++#include ++ + #define L1_CACHE_SHIFT 4 +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #endif /* _ASM_SCORE_CACHE_H */ +diff --git a/arch/score/include/asm/system.h b/arch/score/include/asm/system.h +index 589d5c7..669e274 100644 +--- a/arch/score/include/asm/system.h ++++ b/arch/score/include/asm/system.h +@@ -17,7 +17,7 @@ do { \ + #define finish_arch_switch(prev) do {} while (0) + + typedef void (*vi_handler_t)(void); +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) (x) + + #define mb() barrier() + #define rmb() barrier() +diff --git a/arch/score/kernel/process.c b/arch/score/kernel/process.c +index 25d0803..d6c8e36 100644 +--- a/arch/score/kernel/process.c ++++ b/arch/score/kernel/process.c +@@ -161,8 +161,3 @@ unsigned long get_wchan(struct task_struct *task) + + return task_pt_regs(task)->cp0_epc; + } +- +-unsigned long arch_align_stack(unsigned long sp) +-{ +- return sp; +-} +diff --git a/arch/sh/boards/mach-hp6xx/pm.c b/arch/sh/boards/mach-hp6xx/pm.c +index d936c1a..304a252 100644 +--- a/arch/sh/boards/mach-hp6xx/pm.c ++++ b/arch/sh/boards/mach-hp6xx/pm.c +@@ -143,7 +143,7 @@ static int hp6x0_pm_enter(suspend_state_t state) + return 0; + } + +-static struct platform_suspend_ops hp6x0_pm_ops = { ++static const struct platform_suspend_ops hp6x0_pm_ops = { + .enter = hp6x0_pm_enter, + .valid = suspend_valid_only_mem, + }; +diff --git a/arch/sh/include/asm/cache.h b/arch/sh/include/asm/cache.h +index 02df18e..ae3a793 100644 +--- a/arch/sh/include/asm/cache.h ++++ b/arch/sh/include/asm/cache.h +@@ -9,10 +9,11 @@ + #define __ASM_SH_CACHE_H + #ifdef __KERNEL__ + ++#include + #include + #include + +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define __read_mostly __attribute__((__section__(".data.read_mostly"))) + +diff --git a/arch/sh/kernel/cpu/sh4/sq.c b/arch/sh/kernel/cpu/sh4/sq.c +index 8a8a993..7b3079b 100644 +--- a/arch/sh/kernel/cpu/sh4/sq.c ++++ b/arch/sh/kernel/cpu/sh4/sq.c +@@ -327,7 +327,7 @@ static struct attribute *sq_sysfs_attrs[] = { + NULL, + }; + +-static struct sysfs_ops sq_sysfs_ops = { ++static const struct sysfs_ops sq_sysfs_ops = { + .show = sq_sysfs_show, + .store = sq_sysfs_store, + }; +diff --git a/arch/sh/kernel/cpu/shmobile/pm.c b/arch/sh/kernel/cpu/shmobile/pm.c +index ee3c2aa..c49cee6 100644 +--- a/arch/sh/kernel/cpu/shmobile/pm.c ++++ b/arch/sh/kernel/cpu/shmobile/pm.c +@@ -58,7 +58,7 @@ static int sh_pm_enter(suspend_state_t state) + return 0; + } + +-static struct platform_suspend_ops sh_pm_ops = { ++static const struct platform_suspend_ops sh_pm_ops = { + .enter = sh_pm_enter, + .valid = suspend_valid_only_mem, + }; +diff --git a/arch/sh/kernel/kgdb.c b/arch/sh/kernel/kgdb.c +index 3e532d0..9faa306 100644 +--- a/arch/sh/kernel/kgdb.c ++++ b/arch/sh/kernel/kgdb.c +@@ -271,7 +271,7 @@ void kgdb_arch_exit(void) + { + } + +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + /* Breakpoint instruction: trapa #0x3c */ + #ifdef CONFIG_CPU_LITTLE_ENDIAN + .gdb_bpt_instr = { 0x3c, 0xc3 }, +diff --git a/arch/sh/mm/mmap.c b/arch/sh/mm/mmap.c +index afeb710..d1d1289 100644 +--- a/arch/sh/mm/mmap.c ++++ b/arch/sh/mm/mmap.c +@@ -74,8 +74,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, + addr = PAGE_ALIGN(addr); + + vma = find_vma(mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + +@@ -106,7 +105,7 @@ full_search: + } + return -ENOMEM; + } +- if (likely(!vma || addr + len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* + * Remember the place where we stopped the search: + */ +@@ -157,8 +156,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + addr = PAGE_ALIGN(addr); + + vma = find_vma(mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (TASK_SIZE - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + +@@ -179,7 +177,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + /* make sure it can fit in the remaining address space */ + if (likely(addr > len)) { + vma = find_vma(mm, addr-len); +- if (!vma || addr <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr - len, len)) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr-len); + } +@@ -188,18 +186,18 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + if (unlikely(mm->mmap_base < len)) + goto bottomup; + +- addr = mm->mmap_base-len; +- if (do_colour_align) +- addr = COLOUR_ALIGN_DOWN(addr, pgoff); ++ addr = mm->mmap_base - len; + + do { ++ if (do_colour_align) ++ addr = COLOUR_ALIGN_DOWN(addr, pgoff); + /* + * Lookup failure means no vma is above this address, + * else if new region fits below vma->vm_start, + * return with success: + */ + vma = find_vma(mm, addr); +- if (likely(!vma || addr+len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr); + } +@@ -209,10 +207,8 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + mm->cached_hole_size = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = vma->vm_start-len; +- if (do_colour_align) +- addr = COLOUR_ALIGN_DOWN(addr, pgoff); +- } while (likely(len < vma->vm_start)); ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); + + bottomup: + /* +diff --git a/arch/sparc/Kconfig b/arch/sparc/Kconfig +index 05ef538..dc9c857 100644 +--- a/arch/sparc/Kconfig ++++ b/arch/sparc/Kconfig +@@ -32,6 +32,7 @@ config SPARC + + config SPARC32 + def_bool !64BIT ++ select GENERIC_ATOMIC64 + + config SPARC64 + def_bool 64BIT +diff --git a/arch/sparc/Makefile b/arch/sparc/Makefile +index 113225b..7fd04e7 100644 +--- a/arch/sparc/Makefile ++++ b/arch/sparc/Makefile +@@ -75,7 +75,7 @@ drivers-$(CONFIG_OPROFILE) += arch/sparc/oprofile/ + # Export what is needed by arch/sparc/boot/Makefile + export VMLINUX_INIT VMLINUX_MAIN + VMLINUX_INIT := $(head-y) $(init-y) +-VMLINUX_MAIN := $(core-y) kernel/ mm/ fs/ ipc/ security/ crypto/ block/ ++VMLINUX_MAIN := $(core-y) kernel/ mm/ fs/ ipc/ security/ crypto/ block/ grsecurity/ + VMLINUX_MAIN += $(patsubst %/, %/lib.a, $(libs-y)) $(libs-y) + VMLINUX_MAIN += $(drivers-y) $(net-y) + +diff --git a/arch/sparc/include/asm/atomic_32.h b/arch/sparc/include/asm/atomic_32.h +index f0d343c..cf36e68 100644 +--- a/arch/sparc/include/asm/atomic_32.h ++++ b/arch/sparc/include/asm/atomic_32.h +@@ -13,6 +13,8 @@ + + #include + ++#include ++ + #ifdef __KERNEL__ + + #include +diff --git a/arch/sparc/include/asm/atomic_64.h b/arch/sparc/include/asm/atomic_64.h +index f5cc06f..f858d47 100644 +--- a/arch/sparc/include/asm/atomic_64.h ++++ b/arch/sparc/include/asm/atomic_64.h +@@ -14,18 +14,40 @@ + #define ATOMIC64_INIT(i) { (i) } + + #define atomic_read(v) ((v)->counter) ++static inline int atomic_read_unchecked(const atomic_unchecked_t *v) ++{ ++ return v->counter; ++} + #define atomic64_read(v) ((v)->counter) ++static inline long atomic64_read_unchecked(const atomic64_unchecked_t *v) ++{ ++ return v->counter; ++} + + #define atomic_set(v, i) (((v)->counter) = i) ++static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i) ++{ ++ v->counter = i; ++} + #define atomic64_set(v, i) (((v)->counter) = i) ++static inline void atomic64_set_unchecked(atomic64_unchecked_t *v, long i) ++{ ++ v->counter = i; ++} + + extern void atomic_add(int, atomic_t *); ++extern void atomic_add_unchecked(int, atomic_unchecked_t *); + extern void atomic64_add(long, atomic64_t *); ++extern void atomic64_add_unchecked(long, atomic64_unchecked_t *); + extern void atomic_sub(int, atomic_t *); ++extern void atomic_sub_unchecked(int, atomic_unchecked_t *); + extern void atomic64_sub(long, atomic64_t *); ++extern void atomic64_sub_unchecked(long, atomic64_unchecked_t *); + + extern int atomic_add_ret(int, atomic_t *); ++extern int atomic_add_ret_unchecked(int, atomic_unchecked_t *); + extern long atomic64_add_ret(long, atomic64_t *); ++extern long atomic64_add_ret_unchecked(long, atomic64_unchecked_t *); + extern int atomic_sub_ret(int, atomic_t *); + extern long atomic64_sub_ret(long, atomic64_t *); + +@@ -33,13 +55,29 @@ extern long atomic64_sub_ret(long, atomic64_t *); + #define atomic64_dec_return(v) atomic64_sub_ret(1, v) + + #define atomic_inc_return(v) atomic_add_ret(1, v) ++static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_add_ret_unchecked(1, v); ++} + #define atomic64_inc_return(v) atomic64_add_ret(1, v) ++static inline long atomic64_inc_return_unchecked(atomic64_unchecked_t *v) ++{ ++ return atomic64_add_ret_unchecked(1, v); ++} + + #define atomic_sub_return(i, v) atomic_sub_ret(i, v) + #define atomic64_sub_return(i, v) atomic64_sub_ret(i, v) + + #define atomic_add_return(i, v) atomic_add_ret(i, v) ++static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v) ++{ ++ return atomic_add_ret_unchecked(i, v); ++} + #define atomic64_add_return(i, v) atomic64_add_ret(i, v) ++static inline long atomic64_add_return_unchecked(long i, atomic64_unchecked_t *v) ++{ ++ return atomic64_add_ret_unchecked(i, v); ++} + + /* + * atomic_inc_and_test - increment and test +@@ -50,6 +88,10 @@ extern long atomic64_sub_ret(long, atomic64_t *); + * other cases. + */ + #define atomic_inc_and_test(v) (atomic_inc_return(v) == 0) ++static inline int atomic_inc_and_test_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_inc_return_unchecked(v) == 0; ++} + #define atomic64_inc_and_test(v) (atomic64_inc_return(v) == 0) + + #define atomic_sub_and_test(i, v) (atomic_sub_ret(i, v) == 0) +@@ -59,30 +101,65 @@ extern long atomic64_sub_ret(long, atomic64_t *); + #define atomic64_dec_and_test(v) (atomic64_sub_ret(1, v) == 0) + + #define atomic_inc(v) atomic_add(1, v) ++static inline void atomic_inc_unchecked(atomic_unchecked_t *v) ++{ ++ atomic_add_unchecked(1, v); ++} + #define atomic64_inc(v) atomic64_add(1, v) ++static inline void atomic64_inc_unchecked(atomic64_unchecked_t *v) ++{ ++ atomic64_add_unchecked(1, v); ++} + + #define atomic_dec(v) atomic_sub(1, v) ++static inline void atomic_dec_unchecked(atomic_unchecked_t *v) ++{ ++ atomic_sub_unchecked(1, v); ++} + #define atomic64_dec(v) atomic64_sub(1, v) ++static inline void atomic64_dec_unchecked(atomic64_unchecked_t *v) ++{ ++ atomic64_sub_unchecked(1, v); ++} + + #define atomic_add_negative(i, v) (atomic_add_ret(i, v) < 0) + #define atomic64_add_negative(i, v) (atomic64_add_ret(i, v) < 0) + + #define atomic_cmpxchg(v, o, n) (cmpxchg(&((v)->counter), (o), (n))) ++static inline int atomic_cmpxchg_unchecked(atomic_unchecked_t *v, int old, int new) ++{ ++ return cmpxchg(&v->counter, old, new); ++} + #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) ++static inline int atomic_xchg_unchecked(atomic_unchecked_t *v, int new) ++{ ++ return xchg(&v->counter, new); ++} + + static inline int atomic_add_unless(atomic_t *v, int a, int u) + { +- int c, old; ++ int c, old, new; + c = atomic_read(v); + for (;;) { +- if (unlikely(c == (u))) ++ if (unlikely(c == u)) + break; +- old = atomic_cmpxchg((v), c, c + (a)); ++ ++ asm volatile("addcc %2, %0, %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "tvs %%icc, 6\n" ++#endif ++ ++ : "=r" (new) ++ : "0" (c), "ir" (a) ++ : "cc"); ++ ++ old = atomic_cmpxchg(v, c, new); + if (likely(old == c)) + break; + c = old; + } +- return c != (u); ++ return c != u; + } + + #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) +@@ -90,20 +167,35 @@ static inline int atomic_add_unless(atomic_t *v, int a, int u) + #define atomic64_cmpxchg(v, o, n) \ + ((__typeof__((v)->counter))cmpxchg(&((v)->counter), (o), (n))) + #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) ++static inline long atomic64_xchg_unchecked(atomic64_unchecked_t *v, long new) ++{ ++ return xchg(&v->counter, new); ++} + + static inline long atomic64_add_unless(atomic64_t *v, long a, long u) + { +- long c, old; ++ long c, old, new; + c = atomic64_read(v); + for (;;) { +- if (unlikely(c == (u))) ++ if (unlikely(c == u)) + break; +- old = atomic64_cmpxchg((v), c, c + (a)); ++ ++ asm volatile("addcc %2, %0, %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "tvs %%xcc, 6\n" ++#endif ++ ++ : "=r" (new) ++ : "0" (c), "ir" (a) ++ : "cc"); ++ ++ old = atomic64_cmpxchg(v, c, new); + if (likely(old == c)) + break; + c = old; + } +- return c != (u); ++ return c != u; + } + + #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) +diff --git a/arch/sparc/include/asm/cache.h b/arch/sparc/include/asm/cache.h +index 41f85ae..73b80b5 100644 +--- a/arch/sparc/include/asm/cache.h ++++ b/arch/sparc/include/asm/cache.h +@@ -7,8 +7,10 @@ + #ifndef _SPARC_CACHE_H + #define _SPARC_CACHE_H + ++#include ++ + #define L1_CACHE_SHIFT 5 +-#define L1_CACHE_BYTES 32 ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + #define L1_CACHE_ALIGN(x) ((((x)+(L1_CACHE_BYTES-1))&~(L1_CACHE_BYTES-1))) + + #ifdef CONFIG_SPARC32 +diff --git a/arch/sparc/include/asm/dma-mapping.h b/arch/sparc/include/asm/dma-mapping.h +index 5a8c308..38def92 100644 +--- a/arch/sparc/include/asm/dma-mapping.h ++++ b/arch/sparc/include/asm/dma-mapping.h +@@ -14,10 +14,10 @@ extern int dma_set_mask(struct device *dev, u64 dma_mask); + #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) + #define dma_is_consistent(d, h) (1) + +-extern struct dma_map_ops *dma_ops, pci32_dma_ops; ++extern const struct dma_map_ops *dma_ops, pci32_dma_ops; + extern struct bus_type pci_bus_type; + +-static inline struct dma_map_ops *get_dma_ops(struct device *dev) ++static inline const struct dma_map_ops *get_dma_ops(struct device *dev) + { + #if defined(CONFIG_SPARC32) && defined(CONFIG_PCI) + if (dev->bus == &pci_bus_type) +@@ -31,7 +31,7 @@ static inline struct dma_map_ops *get_dma_ops(struct device *dev) + static inline void *dma_alloc_coherent(struct device *dev, size_t size, + dma_addr_t *dma_handle, gfp_t flag) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + void *cpu_addr; + + cpu_addr = ops->alloc_coherent(dev, size, dma_handle, flag); +@@ -42,7 +42,7 @@ static inline void *dma_alloc_coherent(struct device *dev, size_t size, + static inline void dma_free_coherent(struct device *dev, size_t size, + void *cpu_addr, dma_addr_t dma_handle) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + + debug_dma_free_coherent(dev, size, cpu_addr, dma_handle); + ops->free_coherent(dev, size, cpu_addr, dma_handle); +diff --git a/arch/sparc/include/asm/elf_32.h b/arch/sparc/include/asm/elf_32.h +index 381a1b5..b97e3ff 100644 +--- a/arch/sparc/include/asm/elf_32.h ++++ b/arch/sparc/include/asm/elf_32.h +@@ -116,6 +116,13 @@ typedef struct { + + #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE) + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE 0x10000UL ++ ++#define PAX_DELTA_MMAP_LEN 16 ++#define PAX_DELTA_STACK_LEN 16 ++#endif ++ + /* This yields a mask that user programs can use to figure out what + instruction set this cpu supports. This can NOT be done in userspace + on Sparc. */ +diff --git a/arch/sparc/include/asm/elf_64.h b/arch/sparc/include/asm/elf_64.h +index 9968085..c2106ef 100644 +--- a/arch/sparc/include/asm/elf_64.h ++++ b/arch/sparc/include/asm/elf_64.h +@@ -163,6 +163,12 @@ typedef struct { + #define ELF_ET_DYN_BASE 0x0000010000000000UL + #define COMPAT_ELF_ET_DYN_BASE 0x0000000070000000UL + ++#ifdef CONFIG_PAX_ASLR ++#define PAX_ELF_ET_DYN_BASE (test_thread_flag(TIF_32BIT) ? 0x10000UL : 0x100000UL) ++ ++#define PAX_DELTA_MMAP_LEN (test_thread_flag(TIF_32BIT) ? 14 : 28) ++#define PAX_DELTA_STACK_LEN (test_thread_flag(TIF_32BIT) ? 15 : 29) ++#endif + + /* This yields a mask that user programs can use to figure out what + instruction set this cpu supports. */ +diff --git a/arch/sparc/include/asm/page_32.h b/arch/sparc/include/asm/page_32.h +index 156707b..aefa786 100644 +--- a/arch/sparc/include/asm/page_32.h ++++ b/arch/sparc/include/asm/page_32.h +@@ -8,6 +8,8 @@ + #ifndef _SPARC_PAGE_H + #define _SPARC_PAGE_H + ++#include ++ + #define PAGE_SHIFT 12 + + #ifndef __ASSEMBLY__ +diff --git a/arch/sparc/include/asm/pgtable_32.h b/arch/sparc/include/asm/pgtable_32.h +index e0cabe7..efd60f1 100644 +--- a/arch/sparc/include/asm/pgtable_32.h ++++ b/arch/sparc/include/asm/pgtable_32.h +@@ -43,6 +43,13 @@ BTFIXUPDEF_SIMM13(user_ptrs_per_pgd) + BTFIXUPDEF_INT(page_none) + BTFIXUPDEF_INT(page_copy) + BTFIXUPDEF_INT(page_readonly) ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++BTFIXUPDEF_INT(page_shared_noexec) ++BTFIXUPDEF_INT(page_copy_noexec) ++BTFIXUPDEF_INT(page_readonly_noexec) ++#endif ++ + BTFIXUPDEF_INT(page_kernel) + + #define PMD_SHIFT SUN4C_PMD_SHIFT +@@ -64,6 +71,16 @@ extern pgprot_t PAGE_SHARED; + #define PAGE_COPY __pgprot(BTFIXUP_INT(page_copy)) + #define PAGE_READONLY __pgprot(BTFIXUP_INT(page_readonly)) + ++#ifdef CONFIG_PAX_PAGEEXEC ++extern pgprot_t PAGE_SHARED_NOEXEC; ++# define PAGE_COPY_NOEXEC __pgprot(BTFIXUP_INT(page_copy_noexec)) ++# define PAGE_READONLY_NOEXEC __pgprot(BTFIXUP_INT(page_readonly_noexec)) ++#else ++# define PAGE_SHARED_NOEXEC PAGE_SHARED ++# define PAGE_COPY_NOEXEC PAGE_COPY ++# define PAGE_READONLY_NOEXEC PAGE_READONLY ++#endif ++ + extern unsigned long page_kernel; + + #ifdef MODULE +diff --git a/arch/sparc/include/asm/pgtsrmmu.h b/arch/sparc/include/asm/pgtsrmmu.h +index 1407c07..7e10231 100644 +--- a/arch/sparc/include/asm/pgtsrmmu.h ++++ b/arch/sparc/include/asm/pgtsrmmu.h +@@ -115,6 +115,13 @@ + SRMMU_EXEC | SRMMU_REF) + #define SRMMU_PAGE_RDONLY __pgprot(SRMMU_VALID | SRMMU_CACHE | \ + SRMMU_EXEC | SRMMU_REF) ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++#define SRMMU_PAGE_SHARED_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_WRITE | SRMMU_REF) ++#define SRMMU_PAGE_COPY_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_REF) ++#define SRMMU_PAGE_RDONLY_NOEXEC __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_REF) ++#endif ++ + #define SRMMU_PAGE_KERNEL __pgprot(SRMMU_VALID | SRMMU_CACHE | SRMMU_PRIV | \ + SRMMU_DIRTY | SRMMU_REF) + +diff --git a/arch/sparc/include/asm/spinlock_64.h b/arch/sparc/include/asm/spinlock_64.h +index 43e5147..47622a1 100644 +--- a/arch/sparc/include/asm/spinlock_64.h ++++ b/arch/sparc/include/asm/spinlock_64.h +@@ -92,14 +92,19 @@ static inline void __raw_spin_lock_flags(raw_spinlock_t *lock, unsigned long fla + + /* Multi-reader locks, these are much saner than the 32-bit Sparc ones... */ + +-static void inline arch_read_lock(raw_rwlock_t *lock) ++static inline void arch_read_lock(raw_rwlock_t *lock) + { + unsigned long tmp1, tmp2; + + __asm__ __volatile__ ( + "1: ldsw [%2], %0\n" + " brlz,pn %0, 2f\n" +-"4: add %0, 1, %1\n" ++"4: addcc %0, 1, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" tvs %%icc, 6\n" ++#endif ++ + " cas [%2], %0, %1\n" + " cmp %0, %1\n" + " bne,pn %%icc, 1b\n" +@@ -112,10 +117,10 @@ static void inline arch_read_lock(raw_rwlock_t *lock) + " .previous" + : "=&r" (tmp1), "=&r" (tmp2) + : "r" (lock) +- : "memory"); ++ : "memory", "cc"); + } + +-static int inline arch_read_trylock(raw_rwlock_t *lock) ++static inline int arch_read_trylock(raw_rwlock_t *lock) + { + int tmp1, tmp2; + +@@ -123,7 +128,12 @@ static int inline arch_read_trylock(raw_rwlock_t *lock) + "1: ldsw [%2], %0\n" + " brlz,a,pn %0, 2f\n" + " mov 0, %0\n" +-" add %0, 1, %1\n" ++" addcc %0, 1, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" tvs %%icc, 6\n" ++#endif ++ + " cas [%2], %0, %1\n" + " cmp %0, %1\n" + " bne,pn %%icc, 1b\n" +@@ -136,13 +146,18 @@ static int inline arch_read_trylock(raw_rwlock_t *lock) + return tmp1; + } + +-static void inline arch_read_unlock(raw_rwlock_t *lock) ++static inline void arch_read_unlock(raw_rwlock_t *lock) + { + unsigned long tmp1, tmp2; + + __asm__ __volatile__( + "1: lduw [%2], %0\n" +-" sub %0, 1, %1\n" ++" subcc %0, 1, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++" tvs %%icc, 6\n" ++#endif ++ + " cas [%2], %0, %1\n" + " cmp %0, %1\n" + " bne,pn %%xcc, 1b\n" +@@ -152,7 +167,7 @@ static void inline arch_read_unlock(raw_rwlock_t *lock) + : "memory"); + } + +-static void inline arch_write_lock(raw_rwlock_t *lock) ++static inline void arch_write_lock(raw_rwlock_t *lock) + { + unsigned long mask, tmp1, tmp2; + +@@ -177,7 +192,7 @@ static void inline arch_write_lock(raw_rwlock_t *lock) + : "memory"); + } + +-static void inline arch_write_unlock(raw_rwlock_t *lock) ++static inline void arch_write_unlock(raw_rwlock_t *lock) + { + __asm__ __volatile__( + " stw %%g0, [%0]" +@@ -186,7 +201,7 @@ static void inline arch_write_unlock(raw_rwlock_t *lock) + : "memory"); + } + +-static int inline arch_write_trylock(raw_rwlock_t *lock) ++static inline int arch_write_trylock(raw_rwlock_t *lock) + { + unsigned long mask, tmp1, tmp2, result; + +diff --git a/arch/sparc/include/asm/thread_info_32.h b/arch/sparc/include/asm/thread_info_32.h +index 844d73a..f787fb9 100644 +--- a/arch/sparc/include/asm/thread_info_32.h ++++ b/arch/sparc/include/asm/thread_info_32.h +@@ -50,6 +50,8 @@ struct thread_info { + unsigned long w_saved; + + struct restart_block restart_block; ++ ++ unsigned long lowest_stack; + }; + + /* +diff --git a/arch/sparc/include/asm/thread_info_64.h b/arch/sparc/include/asm/thread_info_64.h +index f78ad9a..9f55fc7 100644 +--- a/arch/sparc/include/asm/thread_info_64.h ++++ b/arch/sparc/include/asm/thread_info_64.h +@@ -68,6 +68,8 @@ struct thread_info { + struct pt_regs *kern_una_regs; + unsigned int kern_una_insn; + ++ unsigned long lowest_stack; ++ + unsigned long fpregs[0] __attribute__ ((aligned(64))); + }; + +diff --git a/arch/sparc/include/asm/uaccess.h b/arch/sparc/include/asm/uaccess.h +index e88fbe5..96b0ce5 100644 +--- a/arch/sparc/include/asm/uaccess.h ++++ b/arch/sparc/include/asm/uaccess.h +@@ -1,5 +1,13 @@ + #ifndef ___ASM_SPARC_UACCESS_H + #define ___ASM_SPARC_UACCESS_H ++ ++#ifdef __KERNEL__ ++#ifndef __ASSEMBLY__ ++#include ++extern void check_object_size(const void *ptr, unsigned long n, bool to); ++#endif ++#endif ++ + #if defined(__sparc__) && defined(__arch64__) + #include + #else +diff --git a/arch/sparc/include/asm/uaccess_32.h b/arch/sparc/include/asm/uaccess_32.h +index 8303ac4..07f333d 100644 +--- a/arch/sparc/include/asm/uaccess_32.h ++++ b/arch/sparc/include/asm/uaccess_32.h +@@ -249,27 +249,46 @@ extern unsigned long __copy_user(void __user *to, const void __user *from, unsig + + static inline unsigned long copy_to_user(void __user *to, const void *from, unsigned long n) + { +- if (n && __access_ok((unsigned long) to, n)) ++ if ((long)n < 0) ++ return n; ++ ++ if (n && __access_ok((unsigned long) to, n)) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); + return __copy_user(to, (__force void __user *) from, n); +- else ++ } else + return n; + } + + static inline unsigned long __copy_to_user(void __user *to, const void *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); ++ + return __copy_user(to, (__force void __user *) from, n); + } + + static inline unsigned long copy_from_user(void *to, const void __user *from, unsigned long n) + { +- if (n && __access_ok((unsigned long) from, n)) ++ if ((long)n < 0) ++ return n; ++ ++ if (n && __access_ok((unsigned long) from, n)) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); + return __copy_user((__force void __user *) to, from, n); +- else ++ } else + return n; + } + + static inline unsigned long __copy_from_user(void *to, const void __user *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + return __copy_user((__force void __user *) to, from, n); + } + +diff --git a/arch/sparc/include/asm/uaccess_64.h b/arch/sparc/include/asm/uaccess_64.h +index 9ea271e..7b8a271 100644 +--- a/arch/sparc/include/asm/uaccess_64.h ++++ b/arch/sparc/include/asm/uaccess_64.h +@@ -9,6 +9,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -212,8 +213,15 @@ extern unsigned long copy_from_user_fixup(void *to, const void __user *from, + static inline unsigned long __must_check + copy_from_user(void *to, const void __user *from, unsigned long size) + { +- unsigned long ret = ___copy_from_user(to, from, size); ++ unsigned long ret; + ++ if ((long)size < 0 || size > INT_MAX) ++ return size; ++ ++ if (!__builtin_constant_p(size)) ++ check_object_size(to, size, false); ++ ++ ret = ___copy_from_user(to, from, size); + if (unlikely(ret)) + ret = copy_from_user_fixup(to, from, size); + return ret; +@@ -228,8 +236,15 @@ extern unsigned long copy_to_user_fixup(void __user *to, const void *from, + static inline unsigned long __must_check + copy_to_user(void __user *to, const void *from, unsigned long size) + { +- unsigned long ret = ___copy_to_user(to, from, size); ++ unsigned long ret; + ++ if ((long)size < 0 || size > INT_MAX) ++ return size; ++ ++ if (!__builtin_constant_p(size)) ++ check_object_size(from, size, true); ++ ++ ret = ___copy_to_user(to, from, size); + if (unlikely(ret)) + ret = copy_to_user_fixup(to, from, size); + return ret; +diff --git a/arch/sparc/kernel/Makefile b/arch/sparc/kernel/Makefile +index 2782681..77ded84 100644 +--- a/arch/sparc/kernel/Makefile ++++ b/arch/sparc/kernel/Makefile +@@ -3,7 +3,7 @@ + # + + asflags-y := -ansi +-ccflags-y := -Werror ++#ccflags-y := -Werror + + extra-y := head_$(BITS).o + extra-y += init_task.o +diff --git a/arch/sparc/kernel/iommu.c b/arch/sparc/kernel/iommu.c +index 7690cc2..ece64c9 100644 +--- a/arch/sparc/kernel/iommu.c ++++ b/arch/sparc/kernel/iommu.c +@@ -826,7 +826,7 @@ static void dma_4u_sync_sg_for_cpu(struct device *dev, + spin_unlock_irqrestore(&iommu->lock, flags); + } + +-static struct dma_map_ops sun4u_dma_ops = { ++static const struct dma_map_ops sun4u_dma_ops = { + .alloc_coherent = dma_4u_alloc_coherent, + .free_coherent = dma_4u_free_coherent, + .map_page = dma_4u_map_page, +@@ -837,7 +837,7 @@ static struct dma_map_ops sun4u_dma_ops = { + .sync_sg_for_cpu = dma_4u_sync_sg_for_cpu, + }; + +-struct dma_map_ops *dma_ops = &sun4u_dma_ops; ++const struct dma_map_ops *dma_ops = &sun4u_dma_ops; + EXPORT_SYMBOL(dma_ops); + + extern int pci64_dma_supported(struct pci_dev *pdev, u64 device_mask); +diff --git a/arch/sparc/kernel/ioport.c b/arch/sparc/kernel/ioport.c +index 9f61fd8..bd048db 100644 +--- a/arch/sparc/kernel/ioport.c ++++ b/arch/sparc/kernel/ioport.c +@@ -392,7 +392,7 @@ static void sbus_sync_sg_for_device(struct device *dev, struct scatterlist *sg, + BUG(); + } + +-struct dma_map_ops sbus_dma_ops = { ++const struct dma_map_ops sbus_dma_ops = { + .alloc_coherent = sbus_alloc_coherent, + .free_coherent = sbus_free_coherent, + .map_page = sbus_map_page, +@@ -403,7 +403,7 @@ struct dma_map_ops sbus_dma_ops = { + .sync_sg_for_device = sbus_sync_sg_for_device, + }; + +-struct dma_map_ops *dma_ops = &sbus_dma_ops; ++const struct dma_map_ops *dma_ops = &sbus_dma_ops; + EXPORT_SYMBOL(dma_ops); + + static int __init sparc_register_ioport(void) +@@ -640,7 +640,7 @@ static void pci32_sync_sg_for_device(struct device *device, struct scatterlist * + } + } + +-struct dma_map_ops pci32_dma_ops = { ++const struct dma_map_ops pci32_dma_ops = { + .alloc_coherent = pci32_alloc_coherent, + .free_coherent = pci32_free_coherent, + .map_page = pci32_map_page, +diff --git a/arch/sparc/kernel/kgdb_32.c b/arch/sparc/kernel/kgdb_32.c +index 04df4ed..55c4b6e 100644 +--- a/arch/sparc/kernel/kgdb_32.c ++++ b/arch/sparc/kernel/kgdb_32.c +@@ -158,7 +158,7 @@ void kgdb_arch_exit(void) + { + } + +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + /* Breakpoint instruction: ta 0x7d */ + .gdb_bpt_instr = { 0x91, 0xd0, 0x20, 0x7d }, + }; +diff --git a/arch/sparc/kernel/kgdb_64.c b/arch/sparc/kernel/kgdb_64.c +index f5a0fd4..d886f71 100644 +--- a/arch/sparc/kernel/kgdb_64.c ++++ b/arch/sparc/kernel/kgdb_64.c +@@ -180,7 +180,7 @@ void kgdb_arch_exit(void) + { + } + +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + /* Breakpoint instruction: ta 0x72 */ + .gdb_bpt_instr = { 0x91, 0xd0, 0x20, 0x72 }, + }; +diff --git a/arch/sparc/kernel/pci_sun4v.c b/arch/sparc/kernel/pci_sun4v.c +index 23c33ff..d137fbd 100644 +--- a/arch/sparc/kernel/pci_sun4v.c ++++ b/arch/sparc/kernel/pci_sun4v.c +@@ -525,7 +525,7 @@ static void dma_4v_unmap_sg(struct device *dev, struct scatterlist *sglist, + spin_unlock_irqrestore(&iommu->lock, flags); + } + +-static struct dma_map_ops sun4v_dma_ops = { ++static const struct dma_map_ops sun4v_dma_ops = { + .alloc_coherent = dma_4v_alloc_coherent, + .free_coherent = dma_4v_free_coherent, + .map_page = dma_4v_map_page, +diff --git a/arch/sparc/kernel/process_32.c b/arch/sparc/kernel/process_32.c +index c49865b..b41a81b 100644 +--- a/arch/sparc/kernel/process_32.c ++++ b/arch/sparc/kernel/process_32.c +@@ -196,7 +196,7 @@ void __show_backtrace(unsigned long fp) + rw->ins[4], rw->ins[5], + rw->ins[6], + rw->ins[7]); +- printk("%pS\n", (void *) rw->ins[7]); ++ printk("%pA\n", (void *) rw->ins[7]); + rw = (struct reg_window32 *) rw->ins[6]; + } + spin_unlock_irqrestore(&sparc_backtrace_lock, flags); +@@ -263,14 +263,14 @@ void show_regs(struct pt_regs *r) + + printk("PSR: %08lx PC: %08lx NPC: %08lx Y: %08lx %s\n", + r->psr, r->pc, r->npc, r->y, print_tainted()); +- printk("PC: <%pS>\n", (void *) r->pc); ++ printk("PC: <%pA>\n", (void *) r->pc); + printk("%%G: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n", + r->u_regs[0], r->u_regs[1], r->u_regs[2], r->u_regs[3], + r->u_regs[4], r->u_regs[5], r->u_regs[6], r->u_regs[7]); + printk("%%O: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n", + r->u_regs[8], r->u_regs[9], r->u_regs[10], r->u_regs[11], + r->u_regs[12], r->u_regs[13], r->u_regs[14], r->u_regs[15]); +- printk("RPC: <%pS>\n", (void *) r->u_regs[15]); ++ printk("RPC: <%pA>\n", (void *) r->u_regs[15]); + + printk("%%L: %08lx %08lx %08lx %08lx %08lx %08lx %08lx %08lx\n", + rw->locals[0], rw->locals[1], rw->locals[2], rw->locals[3], +@@ -305,7 +305,7 @@ void show_stack(struct task_struct *tsk, unsigned long *_ksp) + rw = (struct reg_window32 *) fp; + pc = rw->ins[7]; + printk("[%08lx : ", pc); +- printk("%pS ] ", (void *) pc); ++ printk("%pA ] ", (void *) pc); + fp = rw->ins[6]; + } while (++count < 16); + printk("\n"); +diff --git a/arch/sparc/kernel/process_64.c b/arch/sparc/kernel/process_64.c +index cb70476..3d0c191 100644 +--- a/arch/sparc/kernel/process_64.c ++++ b/arch/sparc/kernel/process_64.c +@@ -180,14 +180,14 @@ static void show_regwindow(struct pt_regs *regs) + printk("i4: %016lx i5: %016lx i6: %016lx i7: %016lx\n", + rwk->ins[4], rwk->ins[5], rwk->ins[6], rwk->ins[7]); + if (regs->tstate & TSTATE_PRIV) +- printk("I7: <%pS>\n", (void *) rwk->ins[7]); ++ printk("I7: <%pA>\n", (void *) rwk->ins[7]); + } + + void show_regs(struct pt_regs *regs) + { + printk("TSTATE: %016lx TPC: %016lx TNPC: %016lx Y: %08x %s\n", regs->tstate, + regs->tpc, regs->tnpc, regs->y, print_tainted()); +- printk("TPC: <%pS>\n", (void *) regs->tpc); ++ printk("TPC: <%pA>\n", (void *) regs->tpc); + printk("g0: %016lx g1: %016lx g2: %016lx g3: %016lx\n", + regs->u_regs[0], regs->u_regs[1], regs->u_regs[2], + regs->u_regs[3]); +@@ -200,7 +200,7 @@ void show_regs(struct pt_regs *regs) + printk("o4: %016lx o5: %016lx sp: %016lx ret_pc: %016lx\n", + regs->u_regs[12], regs->u_regs[13], regs->u_regs[14], + regs->u_regs[15]); +- printk("RPC: <%pS>\n", (void *) regs->u_regs[15]); ++ printk("RPC: <%pA>\n", (void *) regs->u_regs[15]); + show_regwindow(regs); + } + +@@ -284,7 +284,7 @@ void arch_trigger_all_cpu_backtrace(void) + ((tp && tp->task) ? tp->task->pid : -1)); + + if (gp->tstate & TSTATE_PRIV) { +- printk(" TPC[%pS] O7[%pS] I7[%pS] RPC[%pS]\n", ++ printk(" TPC[%pA] O7[%pA] I7[%pA] RPC[%pA]\n", + (void *) gp->tpc, + (void *) gp->o7, + (void *) gp->i7, +diff --git a/arch/sparc/kernel/sigutil_64.c b/arch/sparc/kernel/sigutil_64.c +index 6edc4e5..06a69b4 100644 +--- a/arch/sparc/kernel/sigutil_64.c ++++ b/arch/sparc/kernel/sigutil_64.c +@@ -2,6 +2,7 @@ + #include + #include + #include ++#include + + #include + #include +diff --git a/arch/sparc/kernel/sys_sparc_32.c b/arch/sparc/kernel/sys_sparc_32.c +index 3a82e65..ce0a53a 100644 +--- a/arch/sparc/kernel/sys_sparc_32.c ++++ b/arch/sparc/kernel/sys_sparc_32.c +@@ -57,7 +57,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + if (ARCH_SUN4C && len > 0x20000000) + return -ENOMEM; + if (!addr) +- addr = TASK_UNMAPPED_BASE; ++ addr = current->mm->mmap_base; + + if (flags & MAP_SHARED) + addr = COLOUR_ALIGN(addr); +@@ -72,7 +72,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + } + if (TASK_SIZE - PAGE_SIZE - len < addr) + return -ENOMEM; +- if (!vmm || addr + len <= vmm->vm_start) ++ if (check_heap_stack_gap(vmm, addr, len)) + return addr; + addr = vmm->vm_end; + if (flags & MAP_SHARED) +diff --git a/arch/sparc/kernel/sys_sparc_64.c b/arch/sparc/kernel/sys_sparc_64.c +index cfa0e19..98972ac 100644 +--- a/arch/sparc/kernel/sys_sparc_64.c ++++ b/arch/sparc/kernel/sys_sparc_64.c +@@ -125,7 +125,7 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + /* We do not accept a shared mapping if it would violate + * cache aliasing constraints. + */ +- if ((flags & MAP_SHARED) && ++ if ((filp || (flags & MAP_SHARED)) && + ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) + return -EINVAL; + return addr; +@@ -140,6 +140,10 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + if (filp || (flags & MAP_SHARED)) + do_color_align = 1; + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + if (addr) { + if (do_color_align) + addr = COLOUR_ALIGN(addr, pgoff); +@@ -147,15 +151,14 @@ unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsi + addr = PAGE_ALIGN(addr); + + vma = find_vma(mm, addr); +- if (task_size - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + + if (len > mm->cached_hole_size) { +- start_addr = addr = mm->free_area_cache; ++ start_addr = addr = mm->free_area_cache; + } else { +- start_addr = addr = TASK_UNMAPPED_BASE; ++ start_addr = addr = mm->mmap_base; + mm->cached_hole_size = 0; + } + +@@ -175,14 +178,14 @@ full_search: + vma = find_vma(mm, VA_EXCLUDE_END); + } + if (unlikely(task_size < addr)) { +- if (start_addr != TASK_UNMAPPED_BASE) { +- start_addr = addr = TASK_UNMAPPED_BASE; ++ if (start_addr != mm->mmap_base) { ++ start_addr = addr = mm->mmap_base; + mm->cached_hole_size = 0; + goto full_search; + } + return -ENOMEM; + } +- if (likely(!vma || addr + len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* + * Remember the place where we stopped the search: + */ +@@ -216,7 +219,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + /* We do not accept a shared mapping if it would violate + * cache aliasing constraints. + */ +- if ((flags & MAP_SHARED) && ++ if ((filp || (flags & MAP_SHARED)) && + ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) + return -EINVAL; + return addr; +@@ -237,8 +240,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + addr = PAGE_ALIGN(addr); + + vma = find_vma(mm, addr); +- if (task_size - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + +@@ -259,7 +261,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + /* make sure it can fit in the remaining address space */ + if (likely(addr > len)) { + vma = find_vma(mm, addr-len); +- if (!vma || addr <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr - len, len)) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr-len); + } +@@ -268,18 +270,18 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + if (unlikely(mm->mmap_base < len)) + goto bottomup; + +- addr = mm->mmap_base-len; +- if (do_color_align) +- addr = COLOUR_ALIGN_DOWN(addr, pgoff); ++ addr = mm->mmap_base - len; + + do { ++ if (do_color_align) ++ addr = COLOUR_ALIGN_DOWN(addr, pgoff); + /* + * Lookup failure means no vma is above this address, + * else if new region fits below vma->vm_start, + * return with success: + */ + vma = find_vma(mm, addr); +- if (likely(!vma || addr+len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr); + } +@@ -289,10 +291,8 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + mm->cached_hole_size = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = vma->vm_start-len; +- if (do_color_align) +- addr = COLOUR_ALIGN_DOWN(addr, pgoff); +- } while (likely(len < vma->vm_start)); ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); + + bottomup: + /* +@@ -384,6 +384,12 @@ void arch_pick_mmap_layout(struct mm_struct *mm) + current->signal->rlim[RLIMIT_STACK].rlim_cur == RLIM_INFINITY || + sysctl_legacy_va_layout) { + mm->mmap_base = TASK_UNMAPPED_BASE + random_factor; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area; + mm->unmap_area = arch_unmap_area; + } else { +@@ -398,6 +404,12 @@ void arch_pick_mmap_layout(struct mm_struct *mm) + gap = (task_size / 6 * 5); + + mm->mmap_base = PAGE_ALIGN(task_size - gap - random_factor); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base -= mm->delta_mmap + mm->delta_stack; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area_topdown; + mm->unmap_area = arch_unmap_area_topdown; + } +diff --git a/arch/sparc/kernel/traps_32.c b/arch/sparc/kernel/traps_32.c +index c0490c7..84959d1 100644 +--- a/arch/sparc/kernel/traps_32.c ++++ b/arch/sparc/kernel/traps_32.c +@@ -44,6 +44,8 @@ static void instruction_dump(unsigned long *pc) + #define __SAVE __asm__ __volatile__("save %sp, -0x40, %sp\n\t") + #define __RESTORE __asm__ __volatile__("restore %g0, %g0, %g0\n\t") + ++extern void gr_handle_kernel_exploit(void); ++ + void die_if_kernel(char *str, struct pt_regs *regs) + { + static int die_counter; +@@ -76,15 +78,17 @@ void die_if_kernel(char *str, struct pt_regs *regs) + count++ < 30 && + (((unsigned long) rw) >= PAGE_OFFSET) && + !(((unsigned long) rw) & 0x7)) { +- printk("Caller[%08lx]: %pS\n", rw->ins[7], ++ printk("Caller[%08lx]: %pA\n", rw->ins[7], + (void *) rw->ins[7]); + rw = (struct reg_window32 *)rw->ins[6]; + } + } + printk("Instruction DUMP:"); + instruction_dump ((unsigned long *) regs->pc); +- if(regs->psr & PSR_PS) ++ if(regs->psr & PSR_PS) { ++ gr_handle_kernel_exploit(); + do_exit(SIGKILL); ++ } + do_exit(SIGSEGV); + } + +diff --git a/arch/sparc/kernel/traps_64.c b/arch/sparc/kernel/traps_64.c +index 10f7bb9..cdb6793 100644 +--- a/arch/sparc/kernel/traps_64.c ++++ b/arch/sparc/kernel/traps_64.c +@@ -73,7 +73,7 @@ static void dump_tl1_traplog(struct tl1_traplog *p) + i + 1, + p->trapstack[i].tstate, p->trapstack[i].tpc, + p->trapstack[i].tnpc, p->trapstack[i].tt); +- printk("TRAPLOG: TPC<%pS>\n", (void *) p->trapstack[i].tpc); ++ printk("TRAPLOG: TPC<%pA>\n", (void *) p->trapstack[i].tpc); + } + } + +@@ -93,6 +93,12 @@ void bad_trap(struct pt_regs *regs, long lvl) + + lvl -= 0x100; + if (regs->tstate & TSTATE_PRIV) { ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ if (lvl == 6) ++ pax_report_refcount_overflow(regs); ++#endif ++ + sprintf(buffer, "Kernel bad sw trap %lx", lvl); + die_if_kernel(buffer, regs); + } +@@ -111,11 +117,16 @@ void bad_trap(struct pt_regs *regs, long lvl) + void bad_trap_tl1(struct pt_regs *regs, long lvl) + { + char buffer[32]; +- ++ + if (notify_die(DIE_TRAP_TL1, "bad trap tl1", regs, + 0, lvl, SIGTRAP) == NOTIFY_STOP) + return; + ++#ifdef CONFIG_PAX_REFCOUNT ++ if (lvl == 6) ++ pax_report_refcount_overflow(regs); ++#endif ++ + dump_tl1_traplog((struct tl1_traplog *)(regs + 1)); + + sprintf (buffer, "Bad trap %lx at tl>0", lvl); +@@ -1139,7 +1150,7 @@ static void cheetah_log_errors(struct pt_regs *regs, struct cheetah_err_info *in + regs->tpc, regs->tnpc, regs->u_regs[UREG_I7], regs->tstate); + printk("%s" "ERROR(%d): ", + (recoverable ? KERN_WARNING : KERN_CRIT), smp_processor_id()); +- printk("TPC<%pS>\n", (void *) regs->tpc); ++ printk("TPC<%pA>\n", (void *) regs->tpc); + printk("%s" "ERROR(%d): M_SYND(%lx), E_SYND(%lx)%s%s\n", + (recoverable ? KERN_WARNING : KERN_CRIT), smp_processor_id(), + (afsr & CHAFSR_M_SYNDROME) >> CHAFSR_M_SYNDROME_SHIFT, +@@ -1746,7 +1757,7 @@ void cheetah_plus_parity_error(int type, struct pt_regs *regs) + smp_processor_id(), + (type & 0x1) ? 'I' : 'D', + regs->tpc); +- printk(KERN_EMERG "TPC<%pS>\n", (void *) regs->tpc); ++ printk(KERN_EMERG "TPC<%pA>\n", (void *) regs->tpc); + panic("Irrecoverable Cheetah+ parity error."); + } + +@@ -1754,7 +1765,7 @@ void cheetah_plus_parity_error(int type, struct pt_regs *regs) + smp_processor_id(), + (type & 0x1) ? 'I' : 'D', + regs->tpc); +- printk(KERN_WARNING "TPC<%pS>\n", (void *) regs->tpc); ++ printk(KERN_WARNING "TPC<%pA>\n", (void *) regs->tpc); + } + + struct sun4v_error_entry { +@@ -1961,9 +1972,9 @@ void sun4v_itlb_error_report(struct pt_regs *regs, int tl) + + printk(KERN_EMERG "SUN4V-ITLB: Error at TPC[%lx], tl %d\n", + regs->tpc, tl); +- printk(KERN_EMERG "SUN4V-ITLB: TPC<%pS>\n", (void *) regs->tpc); ++ printk(KERN_EMERG "SUN4V-ITLB: TPC<%pA>\n", (void *) regs->tpc); + printk(KERN_EMERG "SUN4V-ITLB: O7[%lx]\n", regs->u_regs[UREG_I7]); +- printk(KERN_EMERG "SUN4V-ITLB: O7<%pS>\n", ++ printk(KERN_EMERG "SUN4V-ITLB: O7<%pA>\n", + (void *) regs->u_regs[UREG_I7]); + printk(KERN_EMERG "SUN4V-ITLB: vaddr[%lx] ctx[%lx] " + "pte[%lx] error[%lx]\n", +@@ -1985,9 +1996,9 @@ void sun4v_dtlb_error_report(struct pt_regs *regs, int tl) + + printk(KERN_EMERG "SUN4V-DTLB: Error at TPC[%lx], tl %d\n", + regs->tpc, tl); +- printk(KERN_EMERG "SUN4V-DTLB: TPC<%pS>\n", (void *) regs->tpc); ++ printk(KERN_EMERG "SUN4V-DTLB: TPC<%pA>\n", (void *) regs->tpc); + printk(KERN_EMERG "SUN4V-DTLB: O7[%lx]\n", regs->u_regs[UREG_I7]); +- printk(KERN_EMERG "SUN4V-DTLB: O7<%pS>\n", ++ printk(KERN_EMERG "SUN4V-DTLB: O7<%pA>\n", + (void *) regs->u_regs[UREG_I7]); + printk(KERN_EMERG "SUN4V-DTLB: vaddr[%lx] ctx[%lx] " + "pte[%lx] error[%lx]\n", +@@ -2191,7 +2202,7 @@ void show_stack(struct task_struct *tsk, unsigned long *_ksp) + fp = (unsigned long)sf->fp + STACK_BIAS; + } + +- printk(" [%016lx] %pS\n", pc, (void *) pc); ++ printk(" [%016lx] %pA\n", pc, (void *) pc); + } while (++count < 16); + } + +@@ -2233,6 +2244,8 @@ static inline struct reg_window *kernel_stack_up(struct reg_window *rw) + return (struct reg_window *) (fp + STACK_BIAS); + } + ++extern void gr_handle_kernel_exploit(void); ++ + void die_if_kernel(char *str, struct pt_regs *regs) + { + static int die_counter; +@@ -2260,7 +2273,7 @@ void die_if_kernel(char *str, struct pt_regs *regs) + while (rw && + count++ < 30&& + is_kernel_stack(current, rw)) { +- printk("Caller[%016lx]: %pS\n", rw->ins[7], ++ printk("Caller[%016lx]: %pA\n", rw->ins[7], + (void *) rw->ins[7]); + + rw = kernel_stack_up(rw); +@@ -2273,8 +2286,11 @@ void die_if_kernel(char *str, struct pt_regs *regs) + } + user_instruction_dump ((unsigned int __user *) regs->tpc); + } +- if (regs->tstate & TSTATE_PRIV) ++ if (regs->tstate & TSTATE_PRIV) { ++ gr_handle_kernel_exploit(); + do_exit(SIGKILL); ++ } ++ + do_exit(SIGSEGV); + } + EXPORT_SYMBOL(die_if_kernel); +diff --git a/arch/sparc/kernel/una_asm_64.S b/arch/sparc/kernel/una_asm_64.S +index be183fe..1c8d332 100644 +--- a/arch/sparc/kernel/una_asm_64.S ++++ b/arch/sparc/kernel/una_asm_64.S +@@ -127,7 +127,7 @@ do_int_load: + wr %o5, 0x0, %asi + retl + mov 0, %o0 +- .size __do_int_load, .-__do_int_load ++ .size do_int_load, .-do_int_load + + .section __ex_table,"a" + .word 4b, __retl_efault +diff --git a/arch/sparc/kernel/unaligned_64.c b/arch/sparc/kernel/unaligned_64.c +index 3792099..2af17d8 100644 +--- a/arch/sparc/kernel/unaligned_64.c ++++ b/arch/sparc/kernel/unaligned_64.c +@@ -288,7 +288,7 @@ static void log_unaligned(struct pt_regs *regs) + if (count < 5) { + last_time = jiffies; + count++; +- printk("Kernel unaligned access at TPC[%lx] %pS\n", ++ printk("Kernel unaligned access at TPC[%lx] %pA\n", + regs->tpc, (void *) regs->tpc); + } + } +diff --git a/arch/sparc/lib/Makefile b/arch/sparc/lib/Makefile +index e75faf0..24f12f9 100644 +--- a/arch/sparc/lib/Makefile ++++ b/arch/sparc/lib/Makefile +@@ -2,7 +2,7 @@ + # + + asflags-y := -ansi -DST_DIV0=0x02 +-ccflags-y := -Werror ++#ccflags-y := -Werror + + lib-$(CONFIG_SPARC32) += mul.o rem.o sdiv.o udiv.o umul.o urem.o ashrdi3.o + lib-$(CONFIG_SPARC32) += memcpy.o memset.o +diff --git a/arch/sparc/lib/atomic_64.S b/arch/sparc/lib/atomic_64.S +index 0268210..f0291ca 100644 +--- a/arch/sparc/lib/atomic_64.S ++++ b/arch/sparc/lib/atomic_64.S +@@ -18,7 +18,12 @@ + atomic_add: /* %o0 = increment, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: lduw [%o1], %g1 +- add %g1, %o0, %g7 ++ addcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 2f +@@ -28,12 +33,32 @@ atomic_add: /* %o0 = increment, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic_add, .-atomic_add + ++ .globl atomic_add_unchecked ++ .type atomic_add_unchecked,#function ++atomic_add_unchecked: /* %o0 = increment, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: lduw [%o1], %g1 ++ add %g1, %o0, %g7 ++ cas [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %icc, 2f ++ nop ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic_add_unchecked, .-atomic_add_unchecked ++ + .globl atomic_sub + .type atomic_sub,#function + atomic_sub: /* %o0 = decrement, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: lduw [%o1], %g1 +- sub %g1, %o0, %g7 ++ subcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 2f +@@ -43,12 +68,32 @@ atomic_sub: /* %o0 = decrement, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic_sub, .-atomic_sub + ++ .globl atomic_sub_unchecked ++ .type atomic_sub_unchecked,#function ++atomic_sub_unchecked: /* %o0 = decrement, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: lduw [%o1], %g1 ++ sub %g1, %o0, %g7 ++ cas [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %icc, 2f ++ nop ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic_sub_unchecked, .-atomic_sub_unchecked ++ + .globl atomic_add_ret + .type atomic_add_ret,#function + atomic_add_ret: /* %o0 = increment, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: lduw [%o1], %g1 +- add %g1, %o0, %g7 ++ addcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 2f +@@ -59,12 +104,33 @@ atomic_add_ret: /* %o0 = increment, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic_add_ret, .-atomic_add_ret + ++ .globl atomic_add_ret_unchecked ++ .type atomic_add_ret_unchecked,#function ++atomic_add_ret_unchecked: /* %o0 = increment, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: lduw [%o1], %g1 ++ addcc %g1, %o0, %g7 ++ cas [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %icc, 2f ++ add %g7, %o0, %g7 ++ sra %g7, 0, %o0 ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic_add_ret_unchecked, .-atomic_add_ret_unchecked ++ + .globl atomic_sub_ret + .type atomic_sub_ret,#function + atomic_sub_ret: /* %o0 = decrement, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: lduw [%o1], %g1 +- sub %g1, %o0, %g7 ++ subcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 2f +@@ -80,7 +146,12 @@ atomic_sub_ret: /* %o0 = decrement, %o1 = atomic_ptr */ + atomic64_add: /* %o0 = increment, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: ldx [%o1], %g1 +- add %g1, %o0, %g7 ++ addcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %xcc, 6 ++#endif ++ + casx [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %xcc, 2f +@@ -90,12 +161,32 @@ atomic64_add: /* %o0 = increment, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic64_add, .-atomic64_add + ++ .globl atomic64_add_unchecked ++ .type atomic64_add_unchecked,#function ++atomic64_add_unchecked: /* %o0 = increment, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: ldx [%o1], %g1 ++ addcc %g1, %o0, %g7 ++ casx [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %xcc, 2f ++ nop ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic64_add_unchecked, .-atomic64_add_unchecked ++ + .globl atomic64_sub + .type atomic64_sub,#function + atomic64_sub: /* %o0 = decrement, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: ldx [%o1], %g1 +- sub %g1, %o0, %g7 ++ subcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %xcc, 6 ++#endif ++ + casx [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %xcc, 2f +@@ -105,12 +196,32 @@ atomic64_sub: /* %o0 = decrement, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic64_sub, .-atomic64_sub + ++ .globl atomic64_sub_unchecked ++ .type atomic64_sub_unchecked,#function ++atomic64_sub_unchecked: /* %o0 = decrement, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: ldx [%o1], %g1 ++ subcc %g1, %o0, %g7 ++ casx [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %xcc, 2f ++ nop ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic64_sub_unchecked, .-atomic64_sub_unchecked ++ + .globl atomic64_add_ret + .type atomic64_add_ret,#function + atomic64_add_ret: /* %o0 = increment, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: ldx [%o1], %g1 +- add %g1, %o0, %g7 ++ addcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %xcc, 6 ++#endif ++ + casx [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %xcc, 2f +@@ -121,12 +232,33 @@ atomic64_add_ret: /* %o0 = increment, %o1 = atomic_ptr */ + 2: BACKOFF_SPIN(%o2, %o3, 1b) + .size atomic64_add_ret, .-atomic64_add_ret + ++ .globl atomic64_add_ret_unchecked ++ .type atomic64_add_ret_unchecked,#function ++atomic64_add_ret_unchecked: /* %o0 = increment, %o1 = atomic_ptr */ ++ BACKOFF_SETUP(%o2) ++1: ldx [%o1], %g1 ++ addcc %g1, %o0, %g7 ++ casx [%o1], %g1, %g7 ++ cmp %g1, %g7 ++ bne,pn %xcc, 2f ++ add %g7, %o0, %g7 ++ mov %g7, %o0 ++ retl ++ nop ++2: BACKOFF_SPIN(%o2, %o3, 1b) ++ .size atomic64_add_ret_unchecked, .-atomic64_add_ret_unchecked ++ + .globl atomic64_sub_ret + .type atomic64_sub_ret,#function + atomic64_sub_ret: /* %o0 = decrement, %o1 = atomic_ptr */ + BACKOFF_SETUP(%o2) + 1: ldx [%o1], %g1 +- sub %g1, %o0, %g7 ++ subcc %g1, %o0, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %xcc, 6 ++#endif ++ + casx [%o1], %g1, %g7 + cmp %g1, %g7 + bne,pn %xcc, 2f +diff --git a/arch/sparc/lib/ksyms.c b/arch/sparc/lib/ksyms.c +index 704b126..2e79d76 100644 +--- a/arch/sparc/lib/ksyms.c ++++ b/arch/sparc/lib/ksyms.c +@@ -144,12 +144,18 @@ EXPORT_SYMBOL(__downgrade_write); + + /* Atomic counter implementation. */ + EXPORT_SYMBOL(atomic_add); ++EXPORT_SYMBOL(atomic_add_unchecked); + EXPORT_SYMBOL(atomic_add_ret); ++EXPORT_SYMBOL(atomic_add_ret_unchecked); + EXPORT_SYMBOL(atomic_sub); ++EXPORT_SYMBOL(atomic_sub_unchecked); + EXPORT_SYMBOL(atomic_sub_ret); + EXPORT_SYMBOL(atomic64_add); ++EXPORT_SYMBOL(atomic64_add_unchecked); + EXPORT_SYMBOL(atomic64_add_ret); ++EXPORT_SYMBOL(atomic64_add_ret_unchecked); + EXPORT_SYMBOL(atomic64_sub); ++EXPORT_SYMBOL(atomic64_sub_unchecked); + EXPORT_SYMBOL(atomic64_sub_ret); + + /* Atomic bit operations. */ +diff --git a/arch/sparc/lib/rwsem_64.S b/arch/sparc/lib/rwsem_64.S +index 91a7d29..ce75c29 100644 +--- a/arch/sparc/lib/rwsem_64.S ++++ b/arch/sparc/lib/rwsem_64.S +@@ -11,7 +11,12 @@ + .globl __down_read + __down_read: + 1: lduw [%o0], %g1 +- add %g1, 1, %g7 ++ addcc %g1, 1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 1b +@@ -33,7 +38,12 @@ __down_read: + .globl __down_read_trylock + __down_read_trylock: + 1: lduw [%o0], %g1 +- add %g1, 1, %g7 ++ addcc %g1, 1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cmp %g7, 0 + bl,pn %icc, 2f + mov 0, %o1 +@@ -51,7 +61,12 @@ __down_write: + or %g1, %lo(RWSEM_ACTIVE_WRITE_BIAS), %g1 + 1: + lduw [%o0], %g3 +- add %g3, %g1, %g7 ++ addcc %g3, %g1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g3, %g7 + cmp %g3, %g7 + bne,pn %icc, 1b +@@ -77,7 +92,12 @@ __down_write_trylock: + cmp %g3, 0 + bne,pn %icc, 2f + mov 0, %o1 +- add %g3, %g1, %g7 ++ addcc %g3, %g1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g3, %g7 + cmp %g3, %g7 + bne,pn %icc, 1b +@@ -90,7 +110,12 @@ __down_write_trylock: + __up_read: + 1: + lduw [%o0], %g1 +- sub %g1, 1, %g7 ++ subcc %g1, 1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g1, %g7 + cmp %g1, %g7 + bne,pn %icc, 1b +@@ -118,7 +143,12 @@ __up_write: + or %g1, %lo(RWSEM_ACTIVE_WRITE_BIAS), %g1 + 1: + lduw [%o0], %g3 +- sub %g3, %g1, %g7 ++ subcc %g3, %g1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g3, %g7 + cmp %g3, %g7 + bne,pn %icc, 1b +@@ -143,7 +173,12 @@ __downgrade_write: + or %g1, %lo(RWSEM_WAITING_BIAS), %g1 + 1: + lduw [%o0], %g3 +- sub %g3, %g1, %g7 ++ subcc %g3, %g1, %g7 ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ tvs %icc, 6 ++#endif ++ + cas [%o0], %g3, %g7 + cmp %g3, %g7 + bne,pn %icc, 1b +diff --git a/arch/sparc/mm/Makefile b/arch/sparc/mm/Makefile +index 79836a7..62f47a2 100644 +--- a/arch/sparc/mm/Makefile ++++ b/arch/sparc/mm/Makefile +@@ -2,7 +2,7 @@ + # + + asflags-y := -ansi +-ccflags-y := -Werror ++#ccflags-y := -Werror + + obj-$(CONFIG_SPARC64) += ultra.o tlb.o tsb.o + obj-y += fault_$(BITS).o +diff --git a/arch/sparc/mm/fault_32.c b/arch/sparc/mm/fault_32.c +index b99f81c..3453e93 100644 +--- a/arch/sparc/mm/fault_32.c ++++ b/arch/sparc/mm/fault_32.c +@@ -21,6 +21,9 @@ + #include + #include + #include ++#include ++#include ++#include + + #include + #include +@@ -167,6 +170,267 @@ static unsigned long compute_si_addr(struct pt_regs *regs, int text_fault) + return safe_compute_effective_address(regs, insn); + } + ++#ifdef CONFIG_PAX_PAGEEXEC ++#ifdef CONFIG_PAX_DLRESOLVE ++static void pax_emuplt_close(struct vm_area_struct *vma) ++{ ++ vma->vm_mm->call_dl_resolve = 0UL; ++} ++ ++static int pax_emuplt_fault(struct vm_area_struct *vma, struct vm_fault *vmf) ++{ ++ unsigned int *kaddr; ++ ++ vmf->page = alloc_page(GFP_HIGHUSER); ++ if (!vmf->page) ++ return VM_FAULT_OOM; ++ ++ kaddr = kmap(vmf->page); ++ memset(kaddr, 0, PAGE_SIZE); ++ kaddr[0] = 0x9DE3BFA8U; /* save */ ++ flush_dcache_page(vmf->page); ++ kunmap(vmf->page); ++ return VM_FAULT_MAJOR; ++} ++ ++static const struct vm_operations_struct pax_vm_ops = { ++ .close = pax_emuplt_close, ++ .fault = pax_emuplt_fault ++}; ++ ++static int pax_insert_vma(struct vm_area_struct *vma, unsigned long addr) ++{ ++ int ret; ++ ++ vma->vm_mm = current->mm; ++ vma->vm_start = addr; ++ vma->vm_end = addr + PAGE_SIZE; ++ vma->vm_flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYEXEC; ++ vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); ++ vma->vm_ops = &pax_vm_ops; ++ ++ ret = insert_vm_struct(current->mm, vma); ++ if (ret) ++ return ret; ++ ++ ++current->mm->total_vm; ++ return 0; ++} ++#endif ++ ++/* ++ * PaX: decide what to do with offenders (regs->pc = fault address) ++ * ++ * returns 1 when task should be killed ++ * 2 when patched PLT trampoline was detected ++ * 3 when unpatched PLT trampoline was detected ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ ++#ifdef CONFIG_PAX_EMUPLT ++ int err; ++ ++ do { /* PaX: patched PLT emulation #1 */ ++ unsigned int sethi1, sethi2, jmpl; ++ ++ err = get_user(sethi1, (unsigned int *)regs->pc); ++ err |= get_user(sethi2, (unsigned int *)(regs->pc+4)); ++ err |= get_user(jmpl, (unsigned int *)(regs->pc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi1 & 0xFFC00000U) == 0x03000000U && ++ (sethi2 & 0xFFC00000U) == 0x03000000U && ++ (jmpl & 0xFFFFE000U) == 0x81C06000U) ++ { ++ unsigned int addr; ++ ++ regs->u_regs[UREG_G1] = (sethi2 & 0x003FFFFFU) << 10; ++ addr = regs->u_regs[UREG_G1]; ++ addr += (((jmpl | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U); ++ regs->pc = addr; ++ regs->npc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ { /* PaX: patched PLT emulation #2 */ ++ unsigned int ba; ++ ++ err = get_user(ba, (unsigned int *)regs->pc); ++ ++ if (!err && (ba & 0xFFC00000U) == 0x30800000U) { ++ unsigned int addr; ++ ++ addr = regs->pc + ((((ba | 0xFFC00000U) ^ 0x00200000U) + 0x00200000U) << 2); ++ regs->pc = addr; ++ regs->npc = addr+4; ++ return 2; ++ } ++ } ++ ++ do { /* PaX: patched PLT emulation #3 */ ++ unsigned int sethi, jmpl, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->pc); ++ err |= get_user(jmpl, (unsigned int *)(regs->pc+4)); ++ err |= get_user(nop, (unsigned int *)(regs->pc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ (jmpl & 0xFFFFE000U) == 0x81C06000U && ++ nop == 0x01000000U) ++ { ++ unsigned int addr; ++ ++ addr = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G1] = addr; ++ addr += (((jmpl | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U); ++ regs->pc = addr; ++ regs->npc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: unpatched PLT emulation step 1 */ ++ unsigned int sethi, ba, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->pc); ++ err |= get_user(ba, (unsigned int *)(regs->pc+4)); ++ err |= get_user(nop, (unsigned int *)(regs->pc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ ((ba & 0xFFC00000U) == 0x30800000U || (ba & 0xFFF80000U) == 0x30680000U) && ++ nop == 0x01000000U) ++ { ++ unsigned int addr, save, call; ++ ++ if ((ba & 0xFFC00000U) == 0x30800000U) ++ addr = regs->pc + 4 + ((((ba | 0xFFC00000U) ^ 0x00200000U) + 0x00200000U) << 2); ++ else ++ addr = regs->pc + 4 + ((((ba | 0xFFF80000U) ^ 0x00040000U) + 0x00040000U) << 2); ++ ++ err = get_user(save, (unsigned int *)addr); ++ err |= get_user(call, (unsigned int *)(addr+4)); ++ err |= get_user(nop, (unsigned int *)(addr+8)); ++ if (err) ++ break; ++ ++#ifdef CONFIG_PAX_DLRESOLVE ++ if (save == 0x9DE3BFA8U && ++ (call & 0xC0000000U) == 0x40000000U && ++ nop == 0x01000000U) ++ { ++ struct vm_area_struct *vma; ++ unsigned long call_dl_resolve; ++ ++ down_read(¤t->mm->mmap_sem); ++ call_dl_resolve = current->mm->call_dl_resolve; ++ up_read(¤t->mm->mmap_sem); ++ if (likely(call_dl_resolve)) ++ goto emulate; ++ ++ vma = kmem_cache_zalloc(vm_area_cachep, GFP_KERNEL); ++ ++ down_write(¤t->mm->mmap_sem); ++ if (current->mm->call_dl_resolve) { ++ call_dl_resolve = current->mm->call_dl_resolve; ++ up_write(¤t->mm->mmap_sem); ++ if (vma) ++ kmem_cache_free(vm_area_cachep, vma); ++ goto emulate; ++ } ++ ++ call_dl_resolve = get_unmapped_area(NULL, 0UL, PAGE_SIZE, 0UL, MAP_PRIVATE); ++ if (!vma || (call_dl_resolve & ~PAGE_MASK)) { ++ up_write(¤t->mm->mmap_sem); ++ if (vma) ++ kmem_cache_free(vm_area_cachep, vma); ++ return 1; ++ } ++ ++ if (pax_insert_vma(vma, call_dl_resolve)) { ++ up_write(¤t->mm->mmap_sem); ++ kmem_cache_free(vm_area_cachep, vma); ++ return 1; ++ } ++ ++ current->mm->call_dl_resolve = call_dl_resolve; ++ up_write(¤t->mm->mmap_sem); ++ ++emulate: ++ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10; ++ regs->pc = call_dl_resolve; ++ regs->npc = addr+4; ++ return 3; ++ } ++#endif ++ ++ /* PaX: glibc 2.4+ generates sethi/jmpl instead of save/call */ ++ if ((save & 0xFFC00000U) == 0x05000000U && ++ (call & 0xFFFFE000U) == 0x85C0A000U && ++ nop == 0x01000000U) ++ { ++ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G2] = addr + 4; ++ addr = (save & 0x003FFFFFU) << 10; ++ addr += (((call | 0xFFFFE000U) ^ 0x00001000U) + 0x00001000U); ++ regs->pc = addr; ++ regs->npc = addr+4; ++ return 3; ++ } ++ } ++ } while (0); ++ ++ do { /* PaX: unpatched PLT emulation step 2 */ ++ unsigned int save, call, nop; ++ ++ err = get_user(save, (unsigned int *)(regs->pc-4)); ++ err |= get_user(call, (unsigned int *)regs->pc); ++ err |= get_user(nop, (unsigned int *)(regs->pc+4)); ++ if (err) ++ break; ++ ++ if (save == 0x9DE3BFA8U && ++ (call & 0xC0000000U) == 0x40000000U && ++ nop == 0x01000000U) ++ { ++ unsigned int dl_resolve = regs->pc + ((((call | 0xC0000000U) ^ 0x20000000U) + 0x20000000U) << 2); ++ ++ regs->u_regs[UREG_RETPC] = regs->pc; ++ regs->pc = dl_resolve; ++ regs->npc = dl_resolve+4; ++ return 3; ++ } ++ } while (0); ++#endif ++ ++ return 1; ++} ++ ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 8; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + asmlinkage void do_sparc_fault(struct pt_regs *regs, int text_fault, int write, + unsigned long address) + { +@@ -231,6 +495,24 @@ good_area: + if(!(vma->vm_flags & VM_WRITE)) + goto bad_area; + } else { ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && text_fault && !(vma->vm_flags & VM_EXEC)) { ++ up_read(&mm->mmap_sem); ++ switch (pax_handle_fetch_fault(regs)) { ++ ++#ifdef CONFIG_PAX_EMUPLT ++ case 2: ++ case 3: ++ return; ++#endif ++ ++ } ++ pax_report_fault(regs, (void *)regs->pc, (void *)regs->u_regs[UREG_FP]); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + /* Allow reads even for write-only mappings */ + if(!(vma->vm_flags & (VM_READ | VM_EXEC))) + goto bad_area; +diff --git a/arch/sparc/mm/fault_64.c b/arch/sparc/mm/fault_64.c +index 43b0da9..a0b78f9 100644 +--- a/arch/sparc/mm/fault_64.c ++++ b/arch/sparc/mm/fault_64.c +@@ -20,6 +20,9 @@ + #include + #include + #include ++#include ++#include ++#include + + #include + #include +@@ -78,7 +81,7 @@ static void bad_kernel_pc(struct pt_regs *regs, unsigned long vaddr) + printk(KERN_CRIT "OOPS: Bogus kernel PC [%016lx] in fault handler\n", + regs->tpc); + printk(KERN_CRIT "OOPS: RPC [%016lx]\n", regs->u_regs[15]); +- printk("OOPS: RPC <%pS>\n", (void *) regs->u_regs[15]); ++ printk("OOPS: RPC <%pA>\n", (void *) regs->u_regs[15]); + printk(KERN_CRIT "OOPS: Fault was to vaddr[%lx]\n", vaddr); + dump_stack(); + unhandled_fault(regs->tpc, current, regs); +@@ -249,6 +252,456 @@ static void noinline bogus_32bit_fault_address(struct pt_regs *regs, + show_regs(regs); + } + ++#ifdef CONFIG_PAX_PAGEEXEC ++#ifdef CONFIG_PAX_DLRESOLVE ++static void pax_emuplt_close(struct vm_area_struct *vma) ++{ ++ vma->vm_mm->call_dl_resolve = 0UL; ++} ++ ++static int pax_emuplt_fault(struct vm_area_struct *vma, struct vm_fault *vmf) ++{ ++ unsigned int *kaddr; ++ ++ vmf->page = alloc_page(GFP_HIGHUSER); ++ if (!vmf->page) ++ return VM_FAULT_OOM; ++ ++ kaddr = kmap(vmf->page); ++ memset(kaddr, 0, PAGE_SIZE); ++ kaddr[0] = 0x9DE3BFA8U; /* save */ ++ flush_dcache_page(vmf->page); ++ kunmap(vmf->page); ++ return VM_FAULT_MAJOR; ++} ++ ++static const struct vm_operations_struct pax_vm_ops = { ++ .close = pax_emuplt_close, ++ .fault = pax_emuplt_fault ++}; ++ ++static int pax_insert_vma(struct vm_area_struct *vma, unsigned long addr) ++{ ++ int ret; ++ ++ vma->vm_mm = current->mm; ++ vma->vm_start = addr; ++ vma->vm_end = addr + PAGE_SIZE; ++ vma->vm_flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYEXEC; ++ vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); ++ vma->vm_ops = &pax_vm_ops; ++ ++ ret = insert_vm_struct(current->mm, vma); ++ if (ret) ++ return ret; ++ ++ ++current->mm->total_vm; ++ return 0; ++} ++#endif ++ ++/* ++ * PaX: decide what to do with offenders (regs->tpc = fault address) ++ * ++ * returns 1 when task should be killed ++ * 2 when patched PLT trampoline was detected ++ * 3 when unpatched PLT trampoline was detected ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ ++#ifdef CONFIG_PAX_EMUPLT ++ int err; ++ ++ do { /* PaX: patched PLT emulation #1 */ ++ unsigned int sethi1, sethi2, jmpl; ++ ++ err = get_user(sethi1, (unsigned int *)regs->tpc); ++ err |= get_user(sethi2, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(jmpl, (unsigned int *)(regs->tpc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi1 & 0xFFC00000U) == 0x03000000U && ++ (sethi2 & 0xFFC00000U) == 0x03000000U && ++ (jmpl & 0xFFFFE000U) == 0x81C06000U) ++ { ++ unsigned long addr; ++ ++ regs->u_regs[UREG_G1] = (sethi2 & 0x003FFFFFU) << 10; ++ addr = regs->u_regs[UREG_G1]; ++ addr += (((jmpl | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ { /* PaX: patched PLT emulation #2 */ ++ unsigned int ba; ++ ++ err = get_user(ba, (unsigned int *)regs->tpc); ++ ++ if (!err && (ba & 0xFFC00000U) == 0x30800000U) { ++ unsigned long addr; ++ ++ addr = regs->tpc + ((((ba | 0xFFFFFFFFFFC00000UL) ^ 0x00200000UL) + 0x00200000UL) << 2); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } ++ ++ do { /* PaX: patched PLT emulation #3 */ ++ unsigned int sethi, jmpl, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(jmpl, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ (jmpl & 0xFFFFE000U) == 0x81C06000U && ++ nop == 0x01000000U) ++ { ++ unsigned long addr; ++ ++ addr = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G1] = addr; ++ addr += (((jmpl | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: patched PLT emulation #4 */ ++ unsigned int sethi, mov1, call, mov2; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(mov1, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(call, (unsigned int *)(regs->tpc+8)); ++ err |= get_user(mov2, (unsigned int *)(regs->tpc+12)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ mov1 == 0x8210000FU && ++ (call & 0xC0000000U) == 0x40000000U && ++ mov2 == 0x9E100001U) ++ { ++ unsigned long addr; ++ ++ regs->u_regs[UREG_G1] = regs->u_regs[UREG_RETPC]; ++ addr = regs->tpc + 4 + ((((call | 0xFFFFFFFFC0000000UL) ^ 0x20000000UL) + 0x20000000UL) << 2); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: patched PLT emulation #5 */ ++ unsigned int sethi, sethi1, sethi2, or1, or2, sllx, jmpl, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(sethi1, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(sethi2, (unsigned int *)(regs->tpc+8)); ++ err |= get_user(or1, (unsigned int *)(regs->tpc+12)); ++ err |= get_user(or2, (unsigned int *)(regs->tpc+16)); ++ err |= get_user(sllx, (unsigned int *)(regs->tpc+20)); ++ err |= get_user(jmpl, (unsigned int *)(regs->tpc+24)); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+28)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ (sethi1 & 0xFFC00000U) == 0x03000000U && ++ (sethi2 & 0xFFC00000U) == 0x0B000000U && ++ (or1 & 0xFFFFE000U) == 0x82106000U && ++ (or2 & 0xFFFFE000U) == 0x8A116000U && ++ sllx == 0x83287020U && ++ jmpl == 0x81C04005U && ++ nop == 0x01000000U) ++ { ++ unsigned long addr; ++ ++ regs->u_regs[UREG_G1] = ((sethi1 & 0x003FFFFFU) << 10) | (or1 & 0x000003FFU); ++ regs->u_regs[UREG_G1] <<= 32; ++ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or2 & 0x000003FFU); ++ addr = regs->u_regs[UREG_G1] + regs->u_regs[UREG_G5]; ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: patched PLT emulation #6 */ ++ unsigned int sethi, sethi1, sethi2, sllx, or, jmpl, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(sethi1, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(sethi2, (unsigned int *)(regs->tpc+8)); ++ err |= get_user(sllx, (unsigned int *)(regs->tpc+12)); ++ err |= get_user(or, (unsigned int *)(regs->tpc+16)); ++ err |= get_user(jmpl, (unsigned int *)(regs->tpc+20)); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+24)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ (sethi1 & 0xFFC00000U) == 0x03000000U && ++ (sethi2 & 0xFFC00000U) == 0x0B000000U && ++ sllx == 0x83287020U && ++ (or & 0xFFFFE000U) == 0x8A116000U && ++ jmpl == 0x81C04005U && ++ nop == 0x01000000U) ++ { ++ unsigned long addr; ++ ++ regs->u_regs[UREG_G1] = (sethi1 & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G1] <<= 32; ++ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or & 0x3FFU); ++ addr = regs->u_regs[UREG_G1] + regs->u_regs[UREG_G5]; ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: unpatched PLT emulation step 1 */ ++ unsigned int sethi, ba, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(ba, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ ((ba & 0xFFC00000U) == 0x30800000U || (ba & 0xFFF80000U) == 0x30680000U) && ++ nop == 0x01000000U) ++ { ++ unsigned long addr; ++ unsigned int save, call; ++ unsigned int sethi1, sethi2, or1, or2, sllx, add, jmpl; ++ ++ if ((ba & 0xFFC00000U) == 0x30800000U) ++ addr = regs->tpc + 4 + ((((ba | 0xFFFFFFFFFFC00000UL) ^ 0x00200000UL) + 0x00200000UL) << 2); ++ else ++ addr = regs->tpc + 4 + ((((ba | 0xFFFFFFFFFFF80000UL) ^ 0x00040000UL) + 0x00040000UL) << 2); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ err = get_user(save, (unsigned int *)addr); ++ err |= get_user(call, (unsigned int *)(addr+4)); ++ err |= get_user(nop, (unsigned int *)(addr+8)); ++ if (err) ++ break; ++ ++#ifdef CONFIG_PAX_DLRESOLVE ++ if (save == 0x9DE3BFA8U && ++ (call & 0xC0000000U) == 0x40000000U && ++ nop == 0x01000000U) ++ { ++ struct vm_area_struct *vma; ++ unsigned long call_dl_resolve; ++ ++ down_read(¤t->mm->mmap_sem); ++ call_dl_resolve = current->mm->call_dl_resolve; ++ up_read(¤t->mm->mmap_sem); ++ if (likely(call_dl_resolve)) ++ goto emulate; ++ ++ vma = kmem_cache_zalloc(vm_area_cachep, GFP_KERNEL); ++ ++ down_write(¤t->mm->mmap_sem); ++ if (current->mm->call_dl_resolve) { ++ call_dl_resolve = current->mm->call_dl_resolve; ++ up_write(¤t->mm->mmap_sem); ++ if (vma) ++ kmem_cache_free(vm_area_cachep, vma); ++ goto emulate; ++ } ++ ++ call_dl_resolve = get_unmapped_area(NULL, 0UL, PAGE_SIZE, 0UL, MAP_PRIVATE); ++ if (!vma || (call_dl_resolve & ~PAGE_MASK)) { ++ up_write(¤t->mm->mmap_sem); ++ if (vma) ++ kmem_cache_free(vm_area_cachep, vma); ++ return 1; ++ } ++ ++ if (pax_insert_vma(vma, call_dl_resolve)) { ++ up_write(¤t->mm->mmap_sem); ++ kmem_cache_free(vm_area_cachep, vma); ++ return 1; ++ } ++ ++ current->mm->call_dl_resolve = call_dl_resolve; ++ up_write(¤t->mm->mmap_sem); ++ ++emulate: ++ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10; ++ regs->tpc = call_dl_resolve; ++ regs->tnpc = addr+4; ++ return 3; ++ } ++#endif ++ ++ /* PaX: glibc 2.4+ generates sethi/jmpl instead of save/call */ ++ if ((save & 0xFFC00000U) == 0x05000000U && ++ (call & 0xFFFFE000U) == 0x85C0A000U && ++ nop == 0x01000000U) ++ { ++ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G2] = addr + 4; ++ addr = (save & 0x003FFFFFU) << 10; ++ addr += (((call | 0xFFFFFFFFFFFFE000UL) ^ 0x00001000UL) + 0x00001000UL); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 3; ++ } ++ ++ /* PaX: 64-bit PLT stub */ ++ err = get_user(sethi1, (unsigned int *)addr); ++ err |= get_user(sethi2, (unsigned int *)(addr+4)); ++ err |= get_user(or1, (unsigned int *)(addr+8)); ++ err |= get_user(or2, (unsigned int *)(addr+12)); ++ err |= get_user(sllx, (unsigned int *)(addr+16)); ++ err |= get_user(add, (unsigned int *)(addr+20)); ++ err |= get_user(jmpl, (unsigned int *)(addr+24)); ++ err |= get_user(nop, (unsigned int *)(addr+28)); ++ if (err) ++ break; ++ ++ if ((sethi1 & 0xFFC00000U) == 0x09000000U && ++ (sethi2 & 0xFFC00000U) == 0x0B000000U && ++ (or1 & 0xFFFFE000U) == 0x88112000U && ++ (or2 & 0xFFFFE000U) == 0x8A116000U && ++ sllx == 0x89293020U && ++ add == 0x8A010005U && ++ jmpl == 0x89C14000U && ++ nop == 0x01000000U) ++ { ++ regs->u_regs[UREG_G1] = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G4] = ((sethi1 & 0x003FFFFFU) << 10) | (or1 & 0x000003FFU); ++ regs->u_regs[UREG_G4] <<= 32; ++ regs->u_regs[UREG_G5] = ((sethi2 & 0x003FFFFFU) << 10) | (or2 & 0x000003FFU); ++ regs->u_regs[UREG_G5] += regs->u_regs[UREG_G4]; ++ regs->u_regs[UREG_G4] = addr + 24; ++ addr = regs->u_regs[UREG_G5]; ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 3; ++ } ++ } ++ } while (0); ++ ++#ifdef CONFIG_PAX_DLRESOLVE ++ do { /* PaX: unpatched PLT emulation step 2 */ ++ unsigned int save, call, nop; ++ ++ err = get_user(save, (unsigned int *)(regs->tpc-4)); ++ err |= get_user(call, (unsigned int *)regs->tpc); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+4)); ++ if (err) ++ break; ++ ++ if (save == 0x9DE3BFA8U && ++ (call & 0xC0000000U) == 0x40000000U && ++ nop == 0x01000000U) ++ { ++ unsigned long dl_resolve = regs->tpc + ((((call | 0xFFFFFFFFC0000000UL) ^ 0x20000000UL) + 0x20000000UL) << 2); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ dl_resolve &= 0xFFFFFFFFUL; ++ ++ regs->u_regs[UREG_RETPC] = regs->tpc; ++ regs->tpc = dl_resolve; ++ regs->tnpc = dl_resolve+4; ++ return 3; ++ } ++ } while (0); ++#endif ++ ++ do { /* PaX: patched PLT emulation #7, must be AFTER the unpatched PLT emulation */ ++ unsigned int sethi, ba, nop; ++ ++ err = get_user(sethi, (unsigned int *)regs->tpc); ++ err |= get_user(ba, (unsigned int *)(regs->tpc+4)); ++ err |= get_user(nop, (unsigned int *)(regs->tpc+8)); ++ ++ if (err) ++ break; ++ ++ if ((sethi & 0xFFC00000U) == 0x03000000U && ++ (ba & 0xFFF00000U) == 0x30600000U && ++ nop == 0x01000000U) ++ { ++ unsigned long addr; ++ ++ addr = (sethi & 0x003FFFFFU) << 10; ++ regs->u_regs[UREG_G1] = addr; ++ addr = regs->tpc + ((((ba | 0xFFFFFFFFFFF80000UL) ^ 0x00040000UL) + 0x00040000UL) << 2); ++ ++ if (test_thread_flag(TIF_32BIT)) ++ addr &= 0xFFFFFFFFUL; ++ ++ regs->tpc = addr; ++ regs->tnpc = addr+4; ++ return 2; ++ } ++ } while (0); ++ ++#endif ++ ++ return 1; ++} ++ ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ unsigned long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 8; i++) { ++ unsigned int c; ++ if (get_user(c, (unsigned int *)pc+i)) ++ printk(KERN_CONT "???????? "); ++ else ++ printk(KERN_CONT "%08x ", c); ++ } ++ printk("\n"); ++} ++#endif ++ + asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs) + { + struct mm_struct *mm = current->mm; +@@ -315,6 +768,29 @@ asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs) + if (!vma) + goto bad_area; + ++#ifdef CONFIG_PAX_PAGEEXEC ++ /* PaX: detect ITLB misses on non-exec pages */ ++ if ((mm->pax_flags & MF_PAX_PAGEEXEC) && vma->vm_start <= address && ++ !(vma->vm_flags & VM_EXEC) && (fault_code & FAULT_CODE_ITLB)) ++ { ++ if (address != regs->tpc) ++ goto good_area; ++ ++ up_read(&mm->mmap_sem); ++ switch (pax_handle_fetch_fault(regs)) { ++ ++#ifdef CONFIG_PAX_EMUPLT ++ case 2: ++ case 3: ++ return; ++#endif ++ ++ } ++ pax_report_fault(regs, (void *)regs->tpc, (void *)(regs->u_regs[UREG_FP] + STACK_BIAS)); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + /* Pure DTLB misses do not tell us whether the fault causing + * load/store/atomic was a write or not, it only says that there + * was no match. So in such a case we (carefully) read the +diff --git a/arch/sparc/mm/hugetlbpage.c b/arch/sparc/mm/hugetlbpage.c +index f27d103..1b06377 100644 +--- a/arch/sparc/mm/hugetlbpage.c ++++ b/arch/sparc/mm/hugetlbpage.c +@@ -69,7 +69,7 @@ full_search: + } + return -ENOMEM; + } +- if (likely(!vma || addr + len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* + * Remember the place where we stopped the search: + */ +@@ -108,7 +108,7 @@ hugetlb_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + /* make sure it can fit in the remaining address space */ + if (likely(addr > len)) { + vma = find_vma(mm, addr-len); +- if (!vma || addr <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr - len, len)) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr-len); + } +@@ -117,16 +117,17 @@ hugetlb_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + if (unlikely(mm->mmap_base < len)) + goto bottomup; + +- addr = (mm->mmap_base-len) & HPAGE_MASK; ++ addr = mm->mmap_base - len; + + do { ++ addr &= HPAGE_MASK; + /* + * Lookup failure means no vma is above this address, + * else if new region fits below vma->vm_start, + * return with success: + */ + vma = find_vma(mm, addr); +- if (likely(!vma || addr+len <= vma->vm_start)) { ++ if (likely(check_heap_stack_gap(vma, addr, len))) { + /* remember the address as a hint for next time */ + return (mm->free_area_cache = addr); + } +@@ -136,8 +137,8 @@ hugetlb_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + mm->cached_hole_size = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = (vma->vm_start-len) & HPAGE_MASK; +- } while (likely(len < vma->vm_start)); ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); + + bottomup: + /* +@@ -183,8 +184,7 @@ hugetlb_get_unmapped_area(struct file *file, unsigned long addr, + if (addr) { + addr = ALIGN(addr, HPAGE_SIZE); + vma = find_vma(mm, addr); +- if (task_size - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (task_size - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + if (mm->get_unmapped_area == arch_get_unmapped_area) +diff --git a/arch/sparc/mm/init_32.c b/arch/sparc/mm/init_32.c +index dc7c3b1..34c0070 100644 +--- a/arch/sparc/mm/init_32.c ++++ b/arch/sparc/mm/init_32.c +@@ -317,6 +317,9 @@ extern void device_scan(void); + pgprot_t PAGE_SHARED __read_mostly; + EXPORT_SYMBOL(PAGE_SHARED); + ++pgprot_t PAGE_SHARED_NOEXEC __read_mostly; ++EXPORT_SYMBOL(PAGE_SHARED_NOEXEC); ++ + void __init paging_init(void) + { + switch(sparc_cpu_model) { +@@ -345,17 +348,17 @@ void __init paging_init(void) + + /* Initialize the protection map with non-constant, MMU dependent values. */ + protection_map[0] = PAGE_NONE; +- protection_map[1] = PAGE_READONLY; +- protection_map[2] = PAGE_COPY; +- protection_map[3] = PAGE_COPY; ++ protection_map[1] = PAGE_READONLY_NOEXEC; ++ protection_map[2] = PAGE_COPY_NOEXEC; ++ protection_map[3] = PAGE_COPY_NOEXEC; + protection_map[4] = PAGE_READONLY; + protection_map[5] = PAGE_READONLY; + protection_map[6] = PAGE_COPY; + protection_map[7] = PAGE_COPY; + protection_map[8] = PAGE_NONE; +- protection_map[9] = PAGE_READONLY; +- protection_map[10] = PAGE_SHARED; +- protection_map[11] = PAGE_SHARED; ++ protection_map[9] = PAGE_READONLY_NOEXEC; ++ protection_map[10] = PAGE_SHARED_NOEXEC; ++ protection_map[11] = PAGE_SHARED_NOEXEC; + protection_map[12] = PAGE_READONLY; + protection_map[13] = PAGE_READONLY; + protection_map[14] = PAGE_SHARED; +diff --git a/arch/sparc/mm/srmmu.c b/arch/sparc/mm/srmmu.c +index 509b1ff..bfd7118 100644 +--- a/arch/sparc/mm/srmmu.c ++++ b/arch/sparc/mm/srmmu.c +@@ -2200,6 +2200,13 @@ void __init ld_mmu_srmmu(void) + PAGE_SHARED = pgprot_val(SRMMU_PAGE_SHARED); + BTFIXUPSET_INT(page_copy, pgprot_val(SRMMU_PAGE_COPY)); + BTFIXUPSET_INT(page_readonly, pgprot_val(SRMMU_PAGE_RDONLY)); ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ PAGE_SHARED_NOEXEC = pgprot_val(SRMMU_PAGE_SHARED_NOEXEC); ++ BTFIXUPSET_INT(page_copy_noexec, pgprot_val(SRMMU_PAGE_COPY_NOEXEC)); ++ BTFIXUPSET_INT(page_readonly_noexec, pgprot_val(SRMMU_PAGE_RDONLY_NOEXEC)); ++#endif ++ + BTFIXUPSET_INT(page_kernel, pgprot_val(SRMMU_PAGE_KERNEL)); + page_kernel = pgprot_val(SRMMU_PAGE_KERNEL); + +diff --git a/arch/um/Makefile b/arch/um/Makefile +index fc633db..5e1a1c2 100644 +--- a/arch/um/Makefile ++++ b/arch/um/Makefile +@@ -49,6 +49,10 @@ USER_CFLAGS = $(patsubst $(KERNEL_DEFINES),,$(patsubst -D__KERNEL__,,\ + $(patsubst -I%,,$(KBUILD_CFLAGS)))) $(ARCH_INCLUDE) $(MODE_INCLUDE) \ + $(filter -I%,$(CFLAGS)) -D_FILE_OFFSET_BITS=64 + ++ifdef CONSTIFY_PLUGIN ++USER_CFLAGS += $(CONSTIFY_PLUGIN) -fplugin-arg-constify_plugin-no-constify ++endif ++ + include $(srctree)/$(ARCH_DIR)/Makefile-$(SUBARCH) + + #This will adjust *FLAGS accordingly to the platform. +diff --git a/arch/um/include/asm/cache.h b/arch/um/include/asm/cache.h +index 19e1bdd..3665b77 100644 +--- a/arch/um/include/asm/cache.h ++++ b/arch/um/include/asm/cache.h +@@ -1,6 +1,7 @@ + #ifndef __UM_CACHE_H + #define __UM_CACHE_H + ++#include + + #if defined(CONFIG_UML_X86) && !defined(CONFIG_64BIT) + # define L1_CACHE_SHIFT (CONFIG_X86_L1_CACHE_SHIFT) +@@ -12,6 +13,6 @@ + # define L1_CACHE_SHIFT 5 + #endif + +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #endif +diff --git a/arch/um/include/asm/kmap_types.h b/arch/um/include/asm/kmap_types.h +index 6c03acd..a5e0215 100644 +--- a/arch/um/include/asm/kmap_types.h ++++ b/arch/um/include/asm/kmap_types.h +@@ -23,6 +23,7 @@ enum km_type { + KM_IRQ1, + KM_SOFTIRQ0, + KM_SOFTIRQ1, ++ KM_CLEARPAGE, + KM_TYPE_NR + }; + +diff --git a/arch/um/include/asm/page.h b/arch/um/include/asm/page.h +index 4cc9b6c..02e5029 100644 +--- a/arch/um/include/asm/page.h ++++ b/arch/um/include/asm/page.h +@@ -14,6 +14,9 @@ + #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) + #define PAGE_MASK (~(PAGE_SIZE-1)) + ++#define ktla_ktva(addr) (addr) ++#define ktva_ktla(addr) (addr) ++ + #ifndef __ASSEMBLY__ + + struct page; +diff --git a/arch/um/kernel/process.c b/arch/um/kernel/process.c +index 4a28a15..654dc2a 100644 +--- a/arch/um/kernel/process.c ++++ b/arch/um/kernel/process.c +@@ -393,22 +393,6 @@ int singlestepping(void * t) + return 2; + } + +-/* +- * Only x86 and x86_64 have an arch_align_stack(). +- * All other arches have "#define arch_align_stack(x) (x)" +- * in their asm/system.h +- * As this is included in UML from asm-um/system-generic.h, +- * we can use it to behave as the subarch does. +- */ +-#ifndef arch_align_stack +-unsigned long arch_align_stack(unsigned long sp) +-{ +- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space) +- sp -= get_random_int() % 8192; +- return sp & ~0xf; +-} +-#endif +- + unsigned long get_wchan(struct task_struct *p) + { + unsigned long stack_page, sp, ip; +diff --git a/arch/um/sys-i386/shared/sysdep/system.h b/arch/um/sys-i386/shared/sysdep/system.h +index d1b93c4..ae1b7fd 100644 +--- a/arch/um/sys-i386/shared/sysdep/system.h ++++ b/arch/um/sys-i386/shared/sysdep/system.h +@@ -17,7 +17,7 @@ + # define AT_VECTOR_SIZE_ARCH 1 + #endif + +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) ((x) & ~0xfUL) + + void default_idle(void); + +diff --git a/arch/um/sys-i386/syscalls.c b/arch/um/sys-i386/syscalls.c +index 857ca0b..9a2669d 100644 +--- a/arch/um/sys-i386/syscalls.c ++++ b/arch/um/sys-i386/syscalls.c +@@ -11,6 +11,21 @@ + #include "asm/uaccess.h" + #include "asm/unistd.h" + ++int i386_mmap_check(unsigned long addr, unsigned long len, unsigned long flags) ++{ ++ unsigned long pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (current->mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ if (len > pax_task_size || addr > pax_task_size - len) ++ return -EINVAL; ++ ++ return 0; ++} ++ + /* + * Perform the select(nd, in, out, ex, tv) and mmap() system + * calls. Linux/i386 didn't use to be able to handle more than +diff --git a/arch/um/sys-x86_64/shared/sysdep/system.h b/arch/um/sys-x86_64/shared/sysdep/system.h +index d1b93c4..ae1b7fd 100644 +--- a/arch/um/sys-x86_64/shared/sysdep/system.h ++++ b/arch/um/sys-x86_64/shared/sysdep/system.h +@@ -17,7 +17,7 @@ + # define AT_VECTOR_SIZE_ARCH 1 + #endif + +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) ((x) & ~0xfUL) + + void default_idle(void); + +diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig +index 73ae02a..f932de5 100644 +--- a/arch/x86/Kconfig ++++ b/arch/x86/Kconfig +@@ -223,7 +223,7 @@ config X86_TRAMPOLINE + + config X86_32_LAZY_GS + def_bool y +- depends on X86_32 && !CC_STACKPROTECTOR ++ depends on X86_32 && !CC_STACKPROTECTOR && !PAX_MEMORY_UDEREF + + config KTIME_SCALAR + def_bool X86_32 +@@ -1008,7 +1008,7 @@ choice + + config NOHIGHMEM + bool "off" +- depends on !X86_NUMAQ ++ depends on !X86_NUMAQ && !(PAX_PAGEEXEC && PAX_ENABLE_PAE) + ---help--- + Linux can use up to 64 Gigabytes of physical memory on x86 systems. + However, the address space of 32-bit x86 processors is only 4 +@@ -1045,7 +1045,7 @@ config NOHIGHMEM + + config HIGHMEM4G + bool "4GB" +- depends on !X86_NUMAQ ++ depends on !X86_NUMAQ && !(PAX_PAGEEXEC && PAX_ENABLE_PAE) + ---help--- + Select this if you have a 32-bit processor and between 1 and 4 + gigabytes of physical RAM. +@@ -1099,7 +1099,7 @@ config PAGE_OFFSET + hex + default 0xB0000000 if VMSPLIT_3G_OPT + default 0x80000000 if VMSPLIT_2G +- default 0x78000000 if VMSPLIT_2G_OPT ++ default 0x70000000 if VMSPLIT_2G_OPT + default 0x40000000 if VMSPLIT_1G + default 0xC0000000 + depends on X86_32 +@@ -1460,6 +1460,7 @@ config SECCOMP + + config CC_STACKPROTECTOR + bool "Enable -fstack-protector buffer overflow detection (EXPERIMENTAL)" ++ depends on X86_64 || !PAX_MEMORY_UDEREF + ---help--- + This option turns on the -fstack-protector GCC feature. This + feature puts, at the beginning of functions, a canary value on +@@ -1517,6 +1518,7 @@ config KEXEC_JUMP + config PHYSICAL_START + hex "Physical address where the kernel is loaded" if (EMBEDDED || CRASH_DUMP) + default "0x1000000" ++ range 0x400000 0x40000000 + ---help--- + This gives the physical address where the kernel is loaded. + +@@ -1581,6 +1583,7 @@ config PHYSICAL_ALIGN + hex + prompt "Alignment value to which kernel should be aligned" if X86_32 + default "0x1000000" ++ range 0x400000 0x1000000 if PAX_KERNEXEC + range 0x2000 0x1000000 + ---help--- + This value puts the alignment restrictions on physical address +@@ -1612,9 +1615,10 @@ config HOTPLUG_CPU + Say N if you want to disable CPU hotplug. + + config COMPAT_VDSO +- def_bool y ++ def_bool n + prompt "Compat VDSO support" + depends on X86_32 || IA32_EMULATION ++ depends on !PAX_NOEXEC && !PAX_MEMORY_UDEREF + ---help--- + Map the 32-bit VDSO to the predictable old-style address too. + ---help--- +diff --git a/arch/x86/Kconfig.cpu b/arch/x86/Kconfig.cpu +index 0e566103..1a6b57e 100644 +--- a/arch/x86/Kconfig.cpu ++++ b/arch/x86/Kconfig.cpu +@@ -340,7 +340,7 @@ config X86_PPRO_FENCE + + config X86_F00F_BUG + def_bool y +- depends on M586MMX || M586TSC || M586 || M486 || M386 ++ depends on (M586MMX || M586TSC || M586 || M486 || M386) && !PAX_KERNEXEC + + config X86_WP_WORKS_OK + def_bool y +@@ -360,7 +360,7 @@ config X86_POPAD_OK + + config X86_ALIGNMENT_16 + def_bool y +- depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || X86_ELAN || MK6 || M586MMX || M586TSC || M586 || M486 || MVIAC3_2 || MGEODEGX1 ++ depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || X86_ELAN || MK8 || MK7 || MK6 || MCORE2 || MPENTIUM4 || MPENTIUMIII || MPENTIUMII || M686 || M586MMX || M586TSC || M586 || M486 || MVIAC3_2 || MGEODEGX1 + + config X86_INTEL_USERCOPY + def_bool y +@@ -406,7 +406,7 @@ config X86_CMPXCHG64 + # generates cmov. + config X86_CMOV + def_bool y +- depends on (MK8 || MK7 || MCORE2 || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MVIAC3_2 || MVIAC7 || MCRUSOE || MEFFICEON || X86_64 || MATOM) ++ depends on (MK8 || MK7 || MCORE2 || MPSC || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MVIAC3_2 || MVIAC7 || MCRUSOE || MEFFICEON || X86_64 || MATOM) + + config X86_MINIMUM_CPU_FAMILY + int +diff --git a/arch/x86/Kconfig.debug b/arch/x86/Kconfig.debug +index d105f29..c928727 100644 +--- a/arch/x86/Kconfig.debug ++++ b/arch/x86/Kconfig.debug +@@ -99,7 +99,7 @@ config X86_PTDUMP + config DEBUG_RODATA + bool "Write protect kernel read-only data structures" + default y +- depends on DEBUG_KERNEL ++ depends on DEBUG_KERNEL && BROKEN + ---help--- + Mark the kernel read-only data as write-protected in the pagetables, + in order to catch accidental (and incorrect) writes to such const +diff --git a/arch/x86/Makefile b/arch/x86/Makefile +index d2d24c9..0f21f8d 100644 +--- a/arch/x86/Makefile ++++ b/arch/x86/Makefile +@@ -44,6 +44,7 @@ ifeq ($(CONFIG_X86_32),y) + else + BITS := 64 + UTS_MACHINE := x86_64 ++ biarch := $(call cc-option,-m64) + CHECKFLAGS += -D__x86_64__ -m64 + + KBUILD_AFLAGS += -m64 +@@ -189,3 +190,12 @@ define archhelp + echo ' FDARGS="..." arguments for the booted kernel' + echo ' FDINITRD=file initrd for the booted kernel' + endef ++ ++define OLD_LD ++ ++*** ${VERSION}.${PATCHLEVEL} PaX kernels no longer build correctly with old versions of binutils. ++*** Please upgrade your binutils to 2.18 or newer ++endef ++ ++archprepare: ++ $(if $(LDFLAGS_BUILD_ID),,$(error $(OLD_LD))) +diff --git a/arch/x86/boot/Makefile b/arch/x86/boot/Makefile +index ec749c2..bbb5319 100644 +--- a/arch/x86/boot/Makefile ++++ b/arch/x86/boot/Makefile +@@ -69,6 +69,9 @@ KBUILD_CFLAGS := $(LINUXINCLUDE) -g -Os -D_SETUP -D__KERNEL__ \ + $(call cc-option, -fno-stack-protector) \ + $(call cc-option, -mpreferred-stack-boundary=2) + KBUILD_CFLAGS += $(call cc-option, -m32) ++ifdef CONSTIFY_PLUGIN ++KBUILD_CFLAGS += $(CONSTIFY_PLUGIN) -fplugin-arg-constify_plugin-no-constify ++endif + KBUILD_AFLAGS := $(KBUILD_CFLAGS) -D__ASSEMBLY__ + GCOV_PROFILE := n + +diff --git a/arch/x86/boot/bitops.h b/arch/x86/boot/bitops.h +index 878e4b9..20537ab 100644 +--- a/arch/x86/boot/bitops.h ++++ b/arch/x86/boot/bitops.h +@@ -26,7 +26,7 @@ static inline int variable_test_bit(int nr, const void *addr) + u8 v; + const u32 *p = (const u32 *)addr; + +- asm("btl %2,%1; setc %0" : "=qm" (v) : "m" (*p), "Ir" (nr)); ++ asm volatile("btl %2,%1; setc %0" : "=qm" (v) : "m" (*p), "Ir" (nr)); + return v; + } + +@@ -37,7 +37,7 @@ static inline int variable_test_bit(int nr, const void *addr) + + static inline void set_bit(int nr, void *addr) + { +- asm("btsl %1,%0" : "+m" (*(u32 *)addr) : "Ir" (nr)); ++ asm volatile("btsl %1,%0" : "+m" (*(u32 *)addr) : "Ir" (nr)); + } + + #endif /* BOOT_BITOPS_H */ +diff --git a/arch/x86/boot/boot.h b/arch/x86/boot/boot.h +index 98239d2..f40214c 100644 +--- a/arch/x86/boot/boot.h ++++ b/arch/x86/boot/boot.h +@@ -82,7 +82,7 @@ static inline void io_delay(void) + static inline u16 ds(void) + { + u16 seg; +- asm("movw %%ds,%0" : "=rm" (seg)); ++ asm volatile("movw %%ds,%0" : "=rm" (seg)); + return seg; + } + +@@ -178,7 +178,7 @@ static inline void wrgs32(u32 v, addr_t addr) + static inline int memcmp(const void *s1, const void *s2, size_t len) + { + u8 diff; +- asm("repe; cmpsb; setnz %0" ++ asm volatile("repe; cmpsb; setnz %0" + : "=qm" (diff), "+D" (s1), "+S" (s2), "+c" (len)); + return diff; + } +diff --git a/arch/x86/boot/compressed/Makefile b/arch/x86/boot/compressed/Makefile +index f8ed065..5bf5ff3 100644 +--- a/arch/x86/boot/compressed/Makefile ++++ b/arch/x86/boot/compressed/Makefile +@@ -13,6 +13,9 @@ cflags-$(CONFIG_X86_64) := -mcmodel=small + KBUILD_CFLAGS += $(cflags-y) + KBUILD_CFLAGS += $(call cc-option,-ffreestanding) + KBUILD_CFLAGS += $(call cc-option,-fno-stack-protector) ++ifdef CONSTIFY_PLUGIN ++KBUILD_CFLAGS += $(CONSTIFY_PLUGIN) -fplugin-arg-constify_plugin-no-constify ++endif + + KBUILD_AFLAGS := $(KBUILD_CFLAGS) -D__ASSEMBLY__ + GCOV_PROFILE := n +diff --git a/arch/x86/boot/compressed/head_32.S b/arch/x86/boot/compressed/head_32.S +index f543b70..b60fba8 100644 +--- a/arch/x86/boot/compressed/head_32.S ++++ b/arch/x86/boot/compressed/head_32.S +@@ -76,7 +76,7 @@ ENTRY(startup_32) + notl %eax + andl %eax, %ebx + #else +- movl $LOAD_PHYSICAL_ADDR, %ebx ++ movl $____LOAD_PHYSICAL_ADDR, %ebx + #endif + + /* Target address to relocate to for decompression */ +@@ -149,7 +149,7 @@ relocated: + * and where it was actually loaded. + */ + movl %ebp, %ebx +- subl $LOAD_PHYSICAL_ADDR, %ebx ++ subl $____LOAD_PHYSICAL_ADDR, %ebx + jz 2f /* Nothing to be done if loaded at compiled addr. */ + /* + * Process relocations. +@@ -157,8 +157,7 @@ relocated: + + 1: subl $4, %edi + movl (%edi), %ecx +- testl %ecx, %ecx +- jz 2f ++ jecxz 2f + addl %ebx, -__PAGE_OFFSET(%ebx, %ecx) + jmp 1b + 2: +diff --git a/arch/x86/boot/compressed/head_64.S b/arch/x86/boot/compressed/head_64.S +index 077e1b6..2c6b13b 100644 +--- a/arch/x86/boot/compressed/head_64.S ++++ b/arch/x86/boot/compressed/head_64.S +@@ -91,7 +91,7 @@ ENTRY(startup_32) + notl %eax + andl %eax, %ebx + #else +- movl $LOAD_PHYSICAL_ADDR, %ebx ++ movl $____LOAD_PHYSICAL_ADDR, %ebx + #endif + + /* Target address to relocate to for decompression */ +@@ -183,7 +183,7 @@ no_longmode: + hlt + jmp 1b + +-#include "../../kernel/verify_cpu_64.S" ++#include "../../kernel/verify_cpu.S" + + /* + * Be careful here startup_64 needs to be at a predictable +@@ -234,7 +234,7 @@ ENTRY(startup_64) + notq %rax + andq %rax, %rbp + #else +- movq $LOAD_PHYSICAL_ADDR, %rbp ++ movq $____LOAD_PHYSICAL_ADDR, %rbp + #endif + + /* Target address to relocate to for decompression */ +diff --git a/arch/x86/boot/compressed/misc.c b/arch/x86/boot/compressed/misc.c +index 842b2a3..f00178b 100644 +--- a/arch/x86/boot/compressed/misc.c ++++ b/arch/x86/boot/compressed/misc.c +@@ -288,7 +288,7 @@ static void parse_elf(void *output) + case PT_LOAD: + #ifdef CONFIG_RELOCATABLE + dest = output; +- dest += (phdr->p_paddr - LOAD_PHYSICAL_ADDR); ++ dest += (phdr->p_paddr - ____LOAD_PHYSICAL_ADDR); + #else + dest = (void *)(phdr->p_paddr); + #endif +@@ -335,7 +335,7 @@ asmlinkage void decompress_kernel(void *rmode, memptr heap, + error("Destination address too large"); + #endif + #ifndef CONFIG_RELOCATABLE +- if ((unsigned long)output != LOAD_PHYSICAL_ADDR) ++ if ((unsigned long)output != ____LOAD_PHYSICAL_ADDR) + error("Wrong destination address"); + #endif + +diff --git a/arch/x86/boot/compressed/mkpiggy.c b/arch/x86/boot/compressed/mkpiggy.c +index bcbd36c..b1754af 100644 +--- a/arch/x86/boot/compressed/mkpiggy.c ++++ b/arch/x86/boot/compressed/mkpiggy.c +@@ -74,7 +74,7 @@ int main(int argc, char *argv[]) + + offs = (olen > ilen) ? olen - ilen : 0; + offs += olen >> 12; /* Add 8 bytes for each 32K block */ +- offs += 32*1024 + 18; /* Add 32K + 18 bytes slack */ ++ offs += 64*1024; /* Add 64K bytes slack */ + offs = (offs+4095) & ~4095; /* Round to a 4K boundary */ + + printf(".section \".rodata.compressed\",\"a\",@progbits\n"); +diff --git a/arch/x86/boot/compressed/relocs.c b/arch/x86/boot/compressed/relocs.c +index bbeb0c3..f5167ab 100644 +--- a/arch/x86/boot/compressed/relocs.c ++++ b/arch/x86/boot/compressed/relocs.c +@@ -10,8 +10,11 @@ + #define USE_BSD + #include + ++#include "../../../../include/linux/autoconf.h" ++ + #define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0])) + static Elf32_Ehdr ehdr; ++static Elf32_Phdr *phdr; + static unsigned long reloc_count, reloc_idx; + static unsigned long *relocs; + +@@ -37,7 +40,7 @@ static const char* safe_abs_relocs[] = { + + static int is_safe_abs_reloc(const char* sym_name) + { +- int i; ++ unsigned int i; + + for (i = 0; i < ARRAY_SIZE(safe_abs_relocs); i++) { + if (!strcmp(sym_name, safe_abs_relocs[i])) +@@ -245,9 +248,39 @@ static void read_ehdr(FILE *fp) + } + } + ++static void read_phdrs(FILE *fp) ++{ ++ unsigned int i; ++ ++ phdr = calloc(ehdr.e_phnum, sizeof(Elf32_Phdr)); ++ if (!phdr) { ++ die("Unable to allocate %d program headers\n", ++ ehdr.e_phnum); ++ } ++ if (fseek(fp, ehdr.e_phoff, SEEK_SET) < 0) { ++ die("Seek to %d failed: %s\n", ++ ehdr.e_phoff, strerror(errno)); ++ } ++ if (fread(phdr, sizeof(*phdr), ehdr.e_phnum, fp) != ehdr.e_phnum) { ++ die("Cannot read ELF program headers: %s\n", ++ strerror(errno)); ++ } ++ for(i = 0; i < ehdr.e_phnum; i++) { ++ phdr[i].p_type = elf32_to_cpu(phdr[i].p_type); ++ phdr[i].p_offset = elf32_to_cpu(phdr[i].p_offset); ++ phdr[i].p_vaddr = elf32_to_cpu(phdr[i].p_vaddr); ++ phdr[i].p_paddr = elf32_to_cpu(phdr[i].p_paddr); ++ phdr[i].p_filesz = elf32_to_cpu(phdr[i].p_filesz); ++ phdr[i].p_memsz = elf32_to_cpu(phdr[i].p_memsz); ++ phdr[i].p_flags = elf32_to_cpu(phdr[i].p_flags); ++ phdr[i].p_align = elf32_to_cpu(phdr[i].p_align); ++ } ++ ++} ++ + static void read_shdrs(FILE *fp) + { +- int i; ++ unsigned int i; + Elf32_Shdr shdr; + + secs = calloc(ehdr.e_shnum, sizeof(struct section)); +@@ -282,7 +315,7 @@ static void read_shdrs(FILE *fp) + + static void read_strtabs(FILE *fp) + { +- int i; ++ unsigned int i; + for (i = 0; i < ehdr.e_shnum; i++) { + struct section *sec = &secs[i]; + if (sec->shdr.sh_type != SHT_STRTAB) { +@@ -307,7 +340,7 @@ static void read_strtabs(FILE *fp) + + static void read_symtabs(FILE *fp) + { +- int i,j; ++ unsigned int i,j; + for (i = 0; i < ehdr.e_shnum; i++) { + struct section *sec = &secs[i]; + if (sec->shdr.sh_type != SHT_SYMTAB) { +@@ -340,7 +373,9 @@ static void read_symtabs(FILE *fp) + + static void read_relocs(FILE *fp) + { +- int i,j; ++ unsigned int i,j; ++ uint32_t base; ++ + for (i = 0; i < ehdr.e_shnum; i++) { + struct section *sec = &secs[i]; + if (sec->shdr.sh_type != SHT_REL) { +@@ -360,9 +395,18 @@ static void read_relocs(FILE *fp) + die("Cannot read symbol table: %s\n", + strerror(errno)); + } ++ base = 0; ++ for (j = 0; j < ehdr.e_phnum; j++) { ++ if (phdr[j].p_type != PT_LOAD ) ++ continue; ++ if (secs[sec->shdr.sh_info].shdr.sh_offset < phdr[j].p_offset || secs[sec->shdr.sh_info].shdr.sh_offset >= phdr[j].p_offset + phdr[j].p_filesz) ++ continue; ++ base = CONFIG_PAGE_OFFSET + phdr[j].p_paddr - phdr[j].p_vaddr; ++ break; ++ } + for (j = 0; j < sec->shdr.sh_size/sizeof(Elf32_Rel); j++) { + Elf32_Rel *rel = &sec->reltab[j]; +- rel->r_offset = elf32_to_cpu(rel->r_offset); ++ rel->r_offset = elf32_to_cpu(rel->r_offset) + base; + rel->r_info = elf32_to_cpu(rel->r_info); + } + } +@@ -371,14 +415,14 @@ static void read_relocs(FILE *fp) + + static void print_absolute_symbols(void) + { +- int i; ++ unsigned int i; + printf("Absolute symbols\n"); + printf(" Num: Value Size Type Bind Visibility Name\n"); + for (i = 0; i < ehdr.e_shnum; i++) { + struct section *sec = &secs[i]; + char *sym_strtab; + Elf32_Sym *sh_symtab; +- int j; ++ unsigned int j; + + if (sec->shdr.sh_type != SHT_SYMTAB) { + continue; +@@ -406,14 +450,14 @@ static void print_absolute_symbols(void) + + static void print_absolute_relocs(void) + { +- int i, printed = 0; ++ unsigned int i, printed = 0; + + for (i = 0; i < ehdr.e_shnum; i++) { + struct section *sec = &secs[i]; + struct section *sec_applies, *sec_symtab; + char *sym_strtab; + Elf32_Sym *sh_symtab; +- int j; ++ unsigned int j; + if (sec->shdr.sh_type != SHT_REL) { + continue; + } +@@ -474,13 +518,13 @@ static void print_absolute_relocs(void) + + static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym)) + { +- int i; ++ unsigned int i; + /* Walk through the relocations */ + for (i = 0; i < ehdr.e_shnum; i++) { + char *sym_strtab; + Elf32_Sym *sh_symtab; + struct section *sec_applies, *sec_symtab; +- int j; ++ unsigned int j; + struct section *sec = &secs[i]; + + if (sec->shdr.sh_type != SHT_REL) { +@@ -504,6 +548,21 @@ static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym)) + if (sym->st_shndx == SHN_ABS) { + continue; + } ++ /* Don't relocate actual per-cpu variables, they are absolute indices, not addresses */ ++ if (!strcmp(sec_name(sym->st_shndx), ".data.percpu") && strcmp(sym_name(sym_strtab, sym), "__per_cpu_load")) ++ continue; ++ ++#if defined(CONFIG_PAX_KERNEXEC) && defined(CONFIG_X86_32) ++ /* Don't relocate actual code, they are relocated implicitly by the base address of KERNEL_CS */ ++ if (!strcmp(sec_name(sym->st_shndx), ".module.text") && !strcmp(sym_name(sym_strtab, sym), "_etext")) ++ continue; ++ if (!strcmp(sec_name(sym->st_shndx), ".init.text")) ++ continue; ++ if (!strcmp(sec_name(sym->st_shndx), ".exit.text")) ++ continue; ++ if (!strcmp(sec_name(sym->st_shndx), ".text") && strcmp(sym_name(sym_strtab, sym), "__LOAD_PHYSICAL_ADDR")) ++ continue; ++#endif + if (r_type == R_386_NONE || r_type == R_386_PC32) { + /* + * NONE can be ignored and and PC relative +@@ -541,7 +600,7 @@ static int cmp_relocs(const void *va, const void *vb) + + static void emit_relocs(int as_text) + { +- int i; ++ unsigned int i; + /* Count how many relocations I have and allocate space for them. */ + reloc_count = 0; + walk_relocs(count_reloc); +@@ -634,6 +693,7 @@ int main(int argc, char **argv) + fname, strerror(errno)); + } + read_ehdr(fp); ++ read_phdrs(fp); + read_shdrs(fp); + read_strtabs(fp); + read_symtabs(fp); +diff --git a/arch/x86/boot/cpucheck.c b/arch/x86/boot/cpucheck.c +index 4d3ff03..e4972ff 100644 +--- a/arch/x86/boot/cpucheck.c ++++ b/arch/x86/boot/cpucheck.c +@@ -74,7 +74,7 @@ static int has_fpu(void) + u16 fcw = -1, fsw = -1; + u32 cr0; + +- asm("movl %%cr0,%0" : "=r" (cr0)); ++ asm volatile("movl %%cr0,%0" : "=r" (cr0)); + if (cr0 & (X86_CR0_EM|X86_CR0_TS)) { + cr0 &= ~(X86_CR0_EM|X86_CR0_TS); + asm volatile("movl %0,%%cr0" : : "r" (cr0)); +@@ -90,7 +90,7 @@ static int has_eflag(u32 mask) + { + u32 f0, f1; + +- asm("pushfl ; " ++ asm volatile("pushfl ; " + "pushfl ; " + "popl %0 ; " + "movl %0,%1 ; " +@@ -115,7 +115,7 @@ static void get_flags(void) + set_bit(X86_FEATURE_FPU, cpu.flags); + + if (has_eflag(X86_EFLAGS_ID)) { +- asm("cpuid" ++ asm volatile("cpuid" + : "=a" (max_intel_level), + "=b" (cpu_vendor[0]), + "=d" (cpu_vendor[1]), +@@ -124,7 +124,7 @@ static void get_flags(void) + + if (max_intel_level >= 0x00000001 && + max_intel_level <= 0x0000ffff) { +- asm("cpuid" ++ asm volatile("cpuid" + : "=a" (tfms), + "=c" (cpu.flags[4]), + "=d" (cpu.flags[0]) +@@ -136,7 +136,7 @@ static void get_flags(void) + cpu.model += ((tfms >> 16) & 0xf) << 4; + } + +- asm("cpuid" ++ asm volatile("cpuid" + : "=a" (max_amd_level) + : "a" (0x80000000) + : "ebx", "ecx", "edx"); +@@ -144,7 +144,7 @@ static void get_flags(void) + if (max_amd_level >= 0x80000001 && + max_amd_level <= 0x8000ffff) { + u32 eax = 0x80000001; +- asm("cpuid" ++ asm volatile("cpuid" + : "+a" (eax), + "=c" (cpu.flags[6]), + "=d" (cpu.flags[1]) +@@ -203,9 +203,9 @@ int check_cpu(int *cpu_level_ptr, int *req_level_ptr, u32 **err_flags_ptr) + u32 ecx = MSR_K7_HWCR; + u32 eax, edx; + +- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); ++ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); + eax &= ~(1 << 15); +- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); ++ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); + + get_flags(); /* Make sure it really did something */ + err = check_flags(); +@@ -218,9 +218,9 @@ int check_cpu(int *cpu_level_ptr, int *req_level_ptr, u32 **err_flags_ptr) + u32 ecx = MSR_VIA_FCR; + u32 eax, edx; + +- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); ++ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); + eax |= (1<<1)|(1<<7); +- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); ++ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); + + set_bit(X86_FEATURE_CX8, cpu.flags); + err = check_flags(); +@@ -231,12 +231,12 @@ int check_cpu(int *cpu_level_ptr, int *req_level_ptr, u32 **err_flags_ptr) + u32 eax, edx; + u32 level = 1; + +- asm("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); +- asm("wrmsr" : : "a" (~0), "d" (edx), "c" (ecx)); +- asm("cpuid" ++ asm volatile("rdmsr" : "=a" (eax), "=d" (edx) : "c" (ecx)); ++ asm volatile("wrmsr" : : "a" (~0), "d" (edx), "c" (ecx)); ++ asm volatile("cpuid" + : "+a" (level), "=d" (cpu.flags[0]) + : : "ecx", "ebx"); +- asm("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); ++ asm volatile("wrmsr" : : "a" (eax), "d" (edx), "c" (ecx)); + + err = check_flags(); + } +diff --git a/arch/x86/boot/header.S b/arch/x86/boot/header.S +index b31cc54..8d69237 100644 +--- a/arch/x86/boot/header.S ++++ b/arch/x86/boot/header.S +@@ -224,7 +224,7 @@ setup_data: .quad 0 # 64-bit physical pointer to + # single linked list of + # struct setup_data + +-pref_address: .quad LOAD_PHYSICAL_ADDR # preferred load addr ++pref_address: .quad ____LOAD_PHYSICAL_ADDR # preferred load addr + + #define ZO_INIT_SIZE (ZO__end - ZO_startup_32 + ZO_z_extract_offset) + #define VO_INIT_SIZE (VO__end - VO__text) +diff --git a/arch/x86/boot/memory.c b/arch/x86/boot/memory.c +index cae3feb..ff8ff2a 100644 +--- a/arch/x86/boot/memory.c ++++ b/arch/x86/boot/memory.c +@@ -19,7 +19,7 @@ + + static int detect_memory_e820(void) + { +- int count = 0; ++ unsigned int count = 0; + struct biosregs ireg, oreg; + struct e820entry *desc = boot_params.e820_map; + static struct e820entry buf; /* static so it is zeroed */ +diff --git a/arch/x86/boot/video-vesa.c b/arch/x86/boot/video-vesa.c +index 11e8c6e..fdbb1ed 100644 +--- a/arch/x86/boot/video-vesa.c ++++ b/arch/x86/boot/video-vesa.c +@@ -200,6 +200,7 @@ static void vesa_store_pm_info(void) + + boot_params.screen_info.vesapm_seg = oreg.es; + boot_params.screen_info.vesapm_off = oreg.di; ++ boot_params.screen_info.vesapm_size = oreg.cx; + } + + /* +diff --git a/arch/x86/boot/video.c b/arch/x86/boot/video.c +index d42da38..787cdf3 100644 +--- a/arch/x86/boot/video.c ++++ b/arch/x86/boot/video.c +@@ -90,7 +90,7 @@ static void store_mode_params(void) + static unsigned int get_entry(void) + { + char entry_buf[4]; +- int i, len = 0; ++ unsigned int i, len = 0; + int key; + unsigned int v; + +diff --git a/arch/x86/crypto/aes-x86_64-asm_64.S b/arch/x86/crypto/aes-x86_64-asm_64.S +index 5b577d5..3c1fed4 100644 +--- a/arch/x86/crypto/aes-x86_64-asm_64.S ++++ b/arch/x86/crypto/aes-x86_64-asm_64.S +@@ -8,6 +8,8 @@ + * including this sentence is retained in full. + */ + ++#include ++ + .extern crypto_ft_tab + .extern crypto_it_tab + .extern crypto_fl_tab +@@ -71,6 +73,8 @@ FUNC: movq r1,r2; \ + je B192; \ + leaq 32(r9),r9; + ++#define ret pax_force_retaddr 0, 1; ret ++ + #define epilogue(r1,r2,r3,r4,r5,r6,r7,r8,r9) \ + movq r1,r2; \ + movq r3,r4; \ +diff --git a/arch/x86/crypto/aesni-intel_asm.S b/arch/x86/crypto/aesni-intel_asm.S +index eb0566e..e3ebad8 100644 +--- a/arch/x86/crypto/aesni-intel_asm.S ++++ b/arch/x86/crypto/aesni-intel_asm.S +@@ -16,6 +16,7 @@ + */ + + #include ++#include + + .text + +@@ -52,6 +53,7 @@ _key_expansion_256a: + pxor %xmm1, %xmm0 + movaps %xmm0, (%rcx) + add $0x10, %rcx ++ pax_force_retaddr_bts + ret + + _key_expansion_192a: +@@ -75,6 +77,7 @@ _key_expansion_192a: + shufps $0b01001110, %xmm2, %xmm1 + movaps %xmm1, 16(%rcx) + add $0x20, %rcx ++ pax_force_retaddr_bts + ret + + _key_expansion_192b: +@@ -93,6 +96,7 @@ _key_expansion_192b: + + movaps %xmm0, (%rcx) + add $0x10, %rcx ++ pax_force_retaddr_bts + ret + + _key_expansion_256b: +@@ -104,6 +108,7 @@ _key_expansion_256b: + pxor %xmm1, %xmm2 + movaps %xmm2, (%rcx) + add $0x10, %rcx ++ pax_force_retaddr_bts + ret + + /* +@@ -239,7 +244,9 @@ ENTRY(aesni_set_key) + cmp %rcx, %rdi + jb .Ldec_key_loop + xor %rax, %rax ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_set_key) + + /* + * void aesni_enc(struct crypto_aes_ctx *ctx, u8 *dst, const u8 *src) +@@ -249,7 +256,9 @@ ENTRY(aesni_enc) + movups (INP), STATE # input + call _aesni_enc1 + movups STATE, (OUTP) # output ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_enc) + + /* + * _aesni_enc1: internal ABI +@@ -319,6 +328,7 @@ _aesni_enc1: + movaps 0x70(TKEYP), KEY + # aesenclast KEY, STATE # last round + .byte 0x66, 0x0f, 0x38, 0xdd, 0xc2 ++ pax_force_retaddr_bts + ret + + /* +@@ -482,6 +492,7 @@ _aesni_enc4: + .byte 0x66, 0x0f, 0x38, 0xdd, 0xea + # aesenclast KEY, STATE4 + .byte 0x66, 0x0f, 0x38, 0xdd, 0xf2 ++ pax_force_retaddr_bts + ret + + /* +@@ -493,7 +504,9 @@ ENTRY(aesni_dec) + movups (INP), STATE # input + call _aesni_dec1 + movups STATE, (OUTP) #output ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_dec) + + /* + * _aesni_dec1: internal ABI +@@ -563,6 +576,7 @@ _aesni_dec1: + movaps 0x70(TKEYP), KEY + # aesdeclast KEY, STATE # last round + .byte 0x66, 0x0f, 0x38, 0xdf, 0xc2 ++ pax_force_retaddr_bts + ret + + /* +@@ -726,6 +740,7 @@ _aesni_dec4: + .byte 0x66, 0x0f, 0x38, 0xdf, 0xea + # aesdeclast KEY, STATE4 + .byte 0x66, 0x0f, 0x38, 0xdf, 0xf2 ++ pax_force_retaddr_bts + ret + + /* +@@ -769,7 +784,9 @@ ENTRY(aesni_ecb_enc) + cmp $16, LEN + jge .Lecb_enc_loop1 + .Lecb_enc_ret: ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_ecb_enc) + + /* + * void aesni_ecb_dec(struct crypto_aes_ctx *ctx, const u8 *dst, u8 *src, +@@ -813,7 +830,9 @@ ENTRY(aesni_ecb_dec) + cmp $16, LEN + jge .Lecb_dec_loop1 + .Lecb_dec_ret: ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_ecb_dec) + + /* + * void aesni_cbc_enc(struct crypto_aes_ctx *ctx, const u8 *dst, u8 *src, +@@ -837,7 +856,9 @@ ENTRY(aesni_cbc_enc) + jge .Lcbc_enc_loop + movups STATE, (IVP) + .Lcbc_enc_ret: ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_cbc_enc) + + /* + * void aesni_cbc_dec(struct crypto_aes_ctx *ctx, const u8 *dst, u8 *src, +@@ -894,4 +915,6 @@ ENTRY(aesni_cbc_dec) + .Lcbc_dec_ret: + movups IV, (IVP) + .Lcbc_dec_just_ret: ++ pax_force_retaddr 0, 1 + ret ++ENDPROC(aesni_cbc_dec) +diff --git a/arch/x86/crypto/salsa20-x86_64-asm_64.S b/arch/x86/crypto/salsa20-x86_64-asm_64.S +index 6214a9b..1f4fc9a 100644 +--- a/arch/x86/crypto/salsa20-x86_64-asm_64.S ++++ b/arch/x86/crypto/salsa20-x86_64-asm_64.S +@@ -1,3 +1,5 @@ ++#include ++ + # enter ECRYPT_encrypt_bytes + .text + .p2align 5 +@@ -790,6 +792,7 @@ ECRYPT_encrypt_bytes: + add %r11,%rsp + mov %rdi,%rax + mov %rsi,%rdx ++ pax_force_retaddr 0, 1 + ret + # bytesatleast65: + ._bytesatleast65: +@@ -891,6 +894,7 @@ ECRYPT_keysetup: + add %r11,%rsp + mov %rdi,%rax + mov %rsi,%rdx ++ pax_force_retaddr + ret + # enter ECRYPT_ivsetup + .text +@@ -917,4 +921,5 @@ ECRYPT_ivsetup: + add %r11,%rsp + mov %rdi,%rax + mov %rsi,%rdx ++ pax_force_retaddr + ret +diff --git a/arch/x86/crypto/twofish-x86_64-asm_64.S b/arch/x86/crypto/twofish-x86_64-asm_64.S +index 35974a5..5662ae2 100644 +--- a/arch/x86/crypto/twofish-x86_64-asm_64.S ++++ b/arch/x86/crypto/twofish-x86_64-asm_64.S +@@ -21,6 +21,7 @@ + .text + + #include ++#include + + #define a_offset 0 + #define b_offset 4 +@@ -269,6 +270,7 @@ twofish_enc_blk: + + popq R1 + movq $1,%rax ++ pax_force_retaddr 0, 1 + ret + + twofish_dec_blk: +@@ -321,4 +323,5 @@ twofish_dec_blk: + + popq R1 + movq $1,%rax ++ pax_force_retaddr 0, 1 + ret +diff --git a/arch/x86/ia32/ia32_aout.c b/arch/x86/ia32/ia32_aout.c +index 14531ab..bc68a7b 100644 +--- a/arch/x86/ia32/ia32_aout.c ++++ b/arch/x86/ia32/ia32_aout.c +@@ -169,6 +169,8 @@ static int aout_core_dump(long signr, struct pt_regs *regs, struct file *file, + unsigned long dump_start, dump_size; + struct user32 dump; + ++ memset(&dump, 0, sizeof(dump)); ++ + fs = get_fs(); + set_fs(KERNEL_DS); + has_dumped = 1; +@@ -218,12 +220,6 @@ static int aout_core_dump(long signr, struct pt_regs *regs, struct file *file, + dump_size = dump.u_ssize << PAGE_SHIFT; + DUMP_WRITE(dump_start, dump_size); + } +- /* +- * Finally dump the task struct. Not be used by gdb, but +- * could be useful +- */ +- set_fs(KERNEL_DS); +- DUMP_WRITE(current, sizeof(*current)); + end_coredump: + set_fs(fs); + return has_dumped; +@@ -327,6 +323,13 @@ static int load_aout_binary(struct linux_binprm *bprm, struct pt_regs *regs) + current->mm->free_area_cache = TASK_UNMAPPED_BASE; + current->mm->cached_hole_size = 0; + ++ retval = setup_arg_pages(bprm, IA32_STACK_TOP, EXSTACK_DEFAULT); ++ if (retval < 0) { ++ /* Someone check-me: is this error path enough? */ ++ send_sig(SIGKILL, current, 0); ++ return retval; ++ } ++ + install_exec_creds(bprm); + current->flags &= ~PF_FORKNOEXEC; + +@@ -422,13 +425,6 @@ beyond_if: + + set_brk(current->mm->start_brk, current->mm->brk); + +- retval = setup_arg_pages(bprm, IA32_STACK_TOP, EXSTACK_DEFAULT); +- if (retval < 0) { +- /* Someone check-me: is this error path enough? */ +- send_sig(SIGKILL, current, 0); +- return retval; +- } +- + current->mm->start_stack = + (unsigned long)create_aout_tables((char __user *)bprm->p, bprm); + /* start thread */ +diff --git a/arch/x86/ia32/ia32_signal.c b/arch/x86/ia32/ia32_signal.c +index 588a7aa..a3468b0 100644 +--- a/arch/x86/ia32/ia32_signal.c ++++ b/arch/x86/ia32/ia32_signal.c +@@ -167,7 +167,7 @@ asmlinkage long sys32_sigaltstack(const stack_ia32_t __user *uss_ptr, + } + seg = get_fs(); + set_fs(KERNEL_DS); +- ret = do_sigaltstack(uss_ptr ? &uss : NULL, &uoss, regs->sp); ++ ret = do_sigaltstack(uss_ptr ? (const stack_t __force_user *)&uss : NULL, (stack_t __force_user *)&uoss, regs->sp); + set_fs(seg); + if (ret >= 0 && uoss_ptr) { + if (!access_ok(VERIFY_WRITE, uoss_ptr, sizeof(stack_ia32_t))) +@@ -374,7 +374,7 @@ static int ia32_setup_sigcontext(struct sigcontext_ia32 __user *sc, + */ + static void __user *get_sigframe(struct k_sigaction *ka, struct pt_regs *regs, + size_t frame_size, +- void **fpstate) ++ void __user **fpstate) + { + unsigned long sp; + +@@ -395,7 +395,7 @@ static void __user *get_sigframe(struct k_sigaction *ka, struct pt_regs *regs, + + if (used_math()) { + sp = sp - sig_xstate_ia32_size; +- *fpstate = (struct _fpstate_ia32 *) sp; ++ *fpstate = (struct _fpstate_ia32 __user *) sp; + if (save_i387_xstate_ia32(*fpstate) < 0) + return (void __user *) -1L; + } +@@ -403,7 +403,7 @@ static void __user *get_sigframe(struct k_sigaction *ka, struct pt_regs *regs, + sp -= frame_size; + /* Align the stack pointer according to the i386 ABI, + * i.e. so that on function entry ((sp + 4) & 15) == 0. */ +- sp = ((sp + 4) & -16ul) - 4; ++ sp = ((sp - 12) & -16ul) - 4; + return (void __user *) sp; + } + +@@ -461,7 +461,7 @@ int ia32_setup_frame(int sig, struct k_sigaction *ka, + * These are actually not used anymore, but left because some + * gdb versions depend on them as a marker. + */ +- put_user_ex(*((u64 *)&code), (u64 *)frame->retcode); ++ put_user_ex(*((const u64 *)&code), (u64 __user *)frame->retcode); + } put_user_catch(err); + + if (err) +@@ -503,7 +503,7 @@ int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, + 0xb8, + __NR_ia32_rt_sigreturn, + 0x80cd, +- 0, ++ 0 + }; + + frame = get_sigframe(ka, regs, sizeof(*frame), &fpstate); +@@ -533,16 +533,18 @@ int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, + + if (ka->sa.sa_flags & SA_RESTORER) + restorer = ka->sa.sa_restorer; ++ else if (current->mm->context.vdso) ++ /* Return stub is in 32bit vsyscall page */ ++ restorer = VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn); + else +- restorer = VDSO32_SYMBOL(current->mm->context.vdso, +- rt_sigreturn); ++ restorer = &frame->retcode; + put_user_ex(ptr_to_compat(restorer), &frame->pretcode); + + /* + * Not actually used anymore, but left because some gdb + * versions need it. + */ +- put_user_ex(*((u64 *)&code), (u64 *)frame->retcode); ++ put_user_ex(*((const u64 *)&code), (u64 __user *)frame->retcode); + } put_user_catch(err); + + if (err) +diff --git a/arch/x86/ia32/ia32entry.S b/arch/x86/ia32/ia32entry.S +index 4edd8eb..29124b4 100644 +--- a/arch/x86/ia32/ia32entry.S ++++ b/arch/x86/ia32/ia32entry.S +@@ -13,7 +13,9 @@ + #include + #include + #include ++#include + #include ++#include + + /* Avoid __ASSEMBLER__'ifying just for this. */ + #include +@@ -93,6 +95,32 @@ ENTRY(native_irq_enable_sysexit) + ENDPROC(native_irq_enable_sysexit) + #endif + ++ .macro pax_enter_kernel_user ++ pax_set_fptr_mask ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ call pax_enter_kernel_user ++#endif ++ .endm ++ ++ .macro pax_exit_kernel_user ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ call pax_exit_kernel_user ++#endif ++#ifdef CONFIG_PAX_RANDKSTACK ++ pushq %rax ++ pushq %r11 ++ call pax_randomize_kstack ++ popq %r11 ++ popq %rax ++#endif ++ .endm ++ ++.macro pax_erase_kstack ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++ call pax_erase_kstack ++#endif ++.endm ++ + /* + * 32bit SYSENTER instruction entry. + * +@@ -119,12 +147,6 @@ ENTRY(ia32_sysenter_target) + CFI_REGISTER rsp,rbp + SWAPGS_UNSAFE_STACK + movq PER_CPU_VAR(kernel_stack), %rsp +- addq $(KERNEL_STACK_OFFSET),%rsp +- /* +- * No need to follow this irqs on/off section: the syscall +- * disabled irqs, here we enable it straight after entry: +- */ +- ENABLE_INTERRUPTS(CLBR_NONE) + movl %ebp,%ebp /* zero extension */ + pushq $__USER32_DS + CFI_ADJUST_CFA_OFFSET 8 +@@ -135,28 +157,42 @@ ENTRY(ia32_sysenter_target) + pushfq + CFI_ADJUST_CFA_OFFSET 8 + /*CFI_REL_OFFSET rflags,0*/ +- movl 8*3-THREAD_SIZE+TI_sysenter_return(%rsp), %r10d +- CFI_REGISTER rip,r10 ++ orl $X86_EFLAGS_IF,(%rsp) ++ GET_THREAD_INFO(%r11) ++ movl TI_sysenter_return(%r11), %r11d ++ CFI_REGISTER rip,r11 + pushq $__USER32_CS + CFI_ADJUST_CFA_OFFSET 8 + /*CFI_REL_OFFSET cs,0*/ + movl %eax, %eax +- pushq %r10 ++ pushq %r11 + CFI_ADJUST_CFA_OFFSET 8 + CFI_REL_OFFSET rip,0 + pushq %rax + CFI_ADJUST_CFA_OFFSET 8 + cld + SAVE_ARGS 0,0,1 ++ pax_enter_kernel_user ++ /* ++ * No need to follow this irqs on/off section: the syscall ++ * disabled irqs, here we enable it straight after entry: ++ */ ++ ENABLE_INTERRUPTS(CLBR_NONE) + /* no need to do an access_ok check here because rbp has been + 32bit zero extended */ ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ mov $PAX_USER_SHADOW_BASE,%r11 ++ add %r11,%rbp ++#endif ++ + 1: movl (%rbp),%ebp + .section __ex_table,"a" + .quad 1b,ia32_badarg + .previous +- GET_THREAD_INFO(%r10) +- orl $TS_COMPAT,TI_status(%r10) +- testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r10) ++ GET_THREAD_INFO(%r11) ++ orl $TS_COMPAT,TI_status(%r11) ++ testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r11) + CFI_REMEMBER_STATE + jnz sysenter_tracesys + cmpq $(IA32_NR_syscalls-1),%rax +@@ -166,13 +202,15 @@ sysenter_do_call: + sysenter_dispatch: + call *ia32_sys_call_table(,%rax,8) + movq %rax,RAX-ARGOFFSET(%rsp) +- GET_THREAD_INFO(%r10) ++ GET_THREAD_INFO(%r11) + DISABLE_INTERRUPTS(CLBR_NONE) + TRACE_IRQS_OFF +- testl $_TIF_ALLWORK_MASK,TI_flags(%r10) ++ testl $_TIF_ALLWORK_MASK,TI_flags(%r11) + jnz sysexit_audit + sysexit_from_sys_call: +- andl $~TS_COMPAT,TI_status(%r10) ++ pax_exit_kernel_user ++ pax_erase_kstack ++ andl $~TS_COMPAT,TI_status(%r11) + /* clear IF, that popfq doesn't enable interrupts early */ + andl $~0x200,EFLAGS-R11(%rsp) + movl RIP-R11(%rsp),%edx /* User %eip */ +@@ -200,6 +238,9 @@ sysexit_from_sys_call: + movl %eax,%esi /* 2nd arg: syscall number */ + movl $AUDIT_ARCH_I386,%edi /* 1st arg: audit arch */ + call audit_syscall_entry ++ ++ pax_erase_kstack ++ + movl RAX-ARGOFFSET(%rsp),%eax /* reload syscall number */ + cmpq $(IA32_NR_syscalls-1),%rax + ja ia32_badsys +@@ -211,7 +252,7 @@ sysexit_from_sys_call: + .endm + + .macro auditsys_exit exit +- testl $(_TIF_ALLWORK_MASK & ~_TIF_SYSCALL_AUDIT),TI_flags(%r10) ++ testl $(_TIF_ALLWORK_MASK & ~_TIF_SYSCALL_AUDIT),TI_flags(%r11) + jnz ia32_ret_from_sys_call + TRACE_IRQS_ON + sti +@@ -221,12 +262,12 @@ sysexit_from_sys_call: + movzbl %al,%edi /* zero-extend that into %edi */ + inc %edi /* first arg, 0->1(AUDITSC_SUCCESS), 1->2(AUDITSC_FAILURE) */ + call audit_syscall_exit +- GET_THREAD_INFO(%r10) ++ GET_THREAD_INFO(%r11) + movl RAX-ARGOFFSET(%rsp),%eax /* reload syscall return value */ + movl $(_TIF_ALLWORK_MASK & ~_TIF_SYSCALL_AUDIT),%edi + cli + TRACE_IRQS_OFF +- testl %edi,TI_flags(%r10) ++ testl %edi,TI_flags(%r11) + jz \exit + CLEAR_RREGS -ARGOFFSET + jmp int_with_check +@@ -244,7 +285,7 @@ sysexit_audit: + + sysenter_tracesys: + #ifdef CONFIG_AUDITSYSCALL +- testl $(_TIF_WORK_SYSCALL_ENTRY & ~_TIF_SYSCALL_AUDIT),TI_flags(%r10) ++ testl $(_TIF_WORK_SYSCALL_ENTRY & ~_TIF_SYSCALL_AUDIT),TI_flags(%r11) + jz sysenter_auditsys + #endif + SAVE_REST +@@ -252,6 +293,9 @@ sysenter_tracesys: + movq $-ENOSYS,RAX(%rsp)/* ptrace can change this for a bad syscall */ + movq %rsp,%rdi /* &pt_regs -> arg1 */ + call syscall_trace_enter ++ ++ pax_erase_kstack ++ + LOAD_ARGS32 ARGOFFSET /* reload args from stack in case ptrace changed it */ + RESTORE_REST + cmpq $(IA32_NR_syscalls-1),%rax +@@ -283,19 +327,20 @@ ENDPROC(ia32_sysenter_target) + ENTRY(ia32_cstar_target) + CFI_STARTPROC32 simple + CFI_SIGNAL_FRAME +- CFI_DEF_CFA rsp,KERNEL_STACK_OFFSET ++ CFI_DEF_CFA rsp,0 + CFI_REGISTER rip,rcx + /*CFI_REGISTER rflags,r11*/ + SWAPGS_UNSAFE_STACK + movl %esp,%r8d + CFI_REGISTER rsp,r8 + movq PER_CPU_VAR(kernel_stack),%rsp ++ SAVE_ARGS 8*6,1,1 ++ pax_enter_kernel_user + /* + * No need to follow this irqs on/off section: the syscall + * disabled irqs and here we enable it straight after entry: + */ + ENABLE_INTERRUPTS(CLBR_NONE) +- SAVE_ARGS 8,1,1 + movl %eax,%eax /* zero extension */ + movq %rax,ORIG_RAX-ARGOFFSET(%rsp) + movq %rcx,RIP-ARGOFFSET(%rsp) +@@ -311,13 +356,19 @@ ENTRY(ia32_cstar_target) + /* no need to do an access_ok check here because r8 has been + 32bit zero extended */ + /* hardware stack frame is complete now */ ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ mov $PAX_USER_SHADOW_BASE,%r11 ++ add %r11,%r8 ++#endif ++ + 1: movl (%r8),%r9d + .section __ex_table,"a" + .quad 1b,ia32_badarg + .previous +- GET_THREAD_INFO(%r10) +- orl $TS_COMPAT,TI_status(%r10) +- testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r10) ++ GET_THREAD_INFO(%r11) ++ orl $TS_COMPAT,TI_status(%r11) ++ testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r11) + CFI_REMEMBER_STATE + jnz cstar_tracesys + cmpq $IA32_NR_syscalls-1,%rax +@@ -327,13 +378,15 @@ cstar_do_call: + cstar_dispatch: + call *ia32_sys_call_table(,%rax,8) + movq %rax,RAX-ARGOFFSET(%rsp) +- GET_THREAD_INFO(%r10) ++ GET_THREAD_INFO(%r11) + DISABLE_INTERRUPTS(CLBR_NONE) + TRACE_IRQS_OFF +- testl $_TIF_ALLWORK_MASK,TI_flags(%r10) ++ testl $_TIF_ALLWORK_MASK,TI_flags(%r11) + jnz sysretl_audit + sysretl_from_sys_call: +- andl $~TS_COMPAT,TI_status(%r10) ++ pax_exit_kernel_user ++ pax_erase_kstack ++ andl $~TS_COMPAT,TI_status(%r11) + RESTORE_ARGS 1,-ARG_SKIP,1,1,1 + movl RIP-ARGOFFSET(%rsp),%ecx + CFI_REGISTER rip,rcx +@@ -361,7 +414,7 @@ sysretl_audit: + + cstar_tracesys: + #ifdef CONFIG_AUDITSYSCALL +- testl $(_TIF_WORK_SYSCALL_ENTRY & ~_TIF_SYSCALL_AUDIT),TI_flags(%r10) ++ testl $(_TIF_WORK_SYSCALL_ENTRY & ~_TIF_SYSCALL_AUDIT),TI_flags(%r11) + jz cstar_auditsys + #endif + xchgl %r9d,%ebp +@@ -370,6 +423,9 @@ cstar_tracesys: + movq $-ENOSYS,RAX(%rsp) /* ptrace can change this for a bad syscall */ + movq %rsp,%rdi /* &pt_regs -> arg1 */ + call syscall_trace_enter ++ ++ pax_erase_kstack ++ + LOAD_ARGS32 ARGOFFSET, 1 /* reload args from stack in case ptrace changed it */ + RESTORE_REST + xchgl %ebp,%r9d +@@ -415,11 +471,6 @@ ENTRY(ia32_syscall) + CFI_REL_OFFSET rip,RIP-RIP + PARAVIRT_ADJUST_EXCEPTION_FRAME + SWAPGS +- /* +- * No need to follow this irqs on/off section: the syscall +- * disabled irqs and here we enable it straight after entry: +- */ +- ENABLE_INTERRUPTS(CLBR_NONE) + movl %eax,%eax + pushq %rax + CFI_ADJUST_CFA_OFFSET 8 +@@ -427,9 +478,15 @@ ENTRY(ia32_syscall) + /* note the registers are not zero extended to the sf. + this could be a problem. */ + SAVE_ARGS 0,0,1 +- GET_THREAD_INFO(%r10) +- orl $TS_COMPAT,TI_status(%r10) +- testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r10) ++ pax_enter_kernel_user ++ /* ++ * No need to follow this irqs on/off section: the syscall ++ * disabled irqs and here we enable it straight after entry: ++ */ ++ ENABLE_INTERRUPTS(CLBR_NONE) ++ GET_THREAD_INFO(%r11) ++ orl $TS_COMPAT,TI_status(%r11) ++ testl $_TIF_WORK_SYSCALL_ENTRY,TI_flags(%r11) + jnz ia32_tracesys + cmpq $(IA32_NR_syscalls-1),%rax + ja ia32_badsys +@@ -448,6 +505,9 @@ ia32_tracesys: + movq $-ENOSYS,RAX(%rsp) /* ptrace can change this for a bad syscall */ + movq %rsp,%rdi /* &pt_regs -> arg1 */ + call syscall_trace_enter ++ ++ pax_erase_kstack ++ + LOAD_ARGS32 ARGOFFSET /* reload args from stack in case ptrace changed it */ + RESTORE_REST + cmpq $(IA32_NR_syscalls-1),%rax +@@ -462,6 +522,7 @@ ia32_badsys: + + quiet_ni_syscall: + movq $-ENOSYS,%rax ++ pax_force_retaddr + ret + CFI_ENDPROC + +diff --git a/arch/x86/ia32/sys_ia32.c b/arch/x86/ia32/sys_ia32.c +index 016218c..47ccbdd 100644 +--- a/arch/x86/ia32/sys_ia32.c ++++ b/arch/x86/ia32/sys_ia32.c +@@ -69,8 +69,8 @@ asmlinkage long sys32_ftruncate64(unsigned int fd, unsigned long offset_low, + */ + static int cp_stat64(struct stat64 __user *ubuf, struct kstat *stat) + { +- typeof(ubuf->st_uid) uid = 0; +- typeof(ubuf->st_gid) gid = 0; ++ typeof(((struct stat64 *)0)->st_uid) uid = 0; ++ typeof(((struct stat64 *)0)->st_gid) gid = 0; + SET_UID(uid, stat->uid); + SET_GID(gid, stat->gid); + if (!access_ok(VERIFY_WRITE, ubuf, sizeof(struct stat64)) || +@@ -308,8 +308,8 @@ asmlinkage long sys32_rt_sigprocmask(int how, compat_sigset_t __user *set, + } + set_fs(KERNEL_DS); + ret = sys_rt_sigprocmask(how, +- set ? (sigset_t __user *)&s : NULL, +- oset ? (sigset_t __user *)&s : NULL, ++ set ? (sigset_t __force_user *)&s : NULL, ++ oset ? (sigset_t __force_user *)&s : NULL, + sigsetsize); + set_fs(old_fs); + if (ret) +@@ -371,7 +371,7 @@ asmlinkage long sys32_sched_rr_get_interval(compat_pid_t pid, + mm_segment_t old_fs = get_fs(); + + set_fs(KERNEL_DS); +- ret = sys_sched_rr_get_interval(pid, (struct timespec __user *)&t); ++ ret = sys_sched_rr_get_interval(pid, (struct timespec __force_user *)&t); + set_fs(old_fs); + if (put_compat_timespec(&t, interval)) + return -EFAULT; +@@ -387,7 +387,7 @@ asmlinkage long sys32_rt_sigpending(compat_sigset_t __user *set, + mm_segment_t old_fs = get_fs(); + + set_fs(KERNEL_DS); +- ret = sys_rt_sigpending((sigset_t __user *)&s, sigsetsize); ++ ret = sys_rt_sigpending((sigset_t __force_user *)&s, sigsetsize); + set_fs(old_fs); + if (!ret) { + switch (_NSIG_WORDS) { +@@ -412,7 +412,7 @@ asmlinkage long sys32_rt_sigqueueinfo(int pid, int sig, + if (copy_siginfo_from_user32(&info, uinfo)) + return -EFAULT; + set_fs(KERNEL_DS); +- ret = sys_rt_sigqueueinfo(pid, sig, (siginfo_t __user *)&info); ++ ret = sys_rt_sigqueueinfo(pid, sig, (siginfo_t __force_user *)&info); + set_fs(old_fs); + return ret; + } +@@ -513,7 +513,7 @@ asmlinkage long sys32_sendfile(int out_fd, int in_fd, + return -EFAULT; + + set_fs(KERNEL_DS); +- ret = sys_sendfile(out_fd, in_fd, offset ? (off_t __user *)&of : NULL, ++ ret = sys_sendfile(out_fd, in_fd, offset ? (off_t __force_user *)&of : NULL, + count); + set_fs(old_fs); + +diff --git a/arch/x86/include/asm/alternative-asm.h b/arch/x86/include/asm/alternative-asm.h +index e2077d3..17d07ad 100644 +--- a/arch/x86/include/asm/alternative-asm.h ++++ b/arch/x86/include/asm/alternative-asm.h +@@ -8,10 +8,10 @@ + + #ifdef CONFIG_SMP + .macro LOCK_PREFIX +-1: lock ++672: lock + .section .smp_locks,"a" + .align 4 +- X86_ALIGN 1b ++ X86_ALIGN 672b + .previous + .endm + #else +@@ -19,4 +19,43 @@ + .endm + #endif + ++#ifdef KERNEXEC_PLUGIN ++ .macro pax_force_retaddr_bts rip=0 ++ btsq $63,\rip(%rsp) ++ .endm ++#ifdef CONFIG_PAX_KERNEXEC_PLUGIN_METHOD_BTS ++ .macro pax_force_retaddr rip=0, reload=0 ++ btsq $63,\rip(%rsp) ++ .endm ++ .macro pax_force_fptr ptr ++ btsq $63,\ptr ++ .endm ++ .macro pax_set_fptr_mask ++ .endm ++#endif ++#ifdef CONFIG_PAX_KERNEXEC_PLUGIN_METHOD_OR ++ .macro pax_force_retaddr rip=0, reload=0 ++ .if \reload ++ pax_set_fptr_mask ++ .endif ++ orq %r10,\rip(%rsp) ++ .endm ++ .macro pax_force_fptr ptr ++ orq %r10,\ptr ++ .endm ++ .macro pax_set_fptr_mask ++ movabs $0x8000000000000000,%r10 ++ .endm ++#endif ++#else ++ .macro pax_force_retaddr rip=0, reload=0 ++ .endm ++ .macro pax_force_fptr ptr ++ .endm ++ .macro pax_force_retaddr_bts rip=0 ++ .endm ++ .macro pax_set_fptr_mask ++ .endm ++#endif ++ + #endif /* __ASSEMBLY__ */ +diff --git a/arch/x86/include/asm/alternative.h b/arch/x86/include/asm/alternative.h +index c240efc..fdfadf3 100644 +--- a/arch/x86/include/asm/alternative.h ++++ b/arch/x86/include/asm/alternative.h +@@ -85,7 +85,7 @@ static inline void alternatives_smp_switch(int smp) {} + " .byte 662b-661b\n" /* sourcelen */ \ + " .byte 664f-663f\n" /* replacementlen */ \ + ".previous\n" \ +- ".section .altinstr_replacement, \"ax\"\n" \ ++ ".section .altinstr_replacement, \"a\"\n" \ + "663:\n\t" newinstr "\n664:\n" /* replacement */ \ + ".previous" + +diff --git a/arch/x86/include/asm/apic.h b/arch/x86/include/asm/apic.h +index 474d80d..1f97d58 100644 +--- a/arch/x86/include/asm/apic.h ++++ b/arch/x86/include/asm/apic.h +@@ -46,7 +46,7 @@ static inline void generic_apic_probe(void) + + #ifdef CONFIG_X86_LOCAL_APIC + +-extern unsigned int apic_verbosity; ++extern int apic_verbosity; + extern int local_apic_timer_c2_ok; + + extern int disable_apic; +diff --git a/arch/x86/include/asm/apm.h b/arch/x86/include/asm/apm.h +index 20370c6..a2eb9b0 100644 +--- a/arch/x86/include/asm/apm.h ++++ b/arch/x86/include/asm/apm.h +@@ -34,7 +34,7 @@ static inline void apm_bios_call_asm(u32 func, u32 ebx_in, u32 ecx_in, + __asm__ __volatile__(APM_DO_ZERO_SEGS + "pushl %%edi\n\t" + "pushl %%ebp\n\t" +- "lcall *%%cs:apm_bios_entry\n\t" ++ "lcall *%%ss:apm_bios_entry\n\t" + "setc %%al\n\t" + "popl %%ebp\n\t" + "popl %%edi\n\t" +@@ -58,7 +58,7 @@ static inline u8 apm_bios_call_simple_asm(u32 func, u32 ebx_in, + __asm__ __volatile__(APM_DO_ZERO_SEGS + "pushl %%edi\n\t" + "pushl %%ebp\n\t" +- "lcall *%%cs:apm_bios_entry\n\t" ++ "lcall *%%ss:apm_bios_entry\n\t" + "setc %%bl\n\t" + "popl %%ebp\n\t" + "popl %%edi\n\t" +diff --git a/arch/x86/include/asm/atomic_32.h b/arch/x86/include/asm/atomic_32.h +index dc5a667..939040c 100644 +--- a/arch/x86/include/asm/atomic_32.h ++++ b/arch/x86/include/asm/atomic_32.h +@@ -25,6 +25,17 @@ static inline int atomic_read(const atomic_t *v) + } + + /** ++ * atomic_read_unchecked - read atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically reads the value of @v. ++ */ ++static inline int atomic_read_unchecked(const atomic_unchecked_t *v) ++{ ++ return v->counter; ++} ++ ++/** + * atomic_set - set atomic variable + * @v: pointer of type atomic_t + * @i: required value +@@ -37,6 +48,18 @@ static inline void atomic_set(atomic_t *v, int i) + } + + /** ++ * atomic_set_unchecked - set atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * @i: required value ++ * ++ * Atomically sets the value of @v to @i. ++ */ ++static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i) ++{ ++ v->counter = i; ++} ++ ++/** + * atomic_add - add integer to atomic variable + * @i: integer value to add + * @v: pointer of type atomic_t +@@ -45,7 +68,29 @@ static inline void atomic_set(atomic_t *v, int i) + */ + static inline void atomic_add(int i, atomic_t *v) + { +- asm volatile(LOCK_PREFIX "addl %1,%0" ++ asm volatile(LOCK_PREFIX "addl %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subl %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+m" (v->counter) ++ : "ir" (i)); ++} ++ ++/** ++ * atomic_add_unchecked - add integer to atomic variable ++ * @i: integer value to add ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically adds @i to @v. ++ */ ++static inline void atomic_add_unchecked(int i, atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "addl %1,%0\n" + : "+m" (v->counter) + : "ir" (i)); + } +@@ -59,7 +104,29 @@ static inline void atomic_add(int i, atomic_t *v) + */ + static inline void atomic_sub(int i, atomic_t *v) + { +- asm volatile(LOCK_PREFIX "subl %1,%0" ++ asm volatile(LOCK_PREFIX "subl %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addl %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+m" (v->counter) ++ : "ir" (i)); ++} ++ ++/** ++ * atomic_sub_unchecked - subtract integer from atomic variable ++ * @i: integer value to subtract ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically subtracts @i from @v. ++ */ ++static inline void atomic_sub_unchecked(int i, atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "subl %1,%0\n" + : "+m" (v->counter) + : "ir" (i)); + } +@@ -77,7 +144,16 @@ static inline int atomic_sub_and_test(int i, atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "subl %2,%0; sete %1" ++ asm volatile(LOCK_PREFIX "subl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "+m" (v->counter), "=qm" (c) + : "ir" (i) : "memory"); + return c; +@@ -91,7 +167,27 @@ static inline int atomic_sub_and_test(int i, atomic_t *v) + */ + static inline void atomic_inc(atomic_t *v) + { +- asm volatile(LOCK_PREFIX "incl %0" ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+m" (v->counter)); ++} ++ ++/** ++ * atomic_inc_unchecked - increment atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically increments @v by 1. ++ */ ++static inline void atomic_inc_unchecked(atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "incl %0\n" + : "+m" (v->counter)); + } + +@@ -103,7 +199,27 @@ static inline void atomic_inc(atomic_t *v) + */ + static inline void atomic_dec(atomic_t *v) + { +- asm volatile(LOCK_PREFIX "decl %0" ++ asm volatile(LOCK_PREFIX "decl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+m" (v->counter)); ++} ++ ++/** ++ * atomic_dec_unchecked - decrement atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically decrements @v by 1. ++ */ ++static inline void atomic_dec_unchecked(atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "decl %0\n" + : "+m" (v->counter)); + } + +@@ -119,7 +235,16 @@ static inline int atomic_dec_and_test(atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "decl %0; sete %1" ++ asm volatile(LOCK_PREFIX "decl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "+m" (v->counter), "=qm" (c) + : : "memory"); + return c != 0; +@@ -137,7 +262,35 @@ static inline int atomic_inc_and_test(atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "incl %0; sete %1" ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decl %0\n" ++ "into\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" ++ : "+m" (v->counter), "=qm" (c) ++ : : "memory"); ++ return c != 0; ++} ++ ++/** ++ * atomic_inc_and_test_unchecked - increment and test ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically increments @v by 1 ++ * and returns true if the result is zero, or false for all ++ * other cases. ++ */ ++static inline int atomic_inc_and_test_unchecked(atomic_unchecked_t *v) ++{ ++ unsigned char c; ++ ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ "sete %1\n" + : "+m" (v->counter), "=qm" (c) + : : "memory"); + return c != 0; +@@ -156,7 +309,16 @@ static inline int atomic_add_negative(int i, atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "addl %2,%0; sets %1" ++ asm volatile(LOCK_PREFIX "addl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sets %1\n" + : "+m" (v->counter), "=qm" (c) + : "ir" (i) : "memory"); + return c; +@@ -179,7 +341,15 @@ static inline int atomic_add_return(int i, atomic_t *v) + #endif + /* Modern 486+ processor */ + __i = i; +- asm volatile(LOCK_PREFIX "xaddl %0, %1" ++ asm volatile(LOCK_PREFIX "xaddl %0, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "movl %0, %1\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+r" (i), "+m" (v->counter) + : : "memory"); + return i + __i; +@@ -195,6 +365,38 @@ no_xadd: /* Legacy 386 processor */ + } + + /** ++ * atomic_add_return_unchecked - add integer and return ++ * @v: pointer of type atomic_unchecked_t ++ * @i: integer value to add ++ * ++ * Atomically adds @i to @v and returns @i + @v ++ */ ++static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v) ++{ ++ int __i; ++#ifdef CONFIG_M386 ++ unsigned long flags; ++ if (unlikely(boot_cpu_data.x86 <= 3)) ++ goto no_xadd; ++#endif ++ /* Modern 486+ processor */ ++ __i = i; ++ asm volatile(LOCK_PREFIX "xaddl %0, %1" ++ : "+r" (i), "+m" (v->counter) ++ : : "memory"); ++ return i + __i; ++ ++#ifdef CONFIG_M386 ++no_xadd: /* Legacy 386 processor */ ++ local_irq_save(flags); ++ __i = atomic_read_unchecked(v); ++ atomic_set_unchecked(v, i + __i); ++ local_irq_restore(flags); ++ return i + __i; ++#endif ++} ++ ++/** + * atomic_sub_return - subtract integer and return + * @v: pointer of type atomic_t + * @i: integer value to subtract +@@ -211,11 +413,21 @@ static inline int atomic_cmpxchg(atomic_t *v, int old, int new) + return cmpxchg(&v->counter, old, new); + } + ++static inline int atomic_cmpxchg_unchecked(atomic_unchecked_t *v, int old, int new) ++{ ++ return cmpxchg(&v->counter, old, new); ++} ++ + static inline int atomic_xchg(atomic_t *v, int new) + { + return xchg(&v->counter, new); + } + ++static inline int atomic_xchg_unchecked(atomic_unchecked_t *v, int new) ++{ ++ return xchg(&v->counter, new); ++} ++ + /** + * atomic_add_unless - add unless the number is already a given value + * @v: pointer of type atomic_t +@@ -227,22 +439,39 @@ static inline int atomic_xchg(atomic_t *v, int new) + */ + static inline int atomic_add_unless(atomic_t *v, int a, int u) + { +- int c, old; ++ int c, old, new; + c = atomic_read(v); + for (;;) { +- if (unlikely(c == (u))) ++ if (unlikely(c == u)) + break; +- old = atomic_cmpxchg((v), c, c + (a)); ++ ++ asm volatile("addl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "subl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=r" (new) ++ : "0" (c), "ir" (a)); ++ ++ old = atomic_cmpxchg(v, c, new); + if (likely(old == c)) + break; + c = old; + } +- return c != (u); ++ return c != u; + } + + #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) + + #define atomic_inc_return(v) (atomic_add_return(1, v)) ++static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_add_return_unchecked(1, v); ++} + #define atomic_dec_return(v) (atomic_sub_return(1, v)) + + /* These are x86-specific, used by some header files */ +@@ -266,9 +495,18 @@ typedef struct { + u64 __aligned(8) counter; + } atomic64_t; + ++#ifdef CONFIG_PAX_REFCOUNT ++typedef struct { ++ u64 __aligned(8) counter; ++} atomic64_unchecked_t; ++#else ++typedef atomic64_t atomic64_unchecked_t; ++#endif ++ + #define ATOMIC64_INIT(val) { (val) } + + extern u64 atomic64_cmpxchg(atomic64_t *ptr, u64 old_val, u64 new_val); ++extern u64 atomic64_cmpxchg_unchecked(atomic64_unchecked_t *ptr, u64 old_val, u64 new_val); + + /** + * atomic64_xchg - xchg atomic64 variable +@@ -279,6 +517,7 @@ extern u64 atomic64_cmpxchg(atomic64_t *ptr, u64 old_val, u64 new_val); + * the old value. + */ + extern u64 atomic64_xchg(atomic64_t *ptr, u64 new_val); ++extern u64 atomic64_xchg_unchecked(atomic64_unchecked_t *ptr, u64 new_val); + + /** + * atomic64_set - set atomic64 variable +@@ -290,6 +529,15 @@ extern u64 atomic64_xchg(atomic64_t *ptr, u64 new_val); + extern void atomic64_set(atomic64_t *ptr, u64 new_val); + + /** ++ * atomic64_unchecked_set - set atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * @new_val: value to assign ++ * ++ * Atomically sets the value of @ptr to @new_val. ++ */ ++extern void atomic64_set_unchecked(atomic64_unchecked_t *ptr, u64 new_val); ++ ++/** + * atomic64_read - read atomic64 variable + * @ptr: pointer to type atomic64_t + * +@@ -317,7 +565,33 @@ static inline u64 atomic64_read(atomic64_t *ptr) + return res; + } + +-extern u64 atomic64_read(atomic64_t *ptr); ++/** ++ * atomic64_read_unchecked - read atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically reads the value of @ptr and returns it. ++ */ ++static inline u64 atomic64_read_unchecked(atomic64_unchecked_t *ptr) ++{ ++ u64 res; ++ ++ /* ++ * Note, we inline this atomic64_unchecked_t primitive because ++ * it only clobbers EAX/EDX and leaves the others ++ * untouched. We also (somewhat subtly) rely on the ++ * fact that cmpxchg8b returns the current 64-bit value ++ * of the memory location we are touching: ++ */ ++ asm volatile( ++ "mov %%ebx, %%eax\n\t" ++ "mov %%ecx, %%edx\n\t" ++ LOCK_PREFIX "cmpxchg8b %1\n" ++ : "=&A" (res) ++ : "m" (*ptr) ++ ); ++ ++ return res; ++} + + /** + * atomic64_add_return - add and return +@@ -332,8 +606,11 @@ extern u64 atomic64_add_return(u64 delta, atomic64_t *ptr); + * Other variants with different arithmetic operators: + */ + extern u64 atomic64_sub_return(u64 delta, atomic64_t *ptr); ++extern u64 atomic64_sub_return_unchecked(u64 delta, atomic64_unchecked_t *ptr); + extern u64 atomic64_inc_return(atomic64_t *ptr); ++extern u64 atomic64_inc_return_unchecked(atomic64_unchecked_t *ptr); + extern u64 atomic64_dec_return(atomic64_t *ptr); ++extern u64 atomic64_dec_return_unchecked(atomic64_unchecked_t *ptr); + + /** + * atomic64_add - add integer to atomic64 variable +@@ -345,6 +622,15 @@ extern u64 atomic64_dec_return(atomic64_t *ptr); + extern void atomic64_add(u64 delta, atomic64_t *ptr); + + /** ++ * atomic64_add_unchecked - add integer to atomic64 variable ++ * @delta: integer value to add ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically adds @delta to @ptr. ++ */ ++extern void atomic64_add_unchecked(u64 delta, atomic64_unchecked_t *ptr); ++ ++/** + * atomic64_sub - subtract the atomic64 variable + * @delta: integer value to subtract + * @ptr: pointer to type atomic64_t +@@ -354,6 +640,15 @@ extern void atomic64_add(u64 delta, atomic64_t *ptr); + extern void atomic64_sub(u64 delta, atomic64_t *ptr); + + /** ++ * atomic64_sub_unchecked - subtract the atomic64 variable ++ * @delta: integer value to subtract ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically subtracts @delta from @ptr. ++ */ ++extern void atomic64_sub_unchecked(u64 delta, atomic64_unchecked_t *ptr); ++ ++/** + * atomic64_sub_and_test - subtract value from variable and test result + * @delta: integer value to subtract + * @ptr: pointer to type atomic64_t +@@ -373,6 +668,14 @@ extern int atomic64_sub_and_test(u64 delta, atomic64_t *ptr); + extern void atomic64_inc(atomic64_t *ptr); + + /** ++ * atomic64_inc_unchecked - increment atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically increments @ptr by 1. ++ */ ++extern void atomic64_inc_unchecked(atomic64_unchecked_t *ptr); ++ ++/** + * atomic64_dec - decrement atomic64 variable + * @ptr: pointer to type atomic64_t + * +@@ -381,6 +684,14 @@ extern void atomic64_inc(atomic64_t *ptr); + extern void atomic64_dec(atomic64_t *ptr); + + /** ++ * atomic64_dec_unchecked - decrement atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically decrements @ptr by 1. ++ */ ++extern void atomic64_dec_unchecked(atomic64_unchecked_t *ptr); ++ ++/** + * atomic64_dec_and_test - decrement and test + * @ptr: pointer to type atomic64_t + * +diff --git a/arch/x86/include/asm/atomic_64.h b/arch/x86/include/asm/atomic_64.h +index d605dc2..fafd7bd 100644 +--- a/arch/x86/include/asm/atomic_64.h ++++ b/arch/x86/include/asm/atomic_64.h +@@ -24,6 +24,17 @@ static inline int atomic_read(const atomic_t *v) + } + + /** ++ * atomic_read_unchecked - read atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically reads the value of @v. ++ */ ++static inline int atomic_read_unchecked(const atomic_unchecked_t *v) ++{ ++ return v->counter; ++} ++ ++/** + * atomic_set - set atomic variable + * @v: pointer of type atomic_t + * @i: required value +@@ -36,6 +47,18 @@ static inline void atomic_set(atomic_t *v, int i) + } + + /** ++ * atomic_set_unchecked - set atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * @i: required value ++ * ++ * Atomically sets the value of @v to @i. ++ */ ++static inline void atomic_set_unchecked(atomic_unchecked_t *v, int i) ++{ ++ v->counter = i; ++} ++ ++/** + * atomic_add - add integer to atomic variable + * @i: integer value to add + * @v: pointer of type atomic_t +@@ -44,7 +67,29 @@ static inline void atomic_set(atomic_t *v, int i) + */ + static inline void atomic_add(int i, atomic_t *v) + { +- asm volatile(LOCK_PREFIX "addl %1,%0" ++ asm volatile(LOCK_PREFIX "addl %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subl %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "ir" (i), "m" (v->counter)); ++} ++ ++/** ++ * atomic_add_unchecked - add integer to atomic variable ++ * @i: integer value to add ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically adds @i to @v. ++ */ ++static inline void atomic_add_unchecked(int i, atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "addl %1,%0\n" + : "=m" (v->counter) + : "ir" (i), "m" (v->counter)); + } +@@ -58,7 +103,29 @@ static inline void atomic_add(int i, atomic_t *v) + */ + static inline void atomic_sub(int i, atomic_t *v) + { +- asm volatile(LOCK_PREFIX "subl %1,%0" ++ asm volatile(LOCK_PREFIX "subl %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addl %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "ir" (i), "m" (v->counter)); ++} ++ ++/** ++ * atomic_sub_unchecked - subtract the atomic variable ++ * @i: integer value to subtract ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically subtracts @i from @v. ++ */ ++static inline void atomic_sub_unchecked(int i, atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "subl %1,%0\n" + : "=m" (v->counter) + : "ir" (i), "m" (v->counter)); + } +@@ -76,7 +143,16 @@ static inline int atomic_sub_and_test(int i, atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "subl %2,%0; sete %1" ++ asm volatile(LOCK_PREFIX "subl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "ir" (i), "m" (v->counter) : "memory"); + return c; +@@ -90,7 +166,28 @@ static inline int atomic_sub_and_test(int i, atomic_t *v) + */ + static inline void atomic_inc(atomic_t *v) + { +- asm volatile(LOCK_PREFIX "incl %0" ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "m" (v->counter)); ++} ++ ++/** ++ * atomic_inc_unchecked - increment atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically increments @v by 1. ++ */ ++static inline void atomic_inc_unchecked(atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "incl %0\n" + : "=m" (v->counter) + : "m" (v->counter)); + } +@@ -103,7 +200,28 @@ static inline void atomic_inc(atomic_t *v) + */ + static inline void atomic_dec(atomic_t *v) + { +- asm volatile(LOCK_PREFIX "decl %0" ++ asm volatile(LOCK_PREFIX "decl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "m" (v->counter)); ++} ++ ++/** ++ * atomic_dec_unchecked - decrement atomic variable ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically decrements @v by 1. ++ */ ++static inline void atomic_dec_unchecked(atomic_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "decl %0\n" + : "=m" (v->counter) + : "m" (v->counter)); + } +@@ -120,7 +238,16 @@ static inline int atomic_dec_and_test(atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "decl %0; sete %1" ++ asm volatile(LOCK_PREFIX "decl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "m" (v->counter) : "memory"); + return c != 0; +@@ -138,7 +265,35 @@ static inline int atomic_inc_and_test(atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "incl %0; sete %1" ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" ++ : "=m" (v->counter), "=qm" (c) ++ : "m" (v->counter) : "memory"); ++ return c != 0; ++} ++ ++/** ++ * atomic_inc_and_test_unchecked - increment and test ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically increments @v by 1 ++ * and returns true if the result is zero, or false for all ++ * other cases. ++ */ ++static inline int atomic_inc_and_test_unchecked(atomic_unchecked_t *v) ++{ ++ unsigned char c; ++ ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "m" (v->counter) : "memory"); + return c != 0; +@@ -157,7 +312,16 @@ static inline int atomic_add_negative(int i, atomic_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "addl %2,%0; sets %1" ++ asm volatile(LOCK_PREFIX "addl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sets %1\n" + : "=m" (v->counter), "=qm" (c) + : "ir" (i), "m" (v->counter) : "memory"); + return c; +@@ -173,7 +337,31 @@ static inline int atomic_add_negative(int i, atomic_t *v) + static inline int atomic_add_return(int i, atomic_t *v) + { + int __i = i; +- asm volatile(LOCK_PREFIX "xaddl %0, %1" ++ asm volatile(LOCK_PREFIX "xaddl %0, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "movl %0, %1\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+r" (i), "+m" (v->counter) ++ : : "memory"); ++ return i + __i; ++} ++ ++/** ++ * atomic_add_return_unchecked - add and return ++ * @i: integer value to add ++ * @v: pointer of type atomic_unchecked_t ++ * ++ * Atomically adds @i to @v and returns @i + @v ++ */ ++static inline int atomic_add_return_unchecked(int i, atomic_unchecked_t *v) ++{ ++ int __i = i; ++ asm volatile(LOCK_PREFIX "xaddl %0, %1\n" + : "+r" (i), "+m" (v->counter) + : : "memory"); + return i + __i; +@@ -185,6 +373,10 @@ static inline int atomic_sub_return(int i, atomic_t *v) + } + + #define atomic_inc_return(v) (atomic_add_return(1, v)) ++static inline int atomic_inc_return_unchecked(atomic_unchecked_t *v) ++{ ++ return atomic_add_return_unchecked(1, v); ++} + #define atomic_dec_return(v) (atomic_sub_return(1, v)) + + /* The 64-bit atomic type */ +@@ -204,6 +396,18 @@ static inline long atomic64_read(const atomic64_t *v) + } + + /** ++ * atomic64_read_unchecked - read atomic64 variable ++ * @v: pointer of type atomic64_unchecked_t ++ * ++ * Atomically reads the value of @v. ++ * Doesn't imply a read memory barrier. ++ */ ++static inline long atomic64_read_unchecked(const atomic64_unchecked_t *v) ++{ ++ return v->counter; ++} ++ ++/** + * atomic64_set - set atomic64 variable + * @v: pointer to type atomic64_t + * @i: required value +@@ -216,6 +420,18 @@ static inline void atomic64_set(atomic64_t *v, long i) + } + + /** ++ * atomic64_set_unchecked - set atomic64 variable ++ * @v: pointer to type atomic64_unchecked_t ++ * @i: required value ++ * ++ * Atomically sets the value of @v to @i. ++ */ ++static inline void atomic64_set_unchecked(atomic64_unchecked_t *v, long i) ++{ ++ v->counter = i; ++} ++ ++/** + * atomic64_add - add integer to atomic64 variable + * @i: integer value to add + * @v: pointer to type atomic64_t +@@ -224,6 +440,28 @@ static inline void atomic64_set(atomic64_t *v, long i) + */ + static inline void atomic64_add(long i, atomic64_t *v) + { ++ asm volatile(LOCK_PREFIX "addq %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subq %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "er" (i), "m" (v->counter)); ++} ++ ++/** ++ * atomic64_add_unchecked - add integer to atomic64 variable ++ * @i: integer value to add ++ * @v: pointer to type atomic64_unchecked_t ++ * ++ * Atomically adds @i to @v. ++ */ ++static inline void atomic64_add_unchecked(long i, atomic64_unchecked_t *v) ++{ + asm volatile(LOCK_PREFIX "addq %1,%0" + : "=m" (v->counter) + : "er" (i), "m" (v->counter)); +@@ -238,7 +476,15 @@ static inline void atomic64_add(long i, atomic64_t *v) + */ + static inline void atomic64_sub(long i, atomic64_t *v) + { +- asm volatile(LOCK_PREFIX "subq %1,%0" ++ asm volatile(LOCK_PREFIX "subq %1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addq %1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "=m" (v->counter) + : "er" (i), "m" (v->counter)); + } +@@ -256,7 +502,16 @@ static inline int atomic64_sub_and_test(long i, atomic64_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "subq %2,%0; sete %1" ++ asm volatile(LOCK_PREFIX "subq %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "addq %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "er" (i), "m" (v->counter) : "memory"); + return c; +@@ -270,6 +525,27 @@ static inline int atomic64_sub_and_test(long i, atomic64_t *v) + */ + static inline void atomic64_inc(atomic64_t *v) + { ++ asm volatile(LOCK_PREFIX "incq %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decq %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "m" (v->counter)); ++} ++ ++/** ++ * atomic64_inc_unchecked - increment atomic64 variable ++ * @v: pointer to type atomic64_unchecked_t ++ * ++ * Atomically increments @v by 1. ++ */ ++static inline void atomic64_inc_unchecked(atomic64_unchecked_t *v) ++{ + asm volatile(LOCK_PREFIX "incq %0" + : "=m" (v->counter) + : "m" (v->counter)); +@@ -283,7 +559,28 @@ static inline void atomic64_inc(atomic64_t *v) + */ + static inline void atomic64_dec(atomic64_t *v) + { +- asm volatile(LOCK_PREFIX "decq %0" ++ asm volatile(LOCK_PREFIX "decq %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incq %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=m" (v->counter) ++ : "m" (v->counter)); ++} ++ ++/** ++ * atomic64_dec_unchecked - decrement atomic64 variable ++ * @v: pointer to type atomic64_t ++ * ++ * Atomically decrements @v by 1. ++ */ ++static inline void atomic64_dec_unchecked(atomic64_unchecked_t *v) ++{ ++ asm volatile(LOCK_PREFIX "decq %0\n" + : "=m" (v->counter) + : "m" (v->counter)); + } +@@ -300,7 +597,16 @@ static inline int atomic64_dec_and_test(atomic64_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "decq %0; sete %1" ++ asm volatile(LOCK_PREFIX "decq %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "incq %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "m" (v->counter) : "memory"); + return c != 0; +@@ -318,7 +624,16 @@ static inline int atomic64_inc_and_test(atomic64_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "incq %0; sete %1" ++ asm volatile(LOCK_PREFIX "incq %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decq %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "=m" (v->counter), "=qm" (c) + : "m" (v->counter) : "memory"); + return c != 0; +@@ -337,7 +652,16 @@ static inline int atomic64_add_negative(long i, atomic64_t *v) + { + unsigned char c; + +- asm volatile(LOCK_PREFIX "addq %2,%0; sets %1" ++ asm volatile(LOCK_PREFIX "addq %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subq %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sets %1\n" + : "=m" (v->counter), "=qm" (c) + : "er" (i), "m" (v->counter) : "memory"); + return c; +@@ -353,7 +677,31 @@ static inline int atomic64_add_negative(long i, atomic64_t *v) + static inline long atomic64_add_return(long i, atomic64_t *v) + { + long __i = i; +- asm volatile(LOCK_PREFIX "xaddq %0, %1;" ++ asm volatile(LOCK_PREFIX "xaddq %0, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "movq %0, %1\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "+r" (i), "+m" (v->counter) ++ : : "memory"); ++ return i + __i; ++} ++ ++/** ++ * atomic64_add_return_unchecked - add and return ++ * @i: integer value to add ++ * @v: pointer to type atomic64_unchecked_t ++ * ++ * Atomically adds @i to @v and returns @i + @v ++ */ ++static inline long atomic64_add_return_unchecked(long i, atomic64_unchecked_t *v) ++{ ++ long __i = i; ++ asm volatile(LOCK_PREFIX "xaddq %0, %1" + : "+r" (i), "+m" (v->counter) + : : "memory"); + return i + __i; +@@ -365,6 +713,10 @@ static inline long atomic64_sub_return(long i, atomic64_t *v) + } + + #define atomic64_inc_return(v) (atomic64_add_return(1, (v))) ++static inline long atomic64_inc_return_unchecked(atomic64_unchecked_t *v) ++{ ++ return atomic64_add_return_unchecked(1, v); ++} + #define atomic64_dec_return(v) (atomic64_sub_return(1, (v))) + + static inline long atomic64_cmpxchg(atomic64_t *v, long old, long new) +@@ -372,21 +724,41 @@ static inline long atomic64_cmpxchg(atomic64_t *v, long old, long new) + return cmpxchg(&v->counter, old, new); + } + ++static inline long atomic64_cmpxchg_unchecked(atomic64_unchecked_t *v, long old, long new) ++{ ++ return cmpxchg(&v->counter, old, new); ++} ++ + static inline long atomic64_xchg(atomic64_t *v, long new) + { + return xchg(&v->counter, new); + } + ++static inline long atomic64_xchg_unchecked(atomic64_unchecked_t *v, long new) ++{ ++ return xchg(&v->counter, new); ++} ++ + static inline long atomic_cmpxchg(atomic_t *v, int old, int new) + { + return cmpxchg(&v->counter, old, new); + } + ++static inline long atomic_cmpxchg_unchecked(atomic_unchecked_t *v, int old, int new) ++{ ++ return cmpxchg(&v->counter, old, new); ++} ++ + static inline long atomic_xchg(atomic_t *v, int new) + { + return xchg(&v->counter, new); + } + ++static inline long atomic_xchg_unchecked(atomic_unchecked_t *v, int new) ++{ ++ return xchg(&v->counter, new); ++} ++ + /** + * atomic_add_unless - add unless the number is a given value + * @v: pointer of type atomic_t +@@ -398,17 +770,30 @@ static inline long atomic_xchg(atomic_t *v, int new) + */ + static inline int atomic_add_unless(atomic_t *v, int a, int u) + { +- int c, old; ++ int c, old, new; + c = atomic_read(v); + for (;;) { +- if (unlikely(c == (u))) ++ if (unlikely(c == u)) + break; +- old = atomic_cmpxchg((v), c, c + (a)); ++ ++ asm volatile("addl %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "subl %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=r" (new) ++ : "0" (c), "ir" (a)); ++ ++ old = atomic_cmpxchg(v, c, new); + if (likely(old == c)) + break; + c = old; + } +- return c != (u); ++ return c != u; + } + + #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) +@@ -424,17 +809,30 @@ static inline int atomic_add_unless(atomic_t *v, int a, int u) + */ + static inline int atomic64_add_unless(atomic64_t *v, long a, long u) + { +- long c, old; ++ long c, old, new; + c = atomic64_read(v); + for (;;) { +- if (unlikely(c == (u))) ++ if (unlikely(c == u)) + break; +- old = atomic64_cmpxchg((v), c, c + (a)); ++ ++ asm volatile("addq %2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "subq %2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ : "=r" (new) ++ : "0" (c), "er" (a)); ++ ++ old = atomic64_cmpxchg(v, c, new); + if (likely(old == c)) + break; + c = old; + } +- return c != (u); ++ return c != u; + } + + /** +diff --git a/arch/x86/include/asm/bitops.h b/arch/x86/include/asm/bitops.h +index 02b47a6..d5c4b15 100644 +--- a/arch/x86/include/asm/bitops.h ++++ b/arch/x86/include/asm/bitops.h +@@ -38,7 +38,7 @@ + * a mask operation on a byte. + */ + #define IS_IMMEDIATE(nr) (__builtin_constant_p(nr)) +-#define CONST_MASK_ADDR(nr, addr) BITOP_ADDR((void *)(addr) + ((nr)>>3)) ++#define CONST_MASK_ADDR(nr, addr) BITOP_ADDR((volatile void *)(addr) + ((nr)>>3)) + #define CONST_MASK(nr) (1 << ((nr) & 7)) + + /** +diff --git a/arch/x86/include/asm/boot.h b/arch/x86/include/asm/boot.h +index 7a10659..8bbf355 100644 +--- a/arch/x86/include/asm/boot.h ++++ b/arch/x86/include/asm/boot.h +@@ -11,10 +11,15 @@ + #include + + /* Physical address where kernel should be loaded. */ +-#define LOAD_PHYSICAL_ADDR ((CONFIG_PHYSICAL_START \ ++#define ____LOAD_PHYSICAL_ADDR ((CONFIG_PHYSICAL_START \ + + (CONFIG_PHYSICAL_ALIGN - 1)) \ + & ~(CONFIG_PHYSICAL_ALIGN - 1)) + ++#ifndef __ASSEMBLY__ ++extern unsigned char __LOAD_PHYSICAL_ADDR[]; ++#define LOAD_PHYSICAL_ADDR ((unsigned long)__LOAD_PHYSICAL_ADDR) ++#endif ++ + /* Minimum kernel alignment, as a power of two */ + #ifdef CONFIG_X86_64 + #define MIN_KERNEL_ALIGN_LG2 PMD_SHIFT +diff --git a/arch/x86/include/asm/cache.h b/arch/x86/include/asm/cache.h +index 549860d..7d45f68 100644 +--- a/arch/x86/include/asm/cache.h ++++ b/arch/x86/include/asm/cache.h +@@ -5,9 +5,10 @@ + + /* L1 cache line size */ + #define L1_CACHE_SHIFT (CONFIG_X86_L1_CACHE_SHIFT) +-#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) ++#define L1_CACHE_BYTES (_AC(1,UL) << L1_CACHE_SHIFT) + + #define __read_mostly __attribute__((__section__(".data.read_mostly"))) ++#define __read_only __attribute__((__section__(".data.read_only"))) + + #ifdef CONFIG_X86_VSMP + /* vSMP Internode cacheline shift */ +diff --git a/arch/x86/include/asm/cacheflush.h b/arch/x86/include/asm/cacheflush.h +index b54f6af..5b376a6 100644 +--- a/arch/x86/include/asm/cacheflush.h ++++ b/arch/x86/include/asm/cacheflush.h +@@ -60,7 +60,7 @@ PAGEFLAG(WC, WC) + static inline unsigned long get_page_memtype(struct page *pg) + { + if (!PageUncached(pg) && !PageWC(pg)) +- return -1; ++ return ~0UL; + else if (!PageUncached(pg) && PageWC(pg)) + return _PAGE_CACHE_WC; + else if (PageUncached(pg) && !PageWC(pg)) +@@ -85,7 +85,7 @@ static inline void set_page_memtype(struct page *pg, unsigned long memtype) + SetPageWC(pg); + break; + default: +- case -1: ++ case ~0UL: + ClearPageUncached(pg); + ClearPageWC(pg); + break; +diff --git a/arch/x86/include/asm/calling.h b/arch/x86/include/asm/calling.h +index 0e63c9a..ab8d972 100644 +--- a/arch/x86/include/asm/calling.h ++++ b/arch/x86/include/asm/calling.h +@@ -52,32 +52,32 @@ For 32-bit we have the following conventions - kernel is built with + * for assembly code: + */ + +-#define R15 0 +-#define R14 8 +-#define R13 16 +-#define R12 24 +-#define RBP 32 +-#define RBX 40 ++#define R15 (0) ++#define R14 (8) ++#define R13 (16) ++#define R12 (24) ++#define RBP (32) ++#define RBX (40) + + /* arguments: interrupts/non tracing syscalls only save up to here: */ +-#define R11 48 +-#define R10 56 +-#define R9 64 +-#define R8 72 +-#define RAX 80 +-#define RCX 88 +-#define RDX 96 +-#define RSI 104 +-#define RDI 112 +-#define ORIG_RAX 120 /* + error_code */ ++#define R11 (48) ++#define R10 (56) ++#define R9 (64) ++#define R8 (72) ++#define RAX (80) ++#define RCX (88) ++#define RDX (96) ++#define RSI (104) ++#define RDI (112) ++#define ORIG_RAX (120) /* + error_code */ + /* end of arguments */ + + /* cpu exception frame or undefined in case of fast syscall: */ +-#define RIP 128 +-#define CS 136 +-#define EFLAGS 144 +-#define RSP 152 +-#define SS 160 ++#define RIP (128) ++#define CS (136) ++#define EFLAGS (144) ++#define RSP (152) ++#define SS (160) + + #define ARGOFFSET R11 + #define SWFRAME ORIG_RAX +diff --git a/arch/x86/include/asm/checksum_32.h b/arch/x86/include/asm/checksum_32.h +index 46fc474..b02b0f9 100644 +--- a/arch/x86/include/asm/checksum_32.h ++++ b/arch/x86/include/asm/checksum_32.h +@@ -31,6 +31,14 @@ asmlinkage __wsum csum_partial_copy_generic(const void *src, void *dst, + int len, __wsum sum, + int *src_err_ptr, int *dst_err_ptr); + ++asmlinkage __wsum csum_partial_copy_generic_to_user(const void *src, void *dst, ++ int len, __wsum sum, ++ int *src_err_ptr, int *dst_err_ptr); ++ ++asmlinkage __wsum csum_partial_copy_generic_from_user(const void *src, void *dst, ++ int len, __wsum sum, ++ int *src_err_ptr, int *dst_err_ptr); ++ + /* + * Note: when you get a NULL pointer exception here this means someone + * passed in an incorrect kernel address to one of these functions. +@@ -50,7 +58,7 @@ static inline __wsum csum_partial_copy_from_user(const void __user *src, + int *err_ptr) + { + might_sleep(); +- return csum_partial_copy_generic((__force void *)src, dst, ++ return csum_partial_copy_generic_from_user((__force void *)src, dst, + len, sum, err_ptr, NULL); + } + +@@ -178,7 +186,7 @@ static inline __wsum csum_and_copy_to_user(const void *src, + { + might_sleep(); + if (access_ok(VERIFY_WRITE, dst, len)) +- return csum_partial_copy_generic(src, (__force void *)dst, ++ return csum_partial_copy_generic_to_user(src, (__force void *)dst, + len, sum, NULL, err_ptr); + + if (len) +diff --git a/arch/x86/include/asm/desc.h b/arch/x86/include/asm/desc.h +index 617bd56..7b047a1 100644 +--- a/arch/x86/include/asm/desc.h ++++ b/arch/x86/include/asm/desc.h +@@ -4,6 +4,7 @@ + #include + #include + #include ++#include + #include + + static inline void fill_ldt(struct desc_struct *desc, +@@ -15,6 +16,7 @@ static inline void fill_ldt(struct desc_struct *desc, + desc->base1 = (info->base_addr & 0x00ff0000) >> 16; + desc->type = (info->read_exec_only ^ 1) << 1; + desc->type |= info->contents << 2; ++ desc->type |= info->seg_not_present ^ 1; + desc->s = 1; + desc->dpl = 0x3; + desc->p = info->seg_not_present ^ 1; +@@ -31,16 +33,12 @@ static inline void fill_ldt(struct desc_struct *desc, + } + + extern struct desc_ptr idt_descr; +-extern gate_desc idt_table[]; +- +-struct gdt_page { +- struct desc_struct gdt[GDT_ENTRIES]; +-} __attribute__((aligned(PAGE_SIZE))); +-DECLARE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page); ++extern gate_desc idt_table[256]; + ++extern struct desc_struct cpu_gdt_table[NR_CPUS][PAGE_SIZE / sizeof(struct desc_struct)]; + static inline struct desc_struct *get_cpu_gdt_table(unsigned int cpu) + { +- return per_cpu(gdt_page, cpu).gdt; ++ return cpu_gdt_table[cpu]; + } + + #ifdef CONFIG_X86_64 +@@ -65,9 +63,14 @@ static inline void pack_gate(gate_desc *gate, unsigned char type, + unsigned long base, unsigned dpl, unsigned flags, + unsigned short seg) + { +- gate->a = (seg << 16) | (base & 0xffff); +- gate->b = (base & 0xffff0000) | +- (((0x80 | type | (dpl << 5)) & 0xff) << 8); ++ gate->gate.offset_low = base; ++ gate->gate.seg = seg; ++ gate->gate.reserved = 0; ++ gate->gate.type = type; ++ gate->gate.s = 0; ++ gate->gate.dpl = dpl; ++ gate->gate.p = 1; ++ gate->gate.offset_high = base >> 16; + } + + #endif +@@ -115,13 +118,17 @@ static inline void paravirt_free_ldt(struct desc_struct *ldt, unsigned entries) + static inline void native_write_idt_entry(gate_desc *idt, int entry, + const gate_desc *gate) + { ++ pax_open_kernel(); + memcpy(&idt[entry], gate, sizeof(*gate)); ++ pax_close_kernel(); + } + + static inline void native_write_ldt_entry(struct desc_struct *ldt, int entry, + const void *desc) + { ++ pax_open_kernel(); + memcpy(&ldt[entry], desc, 8); ++ pax_close_kernel(); + } + + static inline void native_write_gdt_entry(struct desc_struct *gdt, int entry, +@@ -139,7 +146,10 @@ static inline void native_write_gdt_entry(struct desc_struct *gdt, int entry, + size = sizeof(struct desc_struct); + break; + } ++ ++ pax_open_kernel(); + memcpy(&gdt[entry], desc, size); ++ pax_close_kernel(); + } + + static inline void pack_descriptor(struct desc_struct *desc, unsigned long base, +@@ -211,7 +221,9 @@ static inline void native_set_ldt(const void *addr, unsigned int entries) + + static inline void native_load_tr_desc(void) + { ++ pax_open_kernel(); + asm volatile("ltr %w0"::"q" (GDT_ENTRY_TSS*8)); ++ pax_close_kernel(); + } + + static inline void native_load_gdt(const struct desc_ptr *dtr) +@@ -246,8 +258,10 @@ static inline void native_load_tls(struct thread_struct *t, unsigned int cpu) + unsigned int i; + struct desc_struct *gdt = get_cpu_gdt_table(cpu); + ++ pax_open_kernel(); + for (i = 0; i < GDT_ENTRY_TLS_ENTRIES; i++) + gdt[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i]; ++ pax_close_kernel(); + } + + #define _LDT_empty(info) \ +@@ -309,7 +323,7 @@ static inline void set_desc_limit(struct desc_struct *desc, unsigned long limit) + desc->limit = (limit >> 16) & 0xf; + } + +-static inline void _set_gate(int gate, unsigned type, void *addr, ++static inline void _set_gate(int gate, unsigned type, const void *addr, + unsigned dpl, unsigned ist, unsigned seg) + { + gate_desc s; +@@ -327,7 +341,7 @@ static inline void _set_gate(int gate, unsigned type, void *addr, + * Pentium F0 0F bugfix can have resulted in the mapped + * IDT being write-protected. + */ +-static inline void set_intr_gate(unsigned int n, void *addr) ++static inline void set_intr_gate(unsigned int n, const void *addr) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_INTERRUPT, addr, 0, 0, __KERNEL_CS); +@@ -356,19 +370,19 @@ static inline void alloc_intr_gate(unsigned int n, void *addr) + /* + * This routine sets up an interrupt gate at directory privilege level 3. + */ +-static inline void set_system_intr_gate(unsigned int n, void *addr) ++static inline void set_system_intr_gate(unsigned int n, const void *addr) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_INTERRUPT, addr, 0x3, 0, __KERNEL_CS); + } + +-static inline void set_system_trap_gate(unsigned int n, void *addr) ++static inline void set_system_trap_gate(unsigned int n, const void *addr) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_TRAP, addr, 0x3, 0, __KERNEL_CS); + } + +-static inline void set_trap_gate(unsigned int n, void *addr) ++static inline void set_trap_gate(unsigned int n, const void *addr) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_TRAP, addr, 0, 0, __KERNEL_CS); +@@ -377,19 +391,31 @@ static inline void set_trap_gate(unsigned int n, void *addr) + static inline void set_task_gate(unsigned int n, unsigned int gdt_entry) + { + BUG_ON((unsigned)n > 0xFF); +- _set_gate(n, GATE_TASK, (void *)0, 0, 0, (gdt_entry<<3)); ++ _set_gate(n, GATE_TASK, (const void *)0, 0, 0, (gdt_entry<<3)); + } + +-static inline void set_intr_gate_ist(int n, void *addr, unsigned ist) ++static inline void set_intr_gate_ist(int n, const void *addr, unsigned ist) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_INTERRUPT, addr, 0, ist, __KERNEL_CS); + } + +-static inline void set_system_intr_gate_ist(int n, void *addr, unsigned ist) ++static inline void set_system_intr_gate_ist(int n, const void *addr, unsigned ist) + { + BUG_ON((unsigned)n > 0xFF); + _set_gate(n, GATE_INTERRUPT, addr, 0x3, ist, __KERNEL_CS); + } + ++#ifdef CONFIG_X86_32 ++static inline void set_user_cs(unsigned long base, unsigned long limit, int cpu) ++{ ++ struct desc_struct d; ++ ++ if (likely(limit)) ++ limit = (limit - 1UL) >> PAGE_SHIFT; ++ pack_descriptor(&d, base, limit, 0xFB, 0xC); ++ write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_DEFAULT_USER_CS, &d, DESCTYPE_S); ++} ++#endif ++ + #endif /* _ASM_X86_DESC_H */ +diff --git a/arch/x86/include/asm/desc_defs.h b/arch/x86/include/asm/desc_defs.h +index 9d66848..6b4a691 100644 +--- a/arch/x86/include/asm/desc_defs.h ++++ b/arch/x86/include/asm/desc_defs.h +@@ -31,6 +31,12 @@ struct desc_struct { + unsigned base1: 8, type: 4, s: 1, dpl: 2, p: 1; + unsigned limit: 4, avl: 1, l: 1, d: 1, g: 1, base2: 8; + }; ++ struct { ++ u16 offset_low; ++ u16 seg; ++ unsigned reserved: 8, type: 4, s: 1, dpl: 2, p: 1; ++ unsigned offset_high: 16; ++ } gate; + }; + } __attribute__((packed)); + +diff --git a/arch/x86/include/asm/device.h b/arch/x86/include/asm/device.h +index cee34e9..a7c3fa2 100644 +--- a/arch/x86/include/asm/device.h ++++ b/arch/x86/include/asm/device.h +@@ -6,7 +6,7 @@ struct dev_archdata { + void *acpi_handle; + #endif + #ifdef CONFIG_X86_64 +-struct dma_map_ops *dma_ops; ++ const struct dma_map_ops *dma_ops; + #endif + #ifdef CONFIG_DMAR + void *iommu; /* hook for IOMMU specific extension */ +diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h +index 6a25d5d..786b202 100644 +--- a/arch/x86/include/asm/dma-mapping.h ++++ b/arch/x86/include/asm/dma-mapping.h +@@ -25,9 +25,9 @@ extern int iommu_merge; + extern struct device x86_dma_fallback_dev; + extern int panic_on_overflow; + +-extern struct dma_map_ops *dma_ops; ++extern const struct dma_map_ops *dma_ops; + +-static inline struct dma_map_ops *get_dma_ops(struct device *dev) ++static inline const struct dma_map_ops *get_dma_ops(struct device *dev) + { + #ifdef CONFIG_X86_32 + return dma_ops; +@@ -44,7 +44,7 @@ static inline struct dma_map_ops *get_dma_ops(struct device *dev) + /* Make sure we keep the same behaviour */ + static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + if (ops->mapping_error) + return ops->mapping_error(dev, dma_addr); + +@@ -122,7 +122,7 @@ static inline void * + dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, + gfp_t gfp) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + void *memory; + + gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32); +@@ -149,7 +149,7 @@ dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, + static inline void dma_free_coherent(struct device *dev, size_t size, + void *vaddr, dma_addr_t bus) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + + WARN_ON(irqs_disabled()); /* for portability */ + +diff --git a/arch/x86/include/asm/e820.h b/arch/x86/include/asm/e820.h +index 40b4e61..40d8133 100644 +--- a/arch/x86/include/asm/e820.h ++++ b/arch/x86/include/asm/e820.h +@@ -133,7 +133,7 @@ extern char *default_machine_specific_memory_setup(void); + #define ISA_END_ADDRESS 0x100000 + #define is_ISA_range(s, e) ((s) >= ISA_START_ADDRESS && (e) < ISA_END_ADDRESS) + +-#define BIOS_BEGIN 0x000a0000 ++#define BIOS_BEGIN 0x000c0000 + #define BIOS_END 0x00100000 + + #ifdef __KERNEL__ +diff --git a/arch/x86/include/asm/elf.h b/arch/x86/include/asm/elf.h +index 8ac9d9a..0a6c96e 100644 +--- a/arch/x86/include/asm/elf.h ++++ b/arch/x86/include/asm/elf.h +@@ -257,7 +257,25 @@ extern int force_personality32; + the loader. We need to make sure that it is out of the way of the program + that it will "exec", and that there is sufficient room for the brk. */ + ++#ifdef CONFIG_PAX_SEGMEXEC ++#define ELF_ET_DYN_BASE ((current->mm->pax_flags & MF_PAX_SEGMEXEC) ? SEGMEXEC_TASK_SIZE/3*2 : TASK_SIZE/3*2) ++#else + #define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2) ++#endif ++ ++#ifdef CONFIG_PAX_ASLR ++#ifdef CONFIG_X86_32 ++#define PAX_ELF_ET_DYN_BASE 0x10000000UL ++ ++#define PAX_DELTA_MMAP_LEN (current->mm->pax_flags & MF_PAX_SEGMEXEC ? 15 : 16) ++#define PAX_DELTA_STACK_LEN (current->mm->pax_flags & MF_PAX_SEGMEXEC ? 15 : 16) ++#else ++#define PAX_ELF_ET_DYN_BASE 0x400000UL ++ ++#define PAX_DELTA_MMAP_LEN ((test_thread_flag(TIF_IA32)) ? 16 : TASK_SIZE_MAX_SHIFT - PAGE_SHIFT - 3) ++#define PAX_DELTA_STACK_LEN ((test_thread_flag(TIF_IA32)) ? 16 : TASK_SIZE_MAX_SHIFT - PAGE_SHIFT - 3) ++#endif ++#endif + + /* This yields a mask that user programs can use to figure out what + instruction set this CPU supports. This could be done in user space, +@@ -310,9 +328,7 @@ do { \ + + #define ARCH_DLINFO \ + do { \ +- if (vdso_enabled) \ +- NEW_AUX_ENT(AT_SYSINFO_EHDR, \ +- (unsigned long)current->mm->context.vdso); \ ++ NEW_AUX_ENT(AT_SYSINFO_EHDR, current->mm->context.vdso); \ + } while (0) + + #define AT_SYSINFO 32 +@@ -323,7 +339,7 @@ do { \ + + #endif /* !CONFIG_X86_32 */ + +-#define VDSO_CURRENT_BASE ((unsigned long)current->mm->context.vdso) ++#define VDSO_CURRENT_BASE (current->mm->context.vdso) + + #define VDSO_ENTRY \ + ((unsigned long)VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall)) +@@ -337,7 +353,4 @@ extern int arch_setup_additional_pages(struct linux_binprm *bprm, + extern int syscall32_setup_pages(struct linux_binprm *, int exstack); + #define compat_arch_setup_additional_pages syscall32_setup_pages + +-extern unsigned long arch_randomize_brk(struct mm_struct *mm); +-#define arch_randomize_brk arch_randomize_brk +- + #endif /* _ASM_X86_ELF_H */ +diff --git a/arch/x86/include/asm/emergency-restart.h b/arch/x86/include/asm/emergency-restart.h +index cc70c1c..d96d011 100644 +--- a/arch/x86/include/asm/emergency-restart.h ++++ b/arch/x86/include/asm/emergency-restart.h +@@ -15,6 +15,6 @@ enum reboot_type { + + extern enum reboot_type reboot_type; + +-extern void machine_emergency_restart(void); ++extern void machine_emergency_restart(void) __noreturn; + + #endif /* _ASM_X86_EMERGENCY_RESTART_H */ +diff --git a/arch/x86/include/asm/futex.h b/arch/x86/include/asm/futex.h +index 1f11ce4..7caabd1 100644 +--- a/arch/x86/include/asm/futex.h ++++ b/arch/x86/include/asm/futex.h +@@ -12,16 +12,18 @@ + #include + + #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ ++ typecheck(u32 __user *, uaddr); \ + asm volatile("1:\t" insn "\n" \ + "2:\t.section .fixup,\"ax\"\n" \ + "3:\tmov\t%3, %1\n" \ + "\tjmp\t2b\n" \ + "\t.previous\n" \ + _ASM_EXTABLE(1b, 3b) \ +- : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \ ++ : "=r" (oldval), "=r" (ret), "+m" (*(u32 __user *)____m(uaddr))\ + : "i" (-EFAULT), "0" (oparg), "1" (0)) + + #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ ++ typecheck(u32 __user *, uaddr); \ + asm volatile("1:\tmovl %2, %0\n" \ + "\tmovl\t%0, %3\n" \ + "\t" insn "\n" \ +@@ -34,10 +36,10 @@ + _ASM_EXTABLE(1b, 4b) \ + _ASM_EXTABLE(2b, 4b) \ + : "=&a" (oldval), "=&r" (ret), \ +- "+m" (*uaddr), "=&r" (tem) \ ++ "+m" (*(u32 __user *)____m(uaddr)), "=&r" (tem) \ + : "r" (oparg), "i" (-EFAULT), "1" (0)) + +-static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr) ++static inline int futex_atomic_op_inuser(int encoded_op, u32 __user *uaddr) + { + int op = (encoded_op >> 28) & 7; + int cmp = (encoded_op >> 24) & 15; +@@ -61,10 +63,10 @@ static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr) + + switch (op) { + case FUTEX_OP_SET: +- __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg); ++ __futex_atomic_op1(__copyuser_seg"xchgl %0, %2", ret, oldval, uaddr, oparg); + break; + case FUTEX_OP_ADD: +- __futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval, ++ __futex_atomic_op1(LOCK_PREFIX __copyuser_seg"xaddl %0, %2", ret, oldval, + uaddr, oparg); + break; + case FUTEX_OP_OR: +@@ -109,7 +111,7 @@ static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr) + return ret; + } + +-static inline int futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, ++static inline int futex_atomic_cmpxchg_inatomic(u32 __user *uaddr, int oldval, + int newval) + { + +@@ -119,16 +121,16 @@ static inline int futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, + return -ENOSYS; + #endif + +- if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int))) ++ if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) + return -EFAULT; + +- asm volatile("1:\t" LOCK_PREFIX "cmpxchgl %3, %1\n" ++ asm volatile("1:\t" LOCK_PREFIX __copyuser_seg"cmpxchgl %3, %1\n" + "2:\t.section .fixup, \"ax\"\n" + "3:\tmov %2, %0\n" + "\tjmp 2b\n" + "\t.previous\n" + _ASM_EXTABLE(1b, 3b) +- : "=a" (oldval), "+m" (*uaddr) ++ : "=a" (oldval), "+m" (*(u32 *)____m(uaddr)) + : "i" (-EFAULT), "r" (newval), "0" (oldval) + : "memory" + ); +diff --git a/arch/x86/include/asm/hw_irq.h b/arch/x86/include/asm/hw_irq.h +index ba180d9..3bad351 100644 +--- a/arch/x86/include/asm/hw_irq.h ++++ b/arch/x86/include/asm/hw_irq.h +@@ -92,8 +92,8 @@ extern void setup_ioapic_dest(void); + extern void enable_IO_APIC(void); + + /* Statistics */ +-extern atomic_t irq_err_count; +-extern atomic_t irq_mis_count; ++extern atomic_unchecked_t irq_err_count; ++extern atomic_unchecked_t irq_mis_count; + + /* EISA */ + extern void eisa_set_level_irq(unsigned int irq); +diff --git a/arch/x86/include/asm/i387.h b/arch/x86/include/asm/i387.h +index 0b20bbb..4cb1396 100644 +--- a/arch/x86/include/asm/i387.h ++++ b/arch/x86/include/asm/i387.h +@@ -60,6 +60,11 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx) + { + int err; + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ if ((unsigned long)fx < PAX_USER_SHADOW_BASE) ++ fx = (struct i387_fxsave_struct *)((void *)fx + PAX_USER_SHADOW_BASE); ++#endif ++ + asm volatile("1: rex64/fxrstor (%[fx])\n\t" + "2:\n" + ".section .fixup,\"ax\"\n" +@@ -105,6 +110,11 @@ static inline int fxsave_user(struct i387_fxsave_struct __user *fx) + { + int err; + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ if ((unsigned long)fx < PAX_USER_SHADOW_BASE) ++ fx = (struct i387_fxsave_struct __user *)((void __user *)fx + PAX_USER_SHADOW_BASE); ++#endif ++ + asm volatile("1: rex64/fxsave (%[fx])\n\t" + "2:\n" + ".section .fixup,\"ax\"\n" +@@ -195,13 +205,8 @@ static inline int fxrstor_checking(struct i387_fxsave_struct *fx) + } + + /* We need a safe address that is cheap to find and that is already +- in L1 during context switch. The best choices are unfortunately +- different for UP and SMP */ +-#ifdef CONFIG_SMP +-#define safe_address (__per_cpu_offset[0]) +-#else +-#define safe_address (kstat_cpu(0).cpustat.user) +-#endif ++ in L1 during context switch. */ ++#define safe_address (init_tss[smp_processor_id()].x86_tss.sp0) + + /* + * These must be called with preempt disabled +@@ -291,7 +296,7 @@ static inline void kernel_fpu_begin(void) + struct thread_info *me = current_thread_info(); + preempt_disable(); + if (me->status & TS_USEDFPU) +- __save_init_fpu(me->task); ++ __save_init_fpu(current); + else + clts(); + } +diff --git a/arch/x86/include/asm/io_32.h b/arch/x86/include/asm/io_32.h +index a299900..15c5410 100644 +--- a/arch/x86/include/asm/io_32.h ++++ b/arch/x86/include/asm/io_32.h +@@ -3,6 +3,7 @@ + + #include + #include ++#include + + /* + * This file contains the definitions for the x86 IO instructions +@@ -42,6 +43,17 @@ + + #ifdef __KERNEL__ + ++#define ARCH_HAS_VALID_PHYS_ADDR_RANGE ++static inline int valid_phys_addr_range(unsigned long addr, size_t count) ++{ ++ return ((addr + count + PAGE_SIZE - 1) >> PAGE_SHIFT) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0; ++} ++ ++static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t count) ++{ ++ return (pfn + (count >> PAGE_SHIFT)) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0; ++} ++ + #include + + #include +diff --git a/arch/x86/include/asm/io_64.h b/arch/x86/include/asm/io_64.h +index 2440678..c158b88 100644 +--- a/arch/x86/include/asm/io_64.h ++++ b/arch/x86/include/asm/io_64.h +@@ -140,6 +140,17 @@ __OUTS(l) + + #include + ++#define ARCH_HAS_VALID_PHYS_ADDR_RANGE ++static inline int valid_phys_addr_range(unsigned long addr, size_t count) ++{ ++ return ((addr + count + PAGE_SIZE - 1) >> PAGE_SHIFT) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0; ++} ++ ++static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t count) ++{ ++ return (pfn + (count >> PAGE_SHIFT)) < (1ULL << (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) ? 1 : 0; ++} ++ + #include + + void __memcpy_fromio(void *, unsigned long, unsigned); +diff --git a/arch/x86/include/asm/iommu.h b/arch/x86/include/asm/iommu.h +index fd6d21b..8b13915 100644 +--- a/arch/x86/include/asm/iommu.h ++++ b/arch/x86/include/asm/iommu.h +@@ -3,7 +3,7 @@ + + extern void pci_iommu_shutdown(void); + extern void no_iommu_init(void); +-extern struct dma_map_ops nommu_dma_ops; ++extern const struct dma_map_ops nommu_dma_ops; + extern int force_iommu, no_iommu; + extern int iommu_detected; + extern int iommu_pass_through; +diff --git a/arch/x86/include/asm/irqflags.h b/arch/x86/include/asm/irqflags.h +index 9e2b952..557206e 100644 +--- a/arch/x86/include/asm/irqflags.h ++++ b/arch/x86/include/asm/irqflags.h +@@ -142,6 +142,11 @@ static inline unsigned long __raw_local_irq_save(void) + sti; \ + sysexit + ++#define GET_CR0_INTO_RDI mov %cr0, %rdi ++#define SET_RDI_INTO_CR0 mov %rdi, %cr0 ++#define GET_CR3_INTO_RDI mov %cr3, %rdi ++#define SET_RDI_INTO_CR3 mov %rdi, %cr3 ++ + #else + #define INTERRUPT_RETURN iret + #define ENABLE_INTERRUPTS_SYSEXIT sti; sysexit +diff --git a/arch/x86/include/asm/kprobes.h b/arch/x86/include/asm/kprobes.h +index 4fe681d..bb6d40c 100644 +--- a/arch/x86/include/asm/kprobes.h ++++ b/arch/x86/include/asm/kprobes.h +@@ -34,13 +34,8 @@ typedef u8 kprobe_opcode_t; + #define BREAKPOINT_INSTRUCTION 0xcc + #define RELATIVEJUMP_INSTRUCTION 0xe9 + #define MAX_INSN_SIZE 16 +-#define MAX_STACK_SIZE 64 +-#define MIN_STACK_SIZE(ADDR) \ +- (((MAX_STACK_SIZE) < (((unsigned long)current_thread_info()) + \ +- THREAD_SIZE - (unsigned long)(ADDR))) \ +- ? (MAX_STACK_SIZE) \ +- : (((unsigned long)current_thread_info()) + \ +- THREAD_SIZE - (unsigned long)(ADDR))) ++#define MAX_STACK_SIZE 64UL ++#define MIN_STACK_SIZE(ADDR) min(MAX_STACK_SIZE, current->thread.sp0 - (unsigned long)(ADDR)) + + #define flush_insn_slot(p) do { } while (0) + +diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h +index 08bc2ff..2e88d1f 100644 +--- a/arch/x86/include/asm/kvm_host.h ++++ b/arch/x86/include/asm/kvm_host.h +@@ -534,9 +534,9 @@ struct kvm_x86_ops { + bool (*gb_page_enable)(void); + + const struct trace_print_flags *exit_reasons_str; +-}; ++} __do_const; + +-extern struct kvm_x86_ops *kvm_x86_ops; ++extern const struct kvm_x86_ops *kvm_x86_ops; + + int kvm_mmu_module_init(void); + void kvm_mmu_module_exit(void); +diff --git a/arch/x86/include/asm/local.h b/arch/x86/include/asm/local.h +index 47b9b6f..815aaa1 100644 +--- a/arch/x86/include/asm/local.h ++++ b/arch/x86/include/asm/local.h +@@ -18,26 +18,58 @@ typedef struct { + + static inline void local_inc(local_t *l) + { +- asm volatile(_ASM_INC "%0" ++ asm volatile(_ASM_INC "%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_DEC "%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (l->a.counter)); + } + + static inline void local_dec(local_t *l) + { +- asm volatile(_ASM_DEC "%0" ++ asm volatile(_ASM_DEC "%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_INC "%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (l->a.counter)); + } + + static inline void local_add(long i, local_t *l) + { +- asm volatile(_ASM_ADD "%1,%0" ++ asm volatile(_ASM_ADD "%1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_SUB "%1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (l->a.counter) + : "ir" (i)); + } + + static inline void local_sub(long i, local_t *l) + { +- asm volatile(_ASM_SUB "%1,%0" ++ asm volatile(_ASM_SUB "%1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_ADD "%1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (l->a.counter) + : "ir" (i)); + } +@@ -55,7 +87,16 @@ static inline int local_sub_and_test(long i, local_t *l) + { + unsigned char c; + +- asm volatile(_ASM_SUB "%2,%0; sete %1" ++ asm volatile(_ASM_SUB "%2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_ADD "%2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "+m" (l->a.counter), "=qm" (c) + : "ir" (i) : "memory"); + return c; +@@ -73,7 +114,16 @@ static inline int local_dec_and_test(local_t *l) + { + unsigned char c; + +- asm volatile(_ASM_DEC "%0; sete %1" ++ asm volatile(_ASM_DEC "%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_INC "%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "+m" (l->a.counter), "=qm" (c) + : : "memory"); + return c != 0; +@@ -91,7 +141,16 @@ static inline int local_inc_and_test(local_t *l) + { + unsigned char c; + +- asm volatile(_ASM_INC "%0; sete %1" ++ asm volatile(_ASM_INC "%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_DEC "%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sete %1\n" + : "+m" (l->a.counter), "=qm" (c) + : : "memory"); + return c != 0; +@@ -110,7 +169,16 @@ static inline int local_add_negative(long i, local_t *l) + { + unsigned char c; + +- asm volatile(_ASM_ADD "%2,%0; sets %1" ++ asm volatile(_ASM_ADD "%2,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_SUB "%2,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ "sets %1\n" + : "+m" (l->a.counter), "=qm" (c) + : "ir" (i) : "memory"); + return c; +@@ -133,7 +201,15 @@ static inline long local_add_return(long i, local_t *l) + #endif + /* Modern 486+ processor */ + __i = i; +- asm volatile(_ASM_XADD "%0, %1;" ++ asm volatile(_ASM_XADD "%0, %1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ _ASM_MOV "%0,%1\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+r" (i), "+m" (l->a.counter) + : : "memory"); + return i + __i; +diff --git a/arch/x86/include/asm/microcode.h b/arch/x86/include/asm/microcode.h +index ef51b50..514ba37 100644 +--- a/arch/x86/include/asm/microcode.h ++++ b/arch/x86/include/asm/microcode.h +@@ -12,13 +12,13 @@ struct device; + enum ucode_state { UCODE_ERROR, UCODE_OK, UCODE_NFOUND }; + + struct microcode_ops { +- enum ucode_state (*request_microcode_user) (int cpu, ++ enum ucode_state (* const request_microcode_user) (int cpu, + const void __user *buf, size_t size); + +- enum ucode_state (*request_microcode_fw) (int cpu, ++ enum ucode_state (* const request_microcode_fw) (int cpu, + struct device *device); + +- void (*microcode_fini_cpu) (int cpu); ++ void (* const microcode_fini_cpu) (int cpu); + + /* + * The generic 'microcode_core' part guarantees that +@@ -38,18 +38,18 @@ struct ucode_cpu_info { + extern struct ucode_cpu_info ucode_cpu_info[]; + + #ifdef CONFIG_MICROCODE_INTEL +-extern struct microcode_ops * __init init_intel_microcode(void); ++extern const struct microcode_ops * __init init_intel_microcode(void); + #else +-static inline struct microcode_ops * __init init_intel_microcode(void) ++static inline const struct microcode_ops * __init init_intel_microcode(void) + { + return NULL; + } + #endif /* CONFIG_MICROCODE_INTEL */ + + #ifdef CONFIG_MICROCODE_AMD +-extern struct microcode_ops * __init init_amd_microcode(void); ++extern const struct microcode_ops * __init init_amd_microcode(void); + #else +-static inline struct microcode_ops * __init init_amd_microcode(void) ++static inline const struct microcode_ops * __init init_amd_microcode(void) + { + return NULL; + } +diff --git a/arch/x86/include/asm/mman.h b/arch/x86/include/asm/mman.h +index 593e51d..fa69c9a 100644 +--- a/arch/x86/include/asm/mman.h ++++ b/arch/x86/include/asm/mman.h +@@ -5,4 +5,14 @@ + + #include + ++#ifdef __KERNEL__ ++#ifndef __ASSEMBLY__ ++#ifdef CONFIG_X86_32 ++#define arch_mmap_check i386_mmap_check ++int i386_mmap_check(unsigned long addr, unsigned long len, ++ unsigned long flags); ++#endif ++#endif ++#endif ++ + #endif /* _ASM_X86_MMAN_H */ +diff --git a/arch/x86/include/asm/mmu.h b/arch/x86/include/asm/mmu.h +index 80a1dee..239c67d 100644 +--- a/arch/x86/include/asm/mmu.h ++++ b/arch/x86/include/asm/mmu.h +@@ -9,10 +9,23 @@ + * we put the segment information here. + */ + typedef struct { +- void *ldt; ++ struct desc_struct *ldt; + int size; + struct mutex lock; +- void *vdso; ++ unsigned long vdso; ++ ++#ifdef CONFIG_X86_32 ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++ unsigned long user_cs_base; ++ unsigned long user_cs_limit; ++ ++#if defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP) ++ cpumask_t cpu_user_cs_mask; ++#endif ++ ++#endif ++#endif ++ + } mm_context_t; + + #ifdef CONFIG_SMP +diff --git a/arch/x86/include/asm/mmu_context.h b/arch/x86/include/asm/mmu_context.h +index 8b5393e..8143173 100644 +--- a/arch/x86/include/asm/mmu_context.h ++++ b/arch/x86/include/asm/mmu_context.h +@@ -24,6 +24,18 @@ void destroy_context(struct mm_struct *mm); + + static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) + { ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ unsigned int i; ++ pgd_t *pgd; ++ ++ pax_open_kernel(); ++ pgd = get_cpu_pgd(smp_processor_id()); ++ for (i = USER_PGD_PTRS; i < 2 * USER_PGD_PTRS; ++i) ++ set_pgd_batched(pgd+i, native_make_pgd(0)); ++ pax_close_kernel(); ++#endif ++ + #ifdef CONFIG_SMP + if (percpu_read(cpu_tlbstate.state) == TLBSTATE_OK) + percpu_write(cpu_tlbstate.state, TLBSTATE_LAZY); +@@ -34,16 +46,30 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, + struct task_struct *tsk) + { + unsigned cpu = smp_processor_id(); ++#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)) && defined(CONFIG_SMP) ++ int tlbstate = TLBSTATE_OK; ++#endif + + if (likely(prev != next)) { + #ifdef CONFIG_SMP ++#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)) ++ tlbstate = percpu_read(cpu_tlbstate.state); ++#endif + percpu_write(cpu_tlbstate.state, TLBSTATE_OK); + percpu_write(cpu_tlbstate.active_mm, next); + #endif + cpumask_set_cpu(cpu, mm_cpumask(next)); + + /* Re-load page tables */ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ pax_open_kernel(); ++ __clone_user_pgds(get_cpu_pgd(cpu), next->pgd, USER_PGD_PTRS); ++ __shadow_user_pgds(get_cpu_pgd(cpu) + USER_PGD_PTRS, next->pgd, USER_PGD_PTRS); ++ pax_close_kernel(); ++ load_cr3(get_cpu_pgd(cpu)); ++#else + load_cr3(next->pgd); ++#endif + + /* stop flush ipis for the previous mm */ + cpumask_clear_cpu(cpu, mm_cpumask(prev)); +@@ -53,9 +79,38 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, + */ + if (unlikely(prev->context.ldt != next->context.ldt)) + load_LDT_nolock(&next->context); +- } ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP) ++ if (!nx_enabled) { ++ smp_mb__before_clear_bit(); ++ cpu_clear(cpu, prev->context.cpu_user_cs_mask); ++ smp_mb__after_clear_bit(); ++ cpu_set(cpu, next->context.cpu_user_cs_mask); ++ } ++#endif ++ ++#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)) ++ if (unlikely(prev->context.user_cs_base != next->context.user_cs_base || ++ prev->context.user_cs_limit != next->context.user_cs_limit)) ++ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu); + #ifdef CONFIG_SMP ++ else if (unlikely(tlbstate != TLBSTATE_OK)) ++ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu); ++#endif ++#endif ++ ++ } + else { ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ pax_open_kernel(); ++ __clone_user_pgds(get_cpu_pgd(cpu), next->pgd, USER_PGD_PTRS); ++ __shadow_user_pgds(get_cpu_pgd(cpu) + USER_PGD_PTRS, next->pgd, USER_PGD_PTRS); ++ pax_close_kernel(); ++ load_cr3(get_cpu_pgd(cpu)); ++#endif ++ ++#ifdef CONFIG_SMP + percpu_write(cpu_tlbstate.state, TLBSTATE_OK); + BUG_ON(percpu_read(cpu_tlbstate.active_mm) != next); + +@@ -64,11 +119,28 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, + * tlb flush IPI delivery. We must reload CR3 + * to make sure to use no freed page tables. + */ ++ ++#ifndef CONFIG_PAX_PER_CPU_PGD + load_cr3(next->pgd); ++#endif ++ + load_LDT_nolock(&next->context); ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) ++ if (!nx_enabled) ++ cpu_set(cpu, next->context.cpu_user_cs_mask); ++#endif ++ ++#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC)) ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (!((next->pax_flags & MF_PAX_PAGEEXEC) && nx_enabled)) ++#endif ++ set_user_cs(next->context.user_cs_base, next->context.user_cs_limit, cpu); ++#endif ++ + } ++#endif + } +-#endif + } + + #define activate_mm(prev, next) \ +diff --git a/arch/x86/include/asm/module.h b/arch/x86/include/asm/module.h +index 3e2ce58..caaf478 100644 +--- a/arch/x86/include/asm/module.h ++++ b/arch/x86/include/asm/module.h +@@ -5,6 +5,7 @@ + + #ifdef CONFIG_X86_64 + /* X86_64 does not define MODULE_PROC_FAMILY */ ++#define MODULE_PROC_FAMILY "" + #elif defined CONFIG_M386 + #define MODULE_PROC_FAMILY "386 " + #elif defined CONFIG_M486 +@@ -59,13 +60,26 @@ + #error unknown processor family + #endif + +-#ifdef CONFIG_X86_32 +-# ifdef CONFIG_4KSTACKS +-# define MODULE_STACKSIZE "4KSTACKS " +-# else +-# define MODULE_STACKSIZE "" +-# endif +-# define MODULE_ARCH_VERMAGIC MODULE_PROC_FAMILY MODULE_STACKSIZE ++#if defined(CONFIG_X86_32) && defined(CONFIG_4KSTACKS) ++#define MODULE_STACKSIZE "4KSTACKS " ++#else ++#define MODULE_STACKSIZE "" + #endif + ++#ifdef CONFIG_PAX_KERNEXEC_PLUGIN_METHOD_BTS ++#define MODULE_PAX_KERNEXEC "KERNEXEC_BTS " ++#elif defined(CONFIG_PAX_KERNEXEC_PLUGIN_METHOD_OR) ++#define MODULE_PAX_KERNEXEC "KERNEXEC_OR " ++#else ++#define MODULE_PAX_KERNEXEC "" ++#endif ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++#define MODULE_PAX_UDEREF "UDEREF " ++#else ++#define MODULE_PAX_UDEREF "" ++#endif ++ ++#define MODULE_ARCH_VERMAGIC MODULE_PROC_FAMILY MODULE_STACKSIZE MODULE_PAX_KERNEXEC MODULE_PAX_UDEREF ++ + #endif /* _ASM_X86_MODULE_H */ +diff --git a/arch/x86/include/asm/page_64_types.h b/arch/x86/include/asm/page_64_types.h +index 7639dbf..e08a58c 100644 +--- a/arch/x86/include/asm/page_64_types.h ++++ b/arch/x86/include/asm/page_64_types.h +@@ -56,7 +56,7 @@ void copy_page(void *to, void *from); + + /* duplicated to the one in bootmem.h */ + extern unsigned long max_pfn; +-extern unsigned long phys_base; ++extern const unsigned long phys_base; + + extern unsigned long __phys_addr(unsigned long); + #define __phys_reloc_hide(x) (x) +diff --git a/arch/x86/include/asm/paravirt.h b/arch/x86/include/asm/paravirt.h +index efb3899..ef30687 100644 +--- a/arch/x86/include/asm/paravirt.h ++++ b/arch/x86/include/asm/paravirt.h +@@ -648,6 +648,18 @@ static inline void set_pgd(pgd_t *pgdp, pgd_t pgd) + val); + } + ++static inline void set_pgd_batched(pgd_t *pgdp, pgd_t pgd) ++{ ++ pgdval_t val = native_pgd_val(pgd); ++ ++ if (sizeof(pgdval_t) > sizeof(long)) ++ PVOP_VCALL3(pv_mmu_ops.set_pgd_batched, pgdp, ++ val, (u64)val >> 32); ++ else ++ PVOP_VCALL2(pv_mmu_ops.set_pgd_batched, pgdp, ++ val); ++} ++ + static inline void pgd_clear(pgd_t *pgdp) + { + set_pgd(pgdp, __pgd(0)); +@@ -729,6 +741,21 @@ static inline void __set_fixmap(unsigned /* enum fixed_addresses */ idx, + pv_mmu_ops.set_fixmap(idx, phys, flags); + } + ++#ifdef CONFIG_PAX_KERNEXEC ++static inline unsigned long pax_open_kernel(void) ++{ ++ return PVOP_CALL0(unsigned long, pv_mmu_ops.pax_open_kernel); ++} ++ ++static inline unsigned long pax_close_kernel(void) ++{ ++ return PVOP_CALL0(unsigned long, pv_mmu_ops.pax_close_kernel); ++} ++#else ++static inline unsigned long pax_open_kernel(void) { return 0; } ++static inline unsigned long pax_close_kernel(void) { return 0; } ++#endif ++ + #if defined(CONFIG_SMP) && defined(CONFIG_PARAVIRT_SPINLOCKS) + + static inline int __raw_spin_is_locked(struct raw_spinlock *lock) +@@ -945,7 +972,7 @@ extern void default_banner(void); + + #define PARA_PATCH(struct, off) ((PARAVIRT_PATCH_##struct + (off)) / 4) + #define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .long, 4) +-#define PARA_INDIRECT(addr) *%cs:addr ++#define PARA_INDIRECT(addr) *%ss:addr + #endif + + #define INTERRUPT_RETURN \ +@@ -1022,6 +1049,21 @@ extern void default_banner(void); + PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_irq_enable_sysexit), \ + CLBR_NONE, \ + jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_irq_enable_sysexit)) ++ ++#define GET_CR0_INTO_RDI \ ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0); \ ++ mov %rax,%rdi ++ ++#define SET_RDI_INTO_CR0 \ ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0) ++ ++#define GET_CR3_INTO_RDI \ ++ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_read_cr3); \ ++ mov %rax,%rdi ++ ++#define SET_RDI_INTO_CR3 \ ++ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_write_cr3) ++ + #endif /* CONFIG_X86_32 */ + + #endif /* __ASSEMBLY__ */ +diff --git a/arch/x86/include/asm/paravirt_types.h b/arch/x86/include/asm/paravirt_types.h +index 9357473..aeb2de5 100644 +--- a/arch/x86/include/asm/paravirt_types.h ++++ b/arch/x86/include/asm/paravirt_types.h +@@ -78,19 +78,19 @@ struct pv_init_ops { + */ + unsigned (*patch)(u8 type, u16 clobber, void *insnbuf, + unsigned long addr, unsigned len); +-}; ++} __no_const; + + + struct pv_lazy_ops { + /* Set deferred update mode, used for batching operations. */ + void (*enter)(void); + void (*leave)(void); +-}; ++} __no_const; + + struct pv_time_ops { + unsigned long long (*sched_clock)(void); + unsigned long (*get_tsc_khz)(void); +-}; ++} __no_const; + + struct pv_cpu_ops { + /* hooks for various privileged instructions */ +@@ -186,7 +186,7 @@ struct pv_cpu_ops { + + void (*start_context_switch)(struct task_struct *prev); + void (*end_context_switch)(struct task_struct *next); +-}; ++} __no_const; + + struct pv_irq_ops { + /* +@@ -217,7 +217,7 @@ struct pv_apic_ops { + unsigned long start_eip, + unsigned long start_esp); + #endif +-}; ++} __no_const; + + struct pv_mmu_ops { + unsigned long (*read_cr2)(void); +@@ -301,6 +301,7 @@ struct pv_mmu_ops { + struct paravirt_callee_save make_pud; + + void (*set_pgd)(pgd_t *pudp, pgd_t pgdval); ++ void (*set_pgd_batched)(pgd_t *pudp, pgd_t pgdval); + #endif /* PAGETABLE_LEVELS == 4 */ + #endif /* PAGETABLE_LEVELS >= 3 */ + +@@ -316,6 +317,12 @@ struct pv_mmu_ops { + an mfn. We can tell which is which from the index. */ + void (*set_fixmap)(unsigned /* enum fixed_addresses */ idx, + phys_addr_t phys, pgprot_t flags); ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ unsigned long (*pax_open_kernel)(void); ++ unsigned long (*pax_close_kernel)(void); ++#endif ++ + }; + + struct raw_spinlock; +@@ -326,7 +333,7 @@ struct pv_lock_ops { + void (*spin_lock_flags)(struct raw_spinlock *lock, unsigned long flags); + int (*spin_trylock)(struct raw_spinlock *lock); + void (*spin_unlock)(struct raw_spinlock *lock); +-}; ++} __no_const; + + /* This contains all the paravirt structures: we get a convenient + * number for each function using the offset which we use to indicate +diff --git a/arch/x86/include/asm/pci_x86.h b/arch/x86/include/asm/pci_x86.h +index b399988..3f47c38 100644 +--- a/arch/x86/include/asm/pci_x86.h ++++ b/arch/x86/include/asm/pci_x86.h +@@ -89,16 +89,16 @@ extern int (*pcibios_enable_irq)(struct pci_dev *dev); + extern void (*pcibios_disable_irq)(struct pci_dev *dev); + + struct pci_raw_ops { +- int (*read)(unsigned int domain, unsigned int bus, unsigned int devfn, ++ int (* const read)(unsigned int domain, unsigned int bus, unsigned int devfn, + int reg, int len, u32 *val); +- int (*write)(unsigned int domain, unsigned int bus, unsigned int devfn, ++ int (* const write)(unsigned int domain, unsigned int bus, unsigned int devfn, + int reg, int len, u32 val); + }; + +-extern struct pci_raw_ops *raw_pci_ops; +-extern struct pci_raw_ops *raw_pci_ext_ops; ++extern const struct pci_raw_ops *raw_pci_ops; ++extern const struct pci_raw_ops *raw_pci_ext_ops; + +-extern struct pci_raw_ops pci_direct_conf1; ++extern const struct pci_raw_ops pci_direct_conf1; + extern bool port_cf9_safe; + + /* arch_initcall level */ +diff --git a/arch/x86/include/asm/percpu.h b/arch/x86/include/asm/percpu.h +index b65a36d..50345a4 100644 +--- a/arch/x86/include/asm/percpu.h ++++ b/arch/x86/include/asm/percpu.h +@@ -78,6 +78,7 @@ do { \ + if (0) { \ + T__ tmp__; \ + tmp__ = (val); \ ++ (void)tmp__; \ + } \ + switch (sizeof(var)) { \ + case 1: \ +diff --git a/arch/x86/include/asm/pgalloc.h b/arch/x86/include/asm/pgalloc.h +index 271de94..ef944d6 100644 +--- a/arch/x86/include/asm/pgalloc.h ++++ b/arch/x86/include/asm/pgalloc.h +@@ -63,6 +63,13 @@ static inline void pmd_populate_kernel(struct mm_struct *mm, + pmd_t *pmd, pte_t *pte) + { + paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); ++ set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); ++} ++ ++static inline void pmd_populate_user(struct mm_struct *mm, ++ pmd_t *pmd, pte_t *pte) ++{ ++ paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); + set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); + } + +diff --git a/arch/x86/include/asm/pgtable-2level.h b/arch/x86/include/asm/pgtable-2level.h +index 2334982..70bc412 100644 +--- a/arch/x86/include/asm/pgtable-2level.h ++++ b/arch/x86/include/asm/pgtable-2level.h +@@ -18,7 +18,9 @@ static inline void native_set_pte(pte_t *ptep , pte_t pte) + + static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) + { ++ pax_open_kernel(); + *pmdp = pmd; ++ pax_close_kernel(); + } + + static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) +diff --git a/arch/x86/include/asm/pgtable-3level.h b/arch/x86/include/asm/pgtable-3level.h +index 33927d2..ccde329 100644 +--- a/arch/x86/include/asm/pgtable-3level.h ++++ b/arch/x86/include/asm/pgtable-3level.h +@@ -38,12 +38,16 @@ static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) + + static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) + { ++ pax_open_kernel(); + set_64bit((unsigned long long *)(pmdp), native_pmd_val(pmd)); ++ pax_close_kernel(); + } + + static inline void native_set_pud(pud_t *pudp, pud_t pud) + { ++ pax_open_kernel(); + set_64bit((unsigned long long *)(pudp), native_pud_val(pud)); ++ pax_close_kernel(); + } + + /* +diff --git a/arch/x86/include/asm/pgtable.h b/arch/x86/include/asm/pgtable.h +index af6fd36..867ff74 100644 +--- a/arch/x86/include/asm/pgtable.h ++++ b/arch/x86/include/asm/pgtable.h +@@ -39,6 +39,7 @@ extern struct list_head pgd_list; + + #ifndef __PAGETABLE_PUD_FOLDED + #define set_pgd(pgdp, pgd) native_set_pgd(pgdp, pgd) ++#define set_pgd_batched(pgdp, pgd) native_set_pgd_batched(pgdp, pgd) + #define pgd_clear(pgd) native_pgd_clear(pgd) + #endif + +@@ -74,12 +75,51 @@ extern struct list_head pgd_list; + + #define arch_end_context_switch(prev) do {} while(0) + ++#define pax_open_kernel() native_pax_open_kernel() ++#define pax_close_kernel() native_pax_close_kernel() + #endif /* CONFIG_PARAVIRT */ + ++#define __HAVE_ARCH_PAX_OPEN_KERNEL ++#define __HAVE_ARCH_PAX_CLOSE_KERNEL ++ ++#ifdef CONFIG_PAX_KERNEXEC ++static inline unsigned long native_pax_open_kernel(void) ++{ ++ unsigned long cr0; ++ ++ preempt_disable(); ++ barrier(); ++ cr0 = read_cr0() ^ X86_CR0_WP; ++ BUG_ON(unlikely(cr0 & X86_CR0_WP)); ++ write_cr0(cr0); ++ return cr0 ^ X86_CR0_WP; ++} ++ ++static inline unsigned long native_pax_close_kernel(void) ++{ ++ unsigned long cr0; ++ ++ cr0 = read_cr0() ^ X86_CR0_WP; ++ BUG_ON(unlikely(!(cr0 & X86_CR0_WP))); ++ write_cr0(cr0); ++ barrier(); ++ preempt_enable_no_resched(); ++ return cr0 ^ X86_CR0_WP; ++} ++#else ++static inline unsigned long native_pax_open_kernel(void) { return 0; } ++static inline unsigned long native_pax_close_kernel(void) { return 0; } ++#endif ++ + /* + * The following only work if pte_present() is true. + * Undefined behaviour if not.. + */ ++static inline int pte_user(pte_t pte) ++{ ++ return pte_val(pte) & _PAGE_USER; ++} ++ + static inline int pte_dirty(pte_t pte) + { + return pte_flags(pte) & _PAGE_DIRTY; +@@ -167,9 +207,29 @@ static inline pte_t pte_wrprotect(pte_t pte) + return pte_clear_flags(pte, _PAGE_RW); + } + ++static inline pte_t pte_mkread(pte_t pte) ++{ ++ return __pte(pte_val(pte) | _PAGE_USER); ++} ++ + static inline pte_t pte_mkexec(pte_t pte) + { +- return pte_clear_flags(pte, _PAGE_NX); ++#ifdef CONFIG_X86_PAE ++ if (__supported_pte_mask & _PAGE_NX) ++ return pte_clear_flags(pte, _PAGE_NX); ++ else ++#endif ++ return pte_set_flags(pte, _PAGE_USER); ++} ++ ++static inline pte_t pte_exprotect(pte_t pte) ++{ ++#ifdef CONFIG_X86_PAE ++ if (__supported_pte_mask & _PAGE_NX) ++ return pte_set_flags(pte, _PAGE_NX); ++ else ++#endif ++ return pte_clear_flags(pte, _PAGE_USER); + } + + static inline pte_t pte_mkdirty(pte_t pte) +@@ -302,6 +362,15 @@ pte_t *populate_extra_pte(unsigned long vaddr); + #endif + + #ifndef __ASSEMBLY__ ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++extern pgd_t cpu_pgd[NR_CPUS][PTRS_PER_PGD]; ++static inline pgd_t *get_cpu_pgd(unsigned int cpu) ++{ ++ return cpu_pgd[cpu]; ++} ++#endif ++ + #include + + static inline int pte_none(pte_t pte) +@@ -472,7 +541,7 @@ static inline pud_t *pud_offset(pgd_t *pgd, unsigned long address) + + static inline int pgd_bad(pgd_t pgd) + { +- return (pgd_flags(pgd) & ~_PAGE_USER) != _KERNPG_TABLE; ++ return (pgd_flags(pgd) & ~(_PAGE_USER | _PAGE_NX)) != _KERNPG_TABLE; + } + + static inline int pgd_none(pgd_t pgd) +@@ -495,7 +564,12 @@ static inline int pgd_none(pgd_t pgd) + * pgd_offset() returns a (pgd_t *) + * pgd_index() is used get the offset into the pgd page's array of pgd_t's; + */ +-#define pgd_offset(mm, address) ((mm)->pgd + pgd_index((address))) ++#define pgd_offset(mm, address) ((mm)->pgd + pgd_index(address)) ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++#define pgd_offset_cpu(cpu, address) (get_cpu_pgd(cpu) + pgd_index(address)) ++#endif ++ + /* + * a shortcut which implies the use of the kernel's pgd, instead + * of a process's +@@ -506,6 +580,20 @@ static inline int pgd_none(pgd_t pgd) + #define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET) + #define KERNEL_PGD_PTRS (PTRS_PER_PGD - KERNEL_PGD_BOUNDARY) + ++#ifdef CONFIG_X86_32 ++#define USER_PGD_PTRS KERNEL_PGD_BOUNDARY ++#else ++#define TASK_SIZE_MAX_SHIFT CONFIG_TASK_SIZE_MAX_SHIFT ++#define USER_PGD_PTRS (_AC(1,UL) << (TASK_SIZE_MAX_SHIFT - PGDIR_SHIFT)) ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++#define PAX_USER_SHADOW_BASE (_AC(1,UL) << TASK_SIZE_MAX_SHIFT) ++#else ++#define PAX_USER_SHADOW_BASE (_AC(0,UL)) ++#endif ++ ++#endif ++ + #ifndef __ASSEMBLY__ + + extern int direct_gbpages; +@@ -611,11 +699,23 @@ static inline void ptep_set_wrprotect(struct mm_struct *mm, + * dst and src can be on the same page, but the range must not overlap, + * and must not cross a page boundary. + */ +-static inline void clone_pgd_range(pgd_t *dst, pgd_t *src, int count) ++static inline void clone_pgd_range(pgd_t *dst, const pgd_t *src, int count) + { +- memcpy(dst, src, count * sizeof(pgd_t)); ++ pax_open_kernel(); ++ while (count--) ++ *dst++ = *src++; ++ pax_close_kernel(); + } + ++#ifdef CONFIG_PAX_PER_CPU_PGD ++extern void __clone_user_pgds(pgd_t *dst, const pgd_t *src, int count); ++#endif ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++extern void __shadow_user_pgds(pgd_t *dst, const pgd_t *src, int count); ++#else ++static inline void __shadow_user_pgds(pgd_t *dst, const pgd_t *src, int count) {} ++#endif + + #include + #endif /* __ASSEMBLY__ */ +diff --git a/arch/x86/include/asm/pgtable_32.h b/arch/x86/include/asm/pgtable_32.h +index 750f1bf..971e839 100644 +--- a/arch/x86/include/asm/pgtable_32.h ++++ b/arch/x86/include/asm/pgtable_32.h +@@ -26,9 +26,6 @@ + struct mm_struct; + struct vm_area_struct; + +-extern pgd_t swapper_pg_dir[1024]; +-extern pgd_t trampoline_pg_dir[1024]; +- + static inline void pgtable_cache_init(void) { } + static inline void check_pgt_cache(void) { } + void paging_init(void); +@@ -49,6 +46,12 @@ extern void set_pmd_pfn(unsigned long, unsigned long, pgprot_t); + # include + #endif + ++extern pgd_t swapper_pg_dir[PTRS_PER_PGD]; ++extern pgd_t trampoline_pg_dir[PTRS_PER_PGD]; ++#ifdef CONFIG_X86_PAE ++extern pmd_t swapper_pm_dir[PTRS_PER_PGD][PTRS_PER_PMD]; ++#endif ++ + #if defined(CONFIG_HIGHPTE) + #define __KM_PTE \ + (in_nmi() ? KM_NMI_PTE : \ +@@ -73,7 +76,9 @@ extern void set_pmd_pfn(unsigned long, unsigned long, pgprot_t); + /* Clear a kernel PTE and flush it from the TLB */ + #define kpte_clear_flush(ptep, vaddr) \ + do { \ ++ pax_open_kernel(); \ + pte_clear(&init_mm, (vaddr), (ptep)); \ ++ pax_close_kernel(); \ + __flush_tlb_one((vaddr)); \ + } while (0) + +@@ -85,6 +90,9 @@ do { \ + + #endif /* !__ASSEMBLY__ */ + ++#define HAVE_ARCH_UNMAPPED_AREA ++#define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN ++ + /* + * kern_addr_valid() is (1) for FLATMEM and (0) for + * SPARSEMEM and DISCONTIGMEM +diff --git a/arch/x86/include/asm/pgtable_32_types.h b/arch/x86/include/asm/pgtable_32_types.h +index 5e67c15..12d5c47 100644 +--- a/arch/x86/include/asm/pgtable_32_types.h ++++ b/arch/x86/include/asm/pgtable_32_types.h +@@ -8,7 +8,7 @@ + */ + #ifdef CONFIG_X86_PAE + # include +-# define PMD_SIZE (1UL << PMD_SHIFT) ++# define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) + # define PMD_MASK (~(PMD_SIZE - 1)) + #else + # include +@@ -46,6 +46,19 @@ extern bool __vmalloc_start_set; /* set once high_memory is set */ + # define VMALLOC_END (FIXADDR_START - 2 * PAGE_SIZE) + #endif + ++#ifdef CONFIG_PAX_KERNEXEC ++#ifndef __ASSEMBLY__ ++extern unsigned char MODULES_EXEC_VADDR[]; ++extern unsigned char MODULES_EXEC_END[]; ++#endif ++#include ++#define ktla_ktva(addr) (addr + LOAD_PHYSICAL_ADDR + PAGE_OFFSET) ++#define ktva_ktla(addr) (addr - LOAD_PHYSICAL_ADDR - PAGE_OFFSET) ++#else ++#define ktla_ktva(addr) (addr) ++#define ktva_ktla(addr) (addr) ++#endif ++ + #define MODULES_VADDR VMALLOC_START + #define MODULES_END VMALLOC_END + #define MODULES_LEN (MODULES_VADDR - MODULES_END) +diff --git a/arch/x86/include/asm/pgtable_64.h b/arch/x86/include/asm/pgtable_64.h +index c57a301..6b414ff 100644 +--- a/arch/x86/include/asm/pgtable_64.h ++++ b/arch/x86/include/asm/pgtable_64.h +@@ -16,10 +16,14 @@ + + extern pud_t level3_kernel_pgt[512]; + extern pud_t level3_ident_pgt[512]; ++extern pud_t level3_vmalloc_start_pgt[512]; ++extern pud_t level3_vmalloc_end_pgt[512]; ++extern pud_t level3_vmemmap_pgt[512]; ++extern pud_t level2_vmemmap_pgt[512]; + extern pmd_t level2_kernel_pgt[512]; + extern pmd_t level2_fixmap_pgt[512]; +-extern pmd_t level2_ident_pgt[512]; +-extern pgd_t init_level4_pgt[]; ++extern pmd_t level2_ident_pgt[512*2]; ++extern pgd_t init_level4_pgt[512]; + + #define swapper_pg_dir init_level4_pgt + +@@ -74,7 +78,9 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp) + + static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) + { ++ pax_open_kernel(); + *pmdp = pmd; ++ pax_close_kernel(); + } + + static inline void native_pmd_clear(pmd_t *pmd) +@@ -94,6 +100,13 @@ static inline void native_pud_clear(pud_t *pud) + + static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) + { ++ pax_open_kernel(); ++ *pgdp = pgd; ++ pax_close_kernel(); ++} ++ ++static inline void native_set_pgd_batched(pgd_t *pgdp, pgd_t pgd) ++{ + *pgdp = pgd; + } + +diff --git a/arch/x86/include/asm/pgtable_64_types.h b/arch/x86/include/asm/pgtable_64_types.h +index 766ea16..5b96cb3 100644 +--- a/arch/x86/include/asm/pgtable_64_types.h ++++ b/arch/x86/include/asm/pgtable_64_types.h +@@ -59,5 +59,10 @@ typedef struct { pteval_t pte; } pte_t; + #define MODULES_VADDR _AC(0xffffffffa0000000, UL) + #define MODULES_END _AC(0xffffffffff000000, UL) + #define MODULES_LEN (MODULES_END - MODULES_VADDR) ++#define MODULES_EXEC_VADDR MODULES_VADDR ++#define MODULES_EXEC_END MODULES_END ++ ++#define ktla_ktva(addr) (addr) ++#define ktva_ktla(addr) (addr) + + #endif /* _ASM_X86_PGTABLE_64_DEFS_H */ +diff --git a/arch/x86/include/asm/pgtable_types.h b/arch/x86/include/asm/pgtable_types.h +index d1f4a76..2f46ba1 100644 +--- a/arch/x86/include/asm/pgtable_types.h ++++ b/arch/x86/include/asm/pgtable_types.h +@@ -16,12 +16,11 @@ + #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page */ + #define _PAGE_BIT_PAT 7 /* on 4KB pages */ + #define _PAGE_BIT_GLOBAL 8 /* Global TLB entry PPro+ */ +-#define _PAGE_BIT_UNUSED1 9 /* available for programmer */ ++#define _PAGE_BIT_SPECIAL 9 /* special mappings, no associated struct page */ + #define _PAGE_BIT_IOMAP 10 /* flag used to indicate IO mapping */ + #define _PAGE_BIT_HIDDEN 11 /* hidden by kmemcheck */ + #define _PAGE_BIT_PAT_LARGE 12 /* On 2MB or 1GB pages */ +-#define _PAGE_BIT_SPECIAL _PAGE_BIT_UNUSED1 +-#define _PAGE_BIT_CPA_TEST _PAGE_BIT_UNUSED1 ++#define _PAGE_BIT_CPA_TEST _PAGE_BIT_SPECIAL + #define _PAGE_BIT_NX 63 /* No execute: only valid after cpuid check */ + + /* If _PAGE_BIT_PRESENT is clear, we use these: */ +@@ -39,7 +38,6 @@ + #define _PAGE_DIRTY (_AT(pteval_t, 1) << _PAGE_BIT_DIRTY) + #define _PAGE_PSE (_AT(pteval_t, 1) << _PAGE_BIT_PSE) + #define _PAGE_GLOBAL (_AT(pteval_t, 1) << _PAGE_BIT_GLOBAL) +-#define _PAGE_UNUSED1 (_AT(pteval_t, 1) << _PAGE_BIT_UNUSED1) + #define _PAGE_IOMAP (_AT(pteval_t, 1) << _PAGE_BIT_IOMAP) + #define _PAGE_PAT (_AT(pteval_t, 1) << _PAGE_BIT_PAT) + #define _PAGE_PAT_LARGE (_AT(pteval_t, 1) << _PAGE_BIT_PAT_LARGE) +@@ -55,8 +53,10 @@ + + #if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE) + #define _PAGE_NX (_AT(pteval_t, 1) << _PAGE_BIT_NX) +-#else ++#elif defined(CONFIG_KMEMCHECK) + #define _PAGE_NX (_AT(pteval_t, 0)) ++#else ++#define _PAGE_NX (_AT(pteval_t, 1) << _PAGE_BIT_HIDDEN) + #endif + + #define _PAGE_FILE (_AT(pteval_t, 1) << _PAGE_BIT_FILE) +@@ -93,6 +93,9 @@ + #define PAGE_READONLY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | \ + _PAGE_ACCESSED) + ++#define PAGE_READONLY_NOEXEC PAGE_READONLY ++#define PAGE_SHARED_NOEXEC PAGE_SHARED ++ + #define __PAGE_KERNEL_EXEC \ + (_PAGE_PRESENT | _PAGE_RW | _PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_GLOBAL) + #define __PAGE_KERNEL (__PAGE_KERNEL_EXEC | _PAGE_NX) +@@ -103,8 +106,8 @@ + #define __PAGE_KERNEL_WC (__PAGE_KERNEL | _PAGE_CACHE_WC) + #define __PAGE_KERNEL_NOCACHE (__PAGE_KERNEL | _PAGE_PCD | _PAGE_PWT) + #define __PAGE_KERNEL_UC_MINUS (__PAGE_KERNEL | _PAGE_PCD) +-#define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RX | _PAGE_USER) +-#define __PAGE_KERNEL_VSYSCALL_NOCACHE (__PAGE_KERNEL_VSYSCALL | _PAGE_PCD | _PAGE_PWT) ++#define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RO | _PAGE_USER) ++#define __PAGE_KERNEL_VSYSCALL_NOCACHE (__PAGE_KERNEL_RO | _PAGE_PCD | _PAGE_PWT | _PAGE_USER) + #define __PAGE_KERNEL_LARGE (__PAGE_KERNEL | _PAGE_PSE) + #define __PAGE_KERNEL_LARGE_NOCACHE (__PAGE_KERNEL | _PAGE_CACHE_UC | _PAGE_PSE) + #define __PAGE_KERNEL_LARGE_EXEC (__PAGE_KERNEL_EXEC | _PAGE_PSE) +@@ -163,8 +166,8 @@ + * bits are combined, this will alow user to access the high address mapped + * VDSO in the presence of CONFIG_COMPAT_VDSO + */ +-#define PTE_IDENT_ATTR 0x003 /* PRESENT+RW */ +-#define PDE_IDENT_ATTR 0x067 /* PRESENT+RW+USER+DIRTY+ACCESSED */ ++#define PTE_IDENT_ATTR 0x063 /* PRESENT+RW+DIRTY+ACCESSED */ ++#define PDE_IDENT_ATTR 0x063 /* PRESENT+RW+DIRTY+ACCESSED */ + #define PGD_IDENT_ATTR 0x001 /* PRESENT (no other attributes) */ + #endif + +@@ -202,7 +205,17 @@ static inline pgdval_t pgd_flags(pgd_t pgd) + { + return native_pgd_val(pgd) & PTE_FLAGS_MASK; + } ++#endif + ++#if PAGETABLE_LEVELS == 3 ++#include ++#endif ++ ++#if PAGETABLE_LEVELS == 2 ++#include ++#endif ++ ++#ifndef __ASSEMBLY__ + #if PAGETABLE_LEVELS > 3 + typedef struct { pudval_t pud; } pud_t; + +@@ -216,8 +229,6 @@ static inline pudval_t native_pud_val(pud_t pud) + return pud.pud; + } + #else +-#include +- + static inline pudval_t native_pud_val(pud_t pud) + { + return native_pgd_val(pud.pgd); +@@ -237,8 +248,6 @@ static inline pmdval_t native_pmd_val(pmd_t pmd) + return pmd.pmd; + } + #else +-#include +- + static inline pmdval_t native_pmd_val(pmd_t pmd) + { + return native_pgd_val(pmd.pud.pgd); +@@ -278,7 +287,16 @@ typedef struct page *pgtable_t; + + extern pteval_t __supported_pte_mask; + extern void set_nx(void); ++ ++#ifdef CONFIG_X86_32 ++#ifdef CONFIG_X86_PAE + extern int nx_enabled; ++#else ++#define nx_enabled (0) ++#endif ++#else ++#define nx_enabled (1) ++#endif + + #define pgprot_writecombine pgprot_writecombine + extern pgprot_t pgprot_writecombine(pgprot_t prot); +diff --git a/arch/x86/include/asm/processor.h b/arch/x86/include/asm/processor.h +index fa04dea..5f823fc 100644 +--- a/arch/x86/include/asm/processor.h ++++ b/arch/x86/include/asm/processor.h +@@ -272,7 +272,7 @@ struct tss_struct { + + } ____cacheline_aligned; + +-DECLARE_PER_CPU_SHARED_ALIGNED(struct tss_struct, init_tss); ++extern struct tss_struct init_tss[NR_CPUS]; + + /* + * Save the original ist values for checking stack pointers during debugging +@@ -911,11 +911,18 @@ static inline void spin_lock_prefetch(const void *x) + */ + #define TASK_SIZE PAGE_OFFSET + #define TASK_SIZE_MAX TASK_SIZE ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++#define SEGMEXEC_TASK_SIZE (TASK_SIZE / 2) ++#define STACK_TOP ((current->mm->pax_flags & MF_PAX_SEGMEXEC)?SEGMEXEC_TASK_SIZE:TASK_SIZE) ++#else + #define STACK_TOP TASK_SIZE +-#define STACK_TOP_MAX STACK_TOP ++#endif ++ ++#define STACK_TOP_MAX TASK_SIZE + + #define INIT_THREAD { \ +- .sp0 = sizeof(init_stack) + (long)&init_stack, \ ++ .sp0 = sizeof(init_stack) + (long)&init_stack - 8, \ + .vm86_info = NULL, \ + .sysenter_cs = __KERNEL_CS, \ + .io_bitmap_ptr = NULL, \ +@@ -929,7 +936,7 @@ static inline void spin_lock_prefetch(const void *x) + */ + #define INIT_TSS { \ + .x86_tss = { \ +- .sp0 = sizeof(init_stack) + (long)&init_stack, \ ++ .sp0 = sizeof(init_stack) + (long)&init_stack - 8, \ + .ss0 = __KERNEL_DS, \ + .ss1 = __KERNEL_CS, \ + .io_bitmap_base = INVALID_IO_BITMAP_OFFSET, \ +@@ -940,11 +947,7 @@ static inline void spin_lock_prefetch(const void *x) + extern unsigned long thread_saved_pc(struct task_struct *tsk); + + #define THREAD_SIZE_LONGS (THREAD_SIZE/sizeof(unsigned long)) +-#define KSTK_TOP(info) \ +-({ \ +- unsigned long *__ptr = (unsigned long *)(info); \ +- (unsigned long)(&__ptr[THREAD_SIZE_LONGS]); \ +-}) ++#define KSTK_TOP(info) ((container_of(info, struct task_struct, tinfo))->thread.sp0) + + /* + * The below -8 is to reserve 8 bytes on top of the ring0 stack. +@@ -959,7 +962,7 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); + #define task_pt_regs(task) \ + ({ \ + struct pt_regs *__regs__; \ +- __regs__ = (struct pt_regs *)(KSTK_TOP(task_stack_page(task))-8); \ ++ __regs__ = (struct pt_regs *)((task)->thread.sp0); \ + __regs__ - 1; \ + }) + +@@ -969,13 +972,13 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); + /* + * User space process size. 47bits minus one guard page. + */ +-#define TASK_SIZE_MAX ((1UL << 47) - PAGE_SIZE) ++#define TASK_SIZE_MAX ((1UL << TASK_SIZE_MAX_SHIFT) - PAGE_SIZE) + + /* This decides where the kernel will search for a free chunk of vm + * space during mmap's. + */ + #define IA32_PAGE_OFFSET ((current->personality & ADDR_LIMIT_3GB) ? \ +- 0xc0000000 : 0xFFFFe000) ++ 0xc0000000 : 0xFFFFf000) + + #define TASK_SIZE (test_thread_flag(TIF_IA32) ? \ + IA32_PAGE_OFFSET : TASK_SIZE_MAX) +@@ -986,11 +989,11 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); + #define STACK_TOP_MAX TASK_SIZE_MAX + + #define INIT_THREAD { \ +- .sp0 = (unsigned long)&init_stack + sizeof(init_stack) \ ++ .sp0 = (unsigned long)&init_stack + sizeof(init_stack) - 16 \ + } + + #define INIT_TSS { \ +- .x86_tss.sp0 = (unsigned long)&init_stack + sizeof(init_stack) \ ++ .x86_tss.sp0 = (unsigned long)&init_stack + sizeof(init_stack) - 16 \ + } + + /* +@@ -1012,6 +1015,10 @@ extern void start_thread(struct pt_regs *regs, unsigned long new_ip, + */ + #define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3)) + ++#ifdef CONFIG_PAX_SEGMEXEC ++#define SEGMEXEC_TASK_UNMAPPED_BASE (PAGE_ALIGN(SEGMEXEC_TASK_SIZE / 3)) ++#endif ++ + #define KSTK_EIP(task) (task_pt_regs(task)->ip) + + /* Get/set a process' ability to use the timestamp counter instruction */ +diff --git a/arch/x86/include/asm/ptrace.h b/arch/x86/include/asm/ptrace.h +index 0f0d908..f2e3da2 100644 +--- a/arch/x86/include/asm/ptrace.h ++++ b/arch/x86/include/asm/ptrace.h +@@ -151,28 +151,29 @@ static inline unsigned long regs_return_value(struct pt_regs *regs) + } + + /* +- * user_mode_vm(regs) determines whether a register set came from user mode. ++ * user_mode(regs) determines whether a register set came from user mode. + * This is true if V8086 mode was enabled OR if the register set was from + * protected mode with RPL-3 CS value. This tricky test checks that with + * one comparison. Many places in the kernel can bypass this full check +- * if they have already ruled out V8086 mode, so user_mode(regs) can be used. ++ * if they have already ruled out V8086 mode, so user_mode_novm(regs) can ++ * be used. + */ +-static inline int user_mode(struct pt_regs *regs) ++static inline int user_mode_novm(struct pt_regs *regs) + { + #ifdef CONFIG_X86_32 + return (regs->cs & SEGMENT_RPL_MASK) == USER_RPL; + #else +- return !!(regs->cs & 3); ++ return !!(regs->cs & SEGMENT_RPL_MASK); + #endif + } + +-static inline int user_mode_vm(struct pt_regs *regs) ++static inline int user_mode(struct pt_regs *regs) + { + #ifdef CONFIG_X86_32 + return ((regs->cs & SEGMENT_RPL_MASK) | (regs->flags & X86_VM_MASK)) >= + USER_RPL; + #else +- return user_mode(regs); ++ return user_mode_novm(regs); + #endif + } + +diff --git a/arch/x86/include/asm/reboot.h b/arch/x86/include/asm/reboot.h +index 562d4fd..6e39df1 100644 +--- a/arch/x86/include/asm/reboot.h ++++ b/arch/x86/include/asm/reboot.h +@@ -6,19 +6,19 @@ + struct pt_regs; + + struct machine_ops { +- void (*restart)(char *cmd); +- void (*halt)(void); +- void (*power_off)(void); ++ void (* __noreturn restart)(char *cmd); ++ void (* __noreturn halt)(void); ++ void (* __noreturn power_off)(void); + void (*shutdown)(void); + void (*crash_shutdown)(struct pt_regs *); +- void (*emergency_restart)(void); +-}; ++ void (* __noreturn emergency_restart)(void); ++} __no_const; + + extern struct machine_ops machine_ops; + + void native_machine_crash_shutdown(struct pt_regs *regs); + void native_machine_shutdown(void); +-void machine_real_restart(const unsigned char *code, int length); ++void machine_real_restart(const unsigned char *code, unsigned int length) __noreturn; + + typedef void (*nmi_shootdown_cb)(int, struct die_args*); + void nmi_shootdown_cpus(nmi_shootdown_cb callback); +diff --git a/arch/x86/include/asm/rwsem.h b/arch/x86/include/asm/rwsem.h +index 606ede1..dbfff37 100644 +--- a/arch/x86/include/asm/rwsem.h ++++ b/arch/x86/include/asm/rwsem.h +@@ -118,6 +118,14 @@ static inline void __down_read(struct rw_semaphore *sem) + { + asm volatile("# beginning down_read\n\t" + LOCK_PREFIX _ASM_INC "(%1)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX _ASM_DEC "(%1)\n\t" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + /* adds 0x00000001, returns the old value */ + " jns 1f\n" + " call call_rwsem_down_read_failed\n" +@@ -139,6 +147,14 @@ static inline int __down_read_trylock(struct rw_semaphore *sem) + "1:\n\t" + " mov %1,%2\n\t" + " add %3,%2\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "sub %3,%2\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + " jle 2f\n\t" + LOCK_PREFIX " cmpxchg %2,%0\n\t" + " jnz 1b\n\t" +@@ -160,6 +176,14 @@ static inline void __down_write_nested(struct rw_semaphore *sem, int subclass) + tmp = RWSEM_ACTIVE_WRITE_BIAS; + asm volatile("# beginning down_write\n\t" + LOCK_PREFIX " xadd %1,(%2)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "mov %1,(%2)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + /* subtract 0x0000ffff, returns the old value */ + " test %1,%1\n\t" + /* was the count 0 before? */ +@@ -198,6 +222,14 @@ static inline void __up_read(struct rw_semaphore *sem) + rwsem_count_t tmp = -RWSEM_ACTIVE_READ_BIAS; + asm volatile("# beginning __up_read\n\t" + LOCK_PREFIX " xadd %1,(%2)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "mov %1,(%2)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + /* subtracts 1, returns the old value */ + " jns 1f\n\t" + " call call_rwsem_wake\n" +@@ -216,6 +248,14 @@ static inline void __up_write(struct rw_semaphore *sem) + rwsem_count_t tmp; + asm volatile("# beginning __up_write\n\t" + LOCK_PREFIX " xadd %1,(%2)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "mov %1,(%2)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + /* tries to transition + 0xffff0001 -> 0x00000000 */ + " jz 1f\n" +@@ -234,6 +274,14 @@ static inline void __downgrade_write(struct rw_semaphore *sem) + { + asm volatile("# beginning __downgrade_write\n\t" + LOCK_PREFIX _ASM_ADD "%2,(%1)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX _ASM_SUB "%2,(%1)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + /* + * transitions 0xZZZZ0001 -> 0xYYYY0001 (i386) + * 0xZZZZZZZZ00000001 -> 0xYYYYYYYY00000001 (x86_64) +@@ -253,7 +301,15 @@ static inline void __downgrade_write(struct rw_semaphore *sem) + static inline void rwsem_atomic_add(rwsem_count_t delta, + struct rw_semaphore *sem) + { +- asm volatile(LOCK_PREFIX _ASM_ADD "%1,%0" ++ asm volatile(LOCK_PREFIX _ASM_ADD "%1,%0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX _ASM_SUB "%1,%0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (sem->count) + : "er" (delta)); + } +@@ -266,7 +322,15 @@ static inline rwsem_count_t rwsem_atomic_update(rwsem_count_t delta, + { + rwsem_count_t tmp = delta; + +- asm volatile(LOCK_PREFIX "xadd %0,%1" ++ asm volatile(LOCK_PREFIX "xadd %0,%1\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ "mov %0,%1\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+r" (tmp), "+m" (sem->count) + : : "memory"); + +diff --git a/arch/x86/include/asm/segment.h b/arch/x86/include/asm/segment.h +index 14e0ed8..7f7dd5e 100644 +--- a/arch/x86/include/asm/segment.h ++++ b/arch/x86/include/asm/segment.h +@@ -62,10 +62,15 @@ + * 26 - ESPFIX small SS + * 27 - per-cpu [ offset to per-cpu data area ] + * 28 - stack_canary-20 [ for stack protector ] +- * 29 - unused +- * 30 - unused ++ * 29 - PCI BIOS CS ++ * 30 - PCI BIOS DS + * 31 - TSS for double fault handler + */ ++#define GDT_ENTRY_KERNEXEC_EFI_CS (1) ++#define GDT_ENTRY_KERNEXEC_EFI_DS (2) ++#define __KERNEXEC_EFI_CS (GDT_ENTRY_KERNEXEC_EFI_CS*8) ++#define __KERNEXEC_EFI_DS (GDT_ENTRY_KERNEXEC_EFI_DS*8) ++ + #define GDT_ENTRY_TLS_MIN 6 + #define GDT_ENTRY_TLS_MAX (GDT_ENTRY_TLS_MIN + GDT_ENTRY_TLS_ENTRIES - 1) + +@@ -77,6 +82,8 @@ + + #define GDT_ENTRY_KERNEL_CS (GDT_ENTRY_KERNEL_BASE + 0) + ++#define GDT_ENTRY_KERNEXEC_KERNEL_CS (4) ++ + #define GDT_ENTRY_KERNEL_DS (GDT_ENTRY_KERNEL_BASE + 1) + + #define GDT_ENTRY_TSS (GDT_ENTRY_KERNEL_BASE + 4) +@@ -88,7 +95,7 @@ + #define GDT_ENTRY_ESPFIX_SS (GDT_ENTRY_KERNEL_BASE + 14) + #define __ESPFIX_SS (GDT_ENTRY_ESPFIX_SS * 8) + +-#define GDT_ENTRY_PERCPU (GDT_ENTRY_KERNEL_BASE + 15) ++#define GDT_ENTRY_PERCPU (GDT_ENTRY_KERNEL_BASE + 15) + #ifdef CONFIG_SMP + #define __KERNEL_PERCPU (GDT_ENTRY_PERCPU * 8) + #else +@@ -102,6 +109,12 @@ + #define __KERNEL_STACK_CANARY 0 + #endif + ++#define GDT_ENTRY_PCIBIOS_CS (GDT_ENTRY_KERNEL_BASE + 17) ++#define __PCIBIOS_CS (GDT_ENTRY_PCIBIOS_CS * 8) ++ ++#define GDT_ENTRY_PCIBIOS_DS (GDT_ENTRY_KERNEL_BASE + 18) ++#define __PCIBIOS_DS (GDT_ENTRY_PCIBIOS_DS * 8) ++ + #define GDT_ENTRY_DOUBLEFAULT_TSS 31 + + /* +@@ -139,7 +152,7 @@ + */ + + /* Matches PNP_CS32 and PNP_CS16 (they must be consecutive) */ +-#define SEGMENT_IS_PNP_CODE(x) (((x) & 0xf4) == GDT_ENTRY_PNPBIOS_BASE * 8) ++#define SEGMENT_IS_PNP_CODE(x) (((x) & 0xFFFCU) == PNP_CS32 || ((x) & 0xFFFCU) == PNP_CS16) + + + #else +@@ -163,6 +176,8 @@ + #define __USER32_CS (GDT_ENTRY_DEFAULT_USER32_CS * 8 + 3) + #define __USER32_DS __USER_DS + ++#define GDT_ENTRY_KERNEXEC_KERNEL_CS 7 ++ + #define GDT_ENTRY_TSS 8 /* needs two entries */ + #define GDT_ENTRY_LDT 10 /* needs two entries */ + #define GDT_ENTRY_TLS_MIN 12 +@@ -183,6 +198,7 @@ + #endif + + #define __KERNEL_CS (GDT_ENTRY_KERNEL_CS * 8) ++#define __KERNEXEC_KERNEL_CS (GDT_ENTRY_KERNEXEC_KERNEL_CS * 8) + #define __KERNEL_DS (GDT_ENTRY_KERNEL_DS * 8) + #define __USER_DS (GDT_ENTRY_DEFAULT_USER_DS* 8 + 3) + #define __USER_CS (GDT_ENTRY_DEFAULT_USER_CS* 8 + 3) +diff --git a/arch/x86/include/asm/smp.h b/arch/x86/include/asm/smp.h +index 4c2f63c..5685db2 100644 +--- a/arch/x86/include/asm/smp.h ++++ b/arch/x86/include/asm/smp.h +@@ -24,7 +24,7 @@ extern unsigned int num_processors; + DECLARE_PER_CPU(cpumask_var_t, cpu_sibling_map); + DECLARE_PER_CPU(cpumask_var_t, cpu_core_map); + DECLARE_PER_CPU(u16, cpu_llc_id); +-DECLARE_PER_CPU(int, cpu_number); ++DECLARE_PER_CPU(unsigned int, cpu_number); + + static inline struct cpumask *cpu_sibling_mask(int cpu) + { +@@ -40,10 +40,7 @@ DECLARE_EARLY_PER_CPU(u16, x86_cpu_to_apicid); + DECLARE_EARLY_PER_CPU(u16, x86_bios_cpu_apicid); + + /* Static state in head.S used to set up a CPU */ +-extern struct { +- void *sp; +- unsigned short ss; +-} stack_start; ++extern unsigned long stack_start; /* Initial stack pointer address */ + + struct smp_ops { + void (*smp_prepare_boot_cpu)(void); +@@ -60,7 +57,7 @@ struct smp_ops { + + void (*send_call_func_ipi)(const struct cpumask *mask); + void (*send_call_func_single_ipi)(int cpu); +-}; ++} __no_const; + + /* Globals due to paravirt */ + extern void set_cpu_sibling_map(int cpu); +@@ -175,14 +172,8 @@ extern unsigned disabled_cpus __cpuinitdata; + extern int safe_smp_processor_id(void); + + #elif defined(CONFIG_X86_64_SMP) +-#define raw_smp_processor_id() (percpu_read(cpu_number)) +- +-#define stack_smp_processor_id() \ +-({ \ +- struct thread_info *ti; \ +- __asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \ +- ti->cpu; \ +-}) ++#define raw_smp_processor_id() (percpu_read(cpu_number)) ++#define stack_smp_processor_id() raw_smp_processor_id() + #define safe_smp_processor_id() smp_processor_id() + + #endif +diff --git a/arch/x86/include/asm/spinlock.h b/arch/x86/include/asm/spinlock.h +index 4e77853..4359783 100644 +--- a/arch/x86/include/asm/spinlock.h ++++ b/arch/x86/include/asm/spinlock.h +@@ -249,6 +249,14 @@ static inline int __raw_write_can_lock(raw_rwlock_t *lock) + static inline void __raw_read_lock(raw_rwlock_t *rw) + { + asm volatile(LOCK_PREFIX " subl $1,(%0)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX " addl $1,(%0)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + "jns 1f\n" + "call __read_lock_failed\n\t" + "1:\n" +@@ -258,6 +266,14 @@ static inline void __raw_read_lock(raw_rwlock_t *rw) + static inline void __raw_write_lock(raw_rwlock_t *rw) + { + asm volatile(LOCK_PREFIX " subl %1,(%0)\n\t" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX " addl %1,(%0)\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + "jz 1f\n" + "call __write_lock_failed\n\t" + "1:\n" +@@ -286,12 +302,29 @@ static inline int __raw_write_trylock(raw_rwlock_t *lock) + + static inline void __raw_read_unlock(raw_rwlock_t *rw) + { +- asm volatile(LOCK_PREFIX "incl %0" :"+m" (rw->lock) : : "memory"); ++ asm volatile(LOCK_PREFIX "incl %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "decl %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ ++ :"+m" (rw->lock) : : "memory"); + } + + static inline void __raw_write_unlock(raw_rwlock_t *rw) + { +- asm volatile(LOCK_PREFIX "addl %1, %0" ++ asm volatile(LOCK_PREFIX "addl %1, %0\n" ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ "jno 0f\n" ++ LOCK_PREFIX "subl %1, %0\n" ++ "int $4\n0:\n" ++ _ASM_EXTABLE(0b, 0b) ++#endif ++ + : "+m" (rw->lock) : "i" (RW_LOCK_BIAS) : "memory"); + } + +diff --git a/arch/x86/include/asm/stackprotector.h b/arch/x86/include/asm/stackprotector.h +index 1575177..cb23f52 100644 +--- a/arch/x86/include/asm/stackprotector.h ++++ b/arch/x86/include/asm/stackprotector.h +@@ -48,7 +48,7 @@ + * head_32 for boot CPU and setup_per_cpu_areas() for others. + */ + #define GDT_STACK_CANARY_INIT \ +- [GDT_ENTRY_STACK_CANARY] = GDT_ENTRY_INIT(0x4090, 0, 0x18), ++ [GDT_ENTRY_STACK_CANARY] = GDT_ENTRY_INIT(0x4090, 0, 0x17), + + /* + * Initialize the stackprotector canary value. +@@ -113,7 +113,7 @@ static inline void setup_stack_canary_segment(int cpu) + + static inline void load_stack_canary_segment(void) + { +-#ifdef CONFIG_X86_32 ++#if defined(CONFIG_X86_32) && !defined(CONFIG_PAX_MEMORY_UDEREF) + asm volatile ("mov %0, %%gs" : : "r" (0)); + #endif + } +diff --git a/arch/x86/include/asm/system.h b/arch/x86/include/asm/system.h +index e0fbf29..858ef4a 100644 +--- a/arch/x86/include/asm/system.h ++++ b/arch/x86/include/asm/system.h +@@ -132,7 +132,7 @@ do { \ + "thread_return:\n\t" \ + "movq "__percpu_arg([current_task])",%%rsi\n\t" \ + __switch_canary \ +- "movq %P[thread_info](%%rsi),%%r8\n\t" \ ++ "movq "__percpu_arg([thread_info])",%%r8\n\t" \ + "movq %%rax,%%rdi\n\t" \ + "testl %[_tif_fork],%P[ti_flags](%%r8)\n\t" \ + "jnz ret_from_fork\n\t" \ +@@ -143,7 +143,7 @@ do { \ + [threadrsp] "i" (offsetof(struct task_struct, thread.sp)), \ + [ti_flags] "i" (offsetof(struct thread_info, flags)), \ + [_tif_fork] "i" (_TIF_FORK), \ +- [thread_info] "i" (offsetof(struct task_struct, stack)), \ ++ [thread_info] "m" (per_cpu_var(current_tinfo)), \ + [current_task] "m" (per_cpu_var(current_task)) \ + __switch_canary_iparam \ + : "memory", "cc" __EXTRA_CLOBBER) +@@ -200,7 +200,7 @@ static inline unsigned long get_limit(unsigned long segment) + { + unsigned long __limit; + asm("lsll %1,%0" : "=r" (__limit) : "r" (segment)); +- return __limit + 1; ++ return __limit; + } + + static inline void native_clts(void) +@@ -340,12 +340,12 @@ void enable_hlt(void); + + void cpu_idle_wait(void); + +-extern unsigned long arch_align_stack(unsigned long sp); ++#define arch_align_stack(x) ((x) & ~0xfUL) + extern void free_init_pages(char *what, unsigned long begin, unsigned long end); + + void default_idle(void); + +-void stop_this_cpu(void *dummy); ++void stop_this_cpu(void *dummy) __noreturn; + + /* + * Force strict CPU ordering. +diff --git a/arch/x86/include/asm/thread_info.h b/arch/x86/include/asm/thread_info.h +index 19c3ce4..8962535 100644 +--- a/arch/x86/include/asm/thread_info.h ++++ b/arch/x86/include/asm/thread_info.h +@@ -10,6 +10,7 @@ + #include + #include + #include ++#include + + /* + * low level task data that entry.S needs immediate access to +@@ -24,7 +25,6 @@ struct exec_domain; + #include + + struct thread_info { +- struct task_struct *task; /* main task structure */ + struct exec_domain *exec_domain; /* execution domain */ + __u32 flags; /* low level flags */ + __u32 status; /* thread synchronous flags */ +@@ -34,18 +34,12 @@ struct thread_info { + mm_segment_t addr_limit; + struct restart_block restart_block; + void __user *sysenter_return; +-#ifdef CONFIG_X86_32 +- unsigned long previous_esp; /* ESP of the previous stack in +- case of nested (IRQ) stacks +- */ +- __u8 supervisor_stack[0]; +-#endif ++ unsigned long lowest_stack; + int uaccess_err; + }; + +-#define INIT_THREAD_INFO(tsk) \ ++#define INIT_THREAD_INFO \ + { \ +- .task = &tsk, \ + .exec_domain = &default_exec_domain, \ + .flags = 0, \ + .cpu = 0, \ +@@ -56,7 +50,7 @@ struct thread_info { + }, \ + } + +-#define init_thread_info (init_thread_union.thread_info) ++#define init_thread_info (init_thread_union.stack) + #define init_stack (init_thread_union.stack) + + #else /* !__ASSEMBLY__ */ +@@ -163,45 +157,40 @@ struct thread_info { + #define alloc_thread_info(tsk) \ + ((struct thread_info *)__get_free_pages(THREAD_FLAGS, THREAD_ORDER)) + +-#ifdef CONFIG_X86_32 +- +-#define STACK_WARN (THREAD_SIZE/8) +-/* +- * macros/functions for gaining access to the thread information structure +- * +- * preempt_count needs to be 1 initially, until the scheduler is functional. +- */ +-#ifndef __ASSEMBLY__ +- +- +-/* how to get the current stack pointer from C */ +-register unsigned long current_stack_pointer asm("esp") __used; +- +-/* how to get the thread information struct from C */ +-static inline struct thread_info *current_thread_info(void) +-{ +- return (struct thread_info *) +- (current_stack_pointer & ~(THREAD_SIZE - 1)); +-} +- +-#else /* !__ASSEMBLY__ */ +- ++#ifdef __ASSEMBLY__ + /* how to get the thread information struct from ASM */ + #define GET_THREAD_INFO(reg) \ +- movl $-THREAD_SIZE, reg; \ +- andl %esp, reg ++ mov PER_CPU_VAR(current_tinfo), reg + + /* use this one if reg already contains %esp */ +-#define GET_THREAD_INFO_WITH_ESP(reg) \ +- andl $-THREAD_SIZE, reg ++#define GET_THREAD_INFO_WITH_ESP(reg) GET_THREAD_INFO(reg) ++#else ++/* how to get the thread information struct from C */ ++DECLARE_PER_CPU(struct thread_info *, current_tinfo); ++ ++static __always_inline struct thread_info *current_thread_info(void) ++{ ++ return percpu_read_stable(current_tinfo); ++} ++#endif ++ ++#ifdef CONFIG_X86_32 ++ ++#define STACK_WARN (THREAD_SIZE/8) ++/* ++ * macros/functions for gaining access to the thread information structure ++ * ++ * preempt_count needs to be 1 initially, until the scheduler is functional. ++ */ ++#ifndef __ASSEMBLY__ ++ ++/* how to get the current stack pointer from C */ ++register unsigned long current_stack_pointer asm("esp") __used; + + #endif + + #else /* X86_32 */ + +-#include +-#define KERNEL_STACK_OFFSET (5*8) +- + /* + * macros/functions for gaining access to the thread information structure + * preempt_count needs to be 1 initially, until the scheduler is functional. +@@ -209,21 +198,8 @@ static inline struct thread_info *current_thread_info(void) + #ifndef __ASSEMBLY__ + DECLARE_PER_CPU(unsigned long, kernel_stack); + +-static inline struct thread_info *current_thread_info(void) +-{ +- struct thread_info *ti; +- ti = (void *)(percpu_read_stable(kernel_stack) + +- KERNEL_STACK_OFFSET - THREAD_SIZE); +- return ti; +-} +- +-#else /* !__ASSEMBLY__ */ +- +-/* how to get the thread information struct from ASM */ +-#define GET_THREAD_INFO(reg) \ +- movq PER_CPU_VAR(kernel_stack),reg ; \ +- subq $(THREAD_SIZE-KERNEL_STACK_OFFSET),reg +- ++/* how to get the current stack pointer from C */ ++register unsigned long current_stack_pointer asm("rsp") __used; + #endif + + #endif /* !X86_32 */ +@@ -260,5 +236,16 @@ extern void arch_task_cache_init(void); + extern void free_thread_info(struct thread_info *ti); + extern int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src); + #define arch_task_cache_init arch_task_cache_init ++ ++#define __HAVE_THREAD_FUNCTIONS ++#define task_thread_info(task) (&(task)->tinfo) ++#define task_stack_page(task) ((task)->stack) ++#define setup_thread_stack(p, org) do {} while (0) ++#define end_of_stack(p) ((unsigned long *)task_stack_page(p) + 1) ++ ++#define __HAVE_ARCH_TASK_STRUCT_ALLOCATOR ++extern struct task_struct *alloc_task_struct(void); ++extern void free_task_struct(struct task_struct *); ++ + #endif + #endif /* _ASM_X86_THREAD_INFO_H */ +diff --git a/arch/x86/include/asm/uaccess.h b/arch/x86/include/asm/uaccess.h +index 61c5874..8a046e9 100644 +--- a/arch/x86/include/asm/uaccess.h ++++ b/arch/x86/include/asm/uaccess.h +@@ -8,12 +8,15 @@ + #include + #include + #include ++#include + #include + #include + + #define VERIFY_READ 0 + #define VERIFY_WRITE 1 + ++extern void check_object_size(const void *ptr, unsigned long n, bool to); ++ + /* + * The fs value determines whether argument validity checking should be + * performed or not. If get_fs() == USER_DS, checking is performed, with +@@ -29,7 +32,12 @@ + + #define get_ds() (KERNEL_DS) + #define get_fs() (current_thread_info()->addr_limit) ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF) ++void __set_fs(mm_segment_t x); ++void set_fs(mm_segment_t x); ++#else + #define set_fs(x) (current_thread_info()->addr_limit = (x)) ++#endif + + #define segment_eq(a, b) ((a).seg == (b).seg) + +@@ -77,7 +85,33 @@ + * checks that the pointer is in the user space range - after calling + * this function, memory access functions may still return -EFAULT. + */ +-#define access_ok(type, addr, size) (likely(__range_not_ok(addr, size) == 0)) ++#define __access_ok(type, addr, size) (likely(__range_not_ok(addr, size) == 0)) ++#define access_ok(type, addr, size) \ ++({ \ ++ long __size = size; \ ++ unsigned long __addr = (unsigned long)addr; \ ++ unsigned long __addr_ao = __addr & PAGE_MASK; \ ++ unsigned long __end_ao = __addr + __size - 1; \ ++ bool __ret_ao = __range_not_ok(__addr, __size) == 0; \ ++ if (__ret_ao && unlikely((__end_ao ^ __addr_ao) & PAGE_MASK)) { \ ++ while(__addr_ao <= __end_ao) { \ ++ char __c_ao; \ ++ __addr_ao += PAGE_SIZE; \ ++ if (__size > PAGE_SIZE) \ ++ cond_resched(); \ ++ if (__get_user(__c_ao, (char __user *)__addr)) \ ++ break; \ ++ if (type != VERIFY_WRITE) { \ ++ __addr = __addr_ao; \ ++ continue; \ ++ } \ ++ if (__put_user(__c_ao, (char __user *)__addr)) \ ++ break; \ ++ __addr = __addr_ao; \ ++ } \ ++ } \ ++ __ret_ao; \ ++}) + + /* + * The exception table consists of pairs of addresses: the first is the +@@ -183,12 +217,20 @@ extern int __get_user_bad(void); + asm volatile("call __put_user_" #size : "=a" (__ret_pu) \ + : "0" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx") + +- ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define __copyuser_seg "gs;" ++#define __COPYUSER_SET_ES "pushl %%gs; popl %%es\n" ++#define __COPYUSER_RESTORE_ES "pushl %%ss; popl %%es\n" ++#else ++#define __copyuser_seg ++#define __COPYUSER_SET_ES ++#define __COPYUSER_RESTORE_ES ++#endif + + #ifdef CONFIG_X86_32 + #define __put_user_asm_u64(x, addr, err, errret) \ +- asm volatile("1: movl %%eax,0(%2)\n" \ +- "2: movl %%edx,4(%2)\n" \ ++ asm volatile("1: "__copyuser_seg"movl %%eax,0(%2)\n" \ ++ "2: "__copyuser_seg"movl %%edx,4(%2)\n" \ + "3:\n" \ + ".section .fixup,\"ax\"\n" \ + "4: movl %3,%0\n" \ +@@ -200,8 +242,8 @@ extern int __get_user_bad(void); + : "A" (x), "r" (addr), "i" (errret), "0" (err)) + + #define __put_user_asm_ex_u64(x, addr) \ +- asm volatile("1: movl %%eax,0(%1)\n" \ +- "2: movl %%edx,4(%1)\n" \ ++ asm volatile("1: "__copyuser_seg"movl %%eax,0(%1)\n" \ ++ "2: "__copyuser_seg"movl %%edx,4(%1)\n" \ + "3:\n" \ + _ASM_EXTABLE(1b, 2b - 1b) \ + _ASM_EXTABLE(2b, 3b - 2b) \ +@@ -253,7 +295,7 @@ extern void __put_user_8(void); + __typeof__(*(ptr)) __pu_val; \ + __chk_user_ptr(ptr); \ + might_fault(); \ +- __pu_val = x; \ ++ __pu_val = (x); \ + switch (sizeof(*(ptr))) { \ + case 1: \ + __put_user_x(1, __pu_val, ptr, __ret_pu); \ +@@ -374,7 +416,7 @@ do { \ + } while (0) + + #define __get_user_asm(x, addr, err, itype, rtype, ltype, errret) \ +- asm volatile("1: mov"itype" %2,%"rtype"1\n" \ ++ asm volatile("1: "__copyuser_seg"mov"itype" %2,%"rtype"1\n"\ + "2:\n" \ + ".section .fixup,\"ax\"\n" \ + "3: mov %3,%0\n" \ +@@ -382,7 +424,7 @@ do { \ + " jmp 2b\n" \ + ".previous\n" \ + _ASM_EXTABLE(1b, 3b) \ +- : "=r" (err), ltype(x) \ ++ : "=r" (err), ltype (x) \ + : "m" (__m(addr)), "i" (errret), "0" (err)) + + #define __get_user_size_ex(x, ptr, size) \ +@@ -407,7 +449,7 @@ do { \ + } while (0) + + #define __get_user_asm_ex(x, addr, itype, rtype, ltype) \ +- asm volatile("1: mov"itype" %1,%"rtype"0\n" \ ++ asm volatile("1: "__copyuser_seg"mov"itype" %1,%"rtype"0\n"\ + "2:\n" \ + _ASM_EXTABLE(1b, 2b - 1b) \ + : ltype(x) : "m" (__m(addr))) +@@ -424,13 +466,24 @@ do { \ + int __gu_err; \ + unsigned long __gu_val; \ + __get_user_size(__gu_val, (ptr), (size), __gu_err, -EFAULT); \ +- (x) = (__force __typeof__(*(ptr)))__gu_val; \ ++ (x) = (__typeof__(*(ptr)))__gu_val; \ + __gu_err; \ + }) + + /* FIXME: this hack is definitely wrong -AK */ + struct __large_struct { unsigned long buf[100]; }; +-#define __m(x) (*(struct __large_struct __user *)(x)) ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define ____m(x) \ ++({ \ ++ unsigned long ____x = (unsigned long)(x); \ ++ if (____x < PAX_USER_SHADOW_BASE) \ ++ ____x += PAX_USER_SHADOW_BASE; \ ++ (void __user *)____x; \ ++}) ++#else ++#define ____m(x) (x) ++#endif ++#define __m(x) (*(struct __large_struct __user *)____m(x)) + + /* + * Tell gcc we read from memory instead of writing: this is because +@@ -438,7 +491,7 @@ struct __large_struct { unsigned long buf[100]; }; + * aliasing issues. + */ + #define __put_user_asm(x, addr, err, itype, rtype, ltype, errret) \ +- asm volatile("1: mov"itype" %"rtype"1,%2\n" \ ++ asm volatile("1: "__copyuser_seg"mov"itype" %"rtype"1,%2\n"\ + "2:\n" \ + ".section .fixup,\"ax\"\n" \ + "3: mov %3,%0\n" \ +@@ -446,10 +499,10 @@ struct __large_struct { unsigned long buf[100]; }; + ".previous\n" \ + _ASM_EXTABLE(1b, 3b) \ + : "=r"(err) \ +- : ltype(x), "m" (__m(addr)), "i" (errret), "0" (err)) ++ : ltype (x), "m" (__m(addr)), "i" (errret), "0" (err)) + + #define __put_user_asm_ex(x, addr, itype, rtype, ltype) \ +- asm volatile("1: mov"itype" %"rtype"0,%1\n" \ ++ asm volatile("1: "__copyuser_seg"mov"itype" %"rtype"0,%1\n"\ + "2:\n" \ + _ASM_EXTABLE(1b, 2b - 1b) \ + : : ltype(x), "m" (__m(addr))) +@@ -488,8 +541,12 @@ struct __large_struct { unsigned long buf[100]; }; + * On error, the variable @x is set to zero. + */ + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define __get_user(x, ptr) get_user((x), (ptr)) ++#else + #define __get_user(x, ptr) \ + __get_user_nocheck((x), (ptr), sizeof(*(ptr))) ++#endif + + /** + * __put_user: - Write a simple value into user space, with less checking. +@@ -511,8 +568,12 @@ struct __large_struct { unsigned long buf[100]; }; + * Returns zero on success, or -EFAULT on error. + */ + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define __put_user(x, ptr) put_user((x), (ptr)) ++#else + #define __put_user(x, ptr) \ + __put_user_nocheck((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr))) ++#endif + + #define __get_user_unaligned __get_user + #define __put_user_unaligned __put_user +@@ -530,7 +591,7 @@ struct __large_struct { unsigned long buf[100]; }; + #define get_user_ex(x, ptr) do { \ + unsigned long __gue_val; \ + __get_user_size_ex((__gue_val), (ptr), (sizeof(*(ptr)))); \ +- (x) = (__force __typeof__(*(ptr)))__gue_val; \ ++ (x) = (__typeof__(*(ptr)))__gue_val; \ + } while (0) + + #ifdef CONFIG_X86_WP_WORKS_OK +@@ -567,6 +628,7 @@ extern struct movsl_mask { + + #define ARCH_HAS_NOCACHE_UACCESS 1 + ++#define ARCH_HAS_SORT_EXTABLE + #ifdef CONFIG_X86_32 + # include "uaccess_32.h" + #else +diff --git a/arch/x86/include/asm/uaccess_32.h b/arch/x86/include/asm/uaccess_32.h +index 632fb44..e30e334 100644 +--- a/arch/x86/include/asm/uaccess_32.h ++++ b/arch/x86/include/asm/uaccess_32.h +@@ -44,6 +44,11 @@ unsigned long __must_check __copy_from_user_ll_nocache_nozero + static __always_inline unsigned long __must_check + __copy_to_user_inatomic(void __user *to, const void *from, unsigned long n) + { ++ pax_track_stack(); ++ ++ if ((long)n < 0) ++ return n; ++ + if (__builtin_constant_p(n)) { + unsigned long ret; + +@@ -62,6 +67,8 @@ __copy_to_user_inatomic(void __user *to, const void *from, unsigned long n) + return ret; + } + } ++ if (!__builtin_constant_p(n)) ++ check_object_size(from, n, true); + return __copy_to_user_ll(to, from, n); + } + +@@ -83,12 +90,16 @@ static __always_inline unsigned long __must_check + __copy_to_user(void __user *to, const void *from, unsigned long n) + { + might_fault(); ++ + return __copy_to_user_inatomic(to, from, n); + } + + static __always_inline unsigned long + __copy_from_user_inatomic(void *to, const void __user *from, unsigned long n) + { ++ if ((long)n < 0) ++ return n; ++ + /* Avoid zeroing the tail if the copy fails.. + * If 'n' is constant and 1, 2, or 4, we do still zero on a failure, + * but as the zeroing behaviour is only significant when n is not +@@ -138,6 +149,12 @@ static __always_inline unsigned long + __copy_from_user(void *to, const void __user *from, unsigned long n) + { + might_fault(); ++ ++ pax_track_stack(); ++ ++ if ((long)n < 0) ++ return n; ++ + if (__builtin_constant_p(n)) { + unsigned long ret; + +@@ -153,6 +170,8 @@ __copy_from_user(void *to, const void __user *from, unsigned long n) + return ret; + } + } ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); + return __copy_from_user_ll(to, from, n); + } + +@@ -160,6 +179,10 @@ static __always_inline unsigned long __copy_from_user_nocache(void *to, + const void __user *from, unsigned long n) + { + might_fault(); ++ ++ if ((long)n < 0) ++ return n; ++ + if (__builtin_constant_p(n)) { + unsigned long ret; + +@@ -182,14 +205,62 @@ static __always_inline unsigned long + __copy_from_user_inatomic_nocache(void *to, const void __user *from, + unsigned long n) + { +- return __copy_from_user_ll_nocache_nozero(to, from, n); ++ if ((long)n < 0) ++ return n; ++ ++ return __copy_from_user_ll_nocache_nozero(to, from, n); ++} ++ ++/** ++ * copy_to_user: - Copy a block of data into user space. ++ * @to: Destination address, in user space. ++ * @from: Source address, in kernel space. ++ * @n: Number of bytes to copy. ++ * ++ * Context: User context only. This function may sleep. ++ * ++ * Copy data from kernel space to user space. ++ * ++ * Returns number of bytes that could not be copied. ++ * On success, this will be zero. ++ */ ++static __always_inline unsigned long __must_check ++copy_to_user(void __user *to, const void *from, unsigned long n) ++{ ++ if (access_ok(VERIFY_WRITE, to, n)) ++ n = __copy_to_user(to, from, n); ++ return n; ++} ++ ++/** ++ * copy_from_user: - Copy a block of data from user space. ++ * @to: Destination address, in kernel space. ++ * @from: Source address, in user space. ++ * @n: Number of bytes to copy. ++ * ++ * Context: User context only. This function may sleep. ++ * ++ * Copy data from user space to kernel space. ++ * ++ * Returns number of bytes that could not be copied. ++ * On success, this will be zero. ++ * ++ * If some data could not be copied, this function will pad the copied ++ * data to the requested size using zero bytes. ++ */ ++static __always_inline unsigned long __must_check ++copy_from_user(void *to, const void __user *from, unsigned long n) ++{ ++ if (access_ok(VERIFY_READ, from, n)) ++ n = __copy_from_user(to, from, n); ++ else if ((long)n > 0) { ++ if (!__builtin_constant_p(n)) ++ check_object_size(to, n, false); ++ memset(to, 0, n); ++ } ++ return n; + } + +-unsigned long __must_check copy_to_user(void __user *to, +- const void *from, unsigned long n); +-unsigned long __must_check copy_from_user(void *to, +- const void __user *from, +- unsigned long n); + long __must_check strncpy_from_user(char *dst, const char __user *src, + long count); + long __must_check __strncpy_from_user(char *dst, +diff --git a/arch/x86/include/asm/uaccess_64.h b/arch/x86/include/asm/uaccess_64.h +index db24b21..f595ae7 100644 +--- a/arch/x86/include/asm/uaccess_64.h ++++ b/arch/x86/include/asm/uaccess_64.h +@@ -9,6 +9,9 @@ + #include + #include + #include ++#include ++ ++#define set_fs(x) (current_thread_info()->addr_limit = (x)) + + /* + * Copy To/From Userspace +@@ -16,116 +19,205 @@ + + /* Handles exceptions in both to and from, but doesn't do access_ok */ + __must_check unsigned long +-copy_user_generic(void *to, const void *from, unsigned len); ++copy_user_generic(void *to, const void *from, unsigned long len); + + __must_check unsigned long +-copy_to_user(void __user *to, const void *from, unsigned len); +-__must_check unsigned long +-copy_from_user(void *to, const void __user *from, unsigned len); +-__must_check unsigned long +-copy_in_user(void __user *to, const void __user *from, unsigned len); ++copy_in_user(void __user *to, const void __user *from, unsigned long len); + + static __always_inline __must_check +-int __copy_from_user(void *dst, const void __user *src, unsigned size) ++unsigned long __copy_from_user(void *dst, const void __user *src, unsigned long size) + { +- int ret = 0; ++ unsigned ret = 0; + + might_fault(); +- if (!__builtin_constant_p(size)) +- return copy_user_generic(dst, (__force void *)src, size); ++ ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_READ, src, size)) ++ return size; ++#endif ++ ++ if (!__builtin_constant_p(size)) { ++ check_object_size(dst, size, false); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic(dst, (__force_kernel const void *)src, size); ++ } + switch (size) { +- case 1:__get_user_asm(*(u8 *)dst, (u8 __user *)src, ++ case 1:__get_user_asm(*(u8 *)dst, (const u8 __user *)src, + ret, "b", "b", "=q", 1); + return ret; +- case 2:__get_user_asm(*(u16 *)dst, (u16 __user *)src, ++ case 2:__get_user_asm(*(u16 *)dst, (const u16 __user *)src, + ret, "w", "w", "=r", 2); + return ret; +- case 4:__get_user_asm(*(u32 *)dst, (u32 __user *)src, ++ case 4:__get_user_asm(*(u32 *)dst, (const u32 __user *)src, + ret, "l", "k", "=r", 4); + return ret; +- case 8:__get_user_asm(*(u64 *)dst, (u64 __user *)src, ++ case 8:__get_user_asm(*(u64 *)dst, (const u64 __user *)src, + ret, "q", "", "=r", 8); + return ret; + case 10: +- __get_user_asm(*(u64 *)dst, (u64 __user *)src, ++ __get_user_asm(*(u64 *)dst, (const u64 __user *)src, + ret, "q", "", "=r", 10); + if (unlikely(ret)) + return ret; + __get_user_asm(*(u16 *)(8 + (char *)dst), +- (u16 __user *)(8 + (char __user *)src), ++ (const u16 __user *)(8 + (const char __user *)src), + ret, "w", "w", "=r", 2); + return ret; + case 16: +- __get_user_asm(*(u64 *)dst, (u64 __user *)src, ++ __get_user_asm(*(u64 *)dst, (const u64 __user *)src, + ret, "q", "", "=r", 16); + if (unlikely(ret)) + return ret; + __get_user_asm(*(u64 *)(8 + (char *)dst), +- (u64 __user *)(8 + (char __user *)src), ++ (const u64 __user *)(8 + (const char __user *)src), + ret, "q", "", "=r", 8); + return ret; + default: +- return copy_user_generic(dst, (__force void *)src, size); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic(dst, (__force_kernel const void *)src, size); + } + } + + static __always_inline __must_check +-int __copy_to_user(void __user *dst, const void *src, unsigned size) ++unsigned long __copy_to_user(void __user *dst, const void *src, unsigned long size) + { +- int ret = 0; ++ unsigned ret = 0; + + might_fault(); +- if (!__builtin_constant_p(size)) +- return copy_user_generic((__force void *)dst, src, size); ++ ++ pax_track_stack(); ++ ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_WRITE, dst, size)) ++ return size; ++#endif ++ ++ if (!__builtin_constant_p(size)) { ++ check_object_size(src, size, true); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((__force_kernel void *)dst, src, size); ++ } + switch (size) { +- case 1:__put_user_asm(*(u8 *)src, (u8 __user *)dst, ++ case 1:__put_user_asm(*(const u8 *)src, (u8 __user *)dst, + ret, "b", "b", "iq", 1); + return ret; +- case 2:__put_user_asm(*(u16 *)src, (u16 __user *)dst, ++ case 2:__put_user_asm(*(const u16 *)src, (u16 __user *)dst, + ret, "w", "w", "ir", 2); + return ret; +- case 4:__put_user_asm(*(u32 *)src, (u32 __user *)dst, ++ case 4:__put_user_asm(*(const u32 *)src, (u32 __user *)dst, + ret, "l", "k", "ir", 4); + return ret; +- case 8:__put_user_asm(*(u64 *)src, (u64 __user *)dst, ++ case 8:__put_user_asm(*(const u64 *)src, (u64 __user *)dst, + ret, "q", "", "er", 8); + return ret; + case 10: +- __put_user_asm(*(u64 *)src, (u64 __user *)dst, ++ __put_user_asm(*(const u64 *)src, (u64 __user *)dst, + ret, "q", "", "er", 10); + if (unlikely(ret)) + return ret; + asm("":::"memory"); +- __put_user_asm(4[(u16 *)src], 4 + (u16 __user *)dst, ++ __put_user_asm(4[(const u16 *)src], 4 + (u16 __user *)dst, + ret, "w", "w", "ir", 2); + return ret; + case 16: +- __put_user_asm(*(u64 *)src, (u64 __user *)dst, ++ __put_user_asm(*(const u64 *)src, (u64 __user *)dst, + ret, "q", "", "er", 16); + if (unlikely(ret)) + return ret; + asm("":::"memory"); +- __put_user_asm(1[(u64 *)src], 1 + (u64 __user *)dst, ++ __put_user_asm(1[(const u64 *)src], 1 + (u64 __user *)dst, + ret, "q", "", "er", 8); + return ret; + default: +- return copy_user_generic((__force void *)dst, src, size); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((__force_kernel void *)dst, src, size); ++ } ++} ++ ++static __always_inline __must_check ++unsigned long copy_to_user(void __user *to, const void *from, unsigned long len) ++{ ++ if (access_ok(VERIFY_WRITE, to, len)) ++ len = __copy_to_user(to, from, len); ++ return len; ++} ++ ++static __always_inline __must_check ++unsigned long copy_from_user(void *to, const void __user *from, unsigned long len) ++{ ++ might_fault(); ++ ++ if (access_ok(VERIFY_READ, from, len)) ++ len = __copy_from_user(to, from, len); ++ else if (len < INT_MAX) { ++ if (!__builtin_constant_p(len)) ++ check_object_size(to, len, false); ++ memset(to, 0, len); + } ++ return len; + } + + static __always_inline __must_check +-int __copy_in_user(void __user *dst, const void __user *src, unsigned size) ++unsigned long __copy_in_user(void __user *dst, const void __user *src, unsigned long size) + { +- int ret = 0; ++ unsigned ret = 0; + + might_fault(); +- if (!__builtin_constant_p(size)) +- return copy_user_generic((__force void *)dst, +- (__force void *)src, size); ++ ++ pax_track_stack(); ++ ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_READ, src, size)) ++ return size; ++ if (!__access_ok(VERIFY_WRITE, dst, size)) ++ return size; ++#endif ++ ++ if (!__builtin_constant_p(size)) { ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((__force_kernel void *)dst, ++ (__force_kernel const void *)src, size); ++ } + switch (size) { + case 1: { + u8 tmp; +- __get_user_asm(tmp, (u8 __user *)src, ++ __get_user_asm(tmp, (const u8 __user *)src, + ret, "b", "b", "=q", 1); + if (likely(!ret)) + __put_user_asm(tmp, (u8 __user *)dst, +@@ -134,7 +226,7 @@ int __copy_in_user(void __user *dst, const void __user *src, unsigned size) + } + case 2: { + u16 tmp; +- __get_user_asm(tmp, (u16 __user *)src, ++ __get_user_asm(tmp, (const u16 __user *)src, + ret, "w", "w", "=r", 2); + if (likely(!ret)) + __put_user_asm(tmp, (u16 __user *)dst, +@@ -144,7 +236,7 @@ int __copy_in_user(void __user *dst, const void __user *src, unsigned size) + + case 4: { + u32 tmp; +- __get_user_asm(tmp, (u32 __user *)src, ++ __get_user_asm(tmp, (const u32 __user *)src, + ret, "l", "k", "=r", 4); + if (likely(!ret)) + __put_user_asm(tmp, (u32 __user *)dst, +@@ -153,7 +245,7 @@ int __copy_in_user(void __user *dst, const void __user *src, unsigned size) + } + case 8: { + u64 tmp; +- __get_user_asm(tmp, (u64 __user *)src, ++ __get_user_asm(tmp, (const u64 __user *)src, + ret, "q", "", "=r", 8); + if (likely(!ret)) + __put_user_asm(tmp, (u64 __user *)dst, +@@ -161,8 +253,16 @@ int __copy_in_user(void __user *dst, const void __user *src, unsigned size) + return ret; + } + default: +- return copy_user_generic((__force void *)dst, +- (__force void *)src, size); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((__force_kernel void *)dst, ++ (__force_kernel const void *)src, size); + } + } + +@@ -176,33 +276,75 @@ __must_check long strlen_user(const char __user *str); + __must_check unsigned long clear_user(void __user *mem, unsigned long len); + __must_check unsigned long __clear_user(void __user *mem, unsigned long len); + +-__must_check long __copy_from_user_inatomic(void *dst, const void __user *src, +- unsigned size); ++static __must_check __always_inline unsigned long ++__copy_from_user_inatomic(void *dst, const void __user *src, unsigned long size) ++{ ++ pax_track_stack(); ++ ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_READ, src, size)) ++ return size; + +-static __must_check __always_inline int +-__copy_to_user_inatomic(void __user *dst, const void *src, unsigned size) ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic(dst, (__force_kernel const void *)src, size); ++} ++ ++static __must_check __always_inline unsigned long ++__copy_to_user_inatomic(void __user *dst, const void *src, unsigned long size) + { +- return copy_user_generic((__force void *)dst, src, size); ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_WRITE, dst, size)) ++ return size; ++ ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((__force_kernel void *)dst, src, size); + } + +-extern long __copy_user_nocache(void *dst, const void __user *src, +- unsigned size, int zerorest); ++extern unsigned long __copy_user_nocache(void *dst, const void __user *src, ++ unsigned long size, int zerorest); + +-static inline int +-__copy_from_user_nocache(void *dst, const void __user *src, unsigned size) ++static inline unsigned long __copy_from_user_nocache(void *dst, const void __user *src, unsigned long size) + { + might_sleep(); ++ ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_READ, src, size)) ++ return size; ++#endif ++ + return __copy_user_nocache(dst, src, size, 1); + } + +-static inline int +-__copy_from_user_inatomic_nocache(void *dst, const void __user *src, +- unsigned size) ++static inline unsigned long __copy_from_user_inatomic_nocache(void *dst, const void __user *src, ++ unsigned long size) + { ++ if (size > INT_MAX) ++ return size; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if (!__access_ok(VERIFY_READ, src, size)) ++ return size; ++#endif ++ + return __copy_user_nocache(dst, src, size, 0); + } + +-unsigned long +-copy_user_handle_tail(char *to, char *from, unsigned len, unsigned zerorest); ++extern unsigned long ++copy_user_handle_tail(char __user *to, char __user *from, unsigned long len, unsigned zerorest); + + #endif /* _ASM_X86_UACCESS_64_H */ +diff --git a/arch/x86/include/asm/vdso.h b/arch/x86/include/asm/vdso.h +index 9064052..786cfbc 100644 +--- a/arch/x86/include/asm/vdso.h ++++ b/arch/x86/include/asm/vdso.h +@@ -25,7 +25,7 @@ extern const char VDSO32_PRELINK[]; + #define VDSO32_SYMBOL(base, name) \ + ({ \ + extern const char VDSO32_##name[]; \ +- (void *)(VDSO32_##name - VDSO32_PRELINK + (unsigned long)(base)); \ ++ (void __user *)(VDSO32_##name - VDSO32_PRELINK + (unsigned long)(base)); \ + }) + #endif + +diff --git a/arch/x86/include/asm/vgtod.h b/arch/x86/include/asm/vgtod.h +index 3d61e20..9507180 100644 +--- a/arch/x86/include/asm/vgtod.h ++++ b/arch/x86/include/asm/vgtod.h +@@ -14,6 +14,7 @@ struct vsyscall_gtod_data { + int sysctl_enabled; + struct timezone sys_tz; + struct { /* extract of a clocksource struct */ ++ char name[8]; + cycle_t (*vread)(void); + cycle_t cycle_last; + cycle_t mask; +diff --git a/arch/x86/include/asm/vmi.h b/arch/x86/include/asm/vmi.h +index 61e08c0..b0da582 100644 +--- a/arch/x86/include/asm/vmi.h ++++ b/arch/x86/include/asm/vmi.h +@@ -191,6 +191,7 @@ struct vrom_header { + u8 reserved[96]; /* Reserved for headers */ + char vmi_init[8]; /* VMI_Init jump point */ + char get_reloc[8]; /* VMI_GetRelocationInfo jump point */ ++ char rom_data[8048]; /* rest of the option ROM */ + } __attribute__((packed)); + + struct pnp_header { +diff --git a/arch/x86/include/asm/vmi_time.h b/arch/x86/include/asm/vmi_time.h +index c6e0bee..fcb9f74 100644 +--- a/arch/x86/include/asm/vmi_time.h ++++ b/arch/x86/include/asm/vmi_time.h +@@ -43,7 +43,7 @@ extern struct vmi_timer_ops { + int (*wallclock_updated)(void); + void (*set_alarm)(u32 flags, u64 expiry, u64 period); + void (*cancel_alarm)(u32 flags); +-} vmi_timer_ops; ++} __no_const vmi_timer_ops; + + /* Prototypes */ + extern void __init vmi_time_init(void); +diff --git a/arch/x86/include/asm/vsyscall.h b/arch/x86/include/asm/vsyscall.h +index d0983d2..1f7c9e9 100644 +--- a/arch/x86/include/asm/vsyscall.h ++++ b/arch/x86/include/asm/vsyscall.h +@@ -15,9 +15,10 @@ enum vsyscall_num { + + #ifdef __KERNEL__ + #include ++#include ++#include + + #define __section_vgetcpu_mode __attribute__ ((unused, __section__ (".vgetcpu_mode"), aligned(16))) +-#define __section_jiffies __attribute__ ((unused, __section__ (".jiffies"), aligned(16))) + + /* Definitions for CONFIG_GENERIC_TIME definitions */ + #define __section_vsyscall_gtod_data __attribute__ \ +@@ -31,7 +32,6 @@ enum vsyscall_num { + #define VGETCPU_LSL 2 + + extern int __vgetcpu_mode; +-extern volatile unsigned long __jiffies; + + /* kernel space (writeable) */ + extern int vgetcpu_mode; +@@ -39,6 +39,9 @@ extern struct timezone sys_tz; + + extern void map_vsyscall(void); + ++extern int vgettimeofday(struct timeval * tv, struct timezone * tz); ++extern time_t vtime(time_t *t); ++extern long vgetcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache); + #endif /* __KERNEL__ */ + + #endif /* _ASM_X86_VSYSCALL_H */ +diff --git a/arch/x86/include/asm/x86_init.h b/arch/x86/include/asm/x86_init.h +index 2c756fd..3377e37 100644 +--- a/arch/x86/include/asm/x86_init.h ++++ b/arch/x86/include/asm/x86_init.h +@@ -28,7 +28,7 @@ struct x86_init_mpparse { + void (*mpc_oem_bus_info)(struct mpc_bus *m, char *name); + void (*find_smp_config)(unsigned int reserve); + void (*get_smp_config)(unsigned int early); +-}; ++} __no_const; + + /** + * struct x86_init_resources - platform specific resource related ops +@@ -42,7 +42,7 @@ struct x86_init_resources { + void (*probe_roms)(void); + void (*reserve_resources)(void); + char *(*memory_setup)(void); +-}; ++} __no_const; + + /** + * struct x86_init_irqs - platform specific interrupt setup +@@ -55,7 +55,7 @@ struct x86_init_irqs { + void (*pre_vector_init)(void); + void (*intr_init)(void); + void (*trap_init)(void); +-}; ++} __no_const; + + /** + * struct x86_init_oem - oem platform specific customizing functions +@@ -65,7 +65,7 @@ struct x86_init_irqs { + struct x86_init_oem { + void (*arch_setup)(void); + void (*banner)(void); +-}; ++} __no_const; + + /** + * struct x86_init_paging - platform specific paging functions +@@ -75,7 +75,7 @@ struct x86_init_oem { + struct x86_init_paging { + void (*pagetable_setup_start)(pgd_t *base); + void (*pagetable_setup_done)(pgd_t *base); +-}; ++} __no_const; + + /** + * struct x86_init_timers - platform specific timer setup +@@ -88,7 +88,7 @@ struct x86_init_timers { + void (*setup_percpu_clockev)(void); + void (*tsc_pre_init)(void); + void (*timer_init)(void); +-}; ++} __no_const; + + /** + * struct x86_init_ops - functions for platform specific setup +@@ -101,7 +101,7 @@ struct x86_init_ops { + struct x86_init_oem oem; + struct x86_init_paging paging; + struct x86_init_timers timers; +-}; ++} __no_const; + + /** + * struct x86_cpuinit_ops - platform specific cpu hotplug setups +@@ -109,7 +109,7 @@ struct x86_init_ops { + */ + struct x86_cpuinit_ops { + void (*setup_percpu_clockev)(void); +-}; ++} __no_const; + + /** + * struct x86_platform_ops - platform specific runtime functions +@@ -121,7 +121,7 @@ struct x86_platform_ops { + unsigned long (*calibrate_tsc)(void); + unsigned long (*get_wallclock)(void); + int (*set_wallclock)(unsigned long nowtime); +-}; ++} __no_const; + + extern struct x86_init_ops x86_init; + extern struct x86_cpuinit_ops x86_cpuinit; +diff --git a/arch/x86/include/asm/xsave.h b/arch/x86/include/asm/xsave.h +index 727acc1..554f3eb 100644 +--- a/arch/x86/include/asm/xsave.h ++++ b/arch/x86/include/asm/xsave.h +@@ -56,6 +56,12 @@ static inline int xrstor_checking(struct xsave_struct *fx) + static inline int xsave_user(struct xsave_struct __user *buf) + { + int err; ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ if ((unsigned long)buf < PAX_USER_SHADOW_BASE) ++ buf = (struct xsave_struct __user *)((void __user*)buf + PAX_USER_SHADOW_BASE); ++#endif ++ + __asm__ __volatile__("1: .byte " REX_PREFIX "0x0f,0xae,0x27\n" + "2:\n" + ".section .fixup,\"ax\"\n" +@@ -78,10 +84,15 @@ static inline int xsave_user(struct xsave_struct __user *buf) + static inline int xrestore_user(struct xsave_struct __user *buf, u64 mask) + { + int err; +- struct xsave_struct *xstate = ((__force struct xsave_struct *)buf); ++ struct xsave_struct *xstate = ((__force_kernel struct xsave_struct *)buf); + u32 lmask = mask; + u32 hmask = mask >> 32; + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ if ((unsigned long)xstate < PAX_USER_SHADOW_BASE) ++ xstate = (struct xsave_struct *)((void *)xstate + PAX_USER_SHADOW_BASE); ++#endif ++ + __asm__ __volatile__("1: .byte " REX_PREFIX "0x0f,0xae,0x2f\n" + "2:\n" + ".section .fixup,\"ax\"\n" +diff --git a/arch/x86/kernel/acpi/realmode/Makefile b/arch/x86/kernel/acpi/realmode/Makefile +index 6a564ac..9b1340c 100644 +--- a/arch/x86/kernel/acpi/realmode/Makefile ++++ b/arch/x86/kernel/acpi/realmode/Makefile +@@ -41,6 +41,9 @@ KBUILD_CFLAGS := $(LINUXINCLUDE) -g -Os -D_SETUP -D_WAKEUP -D__KERNEL__ \ + $(call cc-option, -fno-stack-protector) \ + $(call cc-option, -mpreferred-stack-boundary=2) + KBUILD_CFLAGS += $(call cc-option, -m32) ++ifdef CONSTIFY_PLUGIN ++KBUILD_CFLAGS += $(CONSTIFY_PLUGIN) -fplugin-arg-constify_plugin-no-constify ++endif + KBUILD_AFLAGS := $(KBUILD_CFLAGS) -D__ASSEMBLY__ + GCOV_PROFILE := n + +diff --git a/arch/x86/kernel/acpi/realmode/wakeup.S b/arch/x86/kernel/acpi/realmode/wakeup.S +index 580b4e2..d4129e4 100644 +--- a/arch/x86/kernel/acpi/realmode/wakeup.S ++++ b/arch/x86/kernel/acpi/realmode/wakeup.S +@@ -91,6 +91,9 @@ _start: + /* Do any other stuff... */ + + #ifndef CONFIG_64BIT ++ /* Recheck NX bit overrides (64bit path does this in trampoline) */ ++ call verify_cpu ++ + /* This could also be done in C code... */ + movl pmode_cr3, %eax + movl %eax, %cr3 +@@ -104,7 +107,7 @@ _start: + movl %eax, %ecx + orl %edx, %ecx + jz 1f +- movl $0xc0000080, %ecx ++ mov $MSR_EFER, %ecx + wrmsr + 1: + +@@ -114,6 +117,7 @@ _start: + movl pmode_cr0, %eax + movl %eax, %cr0 + jmp pmode_return ++# include "../../verify_cpu.S" + #else + pushw $0 + pushw trampoline_segment +diff --git a/arch/x86/kernel/acpi/sleep.c b/arch/x86/kernel/acpi/sleep.c +index ca93638..7042f24 100644 +--- a/arch/x86/kernel/acpi/sleep.c ++++ b/arch/x86/kernel/acpi/sleep.c +@@ -11,11 +11,12 @@ + #include + #include + #include ++#include + + #include "realmode/wakeup.h" + #include "sleep.h" + +-unsigned long acpi_wakeup_address; ++unsigned long acpi_wakeup_address = 0x2000; + unsigned long acpi_realmode_flags; + + /* address in low memory of the wakeup routine. */ +@@ -98,9 +99,13 @@ int acpi_save_state_mem(void) + #else /* CONFIG_64BIT */ + header->trampoline_segment = setup_trampoline() >> 4; + #ifdef CONFIG_SMP +- stack_start.sp = temp_stack + sizeof(temp_stack); ++ stack_start = (unsigned long)temp_stack + sizeof(temp_stack); ++ ++ pax_open_kernel(); + early_gdt_descr.address = + (unsigned long)get_cpu_gdt_table(smp_processor_id()); ++ pax_close_kernel(); ++ + initial_gs = per_cpu_offset(smp_processor_id()); + #endif + initial_code = (unsigned long)wakeup_long64; +@@ -134,14 +139,8 @@ void __init acpi_reserve_bootmem(void) + return; + } + +- acpi_realmode = (unsigned long)alloc_bootmem_low(WAKEUP_SIZE); +- +- if (!acpi_realmode) { +- printk(KERN_ERR "ACPI: Cannot allocate lowmem, S3 disabled.\n"); +- return; +- } +- +- acpi_wakeup_address = virt_to_phys((void *)acpi_realmode); ++ reserve_early(acpi_wakeup_address, acpi_wakeup_address + WAKEUP_SIZE, "ACPI Wakeup Code"); ++ acpi_realmode = (unsigned long)__va(acpi_wakeup_address);; + } + + +diff --git a/arch/x86/kernel/acpi/wakeup_32.S b/arch/x86/kernel/acpi/wakeup_32.S +index 8ded418..079961e 100644 +--- a/arch/x86/kernel/acpi/wakeup_32.S ++++ b/arch/x86/kernel/acpi/wakeup_32.S +@@ -30,13 +30,11 @@ wakeup_pmode_return: + # and restore the stack ... but you need gdt for this to work + movl saved_context_esp, %esp + +- movl %cs:saved_magic, %eax +- cmpl $0x12345678, %eax ++ cmpl $0x12345678, saved_magic + jne bogus_magic + + # jump to place where we left off +- movl saved_eip, %eax +- jmp *%eax ++ jmp *(saved_eip) + + bogus_magic: + jmp bogus_magic +diff --git a/arch/x86/kernel/alternative.c b/arch/x86/kernel/alternative.c +index de7353c..075da5f 100644 +--- a/arch/x86/kernel/alternative.c ++++ b/arch/x86/kernel/alternative.c +@@ -407,7 +407,7 @@ void __init_or_module apply_paravirt(struct paravirt_patch_site *start, + + BUG_ON(p->len > MAX_PATCH_LEN); + /* prep the buffer with the original instructions */ +- memcpy(insnbuf, p->instr, p->len); ++ memcpy(insnbuf, ktla_ktva(p->instr), p->len); + used = pv_init_ops.patch(p->instrtype, p->clobbers, insnbuf, + (unsigned long)p->instr, p->len); + +@@ -475,7 +475,7 @@ void __init alternative_instructions(void) + if (smp_alt_once) + free_init_pages("SMP alternatives", + (unsigned long)__smp_locks, +- (unsigned long)__smp_locks_end); ++ PAGE_ALIGN((unsigned long)__smp_locks_end)); + + restart_nmi(); + } +@@ -492,13 +492,17 @@ void __init alternative_instructions(void) + * instructions. And on the local CPU you need to be protected again NMI or MCE + * handlers seeing an inconsistent instruction while you patch. + */ +-static void *__init_or_module text_poke_early(void *addr, const void *opcode, ++static void *__kprobes text_poke_early(void *addr, const void *opcode, + size_t len) + { + unsigned long flags; + local_irq_save(flags); +- memcpy(addr, opcode, len); ++ ++ pax_open_kernel(); ++ memcpy(ktla_ktva(addr), opcode, len); + sync_core(); ++ pax_close_kernel(); ++ + local_irq_restore(flags); + /* Could also do a CLFLUSH here to speed up CPU recovery; but + that causes hangs on some VIA CPUs. */ +@@ -520,35 +524,21 @@ static void *__init_or_module text_poke_early(void *addr, const void *opcode, + */ + void *__kprobes text_poke(void *addr, const void *opcode, size_t len) + { +- unsigned long flags; +- char *vaddr; ++ unsigned char *vaddr = ktla_ktva(addr); + struct page *pages[2]; +- int i; ++ size_t i; + + if (!core_kernel_text((unsigned long)addr)) { +- pages[0] = vmalloc_to_page(addr); +- pages[1] = vmalloc_to_page(addr + PAGE_SIZE); ++ pages[0] = vmalloc_to_page(vaddr); ++ pages[1] = vmalloc_to_page(vaddr + PAGE_SIZE); + } else { +- pages[0] = virt_to_page(addr); ++ pages[0] = virt_to_page(vaddr); + WARN_ON(!PageReserved(pages[0])); +- pages[1] = virt_to_page(addr + PAGE_SIZE); ++ pages[1] = virt_to_page(vaddr + PAGE_SIZE); + } + BUG_ON(!pages[0]); +- local_irq_save(flags); +- set_fixmap(FIX_TEXT_POKE0, page_to_phys(pages[0])); +- if (pages[1]) +- set_fixmap(FIX_TEXT_POKE1, page_to_phys(pages[1])); +- vaddr = (char *)fix_to_virt(FIX_TEXT_POKE0); +- memcpy(&vaddr[(unsigned long)addr & ~PAGE_MASK], opcode, len); +- clear_fixmap(FIX_TEXT_POKE0); +- if (pages[1]) +- clear_fixmap(FIX_TEXT_POKE1); +- local_flush_tlb(); +- sync_core(); +- /* Could also do a CLFLUSH here to speed up CPU recovery; but +- that causes hangs on some VIA CPUs. */ ++ text_poke_early(addr, opcode, len); + for (i = 0; i < len; i++) +- BUG_ON(((char *)addr)[i] != ((char *)opcode)[i]); +- local_irq_restore(flags); ++ BUG_ON((vaddr)[i] != ((const unsigned char *)opcode)[i]); + return addr; + } +diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c +index 3a44b75..1601800 100644 +--- a/arch/x86/kernel/amd_iommu.c ++++ b/arch/x86/kernel/amd_iommu.c +@@ -2076,7 +2076,7 @@ static void prealloc_protection_domains(void) + } + } + +-static struct dma_map_ops amd_iommu_dma_ops = { ++static const struct dma_map_ops amd_iommu_dma_ops = { + .alloc_coherent = alloc_coherent, + .free_coherent = free_coherent, + .map_page = map_page, +diff --git a/arch/x86/kernel/apic/apic.c b/arch/x86/kernel/apic/apic.c +index 1d2d670..8e3f477 100644 +--- a/arch/x86/kernel/apic/apic.c ++++ b/arch/x86/kernel/apic/apic.c +@@ -170,7 +170,7 @@ int first_system_vector = 0xfe; + /* + * Debug level, exported for io_apic.c + */ +-unsigned int apic_verbosity; ++int apic_verbosity; + + int pic_mode; + +@@ -1794,7 +1794,7 @@ void smp_error_interrupt(struct pt_regs *regs) + apic_write(APIC_ESR, 0); + v1 = apic_read(APIC_ESR); + ack_APIC_irq(); +- atomic_inc(&irq_err_count); ++ atomic_inc_unchecked(&irq_err_count); + + /* + * Here is what the APIC error bits mean: +@@ -2184,6 +2184,8 @@ static int __cpuinit apic_cluster_num(void) + u16 *bios_cpu_apicid; + DECLARE_BITMAP(clustermap, NUM_APIC_CLUSTERS); + ++ pax_track_stack(); ++ + bios_cpu_apicid = early_per_cpu_ptr(x86_bios_cpu_apicid); + bitmap_zero(clustermap, NUM_APIC_CLUSTERS); + +diff --git a/arch/x86/kernel/apic/io_apic.c b/arch/x86/kernel/apic/io_apic.c +index 8928d97..f799cea 100644 +--- a/arch/x86/kernel/apic/io_apic.c ++++ b/arch/x86/kernel/apic/io_apic.c +@@ -716,7 +716,7 @@ struct IO_APIC_route_entry **alloc_ioapic_entries(void) + ioapic_entries = kzalloc(sizeof(*ioapic_entries) * nr_ioapics, + GFP_ATOMIC); + if (!ioapic_entries) +- return 0; ++ return NULL; + + for (apic = 0; apic < nr_ioapics; apic++) { + ioapic_entries[apic] = +@@ -733,7 +733,7 @@ nomem: + kfree(ioapic_entries[apic]); + kfree(ioapic_entries); + +- return 0; ++ return NULL; + } + + /* +@@ -1150,7 +1150,7 @@ int IO_APIC_get_PCI_irq_vector(int bus, int slot, int pin, + } + EXPORT_SYMBOL(IO_APIC_get_PCI_irq_vector); + +-void lock_vector_lock(void) ++void lock_vector_lock(void) __acquires(vector_lock) + { + /* Used to the online set of cpus does not change + * during assign_irq_vector. +@@ -1158,7 +1158,7 @@ void lock_vector_lock(void) + spin_lock(&vector_lock); + } + +-void unlock_vector_lock(void) ++void unlock_vector_lock(void) __releases(vector_lock) + { + spin_unlock(&vector_lock); + } +@@ -2542,7 +2542,7 @@ static void ack_apic_edge(unsigned int irq) + ack_APIC_irq(); + } + +-atomic_t irq_mis_count; ++atomic_unchecked_t irq_mis_count; + + static void ack_apic_level(unsigned int irq) + { +@@ -2626,7 +2626,7 @@ static void ack_apic_level(unsigned int irq) + + /* Tail end of version 0x11 I/O APIC bug workaround */ + if (!(v & (1 << (i & 0x1f)))) { +- atomic_inc(&irq_mis_count); ++ atomic_inc_unchecked(&irq_mis_count); + spin_lock(&ioapic_lock); + __mask_and_edge_IO_APIC_irq(cfg); + __unmask_and_level_IO_APIC_irq(cfg); +diff --git a/arch/x86/kernel/apm_32.c b/arch/x86/kernel/apm_32.c +index 151ace6..f317474 100644 +--- a/arch/x86/kernel/apm_32.c ++++ b/arch/x86/kernel/apm_32.c +@@ -410,7 +410,7 @@ static DEFINE_SPINLOCK(user_list_lock); + * This is for buggy BIOS's that refer to (real mode) segment 0x40 + * even though they are called in protected mode. + */ +-static struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4092, ++static const struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4093, + (unsigned long)__va(0x400UL), PAGE_SIZE - 0x400 - 1); + + static const char driver_version[] = "1.16ac"; /* no spaces */ +@@ -588,7 +588,10 @@ static long __apm_bios_call(void *_call) + BUG_ON(cpu != 0); + gdt = get_cpu_gdt_table(cpu); + save_desc_40 = gdt[0x40 / 8]; ++ ++ pax_open_kernel(); + gdt[0x40 / 8] = bad_bios_desc; ++ pax_close_kernel(); + + apm_irq_save(flags); + APM_DO_SAVE_SEGS; +@@ -597,7 +600,11 @@ static long __apm_bios_call(void *_call) + &call->esi); + APM_DO_RESTORE_SEGS; + apm_irq_restore(flags); ++ ++ pax_open_kernel(); + gdt[0x40 / 8] = save_desc_40; ++ pax_close_kernel(); ++ + put_cpu(); + + return call->eax & 0xff; +@@ -664,7 +671,10 @@ static long __apm_bios_call_simple(void *_call) + BUG_ON(cpu != 0); + gdt = get_cpu_gdt_table(cpu); + save_desc_40 = gdt[0x40 / 8]; ++ ++ pax_open_kernel(); + gdt[0x40 / 8] = bad_bios_desc; ++ pax_close_kernel(); + + apm_irq_save(flags); + APM_DO_SAVE_SEGS; +@@ -672,7 +682,11 @@ static long __apm_bios_call_simple(void *_call) + &call->eax); + APM_DO_RESTORE_SEGS; + apm_irq_restore(flags); ++ ++ pax_open_kernel(); + gdt[0x40 / 8] = save_desc_40; ++ pax_close_kernel(); ++ + put_cpu(); + return error; + } +@@ -975,7 +989,7 @@ recalc: + + static void apm_power_off(void) + { +- unsigned char po_bios_call[] = { ++ const unsigned char po_bios_call[] = { + 0xb8, 0x00, 0x10, /* movw $0x1000,ax */ + 0x8e, 0xd0, /* movw ax,ss */ + 0xbc, 0x00, 0xf0, /* movw $0xf000,sp */ +@@ -2357,12 +2371,15 @@ static int __init apm_init(void) + * code to that CPU. + */ + gdt = get_cpu_gdt_table(0); ++ ++ pax_open_kernel(); + set_desc_base(&gdt[APM_CS >> 3], + (unsigned long)__va((unsigned long)apm_info.bios.cseg << 4)); + set_desc_base(&gdt[APM_CS_16 >> 3], + (unsigned long)__va((unsigned long)apm_info.bios.cseg_16 << 4)); + set_desc_base(&gdt[APM_DS >> 3], + (unsigned long)__va((unsigned long)apm_info.bios.dseg << 4)); ++ pax_close_kernel(); + + proc_create("apm", 0, NULL, &apm_file_ops); + +diff --git a/arch/x86/kernel/asm-offsets_32.c b/arch/x86/kernel/asm-offsets_32.c +index dfdbf64..9b2b6ce 100644 +--- a/arch/x86/kernel/asm-offsets_32.c ++++ b/arch/x86/kernel/asm-offsets_32.c +@@ -51,7 +51,6 @@ void foo(void) + OFFSET(CPUINFO_x86_vendor_id, cpuinfo_x86, x86_vendor_id); + BLANK(); + +- OFFSET(TI_task, thread_info, task); + OFFSET(TI_exec_domain, thread_info, exec_domain); + OFFSET(TI_flags, thread_info, flags); + OFFSET(TI_status, thread_info, status); +@@ -60,6 +59,8 @@ void foo(void) + OFFSET(TI_restart_block, thread_info, restart_block); + OFFSET(TI_sysenter_return, thread_info, sysenter_return); + OFFSET(TI_cpu, thread_info, cpu); ++ OFFSET(TI_lowest_stack, thread_info, lowest_stack); ++ DEFINE(TI_task_thread_sp0, offsetof(struct task_struct, thread.sp0) - offsetof(struct task_struct, tinfo)); + BLANK(); + + OFFSET(GDS_size, desc_ptr, size); +@@ -99,6 +100,7 @@ void foo(void) + + DEFINE(PAGE_SIZE_asm, PAGE_SIZE); + DEFINE(PAGE_SHIFT_asm, PAGE_SHIFT); ++ DEFINE(THREAD_SIZE_asm, THREAD_SIZE); + DEFINE(PTRS_PER_PTE, PTRS_PER_PTE); + DEFINE(PTRS_PER_PMD, PTRS_PER_PMD); + DEFINE(PTRS_PER_PGD, PTRS_PER_PGD); +@@ -115,6 +117,11 @@ void foo(void) + OFFSET(PV_CPU_iret, pv_cpu_ops, iret); + OFFSET(PV_CPU_irq_enable_sysexit, pv_cpu_ops, irq_enable_sysexit); + OFFSET(PV_CPU_read_cr0, pv_cpu_ops, read_cr0); ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ OFFSET(PV_CPU_write_cr0, pv_cpu_ops, write_cr0); ++#endif ++ + #endif + + #ifdef CONFIG_XEN +diff --git a/arch/x86/kernel/asm-offsets_64.c b/arch/x86/kernel/asm-offsets_64.c +index 4a6aeed..371de20 100644 +--- a/arch/x86/kernel/asm-offsets_64.c ++++ b/arch/x86/kernel/asm-offsets_64.c +@@ -44,6 +44,8 @@ int main(void) + ENTRY(addr_limit); + ENTRY(preempt_count); + ENTRY(status); ++ ENTRY(lowest_stack); ++ DEFINE(TI_task_thread_sp0, offsetof(struct task_struct, thread.sp0) - offsetof(struct task_struct, tinfo)); + #ifdef CONFIG_IA32_EMULATION + ENTRY(sysenter_return); + #endif +@@ -63,6 +65,18 @@ int main(void) + OFFSET(PV_CPU_irq_enable_sysexit, pv_cpu_ops, irq_enable_sysexit); + OFFSET(PV_CPU_swapgs, pv_cpu_ops, swapgs); + OFFSET(PV_MMU_read_cr2, pv_mmu_ops, read_cr2); ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ OFFSET(PV_CPU_read_cr0, pv_cpu_ops, read_cr0); ++ OFFSET(PV_CPU_write_cr0, pv_cpu_ops, write_cr0); ++#endif ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ OFFSET(PV_MMU_read_cr3, pv_mmu_ops, read_cr3); ++ OFFSET(PV_MMU_write_cr3, pv_mmu_ops, write_cr3); ++ OFFSET(PV_MMU_set_pgd_batched, pv_mmu_ops, set_pgd_batched); ++#endif ++ + #endif + + +@@ -115,6 +129,7 @@ int main(void) + ENTRY(cr8); + BLANK(); + #undef ENTRY ++ DEFINE(TSS_size, sizeof(struct tss_struct)); + DEFINE(TSS_ist, offsetof(struct tss_struct, x86_tss.ist)); + BLANK(); + DEFINE(crypto_tfm_ctx_offset, offsetof(struct crypto_tfm, __crt_ctx)); +@@ -130,6 +145,7 @@ int main(void) + + BLANK(); + DEFINE(PAGE_SIZE_asm, PAGE_SIZE); ++ DEFINE(THREAD_SIZE_asm, THREAD_SIZE); + #ifdef CONFIG_XEN + BLANK(); + OFFSET(XEN_vcpu_info_mask, vcpu_info, evtchn_upcall_mask); +diff --git a/arch/x86/kernel/cpu/Makefile b/arch/x86/kernel/cpu/Makefile +index ff502cc..dc5133e 100644 +--- a/arch/x86/kernel/cpu/Makefile ++++ b/arch/x86/kernel/cpu/Makefile +@@ -7,10 +7,6 @@ ifdef CONFIG_FUNCTION_TRACER + CFLAGS_REMOVE_common.o = -pg + endif + +-# Make sure load_percpu_segment has no stackprotector +-nostackp := $(call cc-option, -fno-stack-protector) +-CFLAGS_common.o := $(nostackp) +- + obj-y := intel_cacheinfo.o addon_cpuid_features.o + obj-y += proc.o capflags.o powerflags.o common.o + obj-y += vmware.o hypervisor.o sched.o +diff --git a/arch/x86/kernel/cpu/amd.c b/arch/x86/kernel/cpu/amd.c +index 6e082dc..a0b5f36 100644 +--- a/arch/x86/kernel/cpu/amd.c ++++ b/arch/x86/kernel/cpu/amd.c +@@ -602,7 +602,7 @@ static unsigned int __cpuinit amd_size_cache(struct cpuinfo_x86 *c, + unsigned int size) + { + /* AMD errata T13 (order #21922) */ +- if ((c->x86 == 6)) { ++ if (c->x86 == 6) { + /* Duron Rev A0 */ + if (c->x86_model == 3 && c->x86_mask == 0) + size = 64; +diff --git a/arch/x86/kernel/cpu/common.c b/arch/x86/kernel/cpu/common.c +index 4e34d10..ba6bc97 100644 +--- a/arch/x86/kernel/cpu/common.c ++++ b/arch/x86/kernel/cpu/common.c +@@ -83,60 +83,6 @@ static const struct cpu_dev __cpuinitconst default_cpu = { + + static const struct cpu_dev *this_cpu __cpuinitdata = &default_cpu; + +-DEFINE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page) = { .gdt = { +-#ifdef CONFIG_X86_64 +- /* +- * We need valid kernel segments for data and code in long mode too +- * IRET will check the segment types kkeil 2000/10/28 +- * Also sysret mandates a special GDT layout +- * +- * TLS descriptors are currently at a different place compared to i386. +- * Hopefully nobody expects them at a fixed place (Wine?) +- */ +- [GDT_ENTRY_KERNEL32_CS] = GDT_ENTRY_INIT(0xc09b, 0, 0xfffff), +- [GDT_ENTRY_KERNEL_CS] = GDT_ENTRY_INIT(0xa09b, 0, 0xfffff), +- [GDT_ENTRY_KERNEL_DS] = GDT_ENTRY_INIT(0xc093, 0, 0xfffff), +- [GDT_ENTRY_DEFAULT_USER32_CS] = GDT_ENTRY_INIT(0xc0fb, 0, 0xfffff), +- [GDT_ENTRY_DEFAULT_USER_DS] = GDT_ENTRY_INIT(0xc0f3, 0, 0xfffff), +- [GDT_ENTRY_DEFAULT_USER_CS] = GDT_ENTRY_INIT(0xa0fb, 0, 0xfffff), +-#else +- [GDT_ENTRY_KERNEL_CS] = GDT_ENTRY_INIT(0xc09a, 0, 0xfffff), +- [GDT_ENTRY_KERNEL_DS] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff), +- [GDT_ENTRY_DEFAULT_USER_CS] = GDT_ENTRY_INIT(0xc0fa, 0, 0xfffff), +- [GDT_ENTRY_DEFAULT_USER_DS] = GDT_ENTRY_INIT(0xc0f2, 0, 0xfffff), +- /* +- * Segments used for calling PnP BIOS have byte granularity. +- * They code segments and data segments have fixed 64k limits, +- * the transfer segment sizes are set at run time. +- */ +- /* 32-bit code */ +- [GDT_ENTRY_PNPBIOS_CS32] = GDT_ENTRY_INIT(0x409a, 0, 0xffff), +- /* 16-bit code */ +- [GDT_ENTRY_PNPBIOS_CS16] = GDT_ENTRY_INIT(0x009a, 0, 0xffff), +- /* 16-bit data */ +- [GDT_ENTRY_PNPBIOS_DS] = GDT_ENTRY_INIT(0x0092, 0, 0xffff), +- /* 16-bit data */ +- [GDT_ENTRY_PNPBIOS_TS1] = GDT_ENTRY_INIT(0x0092, 0, 0), +- /* 16-bit data */ +- [GDT_ENTRY_PNPBIOS_TS2] = GDT_ENTRY_INIT(0x0092, 0, 0), +- /* +- * The APM segments have byte granularity and their bases +- * are set at run time. All have 64k limits. +- */ +- /* 32-bit code */ +- [GDT_ENTRY_APMBIOS_BASE] = GDT_ENTRY_INIT(0x409a, 0, 0xffff), +- /* 16-bit code */ +- [GDT_ENTRY_APMBIOS_BASE+1] = GDT_ENTRY_INIT(0x009a, 0, 0xffff), +- /* data */ +- [GDT_ENTRY_APMBIOS_BASE+2] = GDT_ENTRY_INIT(0x4092, 0, 0xffff), +- +- [GDT_ENTRY_ESPFIX_SS] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff), +- [GDT_ENTRY_PERCPU] = GDT_ENTRY_INIT(0xc092, 0, 0xfffff), +- GDT_STACK_CANARY_INIT +-#endif +-} }; +-EXPORT_PER_CPU_SYMBOL_GPL(gdt_page); +- + static int __init x86_xsave_setup(char *s) + { + setup_clear_cpu_cap(X86_FEATURE_XSAVE); +@@ -344,7 +290,7 @@ void switch_to_new_gdt(int cpu) + { + struct desc_ptr gdt_descr; + +- gdt_descr.address = (long)get_cpu_gdt_table(cpu); ++ gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu); + gdt_descr.size = GDT_SIZE - 1; + load_gdt(&gdt_descr); + /* Reload the per-cpu base */ +@@ -798,6 +744,10 @@ static void __cpuinit identify_cpu(struct cpuinfo_x86 *c) + /* Filter out anything that depends on CPUID levels we don't have */ + filter_cpuid_features(c, true); + ++#if defined(CONFIG_X86_32) && (defined(CONFIG_PAX_SEGMEXEC) || defined(CONFIG_PAX_KERNEXEC) || defined(CONFIG_PAX_MEMORY_UDEREF)) ++ setup_clear_cpu_cap(X86_FEATURE_SEP); ++#endif ++ + /* If the model name is still unset, do table lookup. */ + if (!c->x86_model_id[0]) { + const char *p; +@@ -980,6 +930,9 @@ static __init int setup_disablecpuid(char *arg) + } + __setup("clearcpuid=", setup_disablecpuid); + ++DEFINE_PER_CPU(struct thread_info *, current_tinfo) = &init_task.tinfo; ++EXPORT_PER_CPU_SYMBOL(current_tinfo); ++ + #ifdef CONFIG_X86_64 + struct desc_ptr idt_descr = { NR_VECTORS * 16 - 1, (unsigned long) idt_table }; + +@@ -995,7 +948,7 @@ DEFINE_PER_CPU(struct task_struct *, current_task) ____cacheline_aligned = + EXPORT_PER_CPU_SYMBOL(current_task); + + DEFINE_PER_CPU(unsigned long, kernel_stack) = +- (unsigned long)&init_thread_union - KERNEL_STACK_OFFSET + THREAD_SIZE; ++ (unsigned long)&init_thread_union - 16 + THREAD_SIZE; + EXPORT_PER_CPU_SYMBOL(kernel_stack); + + DEFINE_PER_CPU(char *, irq_stack_ptr) = +@@ -1060,7 +1013,7 @@ struct pt_regs * __cpuinit idle_regs(struct pt_regs *regs) + { + memset(regs, 0, sizeof(struct pt_regs)); + regs->fs = __KERNEL_PERCPU; +- regs->gs = __KERNEL_STACK_CANARY; ++ savesegment(gs, regs->gs); + + return regs; + } +@@ -1101,7 +1054,7 @@ void __cpuinit cpu_init(void) + int i; + + cpu = stack_smp_processor_id(); +- t = &per_cpu(init_tss, cpu); ++ t = init_tss + cpu; + orig_ist = &per_cpu(orig_ist, cpu); + + #ifdef CONFIG_NUMA +@@ -1127,7 +1080,7 @@ void __cpuinit cpu_init(void) + switch_to_new_gdt(cpu); + loadsegment(fs, 0); + +- load_idt((const struct desc_ptr *)&idt_descr); ++ load_idt(&idt_descr); + + memset(me->thread.tls_array, 0, GDT_ENTRY_TLS_ENTRIES * 8); + syscall_init(); +@@ -1136,7 +1089,6 @@ void __cpuinit cpu_init(void) + wrmsrl(MSR_KERNEL_GS_BASE, 0); + barrier(); + +- check_efer(); + if (cpu != 0) + enable_x2apic(); + +@@ -1199,7 +1151,7 @@ void __cpuinit cpu_init(void) + { + int cpu = smp_processor_id(); + struct task_struct *curr = current; +- struct tss_struct *t = &per_cpu(init_tss, cpu); ++ struct tss_struct *t = init_tss + cpu; + struct thread_struct *thread = &curr->thread; + + if (cpumask_test_and_set_cpu(cpu, cpu_initialized_mask)) { +diff --git a/arch/x86/kernel/cpu/intel.c b/arch/x86/kernel/cpu/intel.c +index 6a77cca..4f4fca0 100644 +--- a/arch/x86/kernel/cpu/intel.c ++++ b/arch/x86/kernel/cpu/intel.c +@@ -162,7 +162,7 @@ static void __cpuinit trap_init_f00f_bug(void) + * Update the IDT descriptor and reload the IDT so that + * it uses the read-only mapped virtual address. + */ +- idt_descr.address = fix_to_virt(FIX_F00F_IDT); ++ idt_descr.address = (struct desc_struct *)fix_to_virt(FIX_F00F_IDT); + load_idt(&idt_descr); + } + #endif +diff --git a/arch/x86/kernel/cpu/intel_cacheinfo.c b/arch/x86/kernel/cpu/intel_cacheinfo.c +index 417990f..96dc36b 100644 +--- a/arch/x86/kernel/cpu/intel_cacheinfo.c ++++ b/arch/x86/kernel/cpu/intel_cacheinfo.c +@@ -921,7 +921,7 @@ static ssize_t store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops sysfs_ops = { ++static const struct sysfs_ops sysfs_ops = { + .show = show, + .store = store, + }; +diff --git a/arch/x86/kernel/cpu/mcheck/mce-inject.c b/arch/x86/kernel/cpu/mcheck/mce-inject.c +index 472763d..9831e11 100644 +--- a/arch/x86/kernel/cpu/mcheck/mce-inject.c ++++ b/arch/x86/kernel/cpu/mcheck/mce-inject.c +@@ -211,7 +211,9 @@ static ssize_t mce_write(struct file *filp, const char __user *ubuf, + static int inject_init(void) + { + printk(KERN_INFO "Machine check injector initialized\n"); +- mce_chrdev_ops.write = mce_write; ++ pax_open_kernel(); ++ *(void **)&mce_chrdev_ops.write = mce_write; ++ pax_close_kernel(); + register_die_notifier(&mce_raise_nb); + return 0; + } +diff --git a/arch/x86/kernel/cpu/mcheck/mce.c b/arch/x86/kernel/cpu/mcheck/mce.c +index 0f16a2b..21740f5 100644 +--- a/arch/x86/kernel/cpu/mcheck/mce.c ++++ b/arch/x86/kernel/cpu/mcheck/mce.c +@@ -43,6 +43,7 @@ + #include + #include + #include ++#include + + #include "mce-internal.h" + +@@ -187,7 +188,7 @@ static void print_mce(struct mce *m) + !(m->mcgstatus & MCG_STATUS_EIPV) ? " !INEXACT!" : "", + m->cs, m->ip); + +- if (m->cs == __KERNEL_CS) ++ if (m->cs == __KERNEL_CS || m->cs == __KERNEXEC_KERNEL_CS) + print_symbol("{%s}", m->ip); + pr_cont("\n"); + } +@@ -221,10 +222,10 @@ static void print_mce_tail(void) + + #define PANIC_TIMEOUT 5 /* 5 seconds */ + +-static atomic_t mce_paniced; ++static atomic_unchecked_t mce_paniced; + + static int fake_panic; +-static atomic_t mce_fake_paniced; ++static atomic_unchecked_t mce_fake_paniced; + + /* Panic in progress. Enable interrupts and wait for final IPI */ + static void wait_for_panic(void) +@@ -248,7 +249,7 @@ static void mce_panic(char *msg, struct mce *final, char *exp) + /* + * Make sure only one CPU runs in machine check panic + */ +- if (atomic_inc_return(&mce_paniced) > 1) ++ if (atomic_inc_return_unchecked(&mce_paniced) > 1) + wait_for_panic(); + barrier(); + +@@ -256,7 +257,7 @@ static void mce_panic(char *msg, struct mce *final, char *exp) + console_verbose(); + } else { + /* Don't log too much for fake panic */ +- if (atomic_inc_return(&mce_fake_paniced) > 1) ++ if (atomic_inc_return_unchecked(&mce_fake_paniced) > 1) + return; + } + print_mce_head(); +@@ -616,7 +617,7 @@ static int mce_timed_out(u64 *t) + * might have been modified by someone else. + */ + rmb(); +- if (atomic_read(&mce_paniced)) ++ if (atomic_read_unchecked(&mce_paniced)) + wait_for_panic(); + if (!monarch_timeout) + goto out; +@@ -1394,7 +1395,7 @@ static void unexpected_machine_check(struct pt_regs *regs, long error_code) + } + + /* Call the installed machine check handler for this CPU setup. */ +-void (*machine_check_vector)(struct pt_regs *, long error_code) = ++void (*machine_check_vector)(struct pt_regs *, long error_code) __read_only = + unexpected_machine_check; + + /* +@@ -1416,7 +1417,9 @@ void __cpuinit mcheck_init(struct cpuinfo_x86 *c) + return; + } + ++ pax_open_kernel(); + machine_check_vector = do_machine_check; ++ pax_close_kernel(); + + mce_init(); + mce_cpu_features(c); +@@ -1429,14 +1432,14 @@ void __cpuinit mcheck_init(struct cpuinfo_x86 *c) + */ + + static DEFINE_SPINLOCK(mce_state_lock); +-static int open_count; /* #times opened */ ++static local_t open_count; /* #times opened */ + static int open_exclu; /* already open exclusive? */ + + static int mce_open(struct inode *inode, struct file *file) + { + spin_lock(&mce_state_lock); + +- if (open_exclu || (open_count && (file->f_flags & O_EXCL))) { ++ if (open_exclu || (local_read(&open_count) && (file->f_flags & O_EXCL))) { + spin_unlock(&mce_state_lock); + + return -EBUSY; +@@ -1444,7 +1447,7 @@ static int mce_open(struct inode *inode, struct file *file) + + if (file->f_flags & O_EXCL) + open_exclu = 1; +- open_count++; ++ local_inc(&open_count); + + spin_unlock(&mce_state_lock); + +@@ -1455,7 +1458,7 @@ static int mce_release(struct inode *inode, struct file *file) + { + spin_lock(&mce_state_lock); + +- open_count--; ++ local_dec(&open_count); + open_exclu = 0; + + spin_unlock(&mce_state_lock); +@@ -2082,7 +2085,7 @@ struct dentry *mce_get_debugfs_dir(void) + static void mce_reset(void) + { + cpu_missing = 0; +- atomic_set(&mce_fake_paniced, 0); ++ atomic_set_unchecked(&mce_fake_paniced, 0); + atomic_set(&mce_executing, 0); + atomic_set(&mce_callin, 0); + atomic_set(&global_nwo, 0); +diff --git a/arch/x86/kernel/cpu/mcheck/mce_amd.c b/arch/x86/kernel/cpu/mcheck/mce_amd.c +index ef3cd31..9d2f6ab 100644 +--- a/arch/x86/kernel/cpu/mcheck/mce_amd.c ++++ b/arch/x86/kernel/cpu/mcheck/mce_amd.c +@@ -385,7 +385,7 @@ static ssize_t store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops threshold_ops = { ++static const struct sysfs_ops threshold_ops = { + .show = show, + .store = store, + }; +diff --git a/arch/x86/kernel/cpu/mcheck/p5.c b/arch/x86/kernel/cpu/mcheck/p5.c +index 5c0e653..0882b0a 100644 +--- a/arch/x86/kernel/cpu/mcheck/p5.c ++++ b/arch/x86/kernel/cpu/mcheck/p5.c +@@ -12,6 +12,7 @@ + #include + #include + #include ++#include + + /* By default disabled */ + int mce_p5_enabled __read_mostly; +@@ -50,7 +51,9 @@ void intel_p5_mcheck_init(struct cpuinfo_x86 *c) + if (!cpu_has(c, X86_FEATURE_MCE)) + return; + ++ pax_open_kernel(); + machine_check_vector = pentium_machine_check; ++ pax_close_kernel(); + /* Make sure the vector pointer is visible before we enable MCEs: */ + wmb(); + +diff --git a/arch/x86/kernel/cpu/mcheck/winchip.c b/arch/x86/kernel/cpu/mcheck/winchip.c +index 54060f5..c1a7577 100644 +--- a/arch/x86/kernel/cpu/mcheck/winchip.c ++++ b/arch/x86/kernel/cpu/mcheck/winchip.c +@@ -11,6 +11,7 @@ + #include + #include + #include ++#include + + /* Machine check handler for WinChip C6: */ + static void winchip_machine_check(struct pt_regs *regs, long error_code) +@@ -24,7 +25,9 @@ void winchip_mcheck_init(struct cpuinfo_x86 *c) + { + u32 lo, hi; + ++ pax_open_kernel(); + machine_check_vector = winchip_machine_check; ++ pax_close_kernel(); + /* Make sure the vector pointer is visible before we enable MCEs: */ + wmb(); + +diff --git a/arch/x86/kernel/cpu/mtrr/amd.c b/arch/x86/kernel/cpu/mtrr/amd.c +index 33af141..92ba9cd 100644 +--- a/arch/x86/kernel/cpu/mtrr/amd.c ++++ b/arch/x86/kernel/cpu/mtrr/amd.c +@@ -108,7 +108,7 @@ amd_validate_add_page(unsigned long base, unsigned long size, unsigned int type) + return 0; + } + +-static struct mtrr_ops amd_mtrr_ops = { ++static const struct mtrr_ops amd_mtrr_ops = { + .vendor = X86_VENDOR_AMD, + .set = amd_set_mtrr, + .get = amd_get_mtrr, +diff --git a/arch/x86/kernel/cpu/mtrr/centaur.c b/arch/x86/kernel/cpu/mtrr/centaur.c +index de89f14..316fe3e 100644 +--- a/arch/x86/kernel/cpu/mtrr/centaur.c ++++ b/arch/x86/kernel/cpu/mtrr/centaur.c +@@ -110,7 +110,7 @@ centaur_validate_add_page(unsigned long base, unsigned long size, unsigned int t + return 0; + } + +-static struct mtrr_ops centaur_mtrr_ops = { ++static const struct mtrr_ops centaur_mtrr_ops = { + .vendor = X86_VENDOR_CENTAUR, + .set = centaur_set_mcr, + .get = centaur_get_mcr, +diff --git a/arch/x86/kernel/cpu/mtrr/cyrix.c b/arch/x86/kernel/cpu/mtrr/cyrix.c +index 228d982..68a3343 100644 +--- a/arch/x86/kernel/cpu/mtrr/cyrix.c ++++ b/arch/x86/kernel/cpu/mtrr/cyrix.c +@@ -265,7 +265,7 @@ static void cyrix_set_all(void) + post_set(); + } + +-static struct mtrr_ops cyrix_mtrr_ops = { ++static const struct mtrr_ops cyrix_mtrr_ops = { + .vendor = X86_VENDOR_CYRIX, + .set_all = cyrix_set_all, + .set = cyrix_set_arr, +diff --git a/arch/x86/kernel/cpu/mtrr/generic.c b/arch/x86/kernel/cpu/mtrr/generic.c +index 55da0c5..4d75584 100644 +--- a/arch/x86/kernel/cpu/mtrr/generic.c ++++ b/arch/x86/kernel/cpu/mtrr/generic.c +@@ -752,7 +752,7 @@ int positive_have_wrcomb(void) + /* + * Generic structure... + */ +-struct mtrr_ops generic_mtrr_ops = { ++const struct mtrr_ops generic_mtrr_ops = { + .use_intel_if = 1, + .set_all = generic_set_all, + .get = generic_get_mtrr, +diff --git a/arch/x86/kernel/cpu/mtrr/main.c b/arch/x86/kernel/cpu/mtrr/main.c +index fd60f09..c94ef52 100644 +--- a/arch/x86/kernel/cpu/mtrr/main.c ++++ b/arch/x86/kernel/cpu/mtrr/main.c +@@ -60,14 +60,14 @@ static DEFINE_MUTEX(mtrr_mutex); + u64 size_or_mask, size_and_mask; + static bool mtrr_aps_delayed_init; + +-static struct mtrr_ops *mtrr_ops[X86_VENDOR_NUM]; ++static const struct mtrr_ops *mtrr_ops[X86_VENDOR_NUM] __read_only; + +-struct mtrr_ops *mtrr_if; ++const struct mtrr_ops *mtrr_if; + + static void set_mtrr(unsigned int reg, unsigned long base, + unsigned long size, mtrr_type type); + +-void set_mtrr_ops(struct mtrr_ops *ops) ++void set_mtrr_ops(const struct mtrr_ops *ops) + { + if (ops->vendor && ops->vendor < X86_VENDOR_NUM) + mtrr_ops[ops->vendor] = ops; +diff --git a/arch/x86/kernel/cpu/mtrr/mtrr.h b/arch/x86/kernel/cpu/mtrr/mtrr.h +index a501dee..816c719 100644 +--- a/arch/x86/kernel/cpu/mtrr/mtrr.h ++++ b/arch/x86/kernel/cpu/mtrr/mtrr.h +@@ -25,14 +25,14 @@ struct mtrr_ops { + int (*validate_add_page)(unsigned long base, unsigned long size, + unsigned int type); + int (*have_wrcomb)(void); +-}; ++} __do_const; + + extern int generic_get_free_region(unsigned long base, unsigned long size, + int replace_reg); + extern int generic_validate_add_page(unsigned long base, unsigned long size, + unsigned int type); + +-extern struct mtrr_ops generic_mtrr_ops; ++extern const struct mtrr_ops generic_mtrr_ops; + + extern int positive_have_wrcomb(void); + +@@ -53,10 +53,10 @@ void fill_mtrr_var_range(unsigned int index, + u32 base_lo, u32 base_hi, u32 mask_lo, u32 mask_hi); + void get_mtrr_state(void); + +-extern void set_mtrr_ops(struct mtrr_ops *ops); ++extern void set_mtrr_ops(const struct mtrr_ops *ops); + + extern u64 size_or_mask, size_and_mask; +-extern struct mtrr_ops *mtrr_if; ++extern const struct mtrr_ops *mtrr_if; + + #define is_cpu(vnd) (mtrr_if && mtrr_if->vendor == X86_VENDOR_##vnd) + #define use_intel() (mtrr_if && mtrr_if->use_intel_if == 1) +diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c +index 0ff02ca..fc49a60 100644 +--- a/arch/x86/kernel/cpu/perf_event.c ++++ b/arch/x86/kernel/cpu/perf_event.c +@@ -723,10 +723,10 @@ x86_perf_event_update(struct perf_event *event, + * count to the generic event atomically: + */ + again: +- prev_raw_count = atomic64_read(&hwc->prev_count); ++ prev_raw_count = atomic64_read_unchecked(&hwc->prev_count); + rdmsrl(hwc->event_base + idx, new_raw_count); + +- if (atomic64_cmpxchg(&hwc->prev_count, prev_raw_count, ++ if (atomic64_cmpxchg_unchecked(&hwc->prev_count, prev_raw_count, + new_raw_count) != prev_raw_count) + goto again; + +@@ -741,7 +741,7 @@ again: + delta = (new_raw_count << shift) - (prev_raw_count << shift); + delta >>= shift; + +- atomic64_add(delta, &event->count); ++ atomic64_add_unchecked(delta, &event->count); + atomic64_sub(delta, &hwc->period_left); + + return new_raw_count; +@@ -1353,7 +1353,7 @@ x86_perf_event_set_period(struct perf_event *event, + * The hw event starts counting from this event offset, + * mark it to be able to extra future deltas: + */ +- atomic64_set(&hwc->prev_count, (u64)-left); ++ atomic64_set_unchecked(&hwc->prev_count, (u64)-left); + + err = checking_wrmsrl(hwc->event_base + idx, + (u64)(-left) & x86_pmu.event_mask); +@@ -2357,7 +2357,7 @@ perf_callchain_user(struct pt_regs *regs, struct perf_callchain_entry *entry) + break; + + callchain_store(entry, frame.return_address); +- fp = frame.next_frame; ++ fp = (__force const void __user *)frame.next_frame; + } + } + +diff --git a/arch/x86/kernel/cpu/perfctr-watchdog.c b/arch/x86/kernel/cpu/perfctr-watchdog.c +index 898df97..9e82503 100644 +--- a/arch/x86/kernel/cpu/perfctr-watchdog.c ++++ b/arch/x86/kernel/cpu/perfctr-watchdog.c +@@ -30,11 +30,11 @@ struct nmi_watchdog_ctlblk { + + /* Interface defining a CPU specific perfctr watchdog */ + struct wd_ops { +- int (*reserve)(void); +- void (*unreserve)(void); +- int (*setup)(unsigned nmi_hz); +- void (*rearm)(struct nmi_watchdog_ctlblk *wd, unsigned nmi_hz); +- void (*stop)(void); ++ int (* const reserve)(void); ++ void (* const unreserve)(void); ++ int (* const setup)(unsigned nmi_hz); ++ void (* const rearm)(struct nmi_watchdog_ctlblk *wd, unsigned nmi_hz); ++ void (* const stop)(void); + unsigned perfctr; + unsigned evntsel; + u64 checkbit; +@@ -645,6 +645,7 @@ static const struct wd_ops p4_wd_ops = { + #define ARCH_PERFMON_NMI_EVENT_SEL ARCH_PERFMON_UNHALTED_CORE_CYCLES_SEL + #define ARCH_PERFMON_NMI_EVENT_UMASK ARCH_PERFMON_UNHALTED_CORE_CYCLES_UMASK + ++/* cannot be const */ + static struct wd_ops intel_arch_wd_ops; + + static int setup_intel_arch_watchdog(unsigned nmi_hz) +@@ -697,6 +698,7 @@ static int setup_intel_arch_watchdog(unsigned nmi_hz) + return 1; + } + ++/* cannot be const */ + static struct wd_ops intel_arch_wd_ops __read_mostly = { + .reserve = single_msr_reserve, + .unreserve = single_msr_unreserve, +diff --git a/arch/x86/kernel/crash.c b/arch/x86/kernel/crash.c +index ff95824..2ffdcb5 100644 +--- a/arch/x86/kernel/crash.c ++++ b/arch/x86/kernel/crash.c +@@ -41,7 +41,7 @@ static void kdump_nmi_callback(int cpu, struct die_args *args) + regs = args->regs; + + #ifdef CONFIG_X86_32 +- if (!user_mode_vm(regs)) { ++ if (!user_mode(regs)) { + crash_fixup_ss_esp(&fixed_regs, regs); + regs = &fixed_regs; + } +diff --git a/arch/x86/kernel/doublefault_32.c b/arch/x86/kernel/doublefault_32.c +index 37250fe..bf2ec74 100644 +--- a/arch/x86/kernel/doublefault_32.c ++++ b/arch/x86/kernel/doublefault_32.c +@@ -11,7 +11,7 @@ + + #define DOUBLEFAULT_STACKSIZE (1024) + static unsigned long doublefault_stack[DOUBLEFAULT_STACKSIZE]; +-#define STACK_START (unsigned long)(doublefault_stack+DOUBLEFAULT_STACKSIZE) ++#define STACK_START (unsigned long)(doublefault_stack+DOUBLEFAULT_STACKSIZE-2) + + #define ptr_ok(x) ((x) > PAGE_OFFSET && (x) < PAGE_OFFSET + MAXMEM) + +@@ -21,7 +21,7 @@ static void doublefault_fn(void) + unsigned long gdt, tss; + + store_gdt(&gdt_desc); +- gdt = gdt_desc.address; ++ gdt = (unsigned long)gdt_desc.address; + + printk(KERN_EMERG "PANIC: double fault, gdt at %08lx [%d bytes]\n", gdt, gdt_desc.size); + +@@ -58,10 +58,10 @@ struct tss_struct doublefault_tss __cacheline_aligned = { + /* 0x2 bit is always set */ + .flags = X86_EFLAGS_SF | 0x2, + .sp = STACK_START, +- .es = __USER_DS, ++ .es = __KERNEL_DS, + .cs = __KERNEL_CS, + .ss = __KERNEL_DS, +- .ds = __USER_DS, ++ .ds = __KERNEL_DS, + .fs = __KERNEL_PERCPU, + + .__cr3 = __pa_nodebug(swapper_pg_dir), +diff --git a/arch/x86/kernel/dumpstack.c b/arch/x86/kernel/dumpstack.c +index 2d8a371..4fa6ae6 100644 +--- a/arch/x86/kernel/dumpstack.c ++++ b/arch/x86/kernel/dumpstack.c +@@ -2,6 +2,9 @@ + * Copyright (C) 1991, 1992 Linus Torvalds + * Copyright (C) 2000, 2001, 2002 Andi Kleen, SuSE Labs + */ ++#ifdef CONFIG_GRKERNSEC_HIDESYM ++#define __INCLUDED_BY_HIDESYM 1 ++#endif + #include + #include + #include +@@ -28,7 +31,7 @@ static int die_counter; + + void printk_address(unsigned long address, int reliable) + { +- printk(" [<%p>] %s%pS\n", (void *) address, ++ printk(" [<%p>] %s%pA\n", (void *) address, + reliable ? "" : "? ", (void *) address); + } + +@@ -36,9 +39,8 @@ void printk_address(unsigned long address, int reliable) + static void + print_ftrace_graph_addr(unsigned long addr, void *data, + const struct stacktrace_ops *ops, +- struct thread_info *tinfo, int *graph) ++ struct task_struct *task, int *graph) + { +- struct task_struct *task = tinfo->task; + unsigned long ret_addr; + int index = task->curr_ret_stack; + +@@ -59,7 +61,7 @@ print_ftrace_graph_addr(unsigned long addr, void *data, + static inline void + print_ftrace_graph_addr(unsigned long addr, void *data, + const struct stacktrace_ops *ops, +- struct thread_info *tinfo, int *graph) ++ struct task_struct *task, int *graph) + { } + #endif + +@@ -70,10 +72,8 @@ print_ftrace_graph_addr(unsigned long addr, void *data, + * severe exception (double fault, nmi, stack fault, debug, mce) hardware stack + */ + +-static inline int valid_stack_ptr(struct thread_info *tinfo, +- void *p, unsigned int size, void *end) ++static inline int valid_stack_ptr(void *t, void *p, unsigned int size, void *end) + { +- void *t = tinfo; + if (end) { + if (p < end && p >= (end-THREAD_SIZE)) + return 1; +@@ -84,14 +84,14 @@ static inline int valid_stack_ptr(struct thread_info *tinfo, + } + + unsigned long +-print_context_stack(struct thread_info *tinfo, ++print_context_stack(struct task_struct *task, void *stack_start, + unsigned long *stack, unsigned long bp, + const struct stacktrace_ops *ops, void *data, + unsigned long *end, int *graph) + { + struct stack_frame *frame = (struct stack_frame *)bp; + +- while (valid_stack_ptr(tinfo, stack, sizeof(*stack), end)) { ++ while (valid_stack_ptr(stack_start, stack, sizeof(*stack), end)) { + unsigned long addr; + + addr = *stack; +@@ -103,7 +103,7 @@ print_context_stack(struct thread_info *tinfo, + } else { + ops->address(data, addr, 0); + } +- print_ftrace_graph_addr(addr, data, ops, tinfo, graph); ++ print_ftrace_graph_addr(addr, data, ops, task, graph); + } + stack++; + } +@@ -180,7 +180,7 @@ void dump_stack(void) + #endif + + printk("Pid: %d, comm: %.20s %s %s %.*s\n", +- current->pid, current->comm, print_tainted(), ++ task_pid_nr(current), current->comm, print_tainted(), + init_utsname()->release, + (int)strcspn(init_utsname()->version, " "), + init_utsname()->version); +@@ -220,6 +220,8 @@ unsigned __kprobes long oops_begin(void) + return flags; + } + ++extern void gr_handle_kernel_exploit(void); ++ + void __kprobes oops_end(unsigned long flags, struct pt_regs *regs, int signr) + { + if (regs && kexec_should_crash(current)) +@@ -241,7 +243,10 @@ void __kprobes oops_end(unsigned long flags, struct pt_regs *regs, int signr) + panic("Fatal exception in interrupt"); + if (panic_on_oops) + panic("Fatal exception"); +- do_exit(signr); ++ ++ gr_handle_kernel_exploit(); ++ ++ do_group_exit(signr); + } + + int __kprobes __die(const char *str, struct pt_regs *regs, long err) +@@ -295,7 +300,7 @@ void die(const char *str, struct pt_regs *regs, long err) + unsigned long flags = oops_begin(); + int sig = SIGSEGV; + +- if (!user_mode_vm(regs)) ++ if (!user_mode(regs)) + report_bug(regs->ip, regs); + + if (__die(str, regs, err)) +diff --git a/arch/x86/kernel/dumpstack.h b/arch/x86/kernel/dumpstack.h +index 81086c2..13e8b17 100644 +--- a/arch/x86/kernel/dumpstack.h ++++ b/arch/x86/kernel/dumpstack.h +@@ -15,7 +15,7 @@ + #endif + + extern unsigned long +-print_context_stack(struct thread_info *tinfo, ++print_context_stack(struct task_struct *task, void *stack_start, + unsigned long *stack, unsigned long bp, + const struct stacktrace_ops *ops, void *data, + unsigned long *end, int *graph); +diff --git a/arch/x86/kernel/dumpstack_32.c b/arch/x86/kernel/dumpstack_32.c +index f7dd2a7..504f53b 100644 +--- a/arch/x86/kernel/dumpstack_32.c ++++ b/arch/x86/kernel/dumpstack_32.c +@@ -53,16 +53,12 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + #endif + + for (;;) { +- struct thread_info *context; ++ void *stack_start = (void *)((unsigned long)stack & ~(THREAD_SIZE-1)); ++ bp = print_context_stack(task, stack_start, stack, bp, ops, data, NULL, &graph); + +- context = (struct thread_info *) +- ((unsigned long)stack & (~(THREAD_SIZE - 1))); +- bp = print_context_stack(context, stack, bp, ops, +- data, NULL, &graph); +- +- stack = (unsigned long *)context->previous_esp; +- if (!stack) ++ if (stack_start == task_stack_page(task)) + break; ++ stack = *(unsigned long **)stack_start; + if (ops->stack(data, "IRQ") < 0) + break; + touch_nmi_watchdog(); +@@ -112,11 +108,12 @@ void show_registers(struct pt_regs *regs) + * When in-kernel, we also print out the stack and code at the + * time of the fault.. + */ +- if (!user_mode_vm(regs)) { ++ if (!user_mode(regs)) { + unsigned int code_prologue = code_bytes * 43 / 64; + unsigned int code_len = code_bytes; + unsigned char c; + u8 *ip; ++ unsigned long cs_base = get_desc_base(&get_cpu_gdt_table(smp_processor_id())[(0xffff & regs->cs) >> 3]); + + printk(KERN_EMERG "Stack:\n"); + show_stack_log_lvl(NULL, regs, ®s->sp, +@@ -124,10 +121,10 @@ void show_registers(struct pt_regs *regs) + + printk(KERN_EMERG "Code: "); + +- ip = (u8 *)regs->ip - code_prologue; ++ ip = (u8 *)regs->ip - code_prologue + cs_base; + if (ip < (u8 *)PAGE_OFFSET || probe_kernel_address(ip, c)) { + /* try starting at IP */ +- ip = (u8 *)regs->ip; ++ ip = (u8 *)regs->ip + cs_base; + code_len = code_len - code_prologue + 1; + } + for (i = 0; i < code_len; i++, ip++) { +@@ -136,7 +133,7 @@ void show_registers(struct pt_regs *regs) + printk(" Bad EIP value."); + break; + } +- if (ip == (u8 *)regs->ip) ++ if (ip == (u8 *)regs->ip + cs_base) + printk("<%02x> ", c); + else + printk("%02x ", c); +@@ -145,10 +142,23 @@ void show_registers(struct pt_regs *regs) + printk("\n"); + } + ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++void pax_check_alloca(unsigned long size) ++{ ++ unsigned long sp = (unsigned long)&sp, stack_left; ++ ++ /* all kernel stacks are of the same size */ ++ stack_left = sp & (THREAD_SIZE - 1); ++ BUG_ON(stack_left < 256 || size >= stack_left - 256); ++} ++EXPORT_SYMBOL(pax_check_alloca); ++#endif ++ + int is_valid_bugaddr(unsigned long ip) + { + unsigned short ud2; + ++ ip = ktla_ktva(ip); + if (ip < PAGE_OFFSET) + return 0; + if (probe_kernel_address((unsigned short *)ip, ud2)) +diff --git a/arch/x86/kernel/dumpstack_64.c b/arch/x86/kernel/dumpstack_64.c +index a071e6b..36cd585 100644 +--- a/arch/x86/kernel/dumpstack_64.c ++++ b/arch/x86/kernel/dumpstack_64.c +@@ -116,8 +116,8 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + unsigned long *irq_stack_end = + (unsigned long *)per_cpu(irq_stack_ptr, cpu); + unsigned used = 0; +- struct thread_info *tinfo; + int graph = 0; ++ void *stack_start; + + if (!task) + task = current; +@@ -146,10 +146,10 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + * current stack address. If the stacks consist of nested + * exceptions + */ +- tinfo = task_thread_info(task); + for (;;) { + char *id; + unsigned long *estack_end; ++ + estack_end = in_exception_stack(cpu, (unsigned long)stack, + &used, &id); + +@@ -157,7 +157,7 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + if (ops->stack(data, id) < 0) + break; + +- bp = print_context_stack(tinfo, stack, bp, ops, ++ bp = print_context_stack(task, estack_end - EXCEPTION_STKSZ, stack, bp, ops, + data, estack_end, &graph); + ops->stack(data, ""); + /* +@@ -176,7 +176,7 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + if (stack >= irq_stack && stack < irq_stack_end) { + if (ops->stack(data, "IRQ") < 0) + break; +- bp = print_context_stack(tinfo, stack, bp, ++ bp = print_context_stack(task, irq_stack, stack, bp, + ops, data, irq_stack_end, &graph); + /* + * We link to the next stack (which would be +@@ -195,7 +195,8 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, + /* + * This handles the process stack: + */ +- bp = print_context_stack(tinfo, stack, bp, ops, data, NULL, &graph); ++ stack_start = (void *)((unsigned long)stack & ~(THREAD_SIZE-1)); ++ bp = print_context_stack(task, stack_start, stack, bp, ops, data, NULL, &graph); + put_cpu(); + } + EXPORT_SYMBOL(dump_trace); +@@ -304,3 +305,50 @@ int is_valid_bugaddr(unsigned long ip) + return ud2 == 0x0b0f; + } + ++ ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++void pax_check_alloca(unsigned long size) ++{ ++ unsigned long sp = (unsigned long)&sp, stack_start, stack_end; ++ unsigned cpu, used; ++ char *id; ++ ++ /* check the process stack first */ ++ stack_start = (unsigned long)task_stack_page(current); ++ stack_end = stack_start + THREAD_SIZE; ++ if (likely(stack_start <= sp && sp < stack_end)) { ++ unsigned long stack_left = sp & (THREAD_SIZE - 1); ++ BUG_ON(stack_left < 256 || size >= stack_left - 256); ++ return; ++ } ++ ++ cpu = get_cpu(); ++ ++ /* check the irq stacks */ ++ stack_end = (unsigned long)per_cpu(irq_stack_ptr, cpu); ++ stack_start = stack_end - IRQ_STACK_SIZE; ++ if (stack_start <= sp && sp < stack_end) { ++ unsigned long stack_left = sp & (IRQ_STACK_SIZE - 1); ++ put_cpu(); ++ BUG_ON(stack_left < 256 || size >= stack_left - 256); ++ return; ++ } ++ ++ /* check the exception stacks */ ++ used = 0; ++ stack_end = (unsigned long)in_exception_stack(cpu, sp, &used, &id); ++ stack_start = stack_end - EXCEPTION_STKSZ; ++ if (stack_end && stack_start <= sp && sp < stack_end) { ++ unsigned long stack_left = sp & (EXCEPTION_STKSZ - 1); ++ put_cpu(); ++ BUG_ON(stack_left < 256 || size >= stack_left - 256); ++ return; ++ } ++ ++ put_cpu(); ++ ++ /* unknown stack */ ++ BUG(); ++} ++EXPORT_SYMBOL(pax_check_alloca); ++#endif +diff --git a/arch/x86/kernel/e820.c b/arch/x86/kernel/e820.c +index a89739a..95e0c48 100644 +--- a/arch/x86/kernel/e820.c ++++ b/arch/x86/kernel/e820.c +@@ -733,7 +733,7 @@ struct early_res { + }; + static struct early_res early_res[MAX_EARLY_RES] __initdata = { + { 0, PAGE_SIZE, "BIOS data page" }, /* BIOS data page */ +- {} ++ { 0, 0, {0}, 0 } + }; + + static int __init find_overlapped_early(u64 start, u64 end) +diff --git a/arch/x86/kernel/early_printk.c b/arch/x86/kernel/early_printk.c +index b9c830c..1e41a96 100644 +--- a/arch/x86/kernel/early_printk.c ++++ b/arch/x86/kernel/early_printk.c +@@ -7,6 +7,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -170,6 +171,8 @@ asmlinkage void early_printk(const char *fmt, ...) + int n; + va_list ap; + ++ pax_track_stack(); ++ + va_start(ap, fmt); + n = vscnprintf(buf, sizeof(buf), fmt, ap); + early_console->write(early_console, buf, n); +diff --git a/arch/x86/kernel/efi_32.c b/arch/x86/kernel/efi_32.c +index 5cab48e..b025f9b 100644 +--- a/arch/x86/kernel/efi_32.c ++++ b/arch/x86/kernel/efi_32.c +@@ -38,70 +38,56 @@ + */ + + static unsigned long efi_rt_eflags; +-static pgd_t efi_bak_pg_dir_pointer[2]; ++static pgd_t __initdata efi_bak_pg_dir_pointer[KERNEL_PGD_PTRS]; + +-void efi_call_phys_prelog(void) ++void __init efi_call_phys_prelog(void) + { +- unsigned long cr4; +- unsigned long temp; + struct desc_ptr gdt_descr; + ++#ifdef CONFIG_PAX_KERNEXEC ++ struct desc_struct d; ++#endif ++ + local_irq_save(efi_rt_eflags); + +- /* +- * If I don't have PAE, I should just duplicate two entries in page +- * directory. If I have PAE, I just need to duplicate one entry in +- * page directory. +- */ +- cr4 = read_cr4_safe(); +- +- if (cr4 & X86_CR4_PAE) { +- efi_bak_pg_dir_pointer[0].pgd = +- swapper_pg_dir[pgd_index(0)].pgd; +- swapper_pg_dir[0].pgd = +- swapper_pg_dir[pgd_index(PAGE_OFFSET)].pgd; +- } else { +- efi_bak_pg_dir_pointer[0].pgd = +- swapper_pg_dir[pgd_index(0)].pgd; +- efi_bak_pg_dir_pointer[1].pgd = +- swapper_pg_dir[pgd_index(0x400000)].pgd; +- swapper_pg_dir[pgd_index(0)].pgd = +- swapper_pg_dir[pgd_index(PAGE_OFFSET)].pgd; +- temp = PAGE_OFFSET + 0x400000; +- swapper_pg_dir[pgd_index(0x400000)].pgd = +- swapper_pg_dir[pgd_index(temp)].pgd; +- } ++ clone_pgd_range(efi_bak_pg_dir_pointer, swapper_pg_dir, KERNEL_PGD_PTRS); ++ clone_pgd_range(swapper_pg_dir, swapper_pg_dir + KERNEL_PGD_BOUNDARY, ++ min_t(unsigned long, KERNEL_PGD_PTRS, KERNEL_PGD_BOUNDARY)); + + /* + * After the lock is released, the original page table is restored. + */ + __flush_tlb_all(); + ++#ifdef CONFIG_PAX_KERNEXEC ++ pack_descriptor(&d, 0, 0xFFFFF, 0x9B, 0xC); ++ write_gdt_entry(get_cpu_gdt_table(0), GDT_ENTRY_KERNEXEC_EFI_CS, &d, DESCTYPE_S); ++ pack_descriptor(&d, 0, 0xFFFFF, 0x93, 0xC); ++ write_gdt_entry(get_cpu_gdt_table(0), GDT_ENTRY_KERNEXEC_EFI_DS, &d, DESCTYPE_S); ++#endif ++ + gdt_descr.address = __pa(get_cpu_gdt_table(0)); + gdt_descr.size = GDT_SIZE - 1; + load_gdt(&gdt_descr); + } + +-void efi_call_phys_epilog(void) ++void __init efi_call_phys_epilog(void) + { +- unsigned long cr4; + struct desc_ptr gdt_descr; + ++#ifdef CONFIG_PAX_KERNEXEC ++ struct desc_struct d; ++ ++ memset(&d, 0, sizeof d); ++ write_gdt_entry(get_cpu_gdt_table(0), GDT_ENTRY_KERNEXEC_EFI_CS, &d, DESCTYPE_S); ++ write_gdt_entry(get_cpu_gdt_table(0), GDT_ENTRY_KERNEXEC_EFI_DS, &d, DESCTYPE_S); ++#endif ++ + gdt_descr.address = (unsigned long)get_cpu_gdt_table(0); + gdt_descr.size = GDT_SIZE - 1; + load_gdt(&gdt_descr); + +- cr4 = read_cr4_safe(); +- +- if (cr4 & X86_CR4_PAE) { +- swapper_pg_dir[pgd_index(0)].pgd = +- efi_bak_pg_dir_pointer[0].pgd; +- } else { +- swapper_pg_dir[pgd_index(0)].pgd = +- efi_bak_pg_dir_pointer[0].pgd; +- swapper_pg_dir[pgd_index(0x400000)].pgd = +- efi_bak_pg_dir_pointer[1].pgd; +- } ++ clone_pgd_range(swapper_pg_dir, efi_bak_pg_dir_pointer, KERNEL_PGD_PTRS); + + /* + * After the lock is released, the original page table is restored. +diff --git a/arch/x86/kernel/efi_stub_32.S b/arch/x86/kernel/efi_stub_32.S +index fbe66e6..c5c0dd2 100644 +--- a/arch/x86/kernel/efi_stub_32.S ++++ b/arch/x86/kernel/efi_stub_32.S +@@ -6,7 +6,9 @@ + */ + + #include ++#include + #include ++#include + + /* + * efi_call_phys(void *, ...) is a function with variable parameters. +@@ -20,7 +22,7 @@ + * service functions will comply with gcc calling convention, too. + */ + +-.text ++__INIT + ENTRY(efi_call_phys) + /* + * 0. The function can only be called in Linux kernel. So CS has been +@@ -36,9 +38,11 @@ ENTRY(efi_call_phys) + * The mapping of lower virtual memory has been created in prelog and + * epilog. + */ +- movl $1f, %edx +- subl $__PAGE_OFFSET, %edx +- jmp *%edx ++ movl $(__KERNEXEC_EFI_DS), %edx ++ mov %edx, %ds ++ mov %edx, %es ++ mov %edx, %ss ++ ljmp $(__KERNEXEC_EFI_CS),$1f-__PAGE_OFFSET + 1: + + /* +@@ -47,14 +51,8 @@ ENTRY(efi_call_phys) + * parameter 2, ..., param n. To make things easy, we save the return + * address of efi_call_phys in a global variable. + */ +- popl %edx +- movl %edx, saved_return_addr +- /* get the function pointer into ECX*/ +- popl %ecx +- movl %ecx, efi_rt_function_ptr +- movl $2f, %edx +- subl $__PAGE_OFFSET, %edx +- pushl %edx ++ popl (saved_return_addr) ++ popl (efi_rt_function_ptr) + + /* + * 3. Clear PG bit in %CR0. +@@ -73,9 +71,8 @@ ENTRY(efi_call_phys) + /* + * 5. Call the physical function. + */ +- jmp *%ecx ++ call *(efi_rt_function_ptr-__PAGE_OFFSET) + +-2: + /* + * 6. After EFI runtime service returns, control will return to + * following instruction. We'd better readjust stack pointer first. +@@ -88,35 +85,32 @@ ENTRY(efi_call_phys) + movl %cr0, %edx + orl $0x80000000, %edx + movl %edx, %cr0 +- jmp 1f +-1: ++ + /* + * 8. Now restore the virtual mode from flat mode by + * adding EIP with PAGE_OFFSET. + */ +- movl $1f, %edx +- jmp *%edx ++ ljmp $(__KERNEL_CS),$1f+__PAGE_OFFSET + 1: ++ movl $(__KERNEL_DS), %edx ++ mov %edx, %ds ++ mov %edx, %es ++ mov %edx, %ss + + /* + * 9. Balance the stack. And because EAX contain the return value, + * we'd better not clobber it. + */ +- leal efi_rt_function_ptr, %edx +- movl (%edx), %ecx +- pushl %ecx ++ pushl (efi_rt_function_ptr) + + /* +- * 10. Push the saved return address onto the stack and return. ++ * 10. Return to the saved return address. + */ +- leal saved_return_addr, %edx +- movl (%edx), %ecx +- pushl %ecx +- ret ++ jmpl *(saved_return_addr) + ENDPROC(efi_call_phys) + .previous + +-.data ++__INITDATA + saved_return_addr: + .long 0 + efi_rt_function_ptr: +diff --git a/arch/x86/kernel/efi_stub_64.S b/arch/x86/kernel/efi_stub_64.S +index 4c07cca..2c8427d 100644 +--- a/arch/x86/kernel/efi_stub_64.S ++++ b/arch/x86/kernel/efi_stub_64.S +@@ -7,6 +7,7 @@ + */ + + #include ++#include + + #define SAVE_XMM \ + mov %rsp, %rax; \ +@@ -40,6 +41,7 @@ ENTRY(efi_call0) + call *%rdi + addq $32, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call0) + +@@ -50,6 +52,7 @@ ENTRY(efi_call1) + call *%rdi + addq $32, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call1) + +@@ -60,6 +63,7 @@ ENTRY(efi_call2) + call *%rdi + addq $32, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call2) + +@@ -71,6 +75,7 @@ ENTRY(efi_call3) + call *%rdi + addq $32, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call3) + +@@ -83,6 +88,7 @@ ENTRY(efi_call4) + call *%rdi + addq $32, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call4) + +@@ -96,6 +102,7 @@ ENTRY(efi_call5) + call *%rdi + addq $48, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call5) + +@@ -112,5 +119,6 @@ ENTRY(efi_call6) + call *%rdi + addq $48, %rsp + RESTORE_XMM ++ pax_force_retaddr 0, 1 + ret + ENDPROC(efi_call6) +diff --git a/arch/x86/kernel/entry_32.S b/arch/x86/kernel/entry_32.S +index c097e7d..c689cf4 100644 +--- a/arch/x86/kernel/entry_32.S ++++ b/arch/x86/kernel/entry_32.S +@@ -185,13 +185,146 @@ + /*CFI_REL_OFFSET gs, PT_GS*/ + .endm + .macro SET_KERNEL_GS reg ++ ++#ifdef CONFIG_CC_STACKPROTECTOR + movl $(__KERNEL_STACK_CANARY), \reg ++#elif defined(CONFIG_PAX_MEMORY_UDEREF) ++ movl $(__USER_DS), \reg ++#else ++ xorl \reg, \reg ++#endif ++ + movl \reg, %gs + .endm + + #endif /* CONFIG_X86_32_LAZY_GS */ + +-.macro SAVE_ALL ++.macro pax_enter_kernel ++#ifdef CONFIG_PAX_KERNEXEC ++ call pax_enter_kernel ++#endif ++.endm ++ ++.macro pax_exit_kernel ++#ifdef CONFIG_PAX_KERNEXEC ++ call pax_exit_kernel ++#endif ++.endm ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ENTRY(pax_enter_kernel) ++#ifdef CONFIG_PARAVIRT ++ pushl %eax ++ pushl %ecx ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0) ++ mov %eax, %esi ++#else ++ mov %cr0, %esi ++#endif ++ bts $16, %esi ++ jnc 1f ++ mov %cs, %esi ++ cmp $__KERNEL_CS, %esi ++ jz 3f ++ ljmp $__KERNEL_CS, $3f ++1: ljmp $__KERNEXEC_KERNEL_CS, $2f ++2: ++#ifdef CONFIG_PARAVIRT ++ mov %esi, %eax ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0) ++#else ++ mov %esi, %cr0 ++#endif ++3: ++#ifdef CONFIG_PARAVIRT ++ popl %ecx ++ popl %eax ++#endif ++ ret ++ENDPROC(pax_enter_kernel) ++ ++ENTRY(pax_exit_kernel) ++#ifdef CONFIG_PARAVIRT ++ pushl %eax ++ pushl %ecx ++#endif ++ mov %cs, %esi ++ cmp $__KERNEXEC_KERNEL_CS, %esi ++ jnz 2f ++#ifdef CONFIG_PARAVIRT ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0); ++ mov %eax, %esi ++#else ++ mov %cr0, %esi ++#endif ++ btr $16, %esi ++ ljmp $__KERNEL_CS, $1f ++1: ++#ifdef CONFIG_PARAVIRT ++ mov %esi, %eax ++ call PARA_INDIRECT(pv_cpu_ops+PV_CPU_write_cr0); ++#else ++ mov %esi, %cr0 ++#endif ++2: ++#ifdef CONFIG_PARAVIRT ++ popl %ecx ++ popl %eax ++#endif ++ ret ++ENDPROC(pax_exit_kernel) ++#endif ++ ++.macro pax_erase_kstack ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++ call pax_erase_kstack ++#endif ++.endm ++ ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++/* ++ * ebp: thread_info ++ * ecx, edx: can be clobbered ++ */ ++ENTRY(pax_erase_kstack) ++ pushl %edi ++ pushl %eax ++ ++ mov TI_lowest_stack(%ebp), %edi ++ mov $-0xBEEF, %eax ++ std ++ ++1: mov %edi, %ecx ++ and $THREAD_SIZE_asm - 1, %ecx ++ shr $2, %ecx ++ repne scasl ++ jecxz 2f ++ ++ cmp $2*16, %ecx ++ jc 2f ++ ++ mov $2*16, %ecx ++ repe scasl ++ jecxz 2f ++ jne 1b ++ ++2: cld ++ mov %esp, %ecx ++ sub %edi, %ecx ++ shr $2, %ecx ++ rep stosl ++ ++ mov TI_task_thread_sp0(%ebp), %edi ++ sub $128, %edi ++ mov %edi, TI_lowest_stack(%ebp) ++ ++ popl %eax ++ popl %edi ++ ret ++ENDPROC(pax_erase_kstack) ++#endif ++ ++.macro __SAVE_ALL _DS + cld + PUSH_GS + pushl %fs +@@ -224,7 +357,7 @@ + pushl %ebx + CFI_ADJUST_CFA_OFFSET 4 + CFI_REL_OFFSET ebx, 0 +- movl $(__USER_DS), %edx ++ movl $\_DS, %edx + movl %edx, %ds + movl %edx, %es + movl $(__KERNEL_PERCPU), %edx +@@ -232,6 +365,15 @@ + SET_KERNEL_GS %edx + .endm + ++.macro SAVE_ALL ++#if defined(CONFIG_PAX_KERNEXEC) || defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) || defined(CONFIG_PAX_MEMORY_UDEREF) ++ __SAVE_ALL __KERNEL_DS ++ pax_enter_kernel ++#else ++ __SAVE_ALL __USER_DS ++#endif ++.endm ++ + .macro RESTORE_INT_REGS + popl %ebx + CFI_ADJUST_CFA_OFFSET -4 +@@ -331,7 +473,7 @@ ENTRY(ret_from_fork) + CFI_ADJUST_CFA_OFFSET -4 + jmp syscall_exit + CFI_ENDPROC +-END(ret_from_fork) ++ENDPROC(ret_from_fork) + + /* + * Return to user mode is not as complex as all this looks, +@@ -352,7 +494,15 @@ check_userspace: + movb PT_CS(%esp), %al + andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %eax + cmpl $USER_RPL, %eax ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ jae resume_userspace ++ ++ PAX_EXIT_KERNEL ++ jmp resume_kernel ++#else + jb resume_kernel # not returning to v8086 or userspace ++#endif + + ENTRY(resume_userspace) + LOCKDEP_SYS_EXIT +@@ -364,8 +514,8 @@ ENTRY(resume_userspace) + andl $_TIF_WORK_MASK, %ecx # is there any work to be done on + # int/exception return? + jne work_pending +- jmp restore_all +-END(ret_from_exception) ++ jmp restore_all_pax ++ENDPROC(ret_from_exception) + + #ifdef CONFIG_PREEMPT + ENTRY(resume_kernel) +@@ -380,7 +530,7 @@ need_resched: + jz restore_all + call preempt_schedule_irq + jmp need_resched +-END(resume_kernel) ++ENDPROC(resume_kernel) + #endif + CFI_ENDPROC + +@@ -414,25 +564,36 @@ sysenter_past_esp: + /*CFI_REL_OFFSET cs, 0*/ + /* + * Push current_thread_info()->sysenter_return to the stack. +- * A tiny bit of offset fixup is necessary - 4*4 means the 4 words +- * pushed above; +8 corresponds to copy_thread's esp0 setting. + */ +- pushl (TI_sysenter_return-THREAD_SIZE+8+4*4)(%esp) ++ pushl $0 + CFI_ADJUST_CFA_OFFSET 4 + CFI_REL_OFFSET eip, 0 + + pushl %eax + CFI_ADJUST_CFA_OFFSET 4 + SAVE_ALL ++ GET_THREAD_INFO(%ebp) ++ movl TI_sysenter_return(%ebp),%ebp ++ movl %ebp,PT_EIP(%esp) + ENABLE_INTERRUPTS(CLBR_NONE) + + /* + * Load the potential sixth argument from user stack. + * Careful about security. + */ ++ movl PT_OLDESP(%esp),%ebp ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ mov PT_OLDSS(%esp),%ds ++1: movl %ds:(%ebp),%ebp ++ push %ss ++ pop %ds ++#else + cmpl $__PAGE_OFFSET-3,%ebp + jae syscall_fault + 1: movl (%ebp),%ebp ++#endif ++ + movl %ebp,PT_EBP(%esp) + .section __ex_table,"a" + .align 4 +@@ -455,12 +616,24 @@ sysenter_do_call: + testl $_TIF_ALLWORK_MASK, %ecx + jne sysexit_audit + sysenter_exit: ++ ++#ifdef CONFIG_PAX_RANDKSTACK ++ pushl_cfi %eax ++ movl %esp, %eax ++ call pax_randomize_kstack ++ popl_cfi %eax ++#endif ++ ++ pax_erase_kstack ++ + /* if something modifies registers it must also disable sysexit */ + movl PT_EIP(%esp), %edx + movl PT_OLDESP(%esp), %ecx + xorl %ebp,%ebp + TRACE_IRQS_ON + 1: mov PT_FS(%esp), %fs ++2: mov PT_DS(%esp), %ds ++3: mov PT_ES(%esp), %es + PTGS_TO_GS + ENABLE_INTERRUPTS_SYSEXIT + +@@ -477,6 +650,9 @@ sysenter_audit: + movl %eax,%edx /* 2nd arg: syscall number */ + movl $AUDIT_ARCH_I386,%eax /* 1st arg: audit arch */ + call audit_syscall_entry ++ ++ pax_erase_kstack ++ + pushl %ebx + CFI_ADJUST_CFA_OFFSET 4 + movl PT_EAX(%esp),%eax /* reload syscall number */ +@@ -504,11 +680,17 @@ sysexit_audit: + + CFI_ENDPROC + .pushsection .fixup,"ax" +-2: movl $0,PT_FS(%esp) ++4: movl $0,PT_FS(%esp) ++ jmp 1b ++5: movl $0,PT_DS(%esp) ++ jmp 1b ++6: movl $0,PT_ES(%esp) + jmp 1b + .section __ex_table,"a" + .align 4 +- .long 1b,2b ++ .long 1b,4b ++ .long 2b,5b ++ .long 3b,6b + .popsection + PTGS_TO_GS_EX + ENDPROC(ia32_sysenter_target) +@@ -538,6 +720,15 @@ syscall_exit: + testl $_TIF_ALLWORK_MASK, %ecx # current->work + jne syscall_exit_work + ++restore_all_pax: ++ ++#ifdef CONFIG_PAX_RANDKSTACK ++ movl %esp, %eax ++ call pax_randomize_kstack ++#endif ++ ++ pax_erase_kstack ++ + restore_all: + TRACE_IRQS_IRET + restore_all_notrace: +@@ -602,10 +793,29 @@ ldt_ss: + mov PT_OLDESP(%esp), %eax /* load userspace esp */ + mov %dx, %ax /* eax: new kernel esp */ + sub %eax, %edx /* offset (low word is 0) */ +- PER_CPU(gdt_page, %ebx) ++#ifdef CONFIG_SMP ++ movl PER_CPU_VAR(cpu_number), %ebx ++ shll $PAGE_SHIFT_asm, %ebx ++ addl $cpu_gdt_table, %ebx ++#else ++ movl $cpu_gdt_table, %ebx ++#endif + shr $16, %edx ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ mov %cr0, %esi ++ btr $16, %esi ++ mov %esi, %cr0 ++#endif ++ + mov %dl, GDT_ENTRY_ESPFIX_SS * 8 + 4(%ebx) /* bits 16..23 */ + mov %dh, GDT_ENTRY_ESPFIX_SS * 8 + 7(%ebx) /* bits 24..31 */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ bts $16, %esi ++ mov %esi, %cr0 ++#endif ++ + pushl $__ESPFIX_SS + CFI_ADJUST_CFA_OFFSET 4 + push %eax /* new kernel esp */ +@@ -636,36 +846,30 @@ work_resched: + movl TI_flags(%ebp), %ecx + andl $_TIF_WORK_MASK, %ecx # is there any work to be done other + # than syscall tracing? +- jz restore_all ++ jz restore_all_pax + testb $_TIF_NEED_RESCHED, %cl + jnz work_resched + + work_notifysig: # deal with pending signals and + # notify-resume requests ++ movl %esp, %eax + #ifdef CONFIG_VM86 + testl $X86_EFLAGS_VM, PT_EFLAGS(%esp) +- movl %esp, %eax +- jne work_notifysig_v86 # returning to kernel-space or ++ jz 1f # returning to kernel-space or + # vm86-space +- xorl %edx, %edx +- call do_notify_resume +- jmp resume_userspace_sig + +- ALIGN +-work_notifysig_v86: + pushl %ecx # save ti_flags for do_notify_resume + CFI_ADJUST_CFA_OFFSET 4 + call save_v86_state # %eax contains pt_regs pointer + popl %ecx + CFI_ADJUST_CFA_OFFSET -4 + movl %eax, %esp +-#else +- movl %esp, %eax ++1: + #endif + xorl %edx, %edx + call do_notify_resume + jmp resume_userspace_sig +-END(work_pending) ++ENDPROC(work_pending) + + # perform syscall exit tracing + ALIGN +@@ -673,11 +877,14 @@ syscall_trace_entry: + movl $-ENOSYS,PT_EAX(%esp) + movl %esp, %eax + call syscall_trace_enter ++ ++ pax_erase_kstack ++ + /* What it returned is what we'll actually use. */ + cmpl $(nr_syscalls), %eax + jnae syscall_call + jmp syscall_exit +-END(syscall_trace_entry) ++ENDPROC(syscall_trace_entry) + + # perform syscall exit tracing + ALIGN +@@ -690,20 +897,24 @@ syscall_exit_work: + movl %esp, %eax + call syscall_trace_leave + jmp resume_userspace +-END(syscall_exit_work) ++ENDPROC(syscall_exit_work) + CFI_ENDPROC + + RING0_INT_FRAME # can't unwind into user space anyway + syscall_fault: ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ push %ss ++ pop %ds ++#endif + GET_THREAD_INFO(%ebp) + movl $-EFAULT,PT_EAX(%esp) + jmp resume_userspace +-END(syscall_fault) ++ENDPROC(syscall_fault) + + syscall_badsys: + movl $-ENOSYS,PT_EAX(%esp) + jmp resume_userspace +-END(syscall_badsys) ++ENDPROC(syscall_badsys) + CFI_ENDPROC + + /* +@@ -726,6 +937,33 @@ PTREGSCALL(rt_sigreturn) + PTREGSCALL(vm86) + PTREGSCALL(vm86old) + ++ ALIGN; ++ENTRY(kernel_execve) ++ push %ebp ++ sub $PT_OLDSS+4,%esp ++ push %edi ++ push %ecx ++ push %eax ++ lea 3*4(%esp),%edi ++ mov $PT_OLDSS/4+1,%ecx ++ xorl %eax,%eax ++ rep stosl ++ pop %eax ++ pop %ecx ++ pop %edi ++ movl $X86_EFLAGS_IF,PT_EFLAGS(%esp) ++ mov %eax,PT_EBX(%esp) ++ mov %edx,PT_ECX(%esp) ++ mov %ecx,PT_EDX(%esp) ++ mov %esp,%eax ++ call sys_execve ++ GET_THREAD_INFO(%ebp) ++ test %eax,%eax ++ jz syscall_exit ++ add $PT_OLDSS+4,%esp ++ pop %ebp ++ ret ++ + .macro FIXUP_ESPFIX_STACK + /* + * Switch back for ESPFIX stack to the normal zerobased stack +@@ -735,7 +973,13 @@ PTREGSCALL(vm86old) + * normal stack and adjusts ESP with the matching offset. + */ + /* fixup the stack */ +- PER_CPU(gdt_page, %ebx) ++#ifdef CONFIG_SMP ++ movl PER_CPU_VAR(cpu_number), %ebx ++ shll $PAGE_SHIFT_asm, %ebx ++ addl $cpu_gdt_table, %ebx ++#else ++ movl $cpu_gdt_table, %ebx ++#endif + mov GDT_ENTRY_ESPFIX_SS * 8 + 4(%ebx), %al /* bits 16..23 */ + mov GDT_ENTRY_ESPFIX_SS * 8 + 7(%ebx), %ah /* bits 24..31 */ + shl $16, %eax +@@ -793,7 +1037,7 @@ vector=vector+1 + .endr + 2: jmp common_interrupt + .endr +-END(irq_entries_start) ++ENDPROC(irq_entries_start) + + .previous + END(interrupt) +@@ -840,7 +1084,7 @@ ENTRY(coprocessor_error) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(coprocessor_error) ++ENDPROC(coprocessor_error) + + ENTRY(simd_coprocessor_error) + RING0_INT_FRAME +@@ -850,7 +1094,7 @@ ENTRY(simd_coprocessor_error) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(simd_coprocessor_error) ++ENDPROC(simd_coprocessor_error) + + ENTRY(device_not_available) + RING0_INT_FRAME +@@ -860,7 +1104,7 @@ ENTRY(device_not_available) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(device_not_available) ++ENDPROC(device_not_available) + + #ifdef CONFIG_PARAVIRT + ENTRY(native_iret) +@@ -869,12 +1113,12 @@ ENTRY(native_iret) + .align 4 + .long native_iret, iret_exc + .previous +-END(native_iret) ++ENDPROC(native_iret) + + ENTRY(native_irq_enable_sysexit) + sti + sysexit +-END(native_irq_enable_sysexit) ++ENDPROC(native_irq_enable_sysexit) + #endif + + ENTRY(overflow) +@@ -885,7 +1129,7 @@ ENTRY(overflow) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(overflow) ++ENDPROC(overflow) + + ENTRY(bounds) + RING0_INT_FRAME +@@ -895,7 +1139,7 @@ ENTRY(bounds) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(bounds) ++ENDPROC(bounds) + + ENTRY(invalid_op) + RING0_INT_FRAME +@@ -905,7 +1149,7 @@ ENTRY(invalid_op) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(invalid_op) ++ENDPROC(invalid_op) + + ENTRY(coprocessor_segment_overrun) + RING0_INT_FRAME +@@ -915,7 +1159,7 @@ ENTRY(coprocessor_segment_overrun) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(coprocessor_segment_overrun) ++ENDPROC(coprocessor_segment_overrun) + + ENTRY(invalid_TSS) + RING0_EC_FRAME +@@ -923,7 +1167,7 @@ ENTRY(invalid_TSS) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(invalid_TSS) ++ENDPROC(invalid_TSS) + + ENTRY(segment_not_present) + RING0_EC_FRAME +@@ -931,7 +1175,7 @@ ENTRY(segment_not_present) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(segment_not_present) ++ENDPROC(segment_not_present) + + ENTRY(stack_segment) + RING0_EC_FRAME +@@ -939,7 +1183,7 @@ ENTRY(stack_segment) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(stack_segment) ++ENDPROC(stack_segment) + + ENTRY(alignment_check) + RING0_EC_FRAME +@@ -947,7 +1191,7 @@ ENTRY(alignment_check) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(alignment_check) ++ENDPROC(alignment_check) + + ENTRY(divide_error) + RING0_INT_FRAME +@@ -957,7 +1201,7 @@ ENTRY(divide_error) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(divide_error) ++ENDPROC(divide_error) + + #ifdef CONFIG_X86_MCE + ENTRY(machine_check) +@@ -968,7 +1212,7 @@ ENTRY(machine_check) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(machine_check) ++ENDPROC(machine_check) + #endif + + ENTRY(spurious_interrupt_bug) +@@ -979,7 +1223,7 @@ ENTRY(spurious_interrupt_bug) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(spurious_interrupt_bug) ++ENDPROC(spurious_interrupt_bug) + + ENTRY(kernel_thread_helper) + pushl $0 # fake return address for unwinder +@@ -1095,7 +1339,7 @@ ENDPROC(xen_failsafe_callback) + + ENTRY(mcount) + ret +-END(mcount) ++ENDPROC(mcount) + + ENTRY(ftrace_caller) + cmpl $0, function_trace_stop +@@ -1124,7 +1368,7 @@ ftrace_graph_call: + .globl ftrace_stub + ftrace_stub: + ret +-END(ftrace_caller) ++ENDPROC(ftrace_caller) + + #else /* ! CONFIG_DYNAMIC_FTRACE */ + +@@ -1160,7 +1404,7 @@ trace: + popl %ecx + popl %eax + jmp ftrace_stub +-END(mcount) ++ENDPROC(mcount) + #endif /* CONFIG_DYNAMIC_FTRACE */ + #endif /* CONFIG_FUNCTION_TRACER */ + +@@ -1181,7 +1425,7 @@ ENTRY(ftrace_graph_caller) + popl %ecx + popl %eax + ret +-END(ftrace_graph_caller) ++ENDPROC(ftrace_graph_caller) + + .globl return_to_handler + return_to_handler: +@@ -1198,7 +1442,6 @@ return_to_handler: + ret + #endif + +-.section .rodata,"a" + #include "syscall_table_32.S" + + syscall_table_size=(.-sys_call_table) +@@ -1255,15 +1498,18 @@ error_code: + movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart + REG_TO_PTGS %ecx + SET_KERNEL_GS %ecx +- movl $(__USER_DS), %ecx ++ movl $(__KERNEL_DS), %ecx + movl %ecx, %ds + movl %ecx, %es ++ ++ pax_enter_kernel ++ + TRACE_IRQS_OFF + movl %esp,%eax # pt_regs pointer + call *%edi + jmp ret_from_exception + CFI_ENDPROC +-END(page_fault) ++ENDPROC(page_fault) + + /* + * Debug traps and NMI can happen at the one SYSENTER instruction +@@ -1309,7 +1555,7 @@ debug_stack_correct: + call do_debug + jmp ret_from_exception + CFI_ENDPROC +-END(debug) ++ENDPROC(debug) + + /* + * NMI is doubly nasty. It can happen _while_ we're handling +@@ -1351,6 +1597,9 @@ nmi_stack_correct: + xorl %edx,%edx # zero error code + movl %esp,%eax # pt_regs pointer + call do_nmi ++ ++ pax_exit_kernel ++ + jmp restore_all_notrace + CFI_ENDPROC + +@@ -1391,12 +1640,15 @@ nmi_espfix_stack: + FIXUP_ESPFIX_STACK # %eax == %esp + xorl %edx,%edx # zero error code + call do_nmi ++ ++ pax_exit_kernel ++ + RESTORE_REGS + lss 12+4(%esp), %esp # back to espfix stack + CFI_ADJUST_CFA_OFFSET -24 + jmp irq_return + CFI_ENDPROC +-END(nmi) ++ENDPROC(nmi) + + ENTRY(int3) + RING0_INT_FRAME +@@ -1409,7 +1661,7 @@ ENTRY(int3) + call do_int3 + jmp ret_from_exception + CFI_ENDPROC +-END(int3) ++ENDPROC(int3) + + ENTRY(general_protection) + RING0_EC_FRAME +@@ -1417,7 +1669,7 @@ ENTRY(general_protection) + CFI_ADJUST_CFA_OFFSET 4 + jmp error_code + CFI_ENDPROC +-END(general_protection) ++ENDPROC(general_protection) + + /* + * End of kprobes section +diff --git a/arch/x86/kernel/entry_64.S b/arch/x86/kernel/entry_64.S +index 34a56a9..74613c5 100644 +--- a/arch/x86/kernel/entry_64.S ++++ b/arch/x86/kernel/entry_64.S +@@ -53,6 +53,8 @@ + #include + #include + #include ++#include ++#include + + /* Avoid __ASSEMBLER__'ifying just for this. */ + #include +@@ -64,8 +66,9 @@ + #ifdef CONFIG_FUNCTION_TRACER + #ifdef CONFIG_DYNAMIC_FTRACE + ENTRY(mcount) ++ pax_force_retaddr + retq +-END(mcount) ++ENDPROC(mcount) + + ENTRY(ftrace_caller) + cmpl $0, function_trace_stop +@@ -88,8 +91,9 @@ GLOBAL(ftrace_graph_call) + #endif + + GLOBAL(ftrace_stub) ++ pax_force_retaddr + retq +-END(ftrace_caller) ++ENDPROC(ftrace_caller) + + #else /* ! CONFIG_DYNAMIC_FTRACE */ + ENTRY(mcount) +@@ -108,6 +112,7 @@ ENTRY(mcount) + #endif + + GLOBAL(ftrace_stub) ++ pax_force_retaddr + retq + + trace: +@@ -117,12 +122,13 @@ trace: + movq 8(%rbp), %rsi + subq $MCOUNT_INSN_SIZE, %rdi + ++ pax_force_fptr ftrace_trace_function + call *ftrace_trace_function + + MCOUNT_RESTORE_FRAME + + jmp ftrace_stub +-END(mcount) ++ENDPROC(mcount) + #endif /* CONFIG_DYNAMIC_FTRACE */ + #endif /* CONFIG_FUNCTION_TRACER */ + +@@ -142,8 +148,9 @@ ENTRY(ftrace_graph_caller) + + MCOUNT_RESTORE_FRAME + ++ pax_force_retaddr + retq +-END(ftrace_graph_caller) ++ENDPROC(ftrace_graph_caller) + + GLOBAL(return_to_handler) + subq $24, %rsp +@@ -159,6 +166,7 @@ GLOBAL(return_to_handler) + movq 8(%rsp), %rdx + movq (%rsp), %rax + addq $16, %rsp ++ pax_force_retaddr + retq + #endif + +@@ -174,6 +182,282 @@ ENTRY(native_usergs_sysret64) + ENDPROC(native_usergs_sysret64) + #endif /* CONFIG_PARAVIRT */ + ++ .macro ljmpq sel, off ++#if defined(CONFIG_MPSC) || defined(CONFIG_MCORE2) || defined (CONFIG_MATOM) ++ .byte 0x48; ljmp *1234f(%rip) ++ .pushsection .rodata ++ .align 16 ++ 1234: .quad \off; .word \sel ++ .popsection ++#else ++ pushq $\sel ++ pushq $\off ++ lretq ++#endif ++ .endm ++ ++ .macro pax_enter_kernel ++ pax_set_fptr_mask ++#ifdef CONFIG_PAX_KERNEXEC ++ call pax_enter_kernel ++#endif ++ .endm ++ ++ .macro pax_exit_kernel ++#ifdef CONFIG_PAX_KERNEXEC ++ call pax_exit_kernel ++#endif ++ .endm ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ENTRY(pax_enter_kernel) ++ pushq %rdi ++ ++#ifdef CONFIG_PARAVIRT ++ PV_SAVE_REGS(CLBR_RDI) ++#endif ++ ++ GET_CR0_INTO_RDI ++ bts $16,%rdi ++ jnc 3f ++ mov %cs,%edi ++ cmp $__KERNEL_CS,%edi ++ jnz 2f ++1: ++ ++#ifdef CONFIG_PARAVIRT ++ PV_RESTORE_REGS(CLBR_RDI) ++#endif ++ ++ popq %rdi ++ pax_force_retaddr ++ retq ++ ++2: ljmpq __KERNEL_CS,1f ++3: ljmpq __KERNEXEC_KERNEL_CS,4f ++4: SET_RDI_INTO_CR0 ++ jmp 1b ++ENDPROC(pax_enter_kernel) ++ ++ENTRY(pax_exit_kernel) ++ pushq %rdi ++ ++#ifdef CONFIG_PARAVIRT ++ PV_SAVE_REGS(CLBR_RDI) ++#endif ++ ++ mov %cs,%rdi ++ cmp $__KERNEXEC_KERNEL_CS,%edi ++ jz 2f ++1: ++ ++#ifdef CONFIG_PARAVIRT ++ PV_RESTORE_REGS(CLBR_RDI); ++#endif ++ ++ popq %rdi ++ pax_force_retaddr ++ retq ++ ++2: GET_CR0_INTO_RDI ++ btr $16,%rdi ++ ljmpq __KERNEL_CS,3f ++3: SET_RDI_INTO_CR0 ++ jmp 1b ++#ifdef CONFIG_PARAVIRT ++ PV_RESTORE_REGS(CLBR_RDI); ++#endif ++ ++ popq %rdi ++ pax_force_retaddr ++ retq ++ENDPROC(pax_exit_kernel) ++#endif ++ ++ .macro pax_enter_kernel_user ++ pax_set_fptr_mask ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ call pax_enter_kernel_user ++#endif ++ .endm ++ ++ .macro pax_exit_kernel_user ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ call pax_exit_kernel_user ++#endif ++#ifdef CONFIG_PAX_RANDKSTACK ++ pushq %rax ++ call pax_randomize_kstack ++ popq %rax ++#endif ++ .endm ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ENTRY(pax_enter_kernel_user) ++ pushq %rdi ++ pushq %rbx ++ ++#ifdef CONFIG_PARAVIRT ++ PV_SAVE_REGS(CLBR_RDI) ++#endif ++ ++ GET_CR3_INTO_RDI ++ mov %rdi,%rbx ++ add $__START_KERNEL_map,%rbx ++ sub phys_base(%rip),%rbx ++ ++#ifdef CONFIG_PARAVIRT ++ pushq %rdi ++ cmpl $0, pv_info+PARAVIRT_enabled ++ jz 1f ++ i = 0 ++ .rept USER_PGD_PTRS ++ mov i*8(%rbx),%rsi ++ mov $0,%sil ++ lea i*8(%rbx),%rdi ++ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_set_pgd_batched) ++ i = i + 1 ++ .endr ++ jmp 2f ++1: ++#endif ++ ++ i = 0 ++ .rept USER_PGD_PTRS ++ movb $0,i*8(%rbx) ++ i = i + 1 ++ .endr ++ ++#ifdef CONFIG_PARAVIRT ++2: popq %rdi ++#endif ++ SET_RDI_INTO_CR3 ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ GET_CR0_INTO_RDI ++ bts $16,%rdi ++ SET_RDI_INTO_CR0 ++#endif ++ ++#ifdef CONFIG_PARAVIRT ++ PV_RESTORE_REGS(CLBR_RDI) ++#endif ++ ++ popq %rbx ++ popq %rdi ++ pax_force_retaddr ++ retq ++ENDPROC(pax_enter_kernel_user) ++ ++ENTRY(pax_exit_kernel_user) ++ push %rdi ++ ++#ifdef CONFIG_PARAVIRT ++ pushq %rbx ++ PV_SAVE_REGS(CLBR_RDI) ++#endif ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ GET_CR0_INTO_RDI ++ btr $16,%rdi ++ SET_RDI_INTO_CR0 ++#endif ++ ++ GET_CR3_INTO_RDI ++ add $__START_KERNEL_map,%rdi ++ sub phys_base(%rip),%rdi ++ ++#ifdef CONFIG_PARAVIRT ++ cmpl $0, pv_info+PARAVIRT_enabled ++ jz 1f ++ mov %rdi,%rbx ++ i = 0 ++ .rept USER_PGD_PTRS ++ mov i*8(%rbx),%rsi ++ mov $0x67,%sil ++ lea i*8(%rbx),%rdi ++ call PARA_INDIRECT(pv_mmu_ops+PV_MMU_set_pgd_batched) ++ i = i + 1 ++ .endr ++ jmp 2f ++1: ++#endif ++ ++ i = 0 ++ .rept USER_PGD_PTRS ++ movb $0x67,i*8(%rdi) ++ i = i + 1 ++ .endr ++ ++#ifdef CONFIG_PARAVIRT ++2: PV_RESTORE_REGS(CLBR_RDI) ++ popq %rbx ++#endif ++ ++ popq %rdi ++ pax_force_retaddr ++ retq ++ENDPROC(pax_exit_kernel_user) ++#endif ++ ++.macro pax_erase_kstack ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++ call pax_erase_kstack ++#endif ++.endm ++ ++#ifdef CONFIG_PAX_MEMORY_STACKLEAK ++/* ++ * r11: thread_info ++ * rcx, rdx: can be clobbered ++ */ ++ENTRY(pax_erase_kstack) ++ pushq %rdi ++ pushq %rax ++ pushq %r11 ++ ++ GET_THREAD_INFO(%r11) ++ mov TI_lowest_stack(%r11), %rdi ++ mov $-0xBEEF, %rax ++ std ++ ++1: mov %edi, %ecx ++ and $THREAD_SIZE_asm - 1, %ecx ++ shr $3, %ecx ++ repne scasq ++ jecxz 2f ++ ++ cmp $2*8, %ecx ++ jc 2f ++ ++ mov $2*8, %ecx ++ repe scasq ++ jecxz 2f ++ jne 1b ++ ++2: cld ++ mov %esp, %ecx ++ sub %edi, %ecx ++ ++ cmp $THREAD_SIZE_asm, %rcx ++ jb 3f ++ ud2 ++3: ++ ++ shr $3, %ecx ++ rep stosq ++ ++ mov TI_task_thread_sp0(%r11), %rdi ++ sub $256, %rdi ++ mov %rdi, TI_lowest_stack(%r11) ++ ++ popq %r11 ++ popq %rax ++ popq %rdi ++ pax_force_retaddr ++ ret ++ENDPROC(pax_erase_kstack) ++#endif + + .macro TRACE_IRQS_IRETQ offset=ARGOFFSET + #ifdef CONFIG_TRACE_IRQFLAGS +@@ -233,8 +517,8 @@ ENDPROC(native_usergs_sysret64) + .endm + + .macro UNFAKE_STACK_FRAME +- addq $8*6, %rsp +- CFI_ADJUST_CFA_OFFSET -(6*8) ++ addq $8*6 + ARG_SKIP, %rsp ++ CFI_ADJUST_CFA_OFFSET -(6*8 + ARG_SKIP) + .endm + + /* +@@ -317,7 +601,7 @@ ENTRY(save_args) + leaq -ARGOFFSET+16(%rsp),%rdi /* arg1 for handler */ + movq_cfi rbp, 8 /* push %rbp */ + leaq 8(%rsp), %rbp /* mov %rsp, %ebp */ +- testl $3, CS(%rdi) ++ testb $3, CS(%rdi) + je 1f + SWAPGS + /* +@@ -337,9 +621,10 @@ ENTRY(save_args) + * We entered an interrupt context - irqs are off: + */ + 2: TRACE_IRQS_OFF ++ pax_force_retaddr_bts + ret + CFI_ENDPROC +-END(save_args) ++ENDPROC(save_args) + + ENTRY(save_rest) + PARTIAL_FRAME 1 REST_SKIP+8 +@@ -352,9 +637,10 @@ ENTRY(save_rest) + movq_cfi r15, R15+16 + movq %r11, 8(%rsp) /* return address */ + FIXUP_TOP_OF_STACK %r11, 16 ++ pax_force_retaddr + ret + CFI_ENDPROC +-END(save_rest) ++ENDPROC(save_rest) + + /* save complete stack frame */ + .pushsection .kprobes.text, "ax" +@@ -383,9 +669,10 @@ ENTRY(save_paranoid) + js 1f /* negative -> in kernel */ + SWAPGS + xorl %ebx,%ebx +-1: ret ++1: pax_force_retaddr_bts ++ ret + CFI_ENDPROC +-END(save_paranoid) ++ENDPROC(save_paranoid) + .popsection + + /* +@@ -409,7 +696,7 @@ ENTRY(ret_from_fork) + + RESTORE_REST + +- testl $3, CS-ARGOFFSET(%rsp) # from kernel_thread? ++ testb $3, CS-ARGOFFSET(%rsp) # from kernel_thread? + je int_ret_from_sys_call + + testl $_TIF_IA32, TI_flags(%rcx) # 32-bit compat task needs IRET +@@ -419,7 +706,7 @@ ENTRY(ret_from_fork) + jmp ret_from_sys_call # go to the SYSRET fastpath + + CFI_ENDPROC +-END(ret_from_fork) ++ENDPROC(ret_from_fork) + + /* + * System call entry. Upto 6 arguments in registers are supported. +@@ -455,7 +742,7 @@ END(ret_from_fork) + ENTRY(system_call) + CFI_STARTPROC simple + CFI_SIGNAL_FRAME +- CFI_DEF_CFA rsp,KERNEL_STACK_OFFSET ++ CFI_DEF_CFA rsp,0 + CFI_REGISTER rip,rcx + /*CFI_REGISTER rflags,r11*/ + SWAPGS_UNSAFE_STACK +@@ -468,12 +755,13 @@ ENTRY(system_call_after_swapgs) + + movq %rsp,PER_CPU_VAR(old_rsp) + movq PER_CPU_VAR(kernel_stack),%rsp ++ SAVE_ARGS 8*6,1 ++ pax_enter_kernel_user + /* + * No need to follow this irqs off/on section - it's straight + * and short: + */ + ENABLE_INTERRUPTS(CLBR_NONE) +- SAVE_ARGS 8,1 + movq %rax,ORIG_RAX-ARGOFFSET(%rsp) + movq %rcx,RIP-ARGOFFSET(%rsp) + CFI_REL_OFFSET rip,RIP-ARGOFFSET +@@ -483,7 +771,7 @@ ENTRY(system_call_after_swapgs) + system_call_fastpath: + cmpq $__NR_syscall_max,%rax + ja badsys +- movq %r10,%rcx ++ movq R10-ARGOFFSET(%rsp),%rcx + call *sys_call_table(,%rax,8) # XXX: rip relative + movq %rax,RAX-ARGOFFSET(%rsp) + /* +@@ -502,6 +790,8 @@ sysret_check: + andl %edi,%edx + jnz sysret_careful + CFI_REMEMBER_STATE ++ pax_exit_kernel_user ++ pax_erase_kstack + /* + * sysretq will re-enable interrupts: + */ +@@ -555,14 +845,18 @@ badsys: + * jump back to the normal fast path. + */ + auditsys: +- movq %r10,%r9 /* 6th arg: 4th syscall arg */ ++ movq R10-ARGOFFSET(%rsp),%r9 /* 6th arg: 4th syscall arg */ + movq %rdx,%r8 /* 5th arg: 3rd syscall arg */ + movq %rsi,%rcx /* 4th arg: 2nd syscall arg */ + movq %rdi,%rdx /* 3rd arg: 1st syscall arg */ + movq %rax,%rsi /* 2nd arg: syscall number */ + movl $AUDIT_ARCH_X86_64,%edi /* 1st arg: audit arch */ + call audit_syscall_entry ++ ++ pax_erase_kstack ++ + LOAD_ARGS 0 /* reload call-clobbered registers */ ++ pax_set_fptr_mask + jmp system_call_fastpath + + /* +@@ -592,16 +886,20 @@ tracesys: + FIXUP_TOP_OF_STACK %rdi + movq %rsp,%rdi + call syscall_trace_enter ++ ++ pax_erase_kstack ++ + /* + * Reload arg registers from stack in case ptrace changed them. + * We don't reload %rax because syscall_trace_enter() returned + * the value it wants us to use in the table lookup. + */ + LOAD_ARGS ARGOFFSET, 1 ++ pax_set_fptr_mask + RESTORE_REST + cmpq $__NR_syscall_max,%rax + ja int_ret_from_sys_call /* RAX(%rsp) set to -ENOSYS above */ +- movq %r10,%rcx /* fixup for C */ ++ movq R10-ARGOFFSET(%rsp),%rcx /* fixup for C */ + call *sys_call_table(,%rax,8) + movq %rax,RAX-ARGOFFSET(%rsp) + /* Use IRET because user could have changed frame */ +@@ -613,7 +911,7 @@ tracesys: + GLOBAL(int_ret_from_sys_call) + DISABLE_INTERRUPTS(CLBR_NONE) + TRACE_IRQS_OFF +- testl $3,CS-ARGOFFSET(%rsp) ++ testb $3,CS-ARGOFFSET(%rsp) + je retint_restore_args + movl $_TIF_ALLWORK_MASK,%edi + /* edi: mask to check */ +@@ -624,6 +922,7 @@ GLOBAL(int_with_check) + andl %edi,%edx + jnz int_careful + andl $~TS_COMPAT,TI_status(%rcx) ++ pax_erase_kstack + jmp retint_swapgs + + /* Either reschedule or signal or syscall exit tracking needed. */ +@@ -674,7 +973,7 @@ int_restore_rest: + TRACE_IRQS_OFF + jmp int_with_check + CFI_ENDPROC +-END(system_call) ++ENDPROC(system_call) + + /* + * Certain special system calls that need to save a complete full stack frame. +@@ -690,7 +989,7 @@ ENTRY(\label) + call \func + jmp ptregscall_common + CFI_ENDPROC +-END(\label) ++ENDPROC(\label) + .endm + + PTREGSCALL stub_clone, sys_clone, %r8 +@@ -708,9 +1007,10 @@ ENTRY(ptregscall_common) + movq_cfi_restore R12+8, r12 + movq_cfi_restore RBP+8, rbp + movq_cfi_restore RBX+8, rbx ++ pax_force_retaddr + ret $REST_SKIP /* pop extended registers */ + CFI_ENDPROC +-END(ptregscall_common) ++ENDPROC(ptregscall_common) + + ENTRY(stub_execve) + CFI_STARTPROC +@@ -726,7 +1026,7 @@ ENTRY(stub_execve) + RESTORE_REST + jmp int_ret_from_sys_call + CFI_ENDPROC +-END(stub_execve) ++ENDPROC(stub_execve) + + /* + * sigreturn is special because it needs to restore all registers on return. +@@ -744,7 +1044,7 @@ ENTRY(stub_rt_sigreturn) + RESTORE_REST + jmp int_ret_from_sys_call + CFI_ENDPROC +-END(stub_rt_sigreturn) ++ENDPROC(stub_rt_sigreturn) + + /* + * Build the entry stubs and pointer table with some assembler magic. +@@ -780,7 +1080,7 @@ vector=vector+1 + 2: jmp common_interrupt + .endr + CFI_ENDPROC +-END(irq_entries_start) ++ENDPROC(irq_entries_start) + + .previous + END(interrupt) +@@ -800,6 +1100,16 @@ END(interrupt) + CFI_ADJUST_CFA_OFFSET 10*8 + call save_args + PARTIAL_FRAME 0 ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rdi) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + call \func + .endm + +@@ -822,7 +1132,7 @@ ret_from_intr: + CFI_ADJUST_CFA_OFFSET -8 + exit_intr: + GET_THREAD_INFO(%rcx) +- testl $3,CS-ARGOFFSET(%rsp) ++ testb $3,CS-ARGOFFSET(%rsp) + je retint_kernel + + /* Interrupt came from user space */ +@@ -844,12 +1154,15 @@ retint_swapgs: /* return to user-space */ + * The iretq could re-enable interrupts: + */ + DISABLE_INTERRUPTS(CLBR_ANY) ++ pax_exit_kernel_user + TRACE_IRQS_IRETQ + SWAPGS + jmp restore_args + + retint_restore_args: /* return to kernel space */ + DISABLE_INTERRUPTS(CLBR_ANY) ++ pax_exit_kernel ++ pax_force_retaddr RIP-ARGOFFSET + /* + * The iretq could re-enable interrupts: + */ +@@ -940,7 +1253,7 @@ ENTRY(retint_kernel) + #endif + + CFI_ENDPROC +-END(common_interrupt) ++ENDPROC(common_interrupt) + + /* + * APIC interrupts. +@@ -953,7 +1266,7 @@ ENTRY(\sym) + interrupt \do_sym + jmp ret_from_intr + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + #ifdef CONFIG_SMP +@@ -1032,12 +1345,22 @@ ENTRY(\sym) + CFI_ADJUST_CFA_OFFSET 15*8 + call error_entry + DEFAULT_FRAME 0 ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + movq %rsp,%rdi /* pt_regs pointer */ + xorl %esi,%esi /* no error code */ + call \do_sym + jmp error_exit /* %ebx: no swapgs flag */ + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + .macro paranoidzeroentry sym do_sym +@@ -1049,12 +1372,22 @@ ENTRY(\sym) + subq $15*8, %rsp + call save_paranoid + TRACE_IRQS_OFF ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + movq %rsp,%rdi /* pt_regs pointer */ + xorl %esi,%esi /* no error code */ + call \do_sym + jmp paranoid_exit /* %ebx: no swapgs flag */ + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + .macro paranoidzeroentry_ist sym do_sym ist +@@ -1066,15 +1399,30 @@ ENTRY(\sym) + subq $15*8, %rsp + call save_paranoid + TRACE_IRQS_OFF ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + movq %rsp,%rdi /* pt_regs pointer */ + xorl %esi,%esi /* no error code */ +- PER_CPU(init_tss, %rbp) ++#ifdef CONFIG_SMP ++ imul $TSS_size, PER_CPU_VAR(cpu_number), %ebp ++ lea init_tss(%rbp), %rbp ++#else ++ lea init_tss(%rip), %rbp ++#endif + subq $EXCEPTION_STKSZ, TSS_ist + (\ist - 1) * 8(%rbp) + call \do_sym + addq $EXCEPTION_STKSZ, TSS_ist + (\ist - 1) * 8(%rbp) + jmp paranoid_exit /* %ebx: no swapgs flag */ + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + .macro errorentry sym do_sym +@@ -1085,13 +1433,23 @@ ENTRY(\sym) + CFI_ADJUST_CFA_OFFSET 15*8 + call error_entry + DEFAULT_FRAME 0 ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + movq %rsp,%rdi /* pt_regs pointer */ + movq ORIG_RAX(%rsp),%rsi /* get error code */ + movq $-1,ORIG_RAX(%rsp) /* no syscall to restart */ + call \do_sym + jmp error_exit /* %ebx: no swapgs flag */ + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + /* error code is on the stack already */ +@@ -1104,13 +1462,23 @@ ENTRY(\sym) + call save_paranoid + DEFAULT_FRAME 0 + TRACE_IRQS_OFF ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + movq %rsp,%rdi /* pt_regs pointer */ + movq ORIG_RAX(%rsp),%rsi /* get error code */ + movq $-1,ORIG_RAX(%rsp) /* no syscall to restart */ + call \do_sym + jmp paranoid_exit /* %ebx: no swapgs flag */ + CFI_ENDPROC +-END(\sym) ++ENDPROC(\sym) + .endm + + zeroentry divide_error do_divide_error +@@ -1141,9 +1509,10 @@ gs_change: + SWAPGS + popf + CFI_ADJUST_CFA_OFFSET -8 ++ pax_force_retaddr + ret + CFI_ENDPROC +-END(native_load_gs_index) ++ENDPROC(native_load_gs_index) + + .section __ex_table,"a" + .align 8 +@@ -1193,11 +1562,12 @@ ENTRY(kernel_thread) + * of hacks for example to fork off the per-CPU idle tasks. + * [Hopefully no generic code relies on the reschedule -AK] + */ +- RESTORE_ALL ++ RESTORE_REST + UNFAKE_STACK_FRAME ++ pax_force_retaddr + ret + CFI_ENDPROC +-END(kernel_thread) ++ENDPROC(kernel_thread) + + ENTRY(child_rip) + pushq $0 # fake return address +@@ -1208,13 +1578,14 @@ ENTRY(child_rip) + */ + movq %rdi, %rax + movq %rsi, %rdi ++ pax_force_fptr %rax + call *%rax + # exit + mov %eax, %edi + call do_exit + ud2 # padding for call trace + CFI_ENDPROC +-END(child_rip) ++ENDPROC(child_rip) + + /* + * execve(). This function needs to use IRET, not SYSRET, to set up all state properly. +@@ -1241,11 +1612,11 @@ ENTRY(kernel_execve) + RESTORE_REST + testq %rax,%rax + je int_ret_from_sys_call +- RESTORE_ARGS + UNFAKE_STACK_FRAME ++ pax_force_retaddr + ret + CFI_ENDPROC +-END(kernel_execve) ++ENDPROC(kernel_execve) + + /* Call softirq on interrupt stack. Interrupts are off. */ + ENTRY(call_softirq) +@@ -1263,9 +1634,10 @@ ENTRY(call_softirq) + CFI_DEF_CFA_REGISTER rsp + CFI_ADJUST_CFA_OFFSET -8 + decl PER_CPU_VAR(irq_count) ++ pax_force_retaddr + ret + CFI_ENDPROC +-END(call_softirq) ++ENDPROC(call_softirq) + + #ifdef CONFIG_XEN + zeroentry xen_hypervisor_callback xen_do_hypervisor_callback +@@ -1303,7 +1675,7 @@ ENTRY(xen_do_hypervisor_callback) # do_hypervisor_callback(struct *pt_regs) + decl PER_CPU_VAR(irq_count) + jmp error_exit + CFI_ENDPROC +-END(xen_do_hypervisor_callback) ++ENDPROC(xen_do_hypervisor_callback) + + /* + * Hypervisor uses this for application faults while it executes. +@@ -1362,7 +1734,7 @@ ENTRY(xen_failsafe_callback) + SAVE_ALL + jmp error_exit + CFI_ENDPROC +-END(xen_failsafe_callback) ++ENDPROC(xen_failsafe_callback) + + #endif /* CONFIG_XEN */ + +@@ -1405,16 +1777,31 @@ ENTRY(paranoid_exit) + TRACE_IRQS_OFF + testl %ebx,%ebx /* swapgs needed? */ + jnz paranoid_restore +- testl $3,CS(%rsp) ++ testb $3,CS(%rsp) + jnz paranoid_userspace ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pax_exit_kernel ++ TRACE_IRQS_IRETQ 0 ++ SWAPGS_UNSAFE_STACK ++ RESTORE_ALL 8 ++ pax_force_retaddr_bts ++ jmp irq_return ++#endif + paranoid_swapgs: ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pax_exit_kernel_user ++#else ++ pax_exit_kernel ++#endif + TRACE_IRQS_IRETQ 0 + SWAPGS_UNSAFE_STACK + RESTORE_ALL 8 + jmp irq_return + paranoid_restore: ++ pax_exit_kernel + TRACE_IRQS_IRETQ 0 + RESTORE_ALL 8 ++ pax_force_retaddr_bts + jmp irq_return + paranoid_userspace: + GET_THREAD_INFO(%rcx) +@@ -1443,7 +1830,7 @@ paranoid_schedule: + TRACE_IRQS_OFF + jmp paranoid_userspace + CFI_ENDPROC +-END(paranoid_exit) ++ENDPROC(paranoid_exit) + + /* + * Exception entry point. This expects an error code/orig_rax on the stack. +@@ -1470,12 +1857,13 @@ ENTRY(error_entry) + movq_cfi r14, R14+8 + movq_cfi r15, R15+8 + xorl %ebx,%ebx +- testl $3,CS+8(%rsp) ++ testb $3,CS+8(%rsp) + je error_kernelspace + error_swapgs: + SWAPGS + error_sti: + TRACE_IRQS_OFF ++ pax_force_retaddr_bts + ret + CFI_ENDPROC + +@@ -1497,7 +1885,7 @@ error_kernelspace: + cmpq $gs_change,RIP+8(%rsp) + je error_swapgs + jmp error_sti +-END(error_entry) ++ENDPROC(error_entry) + + + /* ebx: no swapgs flag (1: don't need swapgs, 0: need it) */ +@@ -1517,7 +1905,7 @@ ENTRY(error_exit) + jnz retint_careful + jmp retint_swapgs + CFI_ENDPROC +-END(error_exit) ++ENDPROC(error_exit) + + + /* runs on exception stack */ +@@ -1529,6 +1917,16 @@ ENTRY(nmi) + CFI_ADJUST_CFA_OFFSET 15*8 + call save_paranoid + DEFAULT_FRAME 0 ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ testb $3, CS(%rsp) ++ jnz 1f ++ pax_enter_kernel ++ jmp 2f ++1: pax_enter_kernel_user ++2: ++#else ++ pax_enter_kernel ++#endif + /* paranoidentry do_nmi, 0; without TRACE_IRQS_OFF */ + movq %rsp,%rdi + movq $-1,%rsi +@@ -1539,12 +1937,28 @@ ENTRY(nmi) + DISABLE_INTERRUPTS(CLBR_NONE) + testl %ebx,%ebx /* swapgs needed? */ + jnz nmi_restore +- testl $3,CS(%rsp) ++ testb $3,CS(%rsp) + jnz nmi_userspace ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pax_exit_kernel ++ SWAPGS_UNSAFE_STACK ++ RESTORE_ALL 8 ++ pax_force_retaddr_bts ++ jmp irq_return ++#endif + nmi_swapgs: ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pax_exit_kernel_user ++#else ++ pax_exit_kernel ++#endif + SWAPGS_UNSAFE_STACK ++ RESTORE_ALL 8 ++ jmp irq_return + nmi_restore: ++ pax_exit_kernel + RESTORE_ALL 8 ++ pax_force_retaddr_bts + jmp irq_return + nmi_userspace: + GET_THREAD_INFO(%rcx) +@@ -1573,14 +1987,14 @@ nmi_schedule: + jmp paranoid_exit + CFI_ENDPROC + #endif +-END(nmi) ++ENDPROC(nmi) + + ENTRY(ignore_sysret) + CFI_STARTPROC + mov $-ENOSYS,%eax + sysret + CFI_ENDPROC +-END(ignore_sysret) ++ENDPROC(ignore_sysret) + + /* + * End of kprobes section +diff --git a/arch/x86/kernel/ftrace.c b/arch/x86/kernel/ftrace.c +index 9dbb527..7b3615a 100644 +--- a/arch/x86/kernel/ftrace.c ++++ b/arch/x86/kernel/ftrace.c +@@ -103,7 +103,7 @@ static void *mod_code_ip; /* holds the IP to write to */ + static void *mod_code_newcode; /* holds the text to write to the IP */ + + static unsigned nmi_wait_count; +-static atomic_t nmi_update_count = ATOMIC_INIT(0); ++static atomic_unchecked_t nmi_update_count = ATOMIC_INIT(0); + + int ftrace_arch_read_dyn_info(char *buf, int size) + { +@@ -111,7 +111,7 @@ int ftrace_arch_read_dyn_info(char *buf, int size) + + r = snprintf(buf, size, "%u %u", + nmi_wait_count, +- atomic_read(&nmi_update_count)); ++ atomic_read_unchecked(&nmi_update_count)); + return r; + } + +@@ -149,8 +149,10 @@ void ftrace_nmi_enter(void) + { + if (atomic_inc_return(&nmi_running) & MOD_CODE_WRITE_FLAG) { + smp_rmb(); ++ pax_open_kernel(); + ftrace_mod_code(); +- atomic_inc(&nmi_update_count); ++ pax_close_kernel(); ++ atomic_inc_unchecked(&nmi_update_count); + } + /* Must have previous changes seen before executions */ + smp_mb(); +@@ -215,7 +217,7 @@ do_ftrace_mod_code(unsigned long ip, void *new_code) + + + +-static unsigned char ftrace_nop[MCOUNT_INSN_SIZE]; ++static unsigned char ftrace_nop[MCOUNT_INSN_SIZE] __read_only; + + static unsigned char *ftrace_nop_replace(void) + { +@@ -228,6 +230,8 @@ ftrace_modify_code(unsigned long ip, unsigned char *old_code, + { + unsigned char replaced[MCOUNT_INSN_SIZE]; + ++ ip = ktla_ktva(ip); ++ + /* + * Note: Due to modules and __init, code can + * disappear and change, we need to protect against faulting +@@ -284,7 +288,7 @@ int ftrace_update_ftrace_func(ftrace_func_t func) + unsigned char old[MCOUNT_INSN_SIZE], *new; + int ret; + +- memcpy(old, &ftrace_call, MCOUNT_INSN_SIZE); ++ memcpy(old, (void *)ktla_ktva((unsigned long)ftrace_call), MCOUNT_INSN_SIZE); + new = ftrace_call_replace(ip, (unsigned long)func); + ret = ftrace_modify_code(ip, old, new); + +@@ -337,15 +341,15 @@ int __init ftrace_dyn_arch_init(void *data) + switch (faulted) { + case 0: + pr_info("ftrace: converting mcount calls to 0f 1f 44 00 00\n"); +- memcpy(ftrace_nop, ftrace_test_p6nop, MCOUNT_INSN_SIZE); ++ memcpy(ftrace_nop, ktla_ktva(ftrace_test_p6nop), MCOUNT_INSN_SIZE); + break; + case 1: + pr_info("ftrace: converting mcount calls to 66 66 66 66 90\n"); +- memcpy(ftrace_nop, ftrace_test_nop5, MCOUNT_INSN_SIZE); ++ memcpy(ftrace_nop, ktla_ktva(ftrace_test_nop5), MCOUNT_INSN_SIZE); + break; + case 2: + pr_info("ftrace: converting mcount calls to jmp . + 5\n"); +- memcpy(ftrace_nop, ftrace_test_jmp, MCOUNT_INSN_SIZE); ++ memcpy(ftrace_nop, ktla_ktva(ftrace_test_jmp), MCOUNT_INSN_SIZE); + break; + } + +@@ -366,6 +370,8 @@ static int ftrace_mod_jmp(unsigned long ip, + { + unsigned char code[MCOUNT_INSN_SIZE]; + ++ ip = ktla_ktva(ip); ++ + if (probe_kernel_read(code, (void *)ip, MCOUNT_INSN_SIZE)) + return -EFAULT; + +diff --git a/arch/x86/kernel/head32.c b/arch/x86/kernel/head32.c +index 4f8e250..df24706 100644 +--- a/arch/x86/kernel/head32.c ++++ b/arch/x86/kernel/head32.c +@@ -16,6 +16,7 @@ + #include + #include + #include ++#include + + static void __init i386_default_early_setup(void) + { +@@ -31,7 +32,7 @@ void __init i386_start_kernel(void) + { + reserve_trampoline_memory(); + +- reserve_early(__pa_symbol(&_text), __pa_symbol(&__bss_stop), "TEXT DATA BSS"); ++ reserve_early(LOAD_PHYSICAL_ADDR, __pa_symbol(&__bss_stop), "TEXT DATA BSS"); + + #ifdef CONFIG_BLK_DEV_INITRD + /* Reserve INITRD */ +diff --git a/arch/x86/kernel/head_32.S b/arch/x86/kernel/head_32.S +index 34c3308..6fc4e76 100644 +--- a/arch/x86/kernel/head_32.S ++++ b/arch/x86/kernel/head_32.S +@@ -19,10 +19,17 @@ + #include + #include + #include ++#include + + /* Physical address */ + #define pa(X) ((X) - __PAGE_OFFSET) + ++#ifdef CONFIG_PAX_KERNEXEC ++#define ta(X) (X) ++#else ++#define ta(X) ((X) - __PAGE_OFFSET) ++#endif ++ + /* + * References to members of the new_cpu_data structure. + */ +@@ -52,11 +59,7 @@ + * and small than max_low_pfn, otherwise will waste some page table entries + */ + +-#if PTRS_PER_PMD > 1 +-#define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD) +-#else +-#define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PGD) +-#endif ++#define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PTE) + + /* Enough space to fit pagetables for the low memory linear map */ + MAPPING_BEYOND_END = \ +@@ -73,6 +76,12 @@ INIT_MAP_SIZE = PAGE_TABLE_SIZE(KERNEL_PAGES) * PAGE_SIZE_asm + RESERVE_BRK(pagetables, INIT_MAP_SIZE) + + /* ++ * Real beginning of normal "text" segment ++ */ ++ENTRY(stext) ++ENTRY(_stext) ++ ++/* + * 32-bit kernel entrypoint; only used by the boot CPU. On entry, + * %esi points to the real-mode code as a 32-bit pointer. + * CS and DS must be 4 GB flat segments, but we don't depend on +@@ -80,7 +89,16 @@ RESERVE_BRK(pagetables, INIT_MAP_SIZE) + * can. + */ + __HEAD ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ jmp startup_32 ++/* PaX: fill first page in .text with int3 to catch NULL derefs in kernel mode */ ++.fill PAGE_SIZE-5,1,0xcc ++#endif ++ + ENTRY(startup_32) ++ movl pa(stack_start),%ecx ++ + /* test KEEP_SEGMENTS flag to see if the bootloader is asking + us to not reload segments */ + testb $(1<<6), BP_loadflags(%esi) +@@ -95,7 +113,60 @@ ENTRY(startup_32) + movl %eax,%es + movl %eax,%fs + movl %eax,%gs ++ movl %eax,%ss + 2: ++ leal -__PAGE_OFFSET(%ecx),%esp ++ ++#ifdef CONFIG_SMP ++ movl $pa(cpu_gdt_table),%edi ++ movl $__per_cpu_load,%eax ++ movw %ax,__KERNEL_PERCPU + 2(%edi) ++ rorl $16,%eax ++ movb %al,__KERNEL_PERCPU + 4(%edi) ++ movb %ah,__KERNEL_PERCPU + 7(%edi) ++ movl $__per_cpu_end - 1,%eax ++ subl $__per_cpu_start,%eax ++ movw %ax,__KERNEL_PERCPU + 0(%edi) ++#endif ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ movl $NR_CPUS,%ecx ++ movl $pa(cpu_gdt_table),%edi ++1: ++ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c09700),GDT_ENTRY_KERNEL_DS * 8 + 4(%edi) ++ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c0fb00),GDT_ENTRY_DEFAULT_USER_CS * 8 + 4(%edi) ++ movl $((((__PAGE_OFFSET-1) & 0xf0000000) >> 12) | 0x00c0f300),GDT_ENTRY_DEFAULT_USER_DS * 8 + 4(%edi) ++ addl $PAGE_SIZE_asm,%edi ++ loop 1b ++#endif ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ movl $pa(boot_gdt),%edi ++ movl $__LOAD_PHYSICAL_ADDR,%eax ++ movw %ax,__BOOT_CS + 2(%edi) ++ rorl $16,%eax ++ movb %al,__BOOT_CS + 4(%edi) ++ movb %ah,__BOOT_CS + 7(%edi) ++ rorl $16,%eax ++ ++ ljmp $(__BOOT_CS),$1f ++1: ++ ++ movl $NR_CPUS,%ecx ++ movl $pa(cpu_gdt_table),%edi ++ addl $__PAGE_OFFSET,%eax ++1: ++ movw %ax,__KERNEL_CS + 2(%edi) ++ movw %ax,__KERNEXEC_KERNEL_CS + 2(%edi) ++ rorl $16,%eax ++ movb %al,__KERNEL_CS + 4(%edi) ++ movb %al,__KERNEXEC_KERNEL_CS + 4(%edi) ++ movb %ah,__KERNEL_CS + 7(%edi) ++ movb %ah,__KERNEXEC_KERNEL_CS + 7(%edi) ++ rorl $16,%eax ++ addl $PAGE_SIZE_asm,%edi ++ loop 1b ++#endif + + /* + * Clear BSS first so that there are no surprises... +@@ -140,9 +211,7 @@ ENTRY(startup_32) + cmpl $num_subarch_entries, %eax + jae bad_subarch + +- movl pa(subarch_entries)(,%eax,4), %eax +- subl $__PAGE_OFFSET, %eax +- jmp *%eax ++ jmp *pa(subarch_entries)(,%eax,4) + + bad_subarch: + WEAK(lguest_entry) +@@ -154,10 +223,10 @@ WEAK(xen_entry) + __INITDATA + + subarch_entries: +- .long default_entry /* normal x86/PC */ +- .long lguest_entry /* lguest hypervisor */ +- .long xen_entry /* Xen hypervisor */ +- .long default_entry /* Moorestown MID */ ++ .long ta(default_entry) /* normal x86/PC */ ++ .long ta(lguest_entry) /* lguest hypervisor */ ++ .long ta(xen_entry) /* Xen hypervisor */ ++ .long ta(default_entry) /* Moorestown MID */ + num_subarch_entries = (. - subarch_entries) / 4 + .previous + #endif /* CONFIG_PARAVIRT */ +@@ -218,8 +287,11 @@ default_entry: + movl %eax, pa(max_pfn_mapped) + + /* Do early initialization of the fixmap area */ +- movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR,%eax +- movl %eax,pa(swapper_pg_pmd+0x1000*KPMDS-8) ++#ifdef CONFIG_COMPAT_VDSO ++ movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR+_PAGE_USER,pa(swapper_pg_pmd+0x1000*KPMDS-8) ++#else ++ movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR,pa(swapper_pg_pmd+0x1000*KPMDS-8) ++#endif + #else /* Not PAE */ + + page_pde_offset = (__PAGE_OFFSET >> 20); +@@ -249,8 +321,11 @@ page_pde_offset = (__PAGE_OFFSET >> 20); + movl %eax, pa(max_pfn_mapped) + + /* Do early initialization of the fixmap area */ +- movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR,%eax +- movl %eax,pa(swapper_pg_dir+0xffc) ++#ifdef CONFIG_COMPAT_VDSO ++ movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR+_PAGE_USER,pa(swapper_pg_dir+0xffc) ++#else ++ movl $pa(swapper_pg_fixmap)+PDE_IDENT_ATTR,pa(swapper_pg_dir+0xffc) ++#endif + #endif + jmp 3f + /* +@@ -272,6 +347,9 @@ ENTRY(startup_32_smp) + movl %eax,%es + movl %eax,%fs + movl %eax,%gs ++ movl pa(stack_start),%ecx ++ movl %eax,%ss ++ leal -__PAGE_OFFSET(%ecx),%esp + #endif /* CONFIG_SMP */ + 3: + +@@ -297,6 +375,7 @@ ENTRY(startup_32_smp) + orl %edx,%eax + movl %eax,%cr4 + ++#ifdef CONFIG_X86_PAE + btl $5, %eax # check if PAE is enabled + jnc 6f + +@@ -305,6 +384,10 @@ ENTRY(startup_32_smp) + cpuid + cmpl $0x80000000, %eax + jbe 6f ++ ++ /* Clear bogus XD_DISABLE bits */ ++ call verify_cpu ++ + mov $0x80000001, %eax + cpuid + /* Execute Disable bit supported? */ +@@ -312,13 +395,17 @@ ENTRY(startup_32_smp) + jnc 6f + + /* Setup EFER (Extended Feature Enable Register) */ +- movl $0xc0000080, %ecx ++ movl $MSR_EFER, %ecx + rdmsr + + btsl $11, %eax + /* Make changes effective */ + wrmsr + ++ btsl $_PAGE_BIT_NX-32,pa(__supported_pte_mask+4) ++ movl $1,pa(nx_enabled) ++#endif ++ + 6: + + /* +@@ -331,8 +418,8 @@ ENTRY(startup_32_smp) + movl %eax,%cr0 /* ..and set paging (PG) bit */ + ljmp $__BOOT_CS,$1f /* Clear prefetch and normalize %eip */ + 1: +- /* Set up the stack pointer */ +- lss stack_start,%esp ++ /* Shift the stack pointer to a virtual address */ ++ addl $__PAGE_OFFSET, %esp + + /* + * Initialize eflags. Some BIOS's leave bits like NT set. This would +@@ -344,9 +431,7 @@ ENTRY(startup_32_smp) + + #ifdef CONFIG_SMP + cmpb $0, ready +- jz 1f /* Initial CPU cleans BSS */ +- jmp checkCPUtype +-1: ++ jnz checkCPUtype + #endif /* CONFIG_SMP */ + + /* +@@ -424,7 +509,7 @@ is386: movl $2,%ecx # set MP + 1: movl $(__KERNEL_DS),%eax # reload all the segment registers + movl %eax,%ss # after changing gdt. + +- movl $(__USER_DS),%eax # DS/ES contains default USER segment ++# movl $(__KERNEL_DS),%eax # DS/ES contains default KERNEL segment + movl %eax,%ds + movl %eax,%es + +@@ -438,15 +523,22 @@ is386: movl $2,%ecx # set MP + */ + cmpb $0,ready + jne 1f +- movl $per_cpu__gdt_page,%eax ++ movl $cpu_gdt_table,%eax + movl $per_cpu__stack_canary,%ecx ++#ifdef CONFIG_SMP ++ addl $__per_cpu_load,%ecx ++#endif + movw %cx, 8 * GDT_ENTRY_STACK_CANARY + 2(%eax) + shrl $16, %ecx + movb %cl, 8 * GDT_ENTRY_STACK_CANARY + 4(%eax) + movb %ch, 8 * GDT_ENTRY_STACK_CANARY + 7(%eax) + 1: +-#endif + movl $(__KERNEL_STACK_CANARY),%eax ++#elif defined(CONFIG_PAX_MEMORY_UDEREF) ++ movl $(__USER_DS),%eax ++#else ++ xorl %eax,%eax ++#endif + movl %eax,%gs + + xorl %eax,%eax # Clear LDT +@@ -454,14 +546,7 @@ is386: movl $2,%ecx # set MP + + cld # gcc2 wants the direction flag cleared at all times + pushl $0 # fake return address for unwinder +-#ifdef CONFIG_SMP +- movb ready, %cl + movb $1, ready +- cmpb $0,%cl # the first CPU calls start_kernel +- je 1f +- movl (stack_start), %esp +-1: +-#endif /* CONFIG_SMP */ + jmp *(initial_code) + + /* +@@ -546,22 +631,22 @@ early_page_fault: + jmp early_fault + + early_fault: +- cld + #ifdef CONFIG_PRINTK ++ cmpl $1,%ss:early_recursion_flag ++ je hlt_loop ++ incl %ss:early_recursion_flag ++ cld + pusha + movl $(__KERNEL_DS),%eax + movl %eax,%ds + movl %eax,%es +- cmpl $2,early_recursion_flag +- je hlt_loop +- incl early_recursion_flag + movl %cr2,%eax + pushl %eax + pushl %edx /* trapno */ + pushl $fault_msg + call printk ++; call dump_stack + #endif +- call dump_stack + hlt_loop: + hlt + jmp hlt_loop +@@ -569,8 +654,11 @@ hlt_loop: + /* This is the default interrupt "handler" :-) */ + ALIGN + ignore_int: +- cld + #ifdef CONFIG_PRINTK ++ cmpl $2,%ss:early_recursion_flag ++ je hlt_loop ++ incl %ss:early_recursion_flag ++ cld + pushl %eax + pushl %ecx + pushl %edx +@@ -579,9 +667,6 @@ ignore_int: + movl $(__KERNEL_DS),%eax + movl %eax,%ds + movl %eax,%es +- cmpl $2,early_recursion_flag +- je hlt_loop +- incl early_recursion_flag + pushl 16(%esp) + pushl 24(%esp) + pushl 32(%esp) +@@ -600,6 +685,8 @@ ignore_int: + #endif + iret + ++#include "verify_cpu.S" ++ + __REFDATA + .align 4 + ENTRY(initial_code) +@@ -610,31 +697,47 @@ ENTRY(initial_page_table) + /* + * BSS section + */ +-__PAGE_ALIGNED_BSS +- .align PAGE_SIZE_asm + #ifdef CONFIG_X86_PAE ++.section .swapper_pg_pmd,"a",@progbits + swapper_pg_pmd: + .fill 1024*KPMDS,4,0 + #else ++.section .swapper_pg_dir,"a",@progbits + ENTRY(swapper_pg_dir) + .fill 1024,4,0 + #endif ++.section .swapper_pg_fixmap,"a",@progbits + swapper_pg_fixmap: + .fill 1024,4,0 + #ifdef CONFIG_X86_TRAMPOLINE ++.section .trampoline_pg_dir,"a",@progbits + ENTRY(trampoline_pg_dir) ++#ifdef CONFIG_X86_PAE ++ .fill 4,8,0 ++#else + .fill 1024,4,0 + #endif ++#endif ++ ++.section .empty_zero_page,"a",@progbits + ENTRY(empty_zero_page) + .fill 4096,1,0 + + /* ++ * The IDT has to be page-aligned to simplify the Pentium ++ * F0 0F bug workaround.. We have a special link segment ++ * for this. ++ */ ++.section .idt,"a",@progbits ++ENTRY(idt_table) ++ .fill 256,8,0 ++ ++/* + * This starts the data section. + */ + #ifdef CONFIG_X86_PAE +-__PAGE_ALIGNED_DATA +- /* Page-aligned for the benefit of paravirt? */ +- .align PAGE_SIZE_asm ++.section .swapper_pg_dir,"a",@progbits ++ + ENTRY(swapper_pg_dir) + .long pa(swapper_pg_pmd+PGD_IDENT_ATTR),0 /* low identity map */ + # if KPMDS == 3 +@@ -653,15 +756,24 @@ ENTRY(swapper_pg_dir) + # error "Kernel PMDs should be 1, 2 or 3" + # endif + .align PAGE_SIZE_asm /* needs to be page-sized too */ ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ENTRY(cpu_pgd) ++ .rept NR_CPUS ++ .fill 4,8,0 ++ .endr ++#endif ++ + #endif + + .data ++.balign 4 + ENTRY(stack_start) +- .long init_thread_union+THREAD_SIZE +- .long __BOOT_DS ++ .long init_thread_union+THREAD_SIZE-8 + + ready: .byte 0 + ++.section .rodata,"a",@progbits + early_recursion_flag: + .long 0 + +@@ -697,7 +809,7 @@ fault_msg: + .word 0 # 32 bit align gdt_desc.address + boot_gdt_descr: + .word __BOOT_DS+7 +- .long boot_gdt - __PAGE_OFFSET ++ .long pa(boot_gdt) + + .word 0 # 32-bit align idt_desc.address + idt_descr: +@@ -708,7 +820,7 @@ idt_descr: + .word 0 # 32 bit align gdt_desc.address + ENTRY(early_gdt_descr) + .word GDT_ENTRIES*8-1 +- .long per_cpu__gdt_page /* Overwritten for secondary CPUs */ ++ .long cpu_gdt_table /* Overwritten for secondary CPUs */ + + /* + * The boot_gdt must mirror the equivalent in setup.S and is +@@ -717,5 +829,65 @@ ENTRY(early_gdt_descr) + .align L1_CACHE_BYTES + ENTRY(boot_gdt) + .fill GDT_ENTRY_BOOT_CS,8,0 +- .quad 0x00cf9a000000ffff /* kernel 4GB code at 0x00000000 */ +- .quad 0x00cf92000000ffff /* kernel 4GB data at 0x00000000 */ ++ .quad 0x00cf9b000000ffff /* kernel 4GB code at 0x00000000 */ ++ .quad 0x00cf93000000ffff /* kernel 4GB data at 0x00000000 */ ++ ++ .align PAGE_SIZE_asm ++ENTRY(cpu_gdt_table) ++ .rept NR_CPUS ++ .quad 0x0000000000000000 /* NULL descriptor */ ++ .quad 0x0000000000000000 /* 0x0b reserved */ ++ .quad 0x0000000000000000 /* 0x13 reserved */ ++ .quad 0x0000000000000000 /* 0x1b reserved */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ .quad 0x00cf9b000000ffff /* 0x20 alternate kernel 4GB code at 0x00000000 */ ++#else ++ .quad 0x0000000000000000 /* 0x20 unused */ ++#endif ++ ++ .quad 0x0000000000000000 /* 0x28 unused */ ++ .quad 0x0000000000000000 /* 0x33 TLS entry 1 */ ++ .quad 0x0000000000000000 /* 0x3b TLS entry 2 */ ++ .quad 0x0000000000000000 /* 0x43 TLS entry 3 */ ++ .quad 0x0000000000000000 /* 0x4b reserved */ ++ .quad 0x0000000000000000 /* 0x53 reserved */ ++ .quad 0x0000000000000000 /* 0x5b reserved */ ++ ++ .quad 0x00cf9b000000ffff /* 0x60 kernel 4GB code at 0x00000000 */ ++ .quad 0x00cf93000000ffff /* 0x68 kernel 4GB data at 0x00000000 */ ++ .quad 0x00cffb000000ffff /* 0x73 user 4GB code at 0x00000000 */ ++ .quad 0x00cff3000000ffff /* 0x7b user 4GB data at 0x00000000 */ ++ ++ .quad 0x0000000000000000 /* 0x80 TSS descriptor */ ++ .quad 0x0000000000000000 /* 0x88 LDT descriptor */ ++ ++ /* ++ * Segments used for calling PnP BIOS have byte granularity. ++ * The code segments and data segments have fixed 64k limits, ++ * the transfer segment sizes are set at run time. ++ */ ++ .quad 0x00409b000000ffff /* 0x90 32-bit code */ ++ .quad 0x00009b000000ffff /* 0x98 16-bit code */ ++ .quad 0x000093000000ffff /* 0xa0 16-bit data */ ++ .quad 0x0000930000000000 /* 0xa8 16-bit data */ ++ .quad 0x0000930000000000 /* 0xb0 16-bit data */ ++ ++ /* ++ * The APM segments have byte granularity and their bases ++ * are set at run time. All have 64k limits. ++ */ ++ .quad 0x00409b000000ffff /* 0xb8 APM CS code */ ++ .quad 0x00009b000000ffff /* 0xc0 APM CS 16 code (16 bit) */ ++ .quad 0x004093000000ffff /* 0xc8 APM DS data */ ++ ++ .quad 0x00c0930000000000 /* 0xd0 - ESPFIX SS */ ++ .quad 0x0040930000000000 /* 0xd8 - PERCPU */ ++ .quad 0x0040910000000017 /* 0xe0 - STACK_CANARY */ ++ .quad 0x0000000000000000 /* 0xe8 - PCIBIOS_CS */ ++ .quad 0x0000000000000000 /* 0xf0 - PCIBIOS_DS */ ++ .quad 0x0000000000000000 /* 0xf8 - GDT entry 31: double-fault TSS */ ++ ++ /* Be sure this is zeroed to avoid false validations in Xen */ ++ .fill PAGE_SIZE_asm - GDT_SIZE,1,0 ++ .endr +diff --git a/arch/x86/kernel/head_64.S b/arch/x86/kernel/head_64.S +index 780cd92..758b2a6 100644 +--- a/arch/x86/kernel/head_64.S ++++ b/arch/x86/kernel/head_64.S +@@ -19,6 +19,8 @@ + #include + #include + #include ++#include ++#include + + #ifdef CONFIG_PARAVIRT + #include +@@ -38,6 +40,12 @@ L4_PAGE_OFFSET = pgd_index(__PAGE_OFFSET) + L3_PAGE_OFFSET = pud_index(__PAGE_OFFSET) + L4_START_KERNEL = pgd_index(__START_KERNEL_map) + L3_START_KERNEL = pud_index(__START_KERNEL_map) ++L4_VMALLOC_START = pgd_index(VMALLOC_START) ++L3_VMALLOC_START = pud_index(VMALLOC_START) ++L4_VMALLOC_END = pgd_index(VMALLOC_END) ++L3_VMALLOC_END = pud_index(VMALLOC_END) ++L4_VMEMMAP_START = pgd_index(VMEMMAP_START) ++L3_VMEMMAP_START = pud_index(VMEMMAP_START) + + .text + __HEAD +@@ -85,35 +93,23 @@ startup_64: + */ + addq %rbp, init_level4_pgt + 0(%rip) + addq %rbp, init_level4_pgt + (L4_PAGE_OFFSET*8)(%rip) ++ addq %rbp, init_level4_pgt + (L4_VMALLOC_START*8)(%rip) ++ addq %rbp, init_level4_pgt + (L4_VMALLOC_END*8)(%rip) ++ addq %rbp, init_level4_pgt + (L4_VMEMMAP_START*8)(%rip) + addq %rbp, init_level4_pgt + (L4_START_KERNEL*8)(%rip) + + addq %rbp, level3_ident_pgt + 0(%rip) ++#ifndef CONFIG_XEN ++ addq %rbp, level3_ident_pgt + 8(%rip) ++#endif + +- addq %rbp, level3_kernel_pgt + (510*8)(%rip) +- addq %rbp, level3_kernel_pgt + (511*8)(%rip) ++ addq %rbp, level3_vmemmap_pgt + (L3_VMEMMAP_START*8)(%rip) ++ ++ addq %rbp, level3_kernel_pgt + (L3_START_KERNEL*8)(%rip) ++ addq %rbp, level3_kernel_pgt + (L3_START_KERNEL*8+8)(%rip) + + addq %rbp, level2_fixmap_pgt + (506*8)(%rip) +- +- /* Add an Identity mapping if I am above 1G */ +- leaq _text(%rip), %rdi +- andq $PMD_PAGE_MASK, %rdi +- +- movq %rdi, %rax +- shrq $PUD_SHIFT, %rax +- andq $(PTRS_PER_PUD - 1), %rax +- jz ident_complete +- +- leaq (level2_spare_pgt - __START_KERNEL_map + _KERNPG_TABLE)(%rbp), %rdx +- leaq level3_ident_pgt(%rip), %rbx +- movq %rdx, 0(%rbx, %rax, 8) +- +- movq %rdi, %rax +- shrq $PMD_SHIFT, %rax +- andq $(PTRS_PER_PMD - 1), %rax +- leaq __PAGE_KERNEL_IDENT_LARGE_EXEC(%rdi), %rdx +- leaq level2_spare_pgt(%rip), %rbx +- movq %rdx, 0(%rbx, %rax, 8) +-ident_complete: ++ addq %rbp, level2_fixmap_pgt + (507*8)(%rip) + + /* + * Fixup the kernel text+data virtual addresses. Note that +@@ -161,8 +157,8 @@ ENTRY(secondary_startup_64) + * after the boot processor executes this code. + */ + +- /* Enable PAE mode and PGE */ +- movl $(X86_CR4_PAE | X86_CR4_PGE), %eax ++ /* Enable PAE mode and PSE/PGE */ ++ movl $(X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE), %eax + movq %rax, %cr4 + + /* Setup early boot stage 4 level pagetables. */ +@@ -184,9 +180,16 @@ ENTRY(secondary_startup_64) + movl $MSR_EFER, %ecx + rdmsr + btsl $_EFER_SCE, %eax /* Enable System Call */ +- btl $20,%edi /* No Execute supported? */ ++ btl $(X86_FEATURE_NX & 31),%edi /* No Execute supported? */ + jnc 1f + btsl $_EFER_NX, %eax ++ leaq init_level4_pgt(%rip), %rdi ++#ifndef CONFIG_EFI ++ btsq $_PAGE_BIT_NX, 8*L4_PAGE_OFFSET(%rdi) ++#endif ++ btsq $_PAGE_BIT_NX, 8*L4_VMALLOC_START(%rdi) ++ btsq $_PAGE_BIT_NX, 8*L4_VMALLOC_END(%rdi) ++ btsq $_PAGE_BIT_NX, 8*L4_VMEMMAP_START(%rdi) + 1: wrmsr /* Make changes effective */ + + /* Setup cr0 */ +@@ -249,6 +252,7 @@ ENTRY(secondary_startup_64) + * jump. In addition we need to ensure %cs is set so we make this + * a far return. + */ ++ pax_set_fptr_mask + movq initial_code(%rip),%rax + pushq $0 # fake return address to stop unwinder + pushq $__KERNEL_CS # set correct cs +@@ -262,16 +266,16 @@ ENTRY(secondary_startup_64) + .quad x86_64_start_kernel + ENTRY(initial_gs) + .quad INIT_PER_CPU_VAR(irq_stack_union) +- __FINITDATA + + ENTRY(stack_start) + .quad init_thread_union+THREAD_SIZE-8 + .word 0 ++ __FINITDATA + + bad_address: + jmp bad_address + +- .section ".init.text","ax" ++ __INIT + #ifdef CONFIG_EARLY_PRINTK + .globl early_idt_handlers + early_idt_handlers: +@@ -316,18 +320,23 @@ ENTRY(early_idt_handler) + #endif /* EARLY_PRINTK */ + 1: hlt + jmp 1b ++ .previous + + #ifdef CONFIG_EARLY_PRINTK ++ __INITDATA + early_recursion_flag: + .long 0 ++ .previous + ++ .section .rodata,"a",@progbits + early_idt_msg: + .asciz "PANIC: early exception %02lx rip %lx:%lx error %lx cr2 %lx\n" + early_idt_ripmsg: + .asciz "RIP %s\n" ++ .previous + #endif /* CONFIG_EARLY_PRINTK */ +- .previous + ++ .section .rodata,"a",@progbits + #define NEXT_PAGE(name) \ + .balign PAGE_SIZE; \ + ENTRY(name) +@@ -350,13 +359,41 @@ NEXT_PAGE(init_level4_pgt) + .quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE + .org init_level4_pgt + L4_PAGE_OFFSET*8, 0 + .quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE ++ .org init_level4_pgt + L4_VMALLOC_START*8, 0 ++ .quad level3_vmalloc_start_pgt - __START_KERNEL_map + _KERNPG_TABLE ++ .org init_level4_pgt + L4_VMALLOC_END*8, 0 ++ .quad level3_vmalloc_end_pgt - __START_KERNEL_map + _KERNPG_TABLE ++ .org init_level4_pgt + L4_VMEMMAP_START*8, 0 ++ .quad level3_vmemmap_pgt - __START_KERNEL_map + _KERNPG_TABLE + .org init_level4_pgt + L4_START_KERNEL*8, 0 + /* (2^48-(2*1024*1024*1024))/(2^39) = 511 */ + .quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE + ++#ifdef CONFIG_PAX_PER_CPU_PGD ++NEXT_PAGE(cpu_pgd) ++ .rept NR_CPUS ++ .fill 512,8,0 ++ .endr ++#endif ++ + NEXT_PAGE(level3_ident_pgt) + .quad level2_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE ++#ifdef CONFIG_XEN + .fill 511,8,0 ++#else ++ .quad level2_ident_pgt + PAGE_SIZE - __START_KERNEL_map + _KERNPG_TABLE ++ .fill 510,8,0 ++#endif ++ ++NEXT_PAGE(level3_vmalloc_start_pgt) ++ .fill 512,8,0 ++ ++NEXT_PAGE(level3_vmalloc_end_pgt) ++ .fill 512,8,0 ++ ++NEXT_PAGE(level3_vmemmap_pgt) ++ .fill L3_VMEMMAP_START,8,0 ++ .quad level2_vmemmap_pgt - __START_KERNEL_map + _KERNPG_TABLE + + NEXT_PAGE(level3_kernel_pgt) + .fill L3_START_KERNEL,8,0 +@@ -364,20 +401,23 @@ NEXT_PAGE(level3_kernel_pgt) + .quad level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE + .quad level2_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE + ++NEXT_PAGE(level2_vmemmap_pgt) ++ .fill 512,8,0 ++ + NEXT_PAGE(level2_fixmap_pgt) +- .fill 506,8,0 +- .quad level1_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE +- /* 8MB reserved for vsyscalls + a 2MB hole = 4 + 1 entries */ +- .fill 5,8,0 ++ .fill 507,8,0 ++ .quad level1_vsyscall_pgt - __START_KERNEL_map + _PAGE_TABLE ++ /* 6MB reserved for vsyscalls + a 2MB hole = 3 + 1 entries */ ++ .fill 4,8,0 + +-NEXT_PAGE(level1_fixmap_pgt) ++NEXT_PAGE(level1_vsyscall_pgt) + .fill 512,8,0 + +-NEXT_PAGE(level2_ident_pgt) +- /* Since I easily can, map the first 1G. ++ /* Since I easily can, map the first 2G. + * Don't set NX because code runs from these pages. + */ +- PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, PTRS_PER_PMD) ++NEXT_PAGE(level2_ident_pgt) ++ PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, 2*PTRS_PER_PMD) + + NEXT_PAGE(level2_kernel_pgt) + /* +@@ -390,33 +430,55 @@ NEXT_PAGE(level2_kernel_pgt) + * If you want to increase this then increase MODULES_VADDR + * too.) + */ +- PMDS(0, __PAGE_KERNEL_LARGE_EXEC, +- KERNEL_IMAGE_SIZE/PMD_SIZE) +- +-NEXT_PAGE(level2_spare_pgt) +- .fill 512, 8, 0 ++ PMDS(0, __PAGE_KERNEL_LARGE_EXEC, KERNEL_IMAGE_SIZE/PMD_SIZE) + + #undef PMDS + #undef NEXT_PAGE + +- .data ++ .align PAGE_SIZE ++ENTRY(cpu_gdt_table) ++ .rept NR_CPUS ++ .quad 0x0000000000000000 /* NULL descriptor */ ++ .quad 0x00cf9b000000ffff /* __KERNEL32_CS */ ++ .quad 0x00af9b000000ffff /* __KERNEL_CS */ ++ .quad 0x00cf93000000ffff /* __KERNEL_DS */ ++ .quad 0x00cffb000000ffff /* __USER32_CS */ ++ .quad 0x00cff3000000ffff /* __USER_DS, __USER32_DS */ ++ .quad 0x00affb000000ffff /* __USER_CS */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ .quad 0x00af9b000000ffff /* __KERNEXEC_KERNEL_CS */ ++#else ++ .quad 0x0 /* unused */ ++#endif ++ ++ .quad 0,0 /* TSS */ ++ .quad 0,0 /* LDT */ ++ .quad 0,0,0 /* three TLS descriptors */ ++ .quad 0x0000f40000000000 /* node/CPU stored in limit */ ++ /* asm/segment.h:GDT_ENTRIES must match this */ ++ ++ /* zero the remaining page */ ++ .fill PAGE_SIZE / 8 - GDT_ENTRIES,8,0 ++ .endr ++ + .align 16 + .globl early_gdt_descr + early_gdt_descr: + .word GDT_ENTRIES*8-1 + early_gdt_descr_base: +- .quad INIT_PER_CPU_VAR(gdt_page) ++ .quad cpu_gdt_table + + ENTRY(phys_base) + /* This must match the first entry in level2_kernel_pgt */ + .quad 0x0000000000000000 + + #include "../../x86/xen/xen-head.S" +- +- .section .bss, "aw", @nobits ++ ++ .section .rodata,"a",@progbits + .align L1_CACHE_BYTES + ENTRY(idt_table) +- .skip IDT_ENTRIES * 16 ++ .fill 512,8,0 + + __PAGE_ALIGNED_BSS + .align PAGE_SIZE +diff --git a/arch/x86/kernel/i386_ksyms_32.c b/arch/x86/kernel/i386_ksyms_32.c +index 9c3bd4a..e1d9b35 100644 +--- a/arch/x86/kernel/i386_ksyms_32.c ++++ b/arch/x86/kernel/i386_ksyms_32.c +@@ -20,8 +20,12 @@ extern void cmpxchg8b_emu(void); + EXPORT_SYMBOL(cmpxchg8b_emu); + #endif + ++EXPORT_SYMBOL_GPL(cpu_gdt_table); ++ + /* Networking helper routines. */ + EXPORT_SYMBOL(csum_partial_copy_generic); ++EXPORT_SYMBOL(csum_partial_copy_generic_to_user); ++EXPORT_SYMBOL(csum_partial_copy_generic_from_user); + + EXPORT_SYMBOL(__get_user_1); + EXPORT_SYMBOL(__get_user_2); +@@ -36,3 +40,7 @@ EXPORT_SYMBOL(strstr); + + EXPORT_SYMBOL(csum_partial); + EXPORT_SYMBOL(empty_zero_page); ++ ++#ifdef CONFIG_PAX_KERNEXEC ++EXPORT_SYMBOL(__LOAD_PHYSICAL_ADDR); ++#endif +diff --git a/arch/x86/kernel/i8259.c b/arch/x86/kernel/i8259.c +index df89102..a244320 100644 +--- a/arch/x86/kernel/i8259.c ++++ b/arch/x86/kernel/i8259.c +@@ -208,7 +208,7 @@ spurious_8259A_irq: + "spurious 8259A interrupt: IRQ%d.\n", irq); + spurious_irq_mask |= irqmask; + } +- atomic_inc(&irq_err_count); ++ atomic_inc_unchecked(&irq_err_count); + /* + * Theoretically we do not have to handle this IRQ, + * but in Linux this does not cause problems and is +diff --git a/arch/x86/kernel/init_task.c b/arch/x86/kernel/init_task.c +index 3a54dcb..1c22348 100644 +--- a/arch/x86/kernel/init_task.c ++++ b/arch/x86/kernel/init_task.c +@@ -20,8 +20,7 @@ static struct sighand_struct init_sighand = INIT_SIGHAND(init_sighand); + * way process stacks are handled. This is done by having a special + * "init_task" linker map entry.. + */ +-union thread_union init_thread_union __init_task_data = +- { INIT_THREAD_INFO(init_task) }; ++union thread_union init_thread_union __init_task_data; + + /* + * Initial task structure. +@@ -38,5 +37,5 @@ EXPORT_SYMBOL(init_task); + * section. Since TSS's are completely CPU-local, we want them + * on exact cacheline boundaries, to eliminate cacheline ping-pong. + */ +-DEFINE_PER_CPU_SHARED_ALIGNED(struct tss_struct, init_tss) = INIT_TSS; +- ++struct tss_struct init_tss[NR_CPUS] ____cacheline_internodealigned_in_smp = { [0 ... NR_CPUS-1] = INIT_TSS }; ++EXPORT_SYMBOL(init_tss); +diff --git a/arch/x86/kernel/ioport.c b/arch/x86/kernel/ioport.c +index 99c4d30..74c84e9 100644 +--- a/arch/x86/kernel/ioport.c ++++ b/arch/x86/kernel/ioport.c +@@ -6,6 +6,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -41,6 +42,12 @@ asmlinkage long sys_ioperm(unsigned long from, unsigned long num, int turn_on) + + if ((from + num <= from) || (from + num > IO_BITMAP_BITS)) + return -EINVAL; ++#ifdef CONFIG_GRKERNSEC_IO ++ if (turn_on && grsec_disable_privio) { ++ gr_handle_ioperm(); ++ return -EPERM; ++ } ++#endif + if (turn_on && !capable(CAP_SYS_RAWIO)) + return -EPERM; + +@@ -67,7 +74,7 @@ asmlinkage long sys_ioperm(unsigned long from, unsigned long num, int turn_on) + * because the ->io_bitmap_max value must match the bitmap + * contents: + */ +- tss = &per_cpu(init_tss, get_cpu()); ++ tss = init_tss + get_cpu(); + + set_bitmap(t->io_bitmap_ptr, from, num, !turn_on); + +@@ -111,6 +118,12 @@ static int do_iopl(unsigned int level, struct pt_regs *regs) + return -EINVAL; + /* Trying to gain more privileges? */ + if (level > old) { ++#ifdef CONFIG_GRKERNSEC_IO ++ if (grsec_disable_privio) { ++ gr_handle_iopl(); ++ return -EPERM; ++ } ++#endif + if (!capable(CAP_SYS_RAWIO)) + return -EPERM; + } +diff --git a/arch/x86/kernel/irq.c b/arch/x86/kernel/irq.c +index 04bbd52..83a07d9 100644 +--- a/arch/x86/kernel/irq.c ++++ b/arch/x86/kernel/irq.c +@@ -15,7 +15,7 @@ + #include + #include + +-atomic_t irq_err_count; ++atomic_unchecked_t irq_err_count; + + /* Function pointer for generic interrupt vector handling */ + void (*generic_interrupt_extension)(void) = NULL; +@@ -114,9 +114,9 @@ static int show_other_interrupts(struct seq_file *p, int prec) + seq_printf(p, "%10u ", per_cpu(mce_poll_count, j)); + seq_printf(p, " Machine check polls\n"); + #endif +- seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read(&irq_err_count)); ++ seq_printf(p, "%*s: %10u\n", prec, "ERR", atomic_read_unchecked(&irq_err_count)); + #if defined(CONFIG_X86_IO_APIC) +- seq_printf(p, "%*s: %10u\n", prec, "MIS", atomic_read(&irq_mis_count)); ++ seq_printf(p, "%*s: %10u\n", prec, "MIS", atomic_read_unchecked(&irq_mis_count)); + #endif + return 0; + } +@@ -209,10 +209,10 @@ u64 arch_irq_stat_cpu(unsigned int cpu) + + u64 arch_irq_stat(void) + { +- u64 sum = atomic_read(&irq_err_count); ++ u64 sum = atomic_read_unchecked(&irq_err_count); + + #ifdef CONFIG_X86_IO_APIC +- sum += atomic_read(&irq_mis_count); ++ sum += atomic_read_unchecked(&irq_mis_count); + #endif + return sum; + } +diff --git a/arch/x86/kernel/irq_32.c b/arch/x86/kernel/irq_32.c +index 7d35d0f..03f1d52 100644 +--- a/arch/x86/kernel/irq_32.c ++++ b/arch/x86/kernel/irq_32.c +@@ -35,7 +35,7 @@ static int check_stack_overflow(void) + __asm__ __volatile__("andl %%esp,%0" : + "=r" (sp) : "0" (THREAD_SIZE - 1)); + +- return sp < (sizeof(struct thread_info) + STACK_WARN); ++ return sp < STACK_WARN; + } + + static void print_stack_overflow(void) +@@ -54,9 +54,9 @@ static inline void print_stack_overflow(void) { } + * per-CPU IRQ handling contexts (thread information and stack) + */ + union irq_ctx { +- struct thread_info tinfo; +- u32 stack[THREAD_SIZE/sizeof(u32)]; +-} __attribute__((aligned(PAGE_SIZE))); ++ unsigned long previous_esp; ++ u32 stack[THREAD_SIZE/sizeof(u32)]; ++} __attribute__((aligned(THREAD_SIZE))); + + static DEFINE_PER_CPU(union irq_ctx *, hardirq_ctx); + static DEFINE_PER_CPU(union irq_ctx *, softirq_ctx); +@@ -78,10 +78,9 @@ static void call_on_stack(void *func, void *stack) + static inline int + execute_on_irq_stack(int overflow, struct irq_desc *desc, int irq) + { +- union irq_ctx *curctx, *irqctx; ++ union irq_ctx *irqctx; + u32 *isp, arg1, arg2; + +- curctx = (union irq_ctx *) current_thread_info(); + irqctx = __get_cpu_var(hardirq_ctx); + + /* +@@ -90,21 +89,16 @@ execute_on_irq_stack(int overflow, struct irq_desc *desc, int irq) + * handler) we can't do that and just have to keep using the + * current stack (which is the irq stack already after all) + */ +- if (unlikely(curctx == irqctx)) ++ if (unlikely((void *)current_stack_pointer - (void *)irqctx < THREAD_SIZE)) + return 0; + + /* build the stack frame on the IRQ stack */ +- isp = (u32 *) ((char *)irqctx + sizeof(*irqctx)); +- irqctx->tinfo.task = curctx->tinfo.task; +- irqctx->tinfo.previous_esp = current_stack_pointer; ++ isp = (u32 *) ((char *)irqctx + sizeof(*irqctx) - 8); ++ irqctx->previous_esp = current_stack_pointer; + +- /* +- * Copy the softirq bits in preempt_count so that the +- * softirq checks work in the hardirq context. +- */ +- irqctx->tinfo.preempt_count = +- (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | +- (curctx->tinfo.preempt_count & SOFTIRQ_MASK); ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ __set_fs(MAKE_MM_SEG(0)); ++#endif + + if (unlikely(overflow)) + call_on_stack(print_stack_overflow, isp); +@@ -116,6 +110,11 @@ execute_on_irq_stack(int overflow, struct irq_desc *desc, int irq) + : "0" (irq), "1" (desc), "2" (isp), + "D" (desc->handle_irq) + : "memory", "cc", "ecx"); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ __set_fs(current_thread_info()->addr_limit); ++#endif ++ + return 1; + } + +@@ -124,28 +123,11 @@ execute_on_irq_stack(int overflow, struct irq_desc *desc, int irq) + */ + void __cpuinit irq_ctx_init(int cpu) + { +- union irq_ctx *irqctx; +- + if (per_cpu(hardirq_ctx, cpu)) + return; + +- irqctx = &per_cpu(hardirq_stack, cpu); +- irqctx->tinfo.task = NULL; +- irqctx->tinfo.exec_domain = NULL; +- irqctx->tinfo.cpu = cpu; +- irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; +- irqctx->tinfo.addr_limit = MAKE_MM_SEG(0); +- +- per_cpu(hardirq_ctx, cpu) = irqctx; +- +- irqctx = &per_cpu(softirq_stack, cpu); +- irqctx->tinfo.task = NULL; +- irqctx->tinfo.exec_domain = NULL; +- irqctx->tinfo.cpu = cpu; +- irqctx->tinfo.preempt_count = 0; +- irqctx->tinfo.addr_limit = MAKE_MM_SEG(0); +- +- per_cpu(softirq_ctx, cpu) = irqctx; ++ per_cpu(hardirq_ctx, cpu) = &per_cpu(hardirq_stack, cpu); ++ per_cpu(softirq_ctx, cpu) = &per_cpu(softirq_stack, cpu); + + printk(KERN_DEBUG "CPU %u irqstacks, hard=%p soft=%p\n", + cpu, per_cpu(hardirq_ctx, cpu), per_cpu(softirq_ctx, cpu)); +@@ -159,7 +141,6 @@ void irq_ctx_exit(int cpu) + asmlinkage void do_softirq(void) + { + unsigned long flags; +- struct thread_info *curctx; + union irq_ctx *irqctx; + u32 *isp; + +@@ -169,15 +150,22 @@ asmlinkage void do_softirq(void) + local_irq_save(flags); + + if (local_softirq_pending()) { +- curctx = current_thread_info(); + irqctx = __get_cpu_var(softirq_ctx); +- irqctx->tinfo.task = curctx->task; +- irqctx->tinfo.previous_esp = current_stack_pointer; ++ irqctx->previous_esp = current_stack_pointer; + + /* build the stack frame on the softirq stack */ +- isp = (u32 *) ((char *)irqctx + sizeof(*irqctx)); ++ isp = (u32 *) ((char *)irqctx + sizeof(*irqctx) - 8); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ __set_fs(MAKE_MM_SEG(0)); ++#endif + + call_on_stack(__do_softirq, isp); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ __set_fs(current_thread_info()->addr_limit); ++#endif ++ + /* + * Shouldnt happen, we returned above if in_interrupt(): + */ +diff --git a/arch/x86/kernel/kgdb.c b/arch/x86/kernel/kgdb.c +index 8d82a77..0baf312 100644 +--- a/arch/x86/kernel/kgdb.c ++++ b/arch/x86/kernel/kgdb.c +@@ -390,13 +390,13 @@ int kgdb_arch_handle_exception(int e_vector, int signo, int err_code, + + /* clear the trace bit */ + linux_regs->flags &= ~X86_EFLAGS_TF; +- atomic_set(&kgdb_cpu_doing_single_step, -1); ++ atomic_set_unchecked(&kgdb_cpu_doing_single_step, -1); + + /* set the trace bit if we're stepping */ + if (remcomInBuffer[0] == 's') { + linux_regs->flags |= X86_EFLAGS_TF; + kgdb_single_step = 1; +- atomic_set(&kgdb_cpu_doing_single_step, ++ atomic_set_unchecked(&kgdb_cpu_doing_single_step, + raw_smp_processor_id()); + } + +@@ -476,7 +476,7 @@ static int __kgdb_notify(struct die_args *args, unsigned long cmd) + break; + + case DIE_DEBUG: +- if (atomic_read(&kgdb_cpu_doing_single_step) == ++ if (atomic_read_unchecked(&kgdb_cpu_doing_single_step) == + raw_smp_processor_id()) { + if (user_mode(regs)) + return single_step_cont(regs, args); +@@ -573,7 +573,7 @@ unsigned long kgdb_arch_pc(int exception, struct pt_regs *regs) + return instruction_pointer(regs); + } + +-struct kgdb_arch arch_kgdb_ops = { ++const struct kgdb_arch arch_kgdb_ops = { + /* Breakpoint instruction: */ + .gdb_bpt_instr = { 0xcc }, + .flags = KGDB_HW_BREAKPOINT, +diff --git a/arch/x86/kernel/kprobes.c b/arch/x86/kernel/kprobes.c +index 7a67820..70ea187 100644 +--- a/arch/x86/kernel/kprobes.c ++++ b/arch/x86/kernel/kprobes.c +@@ -168,9 +168,13 @@ static void __kprobes set_jmp_op(void *from, void *to) + char op; + s32 raddr; + } __attribute__((packed)) * jop; +- jop = (struct __arch_jmp_op *)from; ++ ++ jop = (struct __arch_jmp_op *)(ktla_ktva(from)); ++ ++ pax_open_kernel(); + jop->raddr = (s32)((long)(to) - ((long)(from) + 5)); + jop->op = RELATIVEJUMP_INSTRUCTION; ++ pax_close_kernel(); + } + + /* +@@ -195,7 +199,7 @@ static int __kprobes can_boost(kprobe_opcode_t *opcodes) + kprobe_opcode_t opcode; + kprobe_opcode_t *orig_opcodes = opcodes; + +- if (search_exception_tables((unsigned long)opcodes)) ++ if (search_exception_tables(ktva_ktla((unsigned long)opcodes))) + return 0; /* Page fault may occur on this address. */ + + retry: +@@ -339,7 +343,9 @@ static void __kprobes fix_riprel(struct kprobe *p) + disp = (u8 *) p->addr + *((s32 *) insn) - + (u8 *) p->ainsn.insn; + BUG_ON((s64) (s32) disp != disp); /* Sanity check. */ ++ pax_open_kernel(); + *(s32 *)insn = (s32) disp; ++ pax_close_kernel(); + } + } + #endif +@@ -347,16 +353,18 @@ static void __kprobes fix_riprel(struct kprobe *p) + + static void __kprobes arch_copy_kprobe(struct kprobe *p) + { +- memcpy(p->ainsn.insn, p->addr, MAX_INSN_SIZE * sizeof(kprobe_opcode_t)); ++ pax_open_kernel(); ++ memcpy(p->ainsn.insn, ktla_ktva(p->addr), MAX_INSN_SIZE * sizeof(kprobe_opcode_t)); ++ pax_close_kernel(); + + fix_riprel(p); + +- if (can_boost(p->addr)) ++ if (can_boost(ktla_ktva(p->addr))) + p->ainsn.boostable = 0; + else + p->ainsn.boostable = -1; + +- p->opcode = *p->addr; ++ p->opcode = *(ktla_ktva(p->addr)); + } + + int __kprobes arch_prepare_kprobe(struct kprobe *p) +@@ -434,7 +442,7 @@ static void __kprobes prepare_singlestep(struct kprobe *p, struct pt_regs *regs) + if (p->opcode == BREAKPOINT_INSTRUCTION) + regs->ip = (unsigned long)p->addr; + else +- regs->ip = (unsigned long)p->ainsn.insn; ++ regs->ip = ktva_ktla((unsigned long)p->ainsn.insn); + } + + void __kprobes arch_prepare_kretprobe(struct kretprobe_instance *ri, +@@ -455,7 +463,7 @@ static void __kprobes setup_singlestep(struct kprobe *p, struct pt_regs *regs, + if (p->ainsn.boostable == 1 && !p->post_handler) { + /* Boost up -- we can execute copied instructions directly */ + reset_current_kprobe(); +- regs->ip = (unsigned long)p->ainsn.insn; ++ regs->ip = ktva_ktla((unsigned long)p->ainsn.insn); + preempt_enable_no_resched(); + return; + } +@@ -525,7 +533,7 @@ static int __kprobes kprobe_handler(struct pt_regs *regs) + struct kprobe_ctlblk *kcb; + + addr = (kprobe_opcode_t *)(regs->ip - sizeof(kprobe_opcode_t)); +- if (*addr != BREAKPOINT_INSTRUCTION) { ++ if (*(kprobe_opcode_t *)ktla_ktva((unsigned long)addr) != BREAKPOINT_INSTRUCTION) { + /* + * The breakpoint instruction was removed right + * after we hit it. Another cpu has removed +@@ -637,6 +645,9 @@ static void __used __kprobes kretprobe_trampoline_holder(void) + /* Skip orig_ax, ip, cs */ + " addq $24, %rsp\n" + " popfq\n" ++#ifdef KERNEXEC_PLUGIN ++ " btsq $63,(%rsp)\n" ++#endif + #else + " pushf\n" + /* +@@ -777,7 +788,7 @@ static void __kprobes resume_execution(struct kprobe *p, + struct pt_regs *regs, struct kprobe_ctlblk *kcb) + { + unsigned long *tos = stack_addr(regs); +- unsigned long copy_ip = (unsigned long)p->ainsn.insn; ++ unsigned long copy_ip = ktva_ktla((unsigned long)p->ainsn.insn); + unsigned long orig_ip = (unsigned long)p->addr; + kprobe_opcode_t *insn = p->ainsn.insn; + +@@ -960,7 +971,7 @@ int __kprobes kprobe_exceptions_notify(struct notifier_block *self, + struct die_args *args = data; + int ret = NOTIFY_DONE; + +- if (args->regs && user_mode_vm(args->regs)) ++ if (args->regs && user_mode(args->regs)) + return ret; + + switch (val) { +diff --git a/arch/x86/kernel/kvm.c b/arch/x86/kernel/kvm.c +index 63b0ec8..6d92227 100644 +--- a/arch/x86/kernel/kvm.c ++++ b/arch/x86/kernel/kvm.c +@@ -216,6 +216,7 @@ static void __init paravirt_ops_setup(void) + pv_mmu_ops.set_pud = kvm_set_pud; + #if PAGETABLE_LEVELS == 4 + pv_mmu_ops.set_pgd = kvm_set_pgd; ++ pv_mmu_ops.set_pgd_batched = kvm_set_pgd; + #endif + #endif + pv_mmu_ops.flush_tlb_user = kvm_flush_tlb; +diff --git a/arch/x86/kernel/ldt.c b/arch/x86/kernel/ldt.c +index ec6ef60..ab2c824 100644 +--- a/arch/x86/kernel/ldt.c ++++ b/arch/x86/kernel/ldt.c +@@ -66,13 +66,13 @@ static int alloc_ldt(mm_context_t *pc, int mincount, int reload) + if (reload) { + #ifdef CONFIG_SMP + preempt_disable(); +- load_LDT(pc); ++ load_LDT_nolock(pc); + if (!cpumask_equal(mm_cpumask(current->mm), + cpumask_of(smp_processor_id()))) + smp_call_function(flush_ldt, current->mm, 1); + preempt_enable(); + #else +- load_LDT(pc); ++ load_LDT_nolock(pc); + #endif + } + if (oldsize) { +@@ -94,7 +94,7 @@ static inline int copy_ldt(mm_context_t *new, mm_context_t *old) + return err; + + for (i = 0; i < old->size; i++) +- write_ldt_entry(new->ldt, i, old->ldt + i * LDT_ENTRY_SIZE); ++ write_ldt_entry(new->ldt, i, old->ldt + i); + return 0; + } + +@@ -115,6 +115,24 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm) + retval = copy_ldt(&mm->context, &old_mm->context); + mutex_unlock(&old_mm->context.lock); + } ++ ++ if (tsk == current) { ++ mm->context.vdso = 0; ++ ++#ifdef CONFIG_X86_32 ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++ mm->context.user_cs_base = 0UL; ++ mm->context.user_cs_limit = ~0UL; ++ ++#if defined(CONFIG_PAX_PAGEEXEC) && defined(CONFIG_SMP) ++ cpus_clear(mm->context.cpu_user_cs_mask); ++#endif ++ ++#endif ++#endif ++ ++ } ++ + return retval; + } + +@@ -229,6 +247,13 @@ static int write_ldt(void __user *ptr, unsigned long bytecount, int oldmode) + } + } + ++#ifdef CONFIG_PAX_SEGMEXEC ++ if ((mm->pax_flags & MF_PAX_SEGMEXEC) && (ldt_info.contents & MODIFY_LDT_CONTENTS_CODE)) { ++ error = -EINVAL; ++ goto out_unlock; ++ } ++#endif ++ + fill_ldt(&ldt, &ldt_info); + if (oldmode) + ldt.avl = 0; +diff --git a/arch/x86/kernel/machine_kexec_32.c b/arch/x86/kernel/machine_kexec_32.c +index c1c429d..f02eaf9 100644 +--- a/arch/x86/kernel/machine_kexec_32.c ++++ b/arch/x86/kernel/machine_kexec_32.c +@@ -26,7 +26,7 @@ + #include + #include + +-static void set_idt(void *newidt, __u16 limit) ++static void set_idt(struct desc_struct *newidt, __u16 limit) + { + struct desc_ptr curidt; + +@@ -38,7 +38,7 @@ static void set_idt(void *newidt, __u16 limit) + } + + +-static void set_gdt(void *newgdt, __u16 limit) ++static void set_gdt(struct desc_struct *newgdt, __u16 limit) + { + struct desc_ptr curgdt; + +@@ -217,7 +217,7 @@ void machine_kexec(struct kimage *image) + } + + control_page = page_address(image->control_code_page); +- memcpy(control_page, relocate_kernel, KEXEC_CONTROL_CODE_MAX_SIZE); ++ memcpy(control_page, (void *)ktla_ktva((unsigned long)relocate_kernel), KEXEC_CONTROL_CODE_MAX_SIZE); + + relocate_kernel_ptr = control_page; + page_list[PA_CONTROL_PAGE] = __pa(control_page); +diff --git a/arch/x86/kernel/microcode_amd.c b/arch/x86/kernel/microcode_amd.c +index 1e47679..e73449d 100644 +--- a/arch/x86/kernel/microcode_amd.c ++++ b/arch/x86/kernel/microcode_amd.c +@@ -364,7 +364,7 @@ static void microcode_fini_cpu_amd(int cpu) + uci->mc = NULL; + } + +-static struct microcode_ops microcode_amd_ops = { ++static const struct microcode_ops microcode_amd_ops = { + .request_microcode_user = request_microcode_user, + .request_microcode_fw = request_microcode_fw, + .collect_cpu_info = collect_cpu_info_amd, +@@ -372,7 +372,7 @@ static struct microcode_ops microcode_amd_ops = { + .microcode_fini_cpu = microcode_fini_cpu_amd, + }; + +-struct microcode_ops * __init init_amd_microcode(void) ++const struct microcode_ops * __init init_amd_microcode(void) + { + return µcode_amd_ops; + } +diff --git a/arch/x86/kernel/microcode_core.c b/arch/x86/kernel/microcode_core.c +index 378e9a8..b5a6ea9 100644 +--- a/arch/x86/kernel/microcode_core.c ++++ b/arch/x86/kernel/microcode_core.c +@@ -90,7 +90,7 @@ MODULE_LICENSE("GPL"); + + #define MICROCODE_VERSION "2.00" + +-static struct microcode_ops *microcode_ops; ++static const struct microcode_ops *microcode_ops; + + /* + * Synchronization. +diff --git a/arch/x86/kernel/microcode_intel.c b/arch/x86/kernel/microcode_intel.c +index 0d334dd..14cedaf 100644 +--- a/arch/x86/kernel/microcode_intel.c ++++ b/arch/x86/kernel/microcode_intel.c +@@ -443,13 +443,13 @@ static enum ucode_state request_microcode_fw(int cpu, struct device *device) + + static int get_ucode_user(void *to, const void *from, size_t n) + { +- return copy_from_user(to, from, n); ++ return copy_from_user(to, (const void __force_user *)from, n); + } + + static enum ucode_state + request_microcode_user(int cpu, const void __user *buf, size_t size) + { +- return generic_load_microcode(cpu, (void *)buf, size, &get_ucode_user); ++ return generic_load_microcode(cpu, (__force_kernel void *)buf, size, &get_ucode_user); + } + + static void microcode_fini_cpu(int cpu) +@@ -460,7 +460,7 @@ static void microcode_fini_cpu(int cpu) + uci->mc = NULL; + } + +-static struct microcode_ops microcode_intel_ops = { ++static const struct microcode_ops microcode_intel_ops = { + .request_microcode_user = request_microcode_user, + .request_microcode_fw = request_microcode_fw, + .collect_cpu_info = collect_cpu_info, +@@ -468,7 +468,7 @@ static struct microcode_ops microcode_intel_ops = { + .microcode_fini_cpu = microcode_fini_cpu, + }; + +-struct microcode_ops * __init init_intel_microcode(void) ++const struct microcode_ops * __init init_intel_microcode(void) + { + return µcode_intel_ops; + } +diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c +index 89f386f..9028f51 100644 +--- a/arch/x86/kernel/module.c ++++ b/arch/x86/kernel/module.c +@@ -34,7 +34,7 @@ + #define DEBUGP(fmt...) + #endif + +-void *module_alloc(unsigned long size) ++static void *__module_alloc(unsigned long size, pgprot_t prot) + { + struct vm_struct *area; + +@@ -48,8 +48,18 @@ void *module_alloc(unsigned long size) + if (!area) + return NULL; + +- return __vmalloc_area(area, GFP_KERNEL | __GFP_HIGHMEM, +- PAGE_KERNEL_EXEC); ++ return __vmalloc_area(area, GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO, prot); ++} ++ ++void *module_alloc(unsigned long size) ++{ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ return __module_alloc(size, PAGE_KERNEL); ++#else ++ return __module_alloc(size, PAGE_KERNEL_EXEC); ++#endif ++ + } + + /* Free memory returned from module_alloc */ +@@ -58,6 +68,40 @@ void module_free(struct module *mod, void *module_region) + vfree(module_region); + } + ++#ifdef CONFIG_PAX_KERNEXEC ++#ifdef CONFIG_X86_32 ++void *module_alloc_exec(unsigned long size) ++{ ++ struct vm_struct *area; ++ ++ if (size == 0) ++ return NULL; ++ ++ area = __get_vm_area(size, VM_ALLOC, (unsigned long)&MODULES_EXEC_VADDR, (unsigned long)&MODULES_EXEC_END); ++ return area ? area->addr : NULL; ++} ++EXPORT_SYMBOL(module_alloc_exec); ++ ++void module_free_exec(struct module *mod, void *module_region) ++{ ++ vunmap(module_region); ++} ++EXPORT_SYMBOL(module_free_exec); ++#else ++void module_free_exec(struct module *mod, void *module_region) ++{ ++ module_free(mod, module_region); ++} ++EXPORT_SYMBOL(module_free_exec); ++ ++void *module_alloc_exec(unsigned long size) ++{ ++ return __module_alloc(size, PAGE_KERNEL_RX); ++} ++EXPORT_SYMBOL(module_alloc_exec); ++#endif ++#endif ++ + /* We don't need anything special. */ + int module_frob_arch_sections(Elf_Ehdr *hdr, + Elf_Shdr *sechdrs, +@@ -77,14 +121,16 @@ int apply_relocate(Elf32_Shdr *sechdrs, + unsigned int i; + Elf32_Rel *rel = (void *)sechdrs[relsec].sh_addr; + Elf32_Sym *sym; +- uint32_t *location; ++ uint32_t *plocation, location; + + DEBUGP("Applying relocate section %u to %u\n", relsec, + sechdrs[relsec].sh_info); + for (i = 0; i < sechdrs[relsec].sh_size / sizeof(*rel); i++) { + /* This is where to make the change */ +- location = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr +- + rel[i].r_offset; ++ plocation = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr + rel[i].r_offset; ++ location = (uint32_t)plocation; ++ if (sechdrs[sechdrs[relsec].sh_info].sh_flags & SHF_EXECINSTR) ++ plocation = ktla_ktva((void *)plocation); + /* This is the symbol it is referring to. Note that all + undefined symbols have been resolved. */ + sym = (Elf32_Sym *)sechdrs[symindex].sh_addr +@@ -93,11 +139,15 @@ int apply_relocate(Elf32_Shdr *sechdrs, + switch (ELF32_R_TYPE(rel[i].r_info)) { + case R_386_32: + /* We add the value into the location given */ +- *location += sym->st_value; ++ pax_open_kernel(); ++ *plocation += sym->st_value; ++ pax_close_kernel(); + break; + case R_386_PC32: + /* Add the value, subtract its postition */ +- *location += sym->st_value - (uint32_t)location; ++ pax_open_kernel(); ++ *plocation += sym->st_value - location; ++ pax_close_kernel(); + break; + default: + printk(KERN_ERR "module %s: Unknown relocation: %u\n", +@@ -153,21 +203,30 @@ int apply_relocate_add(Elf64_Shdr *sechdrs, + case R_X86_64_NONE: + break; + case R_X86_64_64: ++ pax_open_kernel(); + *(u64 *)loc = val; ++ pax_close_kernel(); + break; + case R_X86_64_32: ++ pax_open_kernel(); + *(u32 *)loc = val; ++ pax_close_kernel(); + if (val != *(u32 *)loc) + goto overflow; + break; + case R_X86_64_32S: ++ pax_open_kernel(); + *(s32 *)loc = val; ++ pax_close_kernel(); + if ((s64)val != *(s32 *)loc) + goto overflow; + break; + case R_X86_64_PC32: + val -= (u64)loc; ++ pax_open_kernel(); + *(u32 *)loc = val; ++ pax_close_kernel(); ++ + #if 0 + if ((s64)val != *(s32 *)loc) + goto overflow; +diff --git a/arch/x86/kernel/paravirt-spinlocks.c b/arch/x86/kernel/paravirt-spinlocks.c +index 3a7c5a4..9191528 100644 +--- a/arch/x86/kernel/paravirt-spinlocks.c ++++ b/arch/x86/kernel/paravirt-spinlocks.c +@@ -13,7 +13,7 @@ default_spin_lock_flags(raw_spinlock_t *lock, unsigned long flags) + __raw_spin_lock(lock); + } + +-struct pv_lock_ops pv_lock_ops = { ++struct pv_lock_ops pv_lock_ops __read_only = { + #ifdef CONFIG_SMP + .spin_is_locked = __ticket_spin_is_locked, + .spin_is_contended = __ticket_spin_is_contended, +diff --git a/arch/x86/kernel/paravirt.c b/arch/x86/kernel/paravirt.c +index 1b1739d..dea6077 100644 +--- a/arch/x86/kernel/paravirt.c ++++ b/arch/x86/kernel/paravirt.c +@@ -53,6 +53,9 @@ u64 _paravirt_ident_64(u64 x) + { + return x; + } ++#if defined(CONFIG_X86_32) && defined(CONFIG_X86_PAE) ++PV_CALLEE_SAVE_REGS_THUNK(_paravirt_ident_64); ++#endif + + void __init default_banner(void) + { +@@ -122,7 +125,7 @@ unsigned paravirt_patch_jmp(void *insnbuf, const void *target, + * corresponding structure. */ + static void *get_call_destination(u8 type) + { +- struct paravirt_patch_template tmpl = { ++ const struct paravirt_patch_template tmpl = { + .pv_init_ops = pv_init_ops, + .pv_time_ops = pv_time_ops, + .pv_cpu_ops = pv_cpu_ops, +@@ -133,6 +136,8 @@ static void *get_call_destination(u8 type) + .pv_lock_ops = pv_lock_ops, + #endif + }; ++ ++ pax_track_stack(); + return *((void **)&tmpl + type); + } + +@@ -145,15 +150,19 @@ unsigned paravirt_patch_default(u8 type, u16 clobbers, void *insnbuf, + if (opfunc == NULL) + /* If there's no function, patch it with a ud2a (BUG) */ + ret = paravirt_patch_insns(insnbuf, len, ud2a, ud2a+sizeof(ud2a)); +- else if (opfunc == _paravirt_nop) ++ else if (opfunc == (void *)_paravirt_nop) + /* If the operation is a nop, then nop the callsite */ + ret = paravirt_patch_nop(); + + /* identity functions just return their single argument */ +- else if (opfunc == _paravirt_ident_32) ++ else if (opfunc == (void *)_paravirt_ident_32) + ret = paravirt_patch_ident_32(insnbuf, len); +- else if (opfunc == _paravirt_ident_64) ++ else if (opfunc == (void *)_paravirt_ident_64) + ret = paravirt_patch_ident_64(insnbuf, len); ++#if defined(CONFIG_X86_32) && defined(CONFIG_X86_PAE) ++ else if (opfunc == (void *)__raw_callee_save__paravirt_ident_64) ++ ret = paravirt_patch_ident_64(insnbuf, len); ++#endif + + else if (type == PARAVIRT_PATCH(pv_cpu_ops.iret) || + type == PARAVIRT_PATCH(pv_cpu_ops.irq_enable_sysexit) || +@@ -178,7 +187,7 @@ unsigned paravirt_patch_insns(void *insnbuf, unsigned len, + if (insn_len > len || start == NULL) + insn_len = len; + else +- memcpy(insnbuf, start, insn_len); ++ memcpy(insnbuf, ktla_ktva(start), insn_len); + + return insn_len; + } +@@ -294,22 +303,22 @@ void arch_flush_lazy_mmu_mode(void) + preempt_enable(); + } + +-struct pv_info pv_info = { ++struct pv_info pv_info __read_only = { + .name = "bare hardware", + .paravirt_enabled = 0, + .kernel_rpl = 0, + .shared_kernel_pmd = 1, /* Only used when CONFIG_X86_PAE is set */ + }; + +-struct pv_init_ops pv_init_ops = { ++struct pv_init_ops pv_init_ops __read_only = { + .patch = native_patch, + }; + +-struct pv_time_ops pv_time_ops = { ++struct pv_time_ops pv_time_ops __read_only = { + .sched_clock = native_sched_clock, + }; + +-struct pv_irq_ops pv_irq_ops = { ++struct pv_irq_ops pv_irq_ops __read_only = { + .save_fl = __PV_IS_CALLEE_SAVE(native_save_fl), + .restore_fl = __PV_IS_CALLEE_SAVE(native_restore_fl), + .irq_disable = __PV_IS_CALLEE_SAVE(native_irq_disable), +@@ -321,7 +330,7 @@ struct pv_irq_ops pv_irq_ops = { + #endif + }; + +-struct pv_cpu_ops pv_cpu_ops = { ++struct pv_cpu_ops pv_cpu_ops __read_only = { + .cpuid = native_cpuid, + .get_debugreg = native_get_debugreg, + .set_debugreg = native_set_debugreg, +@@ -382,21 +391,26 @@ struct pv_cpu_ops pv_cpu_ops = { + .end_context_switch = paravirt_nop, + }; + +-struct pv_apic_ops pv_apic_ops = { ++struct pv_apic_ops pv_apic_ops __read_only = { + #ifdef CONFIG_X86_LOCAL_APIC + .startup_ipi_hook = paravirt_nop, + #endif + }; + +-#if defined(CONFIG_X86_32) && !defined(CONFIG_X86_PAE) ++#ifdef CONFIG_X86_32 ++#ifdef CONFIG_X86_PAE ++/* 64-bit pagetable entries */ ++#define PTE_IDENT PV_CALLEE_SAVE(_paravirt_ident_64) ++#else + /* 32-bit pagetable entries */ + #define PTE_IDENT __PV_IS_CALLEE_SAVE(_paravirt_ident_32) ++#endif + #else + /* 64-bit pagetable entries */ + #define PTE_IDENT __PV_IS_CALLEE_SAVE(_paravirt_ident_64) + #endif + +-struct pv_mmu_ops pv_mmu_ops = { ++struct pv_mmu_ops pv_mmu_ops __read_only = { + + .read_cr2 = native_read_cr2, + .write_cr2 = native_write_cr2, +@@ -448,6 +462,7 @@ struct pv_mmu_ops pv_mmu_ops = { + .make_pud = PTE_IDENT, + + .set_pgd = native_set_pgd, ++ .set_pgd_batched = native_set_pgd_batched, + #endif + #endif /* PAGETABLE_LEVELS >= 3 */ + +@@ -467,6 +482,12 @@ struct pv_mmu_ops pv_mmu_ops = { + }, + + .set_fixmap = native_set_fixmap, ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ .pax_open_kernel = native_pax_open_kernel, ++ .pax_close_kernel = native_pax_close_kernel, ++#endif ++ + }; + + EXPORT_SYMBOL_GPL(pv_time_ops); +diff --git a/arch/x86/kernel/pci-calgary_64.c b/arch/x86/kernel/pci-calgary_64.c +index 1a2d4b1..6a0dd55 100644 +--- a/arch/x86/kernel/pci-calgary_64.c ++++ b/arch/x86/kernel/pci-calgary_64.c +@@ -477,7 +477,7 @@ static void calgary_free_coherent(struct device *dev, size_t size, + free_pages((unsigned long)vaddr, get_order(size)); + } + +-static struct dma_map_ops calgary_dma_ops = { ++static const struct dma_map_ops calgary_dma_ops = { + .alloc_coherent = calgary_alloc_coherent, + .free_coherent = calgary_free_coherent, + .map_sg = calgary_map_sg, +diff --git a/arch/x86/kernel/pci-dma.c b/arch/x86/kernel/pci-dma.c +index 6ac3931..42b4414 100644 +--- a/arch/x86/kernel/pci-dma.c ++++ b/arch/x86/kernel/pci-dma.c +@@ -14,7 +14,7 @@ + + static int forbid_dac __read_mostly; + +-struct dma_map_ops *dma_ops; ++const struct dma_map_ops *dma_ops; + EXPORT_SYMBOL(dma_ops); + + static int iommu_sac_force __read_mostly; +@@ -243,7 +243,7 @@ early_param("iommu", iommu_setup); + + int dma_supported(struct device *dev, u64 mask) + { +- struct dma_map_ops *ops = get_dma_ops(dev); ++ const struct dma_map_ops *ops = get_dma_ops(dev); + + #ifdef CONFIG_PCI + if (mask > 0xffffffff && forbid_dac > 0) { +diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c +index 1c76691..e3632db 100644 +--- a/arch/x86/kernel/pci-gart_64.c ++++ b/arch/x86/kernel/pci-gart_64.c +@@ -682,7 +682,7 @@ static __init int init_k8_gatt(struct agp_kern_info *info) + return -1; + } + +-static struct dma_map_ops gart_dma_ops = { ++static const struct dma_map_ops gart_dma_ops = { + .map_sg = gart_map_sg, + .unmap_sg = gart_unmap_sg, + .map_page = gart_map_page, +diff --git a/arch/x86/kernel/pci-nommu.c b/arch/x86/kernel/pci-nommu.c +index a3933d4..c898869 100644 +--- a/arch/x86/kernel/pci-nommu.c ++++ b/arch/x86/kernel/pci-nommu.c +@@ -94,7 +94,7 @@ static void nommu_sync_sg_for_device(struct device *dev, + flush_write_buffers(); + } + +-struct dma_map_ops nommu_dma_ops = { ++const struct dma_map_ops nommu_dma_ops = { + .alloc_coherent = dma_generic_alloc_coherent, + .free_coherent = nommu_free_coherent, + .map_sg = nommu_map_sg, +diff --git a/arch/x86/kernel/pci-swiotlb.c b/arch/x86/kernel/pci-swiotlb.c +index aaa6b78..4de1881 100644 +--- a/arch/x86/kernel/pci-swiotlb.c ++++ b/arch/x86/kernel/pci-swiotlb.c +@@ -25,7 +25,7 @@ static void *x86_swiotlb_alloc_coherent(struct device *hwdev, size_t size, + return swiotlb_alloc_coherent(hwdev, size, dma_handle, flags); + } + +-static struct dma_map_ops swiotlb_dma_ops = { ++static const struct dma_map_ops swiotlb_dma_ops = { + .mapping_error = swiotlb_dma_mapping_error, + .alloc_coherent = x86_swiotlb_alloc_coherent, + .free_coherent = swiotlb_free_coherent, +diff --git a/arch/x86/kernel/process.c b/arch/x86/kernel/process.c +index fc6c84d..0312ca2 100644 +--- a/arch/x86/kernel/process.c ++++ b/arch/x86/kernel/process.c +@@ -51,16 +51,33 @@ void free_thread_xstate(struct task_struct *tsk) + + void free_thread_info(struct thread_info *ti) + { +- free_thread_xstate(ti->task); + free_pages((unsigned long)ti, get_order(THREAD_SIZE)); + } + ++static struct kmem_cache *task_struct_cachep; ++ + void arch_task_cache_init(void) + { +- task_xstate_cachep = +- kmem_cache_create("task_xstate", xstate_size, ++ /* create a slab on which task_structs can be allocated */ ++ task_struct_cachep = ++ kmem_cache_create("task_struct", sizeof(struct task_struct), ++ ARCH_MIN_TASKALIGN, SLAB_PANIC | SLAB_NOTRACK, NULL); ++ ++ task_xstate_cachep = ++ kmem_cache_create("task_xstate", xstate_size, + __alignof__(union thread_xstate), +- SLAB_PANIC | SLAB_NOTRACK, NULL); ++ SLAB_PANIC | SLAB_NOTRACK | SLAB_USERCOPY, NULL); ++} ++ ++struct task_struct *alloc_task_struct(void) ++{ ++ return kmem_cache_alloc(task_struct_cachep, GFP_KERNEL); ++} ++ ++void free_task_struct(struct task_struct *task) ++{ ++ free_thread_xstate(task); ++ kmem_cache_free(task_struct_cachep, task); + } + + /* +@@ -73,7 +90,7 @@ void exit_thread(void) + unsigned long *bp = t->io_bitmap_ptr; + + if (bp) { +- struct tss_struct *tss = &per_cpu(init_tss, get_cpu()); ++ struct tss_struct *tss = init_tss + get_cpu(); + + t->io_bitmap_ptr = NULL; + clear_thread_flag(TIF_IO_BITMAP); +@@ -93,6 +110,9 @@ void flush_thread(void) + + clear_tsk_thread_flag(tsk, TIF_DEBUG); + ++#if defined(CONFIG_X86_32) && !defined(CONFIG_CC_STACKPROTECTOR) && !defined(CONFIG_PAX_MEMORY_UDEREF) ++ loadsegment(gs, 0); ++#endif + tsk->thread.debugreg0 = 0; + tsk->thread.debugreg1 = 0; + tsk->thread.debugreg2 = 0; +@@ -307,7 +327,7 @@ void default_idle(void) + EXPORT_SYMBOL(default_idle); + #endif + +-void stop_this_cpu(void *dummy) ++__noreturn void stop_this_cpu(void *dummy) + { + local_irq_disable(); + /* +@@ -568,16 +588,38 @@ static int __init idle_setup(char *str) + } + early_param("idle", idle_setup); + +-unsigned long arch_align_stack(unsigned long sp) ++#ifdef CONFIG_PAX_RANDKSTACK ++void pax_randomize_kstack(struct pt_regs *regs) + { +- if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space) +- sp -= get_random_int() % 8192; +- return sp & ~0xf; +-} ++ struct thread_struct *thread = ¤t->thread; ++ unsigned long time; + +-unsigned long arch_randomize_brk(struct mm_struct *mm) +-{ +- unsigned long range_end = mm->brk + 0x02000000; +- return randomize_range(mm->brk, range_end, 0) ? : mm->brk; ++ if (!randomize_va_space) ++ return; ++ ++ if (v8086_mode(regs)) ++ return; ++ ++ rdtscl(time); ++ ++ /* P4 seems to return a 0 LSB, ignore it */ ++#ifdef CONFIG_MPENTIUM4 ++ time &= 0x3EUL; ++ time <<= 2; ++#elif defined(CONFIG_X86_64) ++ time &= 0xFUL; ++ time <<= 4; ++#else ++ time &= 0x1FUL; ++ time <<= 3; ++#endif ++ ++ thread->sp0 ^= time; ++ load_sp0(init_tss + smp_processor_id(), thread); ++ ++#ifdef CONFIG_X86_64 ++ percpu_write(kernel_stack, thread->sp0); ++#endif + } ++#endif + +diff --git a/arch/x86/kernel/process_32.c b/arch/x86/kernel/process_32.c +index c40c432..6e1df72 100644 +--- a/arch/x86/kernel/process_32.c ++++ b/arch/x86/kernel/process_32.c +@@ -67,6 +67,7 @@ asmlinkage void ret_from_fork(void) __asm__("ret_from_fork"); + unsigned long thread_saved_pc(struct task_struct *tsk) + { + return ((unsigned long *)tsk->thread.sp)[3]; ++//XXX return tsk->thread.eip; + } + + #ifndef CONFIG_SMP +@@ -129,15 +130,14 @@ void __show_regs(struct pt_regs *regs, int all) + unsigned short ss, gs; + const char *board; + +- if (user_mode_vm(regs)) { ++ if (user_mode(regs)) { + sp = regs->sp; + ss = regs->ss & 0xffff; +- gs = get_user_gs(regs); + } else { + sp = (unsigned long) (®s->sp); + savesegment(ss, ss); +- savesegment(gs, gs); + } ++ gs = get_user_gs(regs); + + printk("\n"); + +@@ -210,10 +210,10 @@ int kernel_thread(int (*fn)(void *), void *arg, unsigned long flags) + regs.bx = (unsigned long) fn; + regs.dx = (unsigned long) arg; + +- regs.ds = __USER_DS; +- regs.es = __USER_DS; ++ regs.ds = __KERNEL_DS; ++ regs.es = __KERNEL_DS; + regs.fs = __KERNEL_PERCPU; +- regs.gs = __KERNEL_STACK_CANARY; ++ savesegment(gs, regs.gs); + regs.orig_ax = -1; + regs.ip = (unsigned long) kernel_thread_helper; + regs.cs = __KERNEL_CS | get_kernel_rpl(); +@@ -247,13 +247,14 @@ int copy_thread(unsigned long clone_flags, unsigned long sp, + struct task_struct *tsk; + int err; + +- childregs = task_pt_regs(p); ++ childregs = task_stack_page(p) + THREAD_SIZE - sizeof(struct pt_regs) - 8; + *childregs = *regs; + childregs->ax = 0; + childregs->sp = sp; + + p->thread.sp = (unsigned long) childregs; + p->thread.sp0 = (unsigned long) (childregs+1); ++ p->tinfo.lowest_stack = (unsigned long)task_stack_page(p); + + p->thread.ip = (unsigned long) ret_from_fork; + +@@ -345,7 +346,7 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + struct thread_struct *prev = &prev_p->thread, + *next = &next_p->thread; + int cpu = smp_processor_id(); +- struct tss_struct *tss = &per_cpu(init_tss, cpu); ++ struct tss_struct *tss = init_tss + cpu; + bool preload_fpu; + + /* never put a printk in __switch_to... printk() calls wake_up*() indirectly */ +@@ -380,6 +381,10 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + */ + lazy_save_gs(prev->gs); + ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ __set_fs(task_thread_info(next_p)->addr_limit); ++#endif ++ + /* + * Load the per-thread Thread-Local Storage descriptor. + */ +@@ -415,6 +420,9 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + */ + arch_end_context_switch(next_p); + ++ percpu_write(current_task, next_p); ++ percpu_write(current_tinfo, &next_p->tinfo); ++ + if (preload_fpu) + __math_state_restore(); + +@@ -424,8 +432,6 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + if (prev->gs | next->gs) + lazy_load_gs(next->gs); + +- percpu_write(current_task, next_p); +- + return prev_p; + } + +@@ -495,4 +501,3 @@ unsigned long get_wchan(struct task_struct *p) + } while (count++ < 16); + return 0; + } +- +diff --git a/arch/x86/kernel/process_64.c b/arch/x86/kernel/process_64.c +index 39493bc..196816d 100644 +--- a/arch/x86/kernel/process_64.c ++++ b/arch/x86/kernel/process_64.c +@@ -91,7 +91,7 @@ static void __exit_idle(void) + void exit_idle(void) + { + /* idle loop has pid 0 */ +- if (current->pid) ++ if (task_pid_nr(current)) + return; + __exit_idle(); + } +@@ -170,7 +170,7 @@ void __show_regs(struct pt_regs *regs, int all) + if (!board) + board = ""; + printk(KERN_INFO "Pid: %d, comm: %.20s %s %s %.*s %s\n", +- current->pid, current->comm, print_tainted(), ++ task_pid_nr(current), current->comm, print_tainted(), + init_utsname()->release, + (int)strcspn(init_utsname()->version, " "), + init_utsname()->version, board); +@@ -280,8 +280,7 @@ int copy_thread(unsigned long clone_flags, unsigned long sp, + struct pt_regs *childregs; + struct task_struct *me = current; + +- childregs = ((struct pt_regs *) +- (THREAD_SIZE + task_stack_page(p))) - 1; ++ childregs = task_stack_page(p) + THREAD_SIZE - sizeof(struct pt_regs) - 16; + *childregs = *regs; + + childregs->ax = 0; +@@ -292,6 +291,7 @@ int copy_thread(unsigned long clone_flags, unsigned long sp, + p->thread.sp = (unsigned long) childregs; + p->thread.sp0 = (unsigned long) (childregs+1); + p->thread.usersp = me->thread.usersp; ++ p->tinfo.lowest_stack = (unsigned long)task_stack_page(p); + + set_tsk_thread_flag(p, TIF_FORK); + +@@ -379,7 +379,7 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + struct thread_struct *prev = &prev_p->thread; + struct thread_struct *next = &next_p->thread; + int cpu = smp_processor_id(); +- struct tss_struct *tss = &per_cpu(init_tss, cpu); ++ struct tss_struct *tss = init_tss + cpu; + unsigned fsindex, gsindex; + bool preload_fpu; + +@@ -475,10 +475,9 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) + prev->usersp = percpu_read(old_rsp); + percpu_write(old_rsp, next->usersp); + percpu_write(current_task, next_p); ++ percpu_write(current_tinfo, &next_p->tinfo); + +- percpu_write(kernel_stack, +- (unsigned long)task_stack_page(next_p) + +- THREAD_SIZE - KERNEL_STACK_OFFSET); ++ percpu_write(kernel_stack, next->sp0); + + /* + * Now maybe reload the debug registers and handle I/O bitmaps +@@ -559,12 +558,11 @@ unsigned long get_wchan(struct task_struct *p) + if (!p || p == current || p->state == TASK_RUNNING) + return 0; + stack = (unsigned long)task_stack_page(p); +- if (p->thread.sp < stack || p->thread.sp >= stack+THREAD_SIZE) ++ if (p->thread.sp < stack || p->thread.sp > stack+THREAD_SIZE-16-sizeof(u64)) + return 0; + fp = *(u64 *)(p->thread.sp); + do { +- if (fp < (unsigned long)stack || +- fp >= (unsigned long)stack+THREAD_SIZE) ++ if (fp < stack || fp > stack+THREAD_SIZE-16-sizeof(u64)) + return 0; + ip = *(u64 *)(fp+8); + if (!in_sched_functions(ip)) +diff --git a/arch/x86/kernel/ptrace.c b/arch/x86/kernel/ptrace.c +index c06acdd..3f5fff5 100644 +--- a/arch/x86/kernel/ptrace.c ++++ b/arch/x86/kernel/ptrace.c +@@ -925,7 +925,7 @@ static const struct user_regset_view user_x86_32_view; /* Initialized below. */ + long arch_ptrace(struct task_struct *child, long request, long addr, long data) + { + int ret; +- unsigned long __user *datap = (unsigned long __user *)data; ++ unsigned long __user *datap = (__force unsigned long __user *)data; + + switch (request) { + /* read the word at location addr in the USER area. */ +@@ -1012,14 +1012,14 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) + if (addr < 0) + return -EIO; + ret = do_get_thread_area(child, addr, +- (struct user_desc __user *) data); ++ (__force struct user_desc __user *) data); + break; + + case PTRACE_SET_THREAD_AREA: + if (addr < 0) + return -EIO; + ret = do_set_thread_area(child, addr, +- (struct user_desc __user *) data, 0); ++ (__force struct user_desc __user *) data, 0); + break; + #endif + +@@ -1038,12 +1038,12 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) + #ifdef CONFIG_X86_PTRACE_BTS + case PTRACE_BTS_CONFIG: + ret = ptrace_bts_config +- (child, data, (struct ptrace_bts_config __user *)addr); ++ (child, data, (__force struct ptrace_bts_config __user *)addr); + break; + + case PTRACE_BTS_STATUS: + ret = ptrace_bts_status +- (child, data, (struct ptrace_bts_config __user *)addr); ++ (child, data, (__force struct ptrace_bts_config __user *)addr); + break; + + case PTRACE_BTS_SIZE: +@@ -1052,7 +1052,7 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) + + case PTRACE_BTS_GET: + ret = ptrace_bts_read_record +- (child, data, (struct bts_struct __user *) addr); ++ (child, data, (__force struct bts_struct __user *) addr); + break; + + case PTRACE_BTS_CLEAR: +@@ -1061,7 +1061,7 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) + + case PTRACE_BTS_DRAIN: + ret = ptrace_bts_drain +- (child, data, (struct bts_struct __user *) addr); ++ (child, data, (__force struct bts_struct __user *) addr); + break; + #endif /* CONFIG_X86_PTRACE_BTS */ + +@@ -1450,7 +1450,7 @@ void send_sigtrap(struct task_struct *tsk, struct pt_regs *regs, + info.si_code = si_code; + + /* User-mode ip? */ +- info.si_addr = user_mode_vm(regs) ? (void __user *) regs->ip : NULL; ++ info.si_addr = user_mode(regs) ? (__force void __user *) regs->ip : NULL; + + /* Send us the fake SIGTRAP */ + force_sig_info(SIGTRAP, &info, tsk); +@@ -1469,7 +1469,7 @@ void send_sigtrap(struct task_struct *tsk, struct pt_regs *regs, + * We must return the syscall number to actually look up in the table. + * This can be -1L to skip running any syscall at all. + */ +-asmregparm long syscall_trace_enter(struct pt_regs *regs) ++long syscall_trace_enter(struct pt_regs *regs) + { + long ret = 0; + +@@ -1514,7 +1514,7 @@ asmregparm long syscall_trace_enter(struct pt_regs *regs) + return ret ?: regs->orig_ax; + } + +-asmregparm void syscall_trace_leave(struct pt_regs *regs) ++void syscall_trace_leave(struct pt_regs *regs) + { + if (unlikely(current->audit_context)) + audit_syscall_exit(AUDITSC_RESULT(regs->ax), regs->ax); +diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c +index cf98100..e76e03d 100644 +--- a/arch/x86/kernel/reboot.c ++++ b/arch/x86/kernel/reboot.c +@@ -33,7 +33,7 @@ void (*pm_power_off)(void); + EXPORT_SYMBOL(pm_power_off); + + static const struct desc_ptr no_idt = {}; +-static int reboot_mode; ++static unsigned short reboot_mode; + enum reboot_type reboot_type = BOOT_KBD; + int reboot_force; + +@@ -292,12 +292,12 @@ core_initcall(reboot_init); + controller to pulse the CPU reset line, which is more thorough, but + doesn't work with at least one type of 486 motherboard. It is easy + to stop this code working; hence the copious comments. */ +-static const unsigned long long +-real_mode_gdt_entries [3] = ++static struct desc_struct ++real_mode_gdt_entries [3] __read_only = + { +- 0x0000000000000000ULL, /* Null descriptor */ +- 0x00009b000000ffffULL, /* 16-bit real-mode 64k code at 0x00000000 */ +- 0x000093000100ffffULL /* 16-bit real-mode 64k data at 0x00000100 */ ++ GDT_ENTRY_INIT(0, 0, 0), /* Null descriptor */ ++ GDT_ENTRY_INIT(0x9b, 0, 0xffff), /* 16-bit real-mode 64k code at 0x00000000 */ ++ GDT_ENTRY_INIT(0x93, 0x100, 0xffff) /* 16-bit real-mode 64k data at 0x00000100 */ + }; + + static const struct desc_ptr +@@ -346,7 +346,7 @@ static const unsigned char jump_to_bios [] = + * specified by the code and length parameters. + * We assume that length will aways be less that 100! + */ +-void machine_real_restart(const unsigned char *code, int length) ++__noreturn void machine_real_restart(const unsigned char *code, unsigned int length) + { + local_irq_disable(); + +@@ -366,8 +366,8 @@ void machine_real_restart(const unsigned char *code, int length) + /* Remap the kernel at virtual address zero, as well as offset zero + from the kernel segment. This assumes the kernel segment starts at + virtual address PAGE_OFFSET. */ +- memcpy(swapper_pg_dir, swapper_pg_dir + KERNEL_PGD_BOUNDARY, +- sizeof(swapper_pg_dir [0]) * KERNEL_PGD_PTRS); ++ clone_pgd_range(swapper_pg_dir, swapper_pg_dir + KERNEL_PGD_BOUNDARY, ++ min_t(unsigned long, KERNEL_PGD_PTRS, KERNEL_PGD_BOUNDARY)); + + /* + * Use `swapper_pg_dir' as our page directory. +@@ -379,16 +379,15 @@ void machine_real_restart(const unsigned char *code, int length) + boot)". This seems like a fairly standard thing that gets set by + REBOOT.COM programs, and the previous reset routine did this + too. */ +- *((unsigned short *)0x472) = reboot_mode; ++ *(unsigned short *)(__va(0x472)) = reboot_mode; + + /* For the switch to real mode, copy some code to low memory. It has + to be in the first 64k because it is running in 16-bit mode, and it + has to have the same physical and virtual address, because it turns + off paging. Copy it near the end of the first page, out of the way + of BIOS variables. */ +- memcpy((void *)(0x1000 - sizeof(real_mode_switch) - 100), +- real_mode_switch, sizeof (real_mode_switch)); +- memcpy((void *)(0x1000 - 100), code, length); ++ memcpy(__va(0x1000 - sizeof (real_mode_switch) - 100), real_mode_switch, sizeof (real_mode_switch)); ++ memcpy(__va(0x1000 - 100), code, length); + + /* Set up the IDT for real mode. */ + load_idt(&real_mode_idt); +@@ -416,6 +415,7 @@ void machine_real_restart(const unsigned char *code, int length) + __asm__ __volatile__ ("ljmp $0x0008,%0" + : + : "i" ((void *)(0x1000 - sizeof (real_mode_switch) - 100))); ++ do { } while (1); + } + #ifdef CONFIG_APM_MODULE + EXPORT_SYMBOL(machine_real_restart); +@@ -544,7 +544,7 @@ void __attribute__((weak)) mach_reboot_fixups(void) + { + } + +-static void native_machine_emergency_restart(void) ++__noreturn static void native_machine_emergency_restart(void) + { + int i; + +@@ -659,13 +659,13 @@ void native_machine_shutdown(void) + #endif + } + +-static void __machine_emergency_restart(int emergency) ++static __noreturn void __machine_emergency_restart(int emergency) + { + reboot_emergency = emergency; + machine_ops.emergency_restart(); + } + +-static void native_machine_restart(char *__unused) ++static __noreturn void native_machine_restart(char *__unused) + { + printk("machine restart\n"); + +@@ -674,7 +674,7 @@ static void native_machine_restart(char *__unused) + __machine_emergency_restart(0); + } + +-static void native_machine_halt(void) ++static __noreturn void native_machine_halt(void) + { + /* stop other cpus and apics */ + machine_shutdown(); +@@ -685,7 +685,7 @@ static void native_machine_halt(void) + stop_this_cpu(NULL); + } + +-static void native_machine_power_off(void) ++__noreturn static void native_machine_power_off(void) + { + if (pm_power_off) { + if (!reboot_force) +@@ -694,6 +694,7 @@ static void native_machine_power_off(void) + } + /* a fallback in case there is no PM info available */ + tboot_shutdown(TB_SHUTDOWN_HALT); ++ do { } while (1); + } + + struct machine_ops machine_ops = { +diff --git a/arch/x86/kernel/relocate_kernel_64.S b/arch/x86/kernel/relocate_kernel_64.S +index 7a6f3b3..976a959 100644 +--- a/arch/x86/kernel/relocate_kernel_64.S ++++ b/arch/x86/kernel/relocate_kernel_64.S +@@ -11,6 +11,7 @@ + #include + #include + #include ++#include + + /* + * Must be relocatable PIC code callable as a C function +@@ -167,6 +168,7 @@ identity_mapped: + xorq %r14, %r14 + xorq %r15, %r15 + ++ pax_force_retaddr 0, 1 + ret + + 1: +diff --git a/arch/x86/kernel/setup.c b/arch/x86/kernel/setup.c +index 5449a26..0b6c759 100644 +--- a/arch/x86/kernel/setup.c ++++ b/arch/x86/kernel/setup.c +@@ -783,14 +783,14 @@ void __init setup_arch(char **cmdline_p) + + if (!boot_params.hdr.root_flags) + root_mountflags &= ~MS_RDONLY; +- init_mm.start_code = (unsigned long) _text; +- init_mm.end_code = (unsigned long) _etext; ++ init_mm.start_code = ktla_ktva((unsigned long) _text); ++ init_mm.end_code = ktla_ktva((unsigned long) _etext); + init_mm.end_data = (unsigned long) _edata; + init_mm.brk = _brk_end; + +- code_resource.start = virt_to_phys(_text); +- code_resource.end = virt_to_phys(_etext)-1; +- data_resource.start = virt_to_phys(_etext); ++ code_resource.start = virt_to_phys(ktla_ktva(_text)); ++ code_resource.end = virt_to_phys(ktla_ktva(_etext))-1; ++ data_resource.start = virt_to_phys(_sdata); + data_resource.end = virt_to_phys(_edata)-1; + bss_resource.start = virt_to_phys(&__bss_start); + bss_resource.end = virt_to_phys(&__bss_stop)-1; +diff --git a/arch/x86/kernel/setup_percpu.c b/arch/x86/kernel/setup_percpu.c +index d559af9..524c6ad 100644 +--- a/arch/x86/kernel/setup_percpu.c ++++ b/arch/x86/kernel/setup_percpu.c +@@ -25,19 +25,17 @@ + # define DBG(x...) + #endif + +-DEFINE_PER_CPU(int, cpu_number); ++#ifdef CONFIG_SMP ++DEFINE_PER_CPU(unsigned int, cpu_number); + EXPORT_PER_CPU_SYMBOL(cpu_number); ++#endif + +-#ifdef CONFIG_X86_64 + #define BOOT_PERCPU_OFFSET ((unsigned long)__per_cpu_load) +-#else +-#define BOOT_PERCPU_OFFSET 0 +-#endif + + DEFINE_PER_CPU(unsigned long, this_cpu_off) = BOOT_PERCPU_OFFSET; + EXPORT_PER_CPU_SYMBOL(this_cpu_off); + +-unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = { ++unsigned long __per_cpu_offset[NR_CPUS] __read_only = { + [0 ... NR_CPUS-1] = BOOT_PERCPU_OFFSET, + }; + EXPORT_SYMBOL(__per_cpu_offset); +@@ -159,10 +157,10 @@ static inline void setup_percpu_segment(int cpu) + { + #ifdef CONFIG_X86_32 + struct desc_struct gdt; ++ unsigned long base = per_cpu_offset(cpu); + +- pack_descriptor(&gdt, per_cpu_offset(cpu), 0xFFFFF, +- 0x2 | DESCTYPE_S, 0x8); +- gdt.s = 1; ++ pack_descriptor(&gdt, base, (VMALLOC_END - base - 1) >> PAGE_SHIFT, ++ 0x83 | DESCTYPE_S, 0xC); + write_gdt_entry(get_cpu_gdt_table(cpu), + GDT_ENTRY_PERCPU, &gdt, DESCTYPE_S); + #endif +@@ -212,6 +210,11 @@ void __init setup_per_cpu_areas(void) + /* alrighty, percpu areas up and running */ + delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start; + for_each_possible_cpu(cpu) { ++#ifdef CONFIG_CC_STACKPROTECTOR ++#ifdef CONFIG_X86_32 ++ unsigned long canary = per_cpu(stack_canary.canary, cpu); ++#endif ++#endif + per_cpu_offset(cpu) = delta + pcpu_unit_offsets[cpu]; + per_cpu(this_cpu_off, cpu) = per_cpu_offset(cpu); + per_cpu(cpu_number, cpu) = cpu; +@@ -239,6 +242,12 @@ void __init setup_per_cpu_areas(void) + early_per_cpu_map(x86_cpu_to_node_map, cpu); + #endif + #endif ++#ifdef CONFIG_CC_STACKPROTECTOR ++#ifdef CONFIG_X86_32 ++ if (!cpu) ++ per_cpu(stack_canary.canary, cpu) = canary; ++#endif ++#endif + /* + * Up to this point, the boot CPU has been using .data.init + * area. Reload any changed state for the boot CPU. +diff --git a/arch/x86/kernel/signal.c b/arch/x86/kernel/signal.c +index 6a44a76..a9287a1 100644 +--- a/arch/x86/kernel/signal.c ++++ b/arch/x86/kernel/signal.c +@@ -197,7 +197,7 @@ static unsigned long align_sigframe(unsigned long sp) + * Align the stack pointer according to the i386 ABI, + * i.e. so that on function entry ((sp + 4) & 15) == 0. + */ +- sp = ((sp + 4) & -16ul) - 4; ++ sp = ((sp - 12) & -16ul) - 4; + #else /* !CONFIG_X86_32 */ + sp = round_down(sp, 16) - 8; + #endif +@@ -248,11 +248,11 @@ get_sigframe(struct k_sigaction *ka, struct pt_regs *regs, size_t frame_size, + * Return an always-bogus address instead so we will die with SIGSEGV. + */ + if (onsigstack && !likely(on_sig_stack(sp))) +- return (void __user *)-1L; ++ return (__force void __user *)-1L; + + /* save i387 state */ + if (used_math() && save_i387_xstate(*fpstate) < 0) +- return (void __user *)-1L; ++ return (__force void __user *)-1L; + + return (void __user *)sp; + } +@@ -307,9 +307,9 @@ __setup_frame(int sig, struct k_sigaction *ka, sigset_t *set, + } + + if (current->mm->context.vdso) +- restorer = VDSO32_SYMBOL(current->mm->context.vdso, sigreturn); ++ restorer = (__force void __user *)VDSO32_SYMBOL(current->mm->context.vdso, sigreturn); + else +- restorer = &frame->retcode; ++ restorer = (void __user *)&frame->retcode; + if (ka->sa.sa_flags & SA_RESTORER) + restorer = ka->sa.sa_restorer; + +@@ -323,7 +323,7 @@ __setup_frame(int sig, struct k_sigaction *ka, sigset_t *set, + * reasons and because gdb uses it as a signature to notice + * signal handler stack frames. + */ +- err |= __put_user(*((u64 *)&retcode), (u64 *)frame->retcode); ++ err |= __put_user(*((u64 *)&retcode), (u64 __user *)frame->retcode); + + if (err) + return -EFAULT; +@@ -377,7 +377,10 @@ static int __setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, + err |= __copy_to_user(&frame->uc.uc_sigmask, set, sizeof(*set)); + + /* Set up to return from userspace. */ +- restorer = VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn); ++ if (current->mm->context.vdso) ++ restorer = (__force void __user *)VDSO32_SYMBOL(current->mm->context.vdso, rt_sigreturn); ++ else ++ restorer = (void __user *)&frame->retcode; + if (ka->sa.sa_flags & SA_RESTORER) + restorer = ka->sa.sa_restorer; + put_user_ex(restorer, &frame->pretcode); +@@ -389,7 +392,7 @@ static int __setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, + * reasons and because gdb uses it as a signature to notice + * signal handler stack frames. + */ +- put_user_ex(*((u64 *)&rt_retcode), (u64 *)frame->retcode); ++ put_user_ex(*((u64 *)&rt_retcode), (u64 __user *)frame->retcode); + } put_user_catch(err); + + if (err) +@@ -782,6 +785,8 @@ static void do_signal(struct pt_regs *regs) + int signr; + sigset_t *oldset; + ++ pax_track_stack(); ++ + /* + * We want the common case to go fast, which is why we may in certain + * cases get here from kernel mode. Just return without doing anything +@@ -789,7 +794,7 @@ static void do_signal(struct pt_regs *regs) + * X86_32: vm86 regs switched out by assembly code before reaching + * here, so testing against kernel CS suffices. + */ +- if (!user_mode(regs)) ++ if (!user_mode_novm(regs)) + return; + + if (current_thread_info()->status & TS_RESTORE_SIGMASK) +diff --git a/arch/x86/kernel/smpboot.c b/arch/x86/kernel/smpboot.c +index 7e8e905..64d5c32 100644 +--- a/arch/x86/kernel/smpboot.c ++++ b/arch/x86/kernel/smpboot.c +@@ -94,14 +94,14 @@ static DEFINE_PER_CPU(struct task_struct *, idle_thread_array); + */ + static DEFINE_MUTEX(x86_cpu_hotplug_driver_mutex); + +-void cpu_hotplug_driver_lock() ++void cpu_hotplug_driver_lock(void) + { +- mutex_lock(&x86_cpu_hotplug_driver_mutex); ++ mutex_lock(&x86_cpu_hotplug_driver_mutex); + } + +-void cpu_hotplug_driver_unlock() ++void cpu_hotplug_driver_unlock(void) + { +- mutex_unlock(&x86_cpu_hotplug_driver_mutex); ++ mutex_unlock(&x86_cpu_hotplug_driver_mutex); + } + + ssize_t arch_cpu_probe(const char *buf, size_t count) { return -1; } +@@ -625,7 +625,7 @@ wakeup_secondary_cpu_via_init(int phys_apicid, unsigned long start_eip) + * target processor state. + */ + startup_ipi_hook(phys_apicid, (unsigned long) start_secondary, +- (unsigned long)stack_start.sp); ++ stack_start); + + /* + * Run STARTUP IPI loop. +@@ -743,6 +743,7 @@ static int __cpuinit do_boot_cpu(int apicid, int cpu) + set_idle_for_cpu(cpu, c_idle.idle); + do_rest: + per_cpu(current_task, cpu) = c_idle.idle; ++ per_cpu(current_tinfo, cpu) = &c_idle.idle->tinfo; + #ifdef CONFIG_X86_32 + /* Stack for startup_32 can be just as for start_secondary onwards */ + irq_ctx_init(cpu); +@@ -750,13 +751,15 @@ do_rest: + #else + clear_tsk_thread_flag(c_idle.idle, TIF_FORK); + initial_gs = per_cpu_offset(cpu); +- per_cpu(kernel_stack, cpu) = +- (unsigned long)task_stack_page(c_idle.idle) - +- KERNEL_STACK_OFFSET + THREAD_SIZE; ++ per_cpu(kernel_stack, cpu) = (unsigned long)task_stack_page(c_idle.idle) - 16 + THREAD_SIZE; + #endif ++ ++ pax_open_kernel(); + early_gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu); ++ pax_close_kernel(); ++ + initial_code = (unsigned long)start_secondary; +- stack_start.sp = (void *) c_idle.idle->thread.sp; ++ stack_start = c_idle.idle->thread.sp; + + /* start_ip had better be page-aligned! */ + start_ip = setup_trampoline(); +@@ -891,6 +894,12 @@ int __cpuinit native_cpu_up(unsigned int cpu) + + per_cpu(cpu_state, cpu) = CPU_UP_PREPARE; + ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ clone_pgd_range(get_cpu_pgd(cpu) + KERNEL_PGD_BOUNDARY, ++ swapper_pg_dir + KERNEL_PGD_BOUNDARY, ++ KERNEL_PGD_PTRS); ++#endif ++ + err = do_boot_cpu(apicid, cpu); + + if (err) { +diff --git a/arch/x86/kernel/step.c b/arch/x86/kernel/step.c +index 3149032..14f1053 100644 +--- a/arch/x86/kernel/step.c ++++ b/arch/x86/kernel/step.c +@@ -27,10 +27,10 @@ unsigned long convert_ip_to_linear(struct task_struct *child, struct pt_regs *re + struct desc_struct *desc; + unsigned long base; + +- seg &= ~7UL; ++ seg >>= 3; + + mutex_lock(&child->mm->context.lock); +- if (unlikely((seg >> 3) >= child->mm->context.size)) ++ if (unlikely(seg >= child->mm->context.size)) + addr = -1L; /* bogus selector, access would fault */ + else { + desc = child->mm->context.ldt + seg; +@@ -42,7 +42,8 @@ unsigned long convert_ip_to_linear(struct task_struct *child, struct pt_regs *re + addr += base; + } + mutex_unlock(&child->mm->context.lock); +- } ++ } else if (seg == __KERNEL_CS || seg == __KERNEXEC_KERNEL_CS) ++ addr = ktla_ktva(addr); + + return addr; + } +@@ -53,6 +54,9 @@ static int is_setting_trap_flag(struct task_struct *child, struct pt_regs *regs) + unsigned char opcode[15]; + unsigned long addr = convert_ip_to_linear(child, regs); + ++ if (addr == -EINVAL) ++ return 0; ++ + copied = access_process_vm(child, addr, opcode, sizeof(opcode), 0); + for (i = 0; i < copied; i++) { + switch (opcode[i]) { +@@ -74,7 +78,7 @@ static int is_setting_trap_flag(struct task_struct *child, struct pt_regs *regs) + + #ifdef CONFIG_X86_64 + case 0x40 ... 0x4f: +- if (regs->cs != __USER_CS) ++ if ((regs->cs & 0xffff) != __USER_CS) + /* 32-bit mode: register increment */ + return 0; + /* 64-bit mode: REX prefix */ +diff --git a/arch/x86/kernel/sys_i386_32.c b/arch/x86/kernel/sys_i386_32.c +index dee1ff7..a397f7f 100644 +--- a/arch/x86/kernel/sys_i386_32.c ++++ b/arch/x86/kernel/sys_i386_32.c +@@ -24,6 +24,21 @@ + + #include + ++int i386_mmap_check(unsigned long addr, unsigned long len, unsigned long flags) ++{ ++ unsigned long pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (current->mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ if (len > pax_task_size || addr > pax_task_size - len) ++ return -EINVAL; ++ ++ return 0; ++} ++ + /* + * Perform the select(nd, in, out, ex, tv) and mmap() system + * calls. Linux/i386 didn't use to be able to handle more than +@@ -58,6 +73,212 @@ out: + return err; + } + ++unsigned long ++arch_get_unmapped_area(struct file *filp, unsigned long addr, ++ unsigned long len, unsigned long pgoff, unsigned long flags) ++{ ++ struct mm_struct *mm = current->mm; ++ struct vm_area_struct *vma; ++ unsigned long start_addr, pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ pax_task_size -= PAGE_SIZE; ++ ++ if (len > pax_task_size) ++ return -ENOMEM; ++ ++ if (flags & MAP_FIXED) ++ return addr; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ ++ if (addr) { ++ addr = PAGE_ALIGN(addr); ++ if (pax_task_size - len >= addr) { ++ vma = find_vma(mm, addr); ++ if (check_heap_stack_gap(vma, addr, len)) ++ return addr; ++ } ++ } ++ if (len > mm->cached_hole_size) { ++ start_addr = addr = mm->free_area_cache; ++ } else { ++ start_addr = addr = mm->mmap_base; ++ mm->cached_hole_size = 0; ++ } ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (!nx_enabled && (mm->pax_flags & MF_PAX_PAGEEXEC) && (flags & MAP_EXECUTABLE) && start_addr >= mm->mmap_base) { ++ start_addr = 0x00110000UL; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ start_addr += mm->delta_mmap & 0x03FFF000UL; ++#endif ++ ++ if (mm->start_brk <= start_addr && start_addr < mm->mmap_base) ++ start_addr = addr = mm->mmap_base; ++ else ++ addr = start_addr; ++ } ++#endif ++ ++full_search: ++ for (vma = find_vma(mm, addr); ; vma = vma->vm_next) { ++ /* At this point: (!vma || addr < vma->vm_end). */ ++ if (pax_task_size - len < addr) { ++ /* ++ * Start a new search - just in case we missed ++ * some holes. ++ */ ++ if (start_addr != mm->mmap_base) { ++ start_addr = addr = mm->mmap_base; ++ mm->cached_hole_size = 0; ++ goto full_search; ++ } ++ return -ENOMEM; ++ } ++ if (check_heap_stack_gap(vma, addr, len)) ++ break; ++ if (addr + mm->cached_hole_size < vma->vm_start) ++ mm->cached_hole_size = vma->vm_start - addr; ++ addr = vma->vm_end; ++ if (mm->start_brk <= addr && addr < mm->mmap_base) { ++ start_addr = addr = mm->mmap_base; ++ mm->cached_hole_size = 0; ++ goto full_search; ++ } ++ } ++ ++ /* ++ * Remember the place where we stopped the search: ++ */ ++ mm->free_area_cache = addr + len; ++ return addr; ++} ++ ++unsigned long ++arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, ++ const unsigned long len, const unsigned long pgoff, ++ const unsigned long flags) ++{ ++ struct vm_area_struct *vma; ++ struct mm_struct *mm = current->mm; ++ unsigned long base = mm->mmap_base, addr = addr0, pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ pax_task_size -= PAGE_SIZE; ++ ++ /* requested length too big for entire address space */ ++ if (len > pax_task_size) ++ return -ENOMEM; ++ ++ if (flags & MAP_FIXED) ++ return addr; ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (!nx_enabled && (mm->pax_flags & MF_PAX_PAGEEXEC) && (flags & MAP_EXECUTABLE)) ++ goto bottomup; ++#endif ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ ++ /* requesting a specific address */ ++ if (addr) { ++ addr = PAGE_ALIGN(addr); ++ if (pax_task_size - len >= addr) { ++ vma = find_vma(mm, addr); ++ if (check_heap_stack_gap(vma, addr, len)) ++ return addr; ++ } ++ } ++ ++ /* check if free_area_cache is useful for us */ ++ if (len <= mm->cached_hole_size) { ++ mm->cached_hole_size = 0; ++ mm->free_area_cache = mm->mmap_base; ++ } ++ ++ /* either no address requested or can't fit in requested address hole */ ++ addr = mm->free_area_cache; ++ ++ /* make sure it can fit in the remaining address space */ ++ if (addr > len) { ++ vma = find_vma(mm, addr-len); ++ if (check_heap_stack_gap(vma, addr - len, len)) ++ /* remember the address as a hint for next time */ ++ return (mm->free_area_cache = addr-len); ++ } ++ ++ if (mm->mmap_base < len) ++ goto bottomup; ++ ++ addr = mm->mmap_base-len; ++ ++ do { ++ /* ++ * Lookup failure means no vma is above this address, ++ * else if new region fits below vma->vm_start, ++ * return with success: ++ */ ++ vma = find_vma(mm, addr); ++ if (check_heap_stack_gap(vma, addr, len)) ++ /* remember the address as a hint for next time */ ++ return (mm->free_area_cache = addr); ++ ++ /* remember the largest hole we saw so far */ ++ if (addr + mm->cached_hole_size < vma->vm_start) ++ mm->cached_hole_size = vma->vm_start - addr; ++ ++ /* try just below the current vma->vm_start */ ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); ++ ++bottomup: ++ /* ++ * A failed mmap() very likely causes application failure, ++ * so fall back to the bottom-up function here. This scenario ++ * can happen with large stack limits and large mmap() ++ * allocations. ++ */ ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ mm->mmap_base = SEGMEXEC_TASK_UNMAPPED_BASE; ++ else ++#endif ++ ++ mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ ++ mm->free_area_cache = mm->mmap_base; ++ mm->cached_hole_size = ~0UL; ++ addr = arch_get_unmapped_area(filp, addr0, len, pgoff, flags); ++ /* ++ * Restore the topdown base: ++ */ ++ mm->mmap_base = base; ++ mm->free_area_cache = base; ++ mm->cached_hole_size = ~0UL; ++ ++ return addr; ++} + + struct sel_arg_struct { + unsigned long n; +@@ -93,7 +314,7 @@ asmlinkage int sys_ipc(uint call, int first, int second, + return sys_semtimedop(first, (struct sembuf __user *)ptr, second, NULL); + case SEMTIMEDOP: + return sys_semtimedop(first, (struct sembuf __user *)ptr, second, +- (const struct timespec __user *)fifth); ++ (__force const struct timespec __user *)fifth); + + case SEMGET: + return sys_semget(first, second, third); +@@ -140,7 +361,7 @@ asmlinkage int sys_ipc(uint call, int first, int second, + ret = do_shmat(first, (char __user *) ptr, second, &raddr); + if (ret) + return ret; +- return put_user(raddr, (ulong __user *) third); ++ return put_user(raddr, (__force ulong __user *) third); + } + case 1: /* iBCS2 emulator entry point */ + if (!segment_eq(get_fs(), get_ds())) +@@ -207,17 +428,3 @@ asmlinkage int sys_olduname(struct oldold_utsname __user *name) + + return error; + } +- +- +-/* +- * Do a system call from kernel instead of calling sys_execve so we +- * end up with proper pt_regs. +- */ +-int kernel_execve(const char *filename, char *const argv[], char *const envp[]) +-{ +- long __res; +- asm volatile ("push %%ebx ; movl %2,%%ebx ; int $0x80 ; pop %%ebx" +- : "=a" (__res) +- : "0" (__NR_execve), "ri" (filename), "c" (argv), "d" (envp) : "memory"); +- return __res; +-} +diff --git a/arch/x86/kernel/sys_x86_64.c b/arch/x86/kernel/sys_x86_64.c +index 8aa2057..b604bc1 100644 +--- a/arch/x86/kernel/sys_x86_64.c ++++ b/arch/x86/kernel/sys_x86_64.c +@@ -32,8 +32,8 @@ out: + return error; + } + +-static void find_start_end(unsigned long flags, unsigned long *begin, +- unsigned long *end) ++static void find_start_end(struct mm_struct *mm, unsigned long flags, ++ unsigned long *begin, unsigned long *end) + { + if (!test_thread_flag(TIF_IA32) && (flags & MAP_32BIT)) { + unsigned long new_begin; +@@ -52,7 +52,7 @@ static void find_start_end(unsigned long flags, unsigned long *begin, + *begin = new_begin; + } + } else { +- *begin = TASK_UNMAPPED_BASE; ++ *begin = mm->mmap_base; + *end = TASK_SIZE; + } + } +@@ -69,16 +69,19 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, + if (flags & MAP_FIXED) + return addr; + +- find_start_end(flags, &begin, &end); ++ find_start_end(mm, flags, &begin, &end); + + if (len > end) + return -ENOMEM; + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + if (addr) { + addr = PAGE_ALIGN(addr); + vma = find_vma(mm, addr); +- if (end - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (end - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + if (((flags & MAP_32BIT) || test_thread_flag(TIF_IA32)) +@@ -106,7 +109,7 @@ full_search: + } + return -ENOMEM; + } +- if (!vma || addr + len <= vma->vm_start) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* + * Remember the place where we stopped the search: + */ +@@ -128,7 +131,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + { + struct vm_area_struct *vma; + struct mm_struct *mm = current->mm; +- unsigned long addr = addr0; ++ unsigned long base = mm->mmap_base, addr = addr0; + + /* requested length too big for entire address space */ + if (len > TASK_SIZE) +@@ -141,13 +144,18 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + if (!test_thread_flag(TIF_IA32) && (flags & MAP_32BIT)) + goto bottomup; + ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + /* requesting a specific address */ + if (addr) { + addr = PAGE_ALIGN(addr); +- vma = find_vma(mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) +- return addr; ++ if (TASK_SIZE - len >= addr) { ++ vma = find_vma(mm, addr); ++ if (check_heap_stack_gap(vma, addr, len)) ++ return addr; ++ } + } + + /* check if free_area_cache is useful for us */ +@@ -162,7 +170,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + /* make sure it can fit in the remaining address space */ + if (addr > len) { + vma = find_vma(mm, addr-len); +- if (!vma || addr <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr - len, len)) + /* remember the address as a hint for next time */ + return mm->free_area_cache = addr-len; + } +@@ -179,7 +187,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + * return with success: + */ + vma = find_vma(mm, addr); +- if (!vma || addr+len <= vma->vm_start) ++ if (check_heap_stack_gap(vma, addr, len)) + /* remember the address as a hint for next time */ + return mm->free_area_cache = addr; + +@@ -188,8 +196,8 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, + mm->cached_hole_size = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = vma->vm_start-len; +- } while (len < vma->vm_start); ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); + + bottomup: + /* +@@ -198,13 +206,21 @@ bottomup: + * can happen with large stack limits and large mmap() + * allocations. + */ ++ mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ ++ mm->free_area_cache = mm->mmap_base; + mm->cached_hole_size = ~0UL; +- mm->free_area_cache = TASK_UNMAPPED_BASE; + addr = arch_get_unmapped_area(filp, addr0, len, pgoff, flags); + /* + * Restore the topdown base: + */ +- mm->free_area_cache = mm->mmap_base; ++ mm->mmap_base = base; ++ mm->free_area_cache = base; + mm->cached_hole_size = ~0UL; + + return addr; +diff --git a/arch/x86/kernel/syscall_table_32.S b/arch/x86/kernel/syscall_table_32.S +index 76d70a4..4c94a44 100644 +--- a/arch/x86/kernel/syscall_table_32.S ++++ b/arch/x86/kernel/syscall_table_32.S +@@ -1,3 +1,4 @@ ++.section .rodata,"a",@progbits + ENTRY(sys_call_table) + .long sys_restart_syscall /* 0 - old "setup()" system call, used for restarting */ + .long sys_exit +diff --git a/arch/x86/kernel/tboot.c b/arch/x86/kernel/tboot.c +index 46b8277..3349d55 100644 +--- a/arch/x86/kernel/tboot.c ++++ b/arch/x86/kernel/tboot.c +@@ -216,7 +216,7 @@ static int tboot_setup_sleep(void) + + void tboot_shutdown(u32 shutdown_type) + { +- void (*shutdown)(void); ++ void (* __noreturn shutdown)(void); + + if (!tboot_enabled()) + return; +@@ -238,7 +238,7 @@ void tboot_shutdown(u32 shutdown_type) + + switch_to_tboot_pt(); + +- shutdown = (void(*)(void))(unsigned long)tboot->shutdown_entry; ++ shutdown = (void *)tboot->shutdown_entry; + shutdown(); + + /* should not reach here */ +@@ -295,7 +295,7 @@ void tboot_sleep(u8 sleep_state, u32 pm1a_control, u32 pm1b_control) + tboot_shutdown(acpi_shutdown_map[sleep_state]); + } + +-static atomic_t ap_wfs_count; ++static atomic_unchecked_t ap_wfs_count; + + static int tboot_wait_for_aps(int num_aps) + { +@@ -319,9 +319,9 @@ static int __cpuinit tboot_cpu_callback(struct notifier_block *nfb, + { + switch (action) { + case CPU_DYING: +- atomic_inc(&ap_wfs_count); ++ atomic_inc_unchecked(&ap_wfs_count); + if (num_online_cpus() == 1) +- if (tboot_wait_for_aps(atomic_read(&ap_wfs_count))) ++ if (tboot_wait_for_aps(atomic_read_unchecked(&ap_wfs_count))) + return NOTIFY_BAD; + break; + } +@@ -340,7 +340,7 @@ static __init int tboot_late_init(void) + + tboot_create_trampoline(); + +- atomic_set(&ap_wfs_count, 0); ++ atomic_set_unchecked(&ap_wfs_count, 0); + register_hotcpu_notifier(&tboot_cpu_notifier); + return 0; + } +diff --git a/arch/x86/kernel/time.c b/arch/x86/kernel/time.c +index be25734..87fe232 100644 +--- a/arch/x86/kernel/time.c ++++ b/arch/x86/kernel/time.c +@@ -26,17 +26,13 @@ + int timer_ack; + #endif + +-#ifdef CONFIG_X86_64 +-volatile unsigned long __jiffies __section_jiffies = INITIAL_JIFFIES; +-#endif +- + unsigned long profile_pc(struct pt_regs *regs) + { + unsigned long pc = instruction_pointer(regs); + +- if (!user_mode_vm(regs) && in_lock_functions(pc)) { ++ if (!user_mode(regs) && in_lock_functions(pc)) { + #ifdef CONFIG_FRAME_POINTER +- return *(unsigned long *)(regs->bp + sizeof(long)); ++ return ktla_ktva(*(unsigned long *)(regs->bp + sizeof(long))); + #else + unsigned long *sp = + (unsigned long *)kernel_stack_pointer(regs); +@@ -45,11 +41,17 @@ unsigned long profile_pc(struct pt_regs *regs) + * or above a saved flags. Eflags has bits 22-31 zero, + * kernel addresses don't. + */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ return ktla_ktva(sp[0]); ++#else + if (sp[0] >> 22) + return sp[0]; + if (sp[1] >> 22) + return sp[1]; + #endif ++ ++#endif + } + return pc; + } +diff --git a/arch/x86/kernel/tls.c b/arch/x86/kernel/tls.c +index 6bb7b85..dd853e1 100644 +--- a/arch/x86/kernel/tls.c ++++ b/arch/x86/kernel/tls.c +@@ -85,6 +85,11 @@ int do_set_thread_area(struct task_struct *p, int idx, + if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) + return -EINVAL; + ++#ifdef CONFIG_PAX_SEGMEXEC ++ if ((p->mm->pax_flags & MF_PAX_SEGMEXEC) && (info.contents & MODIFY_LDT_CONTENTS_CODE)) ++ return -EINVAL; ++#endif ++ + set_tls_desc(p, idx, &info, 1); + + return 0; +diff --git a/arch/x86/kernel/trampoline_32.S b/arch/x86/kernel/trampoline_32.S +index 8508237..229b664 100644 +--- a/arch/x86/kernel/trampoline_32.S ++++ b/arch/x86/kernel/trampoline_32.S +@@ -32,6 +32,12 @@ + #include + #include + ++#ifdef CONFIG_PAX_KERNEXEC ++#define ta(X) (X) ++#else ++#define ta(X) ((X) - __PAGE_OFFSET) ++#endif ++ + /* We can free up trampoline after bootup if cpu hotplug is not supported. */ + __CPUINITRODATA + .code16 +@@ -60,7 +66,7 @@ r_base = . + inc %ax # protected mode (PE) bit + lmsw %ax # into protected mode + # flush prefetch and jump to startup_32_smp in arch/i386/kernel/head.S +- ljmpl $__BOOT_CS, $(startup_32_smp-__PAGE_OFFSET) ++ ljmpl $__BOOT_CS, $ta(startup_32_smp) + + # These need to be in the same 64K segment as the above; + # hence we don't use the boot_gdt_descr defined in head.S +diff --git a/arch/x86/kernel/trampoline_64.S b/arch/x86/kernel/trampoline_64.S +index 3af2dff..ba8aa49 100644 +--- a/arch/x86/kernel/trampoline_64.S ++++ b/arch/x86/kernel/trampoline_64.S +@@ -91,7 +91,7 @@ startup_32: + movl $__KERNEL_DS, %eax # Initialize the %ds segment register + movl %eax, %ds + +- movl $X86_CR4_PAE, %eax ++ movl $(X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE), %eax + movl %eax, %cr4 # Enable PAE mode + + # Setup trampoline 4 level pagetables +@@ -127,7 +127,7 @@ startup_64: + no_longmode: + hlt + jmp no_longmode +-#include "verify_cpu_64.S" ++#include "verify_cpu.S" + + # Careful these need to be in the same 64K segment as the above; + tidt: +@@ -138,7 +138,7 @@ tidt: + # so the kernel can live anywhere + .balign 4 + tgdt: +- .short tgdt_end - tgdt # gdt limit ++ .short tgdt_end - tgdt - 1 # gdt limit + .long tgdt - r_base + .short 0 + .quad 0x00cf9b000000ffff # __KERNEL32_CS +diff --git a/arch/x86/kernel/traps.c b/arch/x86/kernel/traps.c +index 7e37dce..ec3f8e5 100644 +--- a/arch/x86/kernel/traps.c ++++ b/arch/x86/kernel/traps.c +@@ -69,12 +69,6 @@ asmlinkage int system_call(void); + + /* Do we ignore FPU interrupts ? */ + char ignore_fpu_irq; +- +-/* +- * The IDT has to be page-aligned to simplify the Pentium +- * F0 0F bug workaround. +- */ +-gate_desc idt_table[NR_VECTORS] __page_aligned_data = { { { { 0, 0 } } }, }; + #endif + + DECLARE_BITMAP(used_vectors, NR_VECTORS); +@@ -112,19 +106,19 @@ static inline void preempt_conditional_cli(struct pt_regs *regs) + static inline void + die_if_kernel(const char *str, struct pt_regs *regs, long err) + { +- if (!user_mode_vm(regs)) ++ if (!user_mode(regs)) + die(str, regs, err); + } + #endif + + static void __kprobes +-do_trap(int trapnr, int signr, char *str, struct pt_regs *regs, ++do_trap(int trapnr, int signr, const char *str, struct pt_regs *regs, + long error_code, siginfo_t *info) + { + struct task_struct *tsk = current; + + #ifdef CONFIG_X86_32 +- if (regs->flags & X86_VM_MASK) { ++ if (v8086_mode(regs)) { + /* + * traps 0, 1, 3, 4, and 5 should be forwarded to vm86. + * On nmi (interrupt 2), do_trap should not be called. +@@ -135,7 +129,7 @@ do_trap(int trapnr, int signr, char *str, struct pt_regs *regs, + } + #endif + +- if (!user_mode(regs)) ++ if (!user_mode_novm(regs)) + goto kernel_trap; + + #ifdef CONFIG_X86_32 +@@ -158,7 +152,7 @@ trap_signal: + printk_ratelimit()) { + printk(KERN_INFO + "%s[%d] trap %s ip:%lx sp:%lx error:%lx", +- tsk->comm, tsk->pid, str, ++ tsk->comm, task_pid_nr(tsk), str, + regs->ip, regs->sp, error_code); + print_vma_addr(" in ", regs->ip); + printk("\n"); +@@ -175,8 +169,20 @@ kernel_trap: + if (!fixup_exception(regs)) { + tsk->thread.error_code = error_code; + tsk->thread.trap_no = trapnr; ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ if (trapnr == 12 && ((regs->cs & 0xFFFF) == __KERNEL_CS || (regs->cs & 0xFFFF) == __KERNEXEC_KERNEL_CS)) ++ str = "PAX: suspicious stack segment fault"; ++#endif ++ + die(str, regs, error_code); + } ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ if (trapnr == 4) ++ pax_report_refcount_overflow(regs); ++#endif ++ + return; + + #ifdef CONFIG_X86_32 +@@ -265,14 +271,30 @@ do_general_protection(struct pt_regs *regs, long error_code) + conditional_sti(regs); + + #ifdef CONFIG_X86_32 +- if (regs->flags & X86_VM_MASK) ++ if (v8086_mode(regs)) + goto gp_in_vm86; + #endif + + tsk = current; +- if (!user_mode(regs)) ++ if (!user_mode_novm(regs)) + goto gp_in_kernel; + ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) ++ if (!nx_enabled && tsk->mm && (tsk->mm->pax_flags & MF_PAX_PAGEEXEC)) { ++ struct mm_struct *mm = tsk->mm; ++ unsigned long limit; ++ ++ down_write(&mm->mmap_sem); ++ limit = mm->context.user_cs_limit; ++ if (limit < TASK_SIZE) { ++ track_exec_limit(mm, limit, TASK_SIZE, VM_EXEC); ++ up_write(&mm->mmap_sem); ++ return; ++ } ++ up_write(&mm->mmap_sem); ++ } ++#endif ++ + tsk->thread.error_code = error_code; + tsk->thread.trap_no = 13; + +@@ -305,6 +327,13 @@ gp_in_kernel: + if (notify_die(DIE_GPF, "general protection fault", regs, + error_code, 13, SIGSEGV) == NOTIFY_STOP) + return; ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ if ((regs->cs & 0xFFFF) == __KERNEL_CS || (regs->cs & 0xFFFF) == __KERNEXEC_KERNEL_CS) ++ die("PAX: suspicious general protection fault", regs, error_code); ++ else ++#endif ++ + die("general protection fault", regs, error_code); + } + +@@ -435,6 +464,17 @@ static notrace __kprobes void default_do_nmi(struct pt_regs *regs) + dotraplinkage notrace __kprobes void + do_nmi(struct pt_regs *regs, long error_code) + { ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ if (!user_mode(regs)) { ++ unsigned long cs = regs->cs & 0xFFFF; ++ unsigned long ip = ktva_ktla(regs->ip); ++ ++ if ((cs == __KERNEL_CS || cs == __KERNEXEC_KERNEL_CS) && ip <= (unsigned long)_etext) ++ regs->ip = ip; ++ } ++#endif ++ + nmi_enter(); + + inc_irq_stat(__nmi_count); +@@ -558,7 +598,7 @@ dotraplinkage void __kprobes do_debug(struct pt_regs *regs, long error_code) + } + + #ifdef CONFIG_X86_32 +- if (regs->flags & X86_VM_MASK) ++ if (v8086_mode(regs)) + goto debug_vm86; + #endif + +@@ -570,7 +610,7 @@ dotraplinkage void __kprobes do_debug(struct pt_regs *regs, long error_code) + * kernel space (but re-enable TF when returning to user mode). + */ + if (condition & DR_STEP) { +- if (!user_mode(regs)) ++ if (!user_mode_novm(regs)) + goto clear_TF_reenable; + } + +@@ -757,7 +797,7 @@ do_simd_coprocessor_error(struct pt_regs *regs, long error_code) + * Handle strange cache flush from user space exception + * in all other cases. This is undocumented behaviour. + */ +- if (regs->flags & X86_VM_MASK) { ++ if (v8086_mode(regs)) { + handle_vm86_fault((struct kernel_vm86_regs *)regs, error_code); + return; + } +@@ -798,7 +838,7 @@ asmlinkage void __attribute__((weak)) smp_threshold_interrupt(void) + void __math_state_restore(void) + { + struct thread_info *thread = current_thread_info(); +- struct task_struct *tsk = thread->task; ++ struct task_struct *tsk = current; + + /* + * Paranoid restore. send a SIGSEGV if we fail to restore the state. +@@ -825,8 +865,7 @@ void __math_state_restore(void) + */ + asmlinkage void math_state_restore(void) + { +- struct thread_info *thread = current_thread_info(); +- struct task_struct *tsk = thread->task; ++ struct task_struct *tsk = current; + + if (!tsk_used_math(tsk)) { + local_irq_enable(); +diff --git a/arch/x86/kernel/verify_cpu.S b/arch/x86/kernel/verify_cpu.S +new file mode 100644 +index 0000000..50c5edd +--- /dev/null ++++ b/arch/x86/kernel/verify_cpu.S +@@ -0,0 +1,140 @@ ++/* ++ * ++ * verify_cpu.S - Code for cpu long mode and SSE verification. This ++ * code has been borrowed from boot/setup.S and was introduced by ++ * Andi Kleen. ++ * ++ * Copyright (c) 2007 Andi Kleen (ak@suse.de) ++ * Copyright (c) 2007 Eric Biederman (ebiederm@xmission.com) ++ * Copyright (c) 2007 Vivek Goyal (vgoyal@in.ibm.com) ++ * Copyright (c) 2010 Kees Cook (kees.cook@canonical.com) ++ * ++ * This source code is licensed under the GNU General Public License, ++ * Version 2. See the file COPYING for more details. ++ * ++ * This is a common code for verification whether CPU supports ++ * long mode and SSE or not. It is not called directly instead this ++ * file is included at various places and compiled in that context. ++ * This file is expected to run in 32bit code. Currently: ++ * ++ * arch/x86/boot/compressed/head_64.S: Boot cpu verification ++ * arch/x86/kernel/trampoline_64.S: secondary processor verification ++ * arch/x86/kernel/head_32.S: processor startup ++ * arch/x86/kernel/acpi/realmode/wakeup.S: 32bit processor resume ++ * ++ * verify_cpu, returns the status of longmode and SSE in register %eax. ++ * 0: Success 1: Failure ++ * ++ * On Intel, the XD_DISABLE flag will be cleared as a side-effect. ++ * ++ * The caller needs to check for the error code and take the action ++ * appropriately. Either display a message or halt. ++ */ ++ ++#include ++#include ++ ++verify_cpu: ++ pushfl # Save caller passed flags ++ pushl $0 # Kill any dangerous flags ++ popfl ++ ++ pushfl # standard way to check for cpuid ++ popl %eax ++ movl %eax,%ebx ++ xorl $0x200000,%eax ++ pushl %eax ++ popfl ++ pushfl ++ popl %eax ++ cmpl %eax,%ebx ++ jz verify_cpu_no_longmode # cpu has no cpuid ++ ++ movl $0x0,%eax # See if cpuid 1 is implemented ++ cpuid ++ cmpl $0x1,%eax ++ jb verify_cpu_no_longmode # no cpuid 1 ++ ++ xor %di,%di ++ cmpl $0x68747541,%ebx # AuthenticAMD ++ jnz verify_cpu_noamd ++ cmpl $0x69746e65,%edx ++ jnz verify_cpu_noamd ++ cmpl $0x444d4163,%ecx ++ jnz verify_cpu_noamd ++ mov $1,%di # cpu is from AMD ++ jmp verify_cpu_check ++ ++verify_cpu_noamd: ++ cmpl $0x756e6547,%ebx # GenuineIntel? ++ jnz verify_cpu_check ++ cmpl $0x49656e69,%edx ++ jnz verify_cpu_check ++ cmpl $0x6c65746e,%ecx ++ jnz verify_cpu_check ++ ++ # only call IA32_MISC_ENABLE when: ++ # family > 6 || (family == 6 && model >= 0xd) ++ movl $0x1, %eax # check CPU family and model ++ cpuid ++ movl %eax, %ecx ++ ++ andl $0x0ff00f00, %eax # mask family and extended family ++ shrl $8, %eax ++ cmpl $6, %eax ++ ja verify_cpu_clear_xd # family > 6, ok ++ jb verify_cpu_check # family < 6, skip ++ ++ andl $0x000f00f0, %ecx # mask model and extended model ++ shrl $4, %ecx ++ cmpl $0xd, %ecx ++ jb verify_cpu_check # family == 6, model < 0xd, skip ++ ++verify_cpu_clear_xd: ++ movl $MSR_IA32_MISC_ENABLE, %ecx ++ rdmsr ++ btrl $2, %edx # clear MSR_IA32_MISC_ENABLE_XD_DISABLE ++ jnc verify_cpu_check # only write MSR if bit was changed ++ wrmsr ++ ++verify_cpu_check: ++ movl $0x1,%eax # Does the cpu have what it takes ++ cpuid ++ andl $REQUIRED_MASK0,%edx ++ xorl $REQUIRED_MASK0,%edx ++ jnz verify_cpu_no_longmode ++ ++ movl $0x80000000,%eax # See if extended cpuid is implemented ++ cpuid ++ cmpl $0x80000001,%eax ++ jb verify_cpu_no_longmode # no extended cpuid ++ ++ movl $0x80000001,%eax # Does the cpu have what it takes ++ cpuid ++ andl $REQUIRED_MASK1,%edx ++ xorl $REQUIRED_MASK1,%edx ++ jnz verify_cpu_no_longmode ++ ++verify_cpu_sse_test: ++ movl $1,%eax ++ cpuid ++ andl $SSE_MASK,%edx ++ cmpl $SSE_MASK,%edx ++ je verify_cpu_sse_ok ++ test %di,%di ++ jz verify_cpu_no_longmode # only try to force SSE on AMD ++ movl $MSR_K7_HWCR,%ecx ++ rdmsr ++ btr $15,%eax # enable SSE ++ wrmsr ++ xor %di,%di # don't loop ++ jmp verify_cpu_sse_test # try again ++ ++verify_cpu_no_longmode: ++ popfl # Restore caller passed flags ++ movl $1,%eax ++ ret ++verify_cpu_sse_ok: ++ popfl # Restore caller passed flags ++ xorl %eax, %eax ++ ret +diff --git a/arch/x86/kernel/verify_cpu_64.S b/arch/x86/kernel/verify_cpu_64.S +deleted file mode 100644 +index 45b6f8a..0000000 +--- a/arch/x86/kernel/verify_cpu_64.S ++++ /dev/null +@@ -1,105 +0,0 @@ +-/* +- * +- * verify_cpu.S - Code for cpu long mode and SSE verification. This +- * code has been borrowed from boot/setup.S and was introduced by +- * Andi Kleen. +- * +- * Copyright (c) 2007 Andi Kleen (ak@suse.de) +- * Copyright (c) 2007 Eric Biederman (ebiederm@xmission.com) +- * Copyright (c) 2007 Vivek Goyal (vgoyal@in.ibm.com) +- * +- * This source code is licensed under the GNU General Public License, +- * Version 2. See the file COPYING for more details. +- * +- * This is a common code for verification whether CPU supports +- * long mode and SSE or not. It is not called directly instead this +- * file is included at various places and compiled in that context. +- * Following are the current usage. +- * +- * This file is included by both 16bit and 32bit code. +- * +- * arch/x86_64/boot/setup.S : Boot cpu verification (16bit) +- * arch/x86_64/boot/compressed/head.S: Boot cpu verification (32bit) +- * arch/x86_64/kernel/trampoline.S: secondary processor verfication (16bit) +- * arch/x86_64/kernel/acpi/wakeup.S:Verfication at resume (16bit) +- * +- * verify_cpu, returns the status of cpu check in register %eax. +- * 0: Success 1: Failure +- * +- * The caller needs to check for the error code and take the action +- * appropriately. Either display a message or halt. +- */ +- +-#include +- +-verify_cpu: +- pushfl # Save caller passed flags +- pushl $0 # Kill any dangerous flags +- popfl +- +- pushfl # standard way to check for cpuid +- popl %eax +- movl %eax,%ebx +- xorl $0x200000,%eax +- pushl %eax +- popfl +- pushfl +- popl %eax +- cmpl %eax,%ebx +- jz verify_cpu_no_longmode # cpu has no cpuid +- +- movl $0x0,%eax # See if cpuid 1 is implemented +- cpuid +- cmpl $0x1,%eax +- jb verify_cpu_no_longmode # no cpuid 1 +- +- xor %di,%di +- cmpl $0x68747541,%ebx # AuthenticAMD +- jnz verify_cpu_noamd +- cmpl $0x69746e65,%edx +- jnz verify_cpu_noamd +- cmpl $0x444d4163,%ecx +- jnz verify_cpu_noamd +- mov $1,%di # cpu is from AMD +- +-verify_cpu_noamd: +- movl $0x1,%eax # Does the cpu have what it takes +- cpuid +- andl $REQUIRED_MASK0,%edx +- xorl $REQUIRED_MASK0,%edx +- jnz verify_cpu_no_longmode +- +- movl $0x80000000,%eax # See if extended cpuid is implemented +- cpuid +- cmpl $0x80000001,%eax +- jb verify_cpu_no_longmode # no extended cpuid +- +- movl $0x80000001,%eax # Does the cpu have what it takes +- cpuid +- andl $REQUIRED_MASK1,%edx +- xorl $REQUIRED_MASK1,%edx +- jnz verify_cpu_no_longmode +- +-verify_cpu_sse_test: +- movl $1,%eax +- cpuid +- andl $SSE_MASK,%edx +- cmpl $SSE_MASK,%edx +- je verify_cpu_sse_ok +- test %di,%di +- jz verify_cpu_no_longmode # only try to force SSE on AMD +- movl $0xc0010015,%ecx # HWCR +- rdmsr +- btr $15,%eax # enable SSE +- wrmsr +- xor %di,%di # don't loop +- jmp verify_cpu_sse_test # try again +- +-verify_cpu_no_longmode: +- popfl # Restore caller passed flags +- movl $1,%eax +- ret +-verify_cpu_sse_ok: +- popfl # Restore caller passed flags +- xorl %eax, %eax +- ret +diff --git a/arch/x86/kernel/vm86_32.c b/arch/x86/kernel/vm86_32.c +index 9c4e625..c992817 100644 +--- a/arch/x86/kernel/vm86_32.c ++++ b/arch/x86/kernel/vm86_32.c +@@ -41,6 +41,7 @@ + #include + #include + #include ++#include + + #include + #include +@@ -148,7 +149,7 @@ struct pt_regs *save_v86_state(struct kernel_vm86_regs *regs) + do_exit(SIGSEGV); + } + +- tss = &per_cpu(init_tss, get_cpu()); ++ tss = init_tss + get_cpu(); + current->thread.sp0 = current->thread.saved_sp0; + current->thread.sysenter_cs = __KERNEL_CS; + load_sp0(tss, ¤t->thread); +@@ -208,6 +209,13 @@ int sys_vm86old(struct pt_regs *regs) + struct task_struct *tsk; + int tmp, ret = -EPERM; + ++#ifdef CONFIG_GRKERNSEC_VM86 ++ if (!capable(CAP_SYS_RAWIO)) { ++ gr_handle_vm86(); ++ goto out; ++ } ++#endif ++ + tsk = current; + if (tsk->thread.saved_sp0) + goto out; +@@ -238,6 +246,14 @@ int sys_vm86(struct pt_regs *regs) + int tmp, ret; + struct vm86plus_struct __user *v86; + ++#ifdef CONFIG_GRKERNSEC_VM86 ++ if (!capable(CAP_SYS_RAWIO)) { ++ gr_handle_vm86(); ++ ret = -EPERM; ++ goto out; ++ } ++#endif ++ + tsk = current; + switch (regs->bx) { + case VM86_REQUEST_IRQ: +@@ -324,7 +340,7 @@ static void do_sys_vm86(struct kernel_vm86_struct *info, struct task_struct *tsk + tsk->thread.saved_fs = info->regs32->fs; + tsk->thread.saved_gs = get_user_gs(info->regs32); + +- tss = &per_cpu(init_tss, get_cpu()); ++ tss = init_tss + get_cpu(); + tsk->thread.sp0 = (unsigned long) &info->VM86_TSS_ESP0; + if (cpu_has_sep) + tsk->thread.sysenter_cs = 0; +@@ -529,7 +545,7 @@ static void do_int(struct kernel_vm86_regs *regs, int i, + goto cannot_handle; + if (i == 0x21 && is_revectored(AH(regs), &KVM86->int21_revectored)) + goto cannot_handle; +- intr_ptr = (unsigned long __user *) (i << 2); ++ intr_ptr = (__force unsigned long __user *) (i << 2); + if (get_user(segoffs, intr_ptr)) + goto cannot_handle; + if ((segoffs >> 16) == BIOSSEG) +diff --git a/arch/x86/kernel/vmi_32.c b/arch/x86/kernel/vmi_32.c +index d430e4c..831f817 100644 +--- a/arch/x86/kernel/vmi_32.c ++++ b/arch/x86/kernel/vmi_32.c +@@ -44,12 +44,17 @@ typedef u32 __attribute__((regparm(1))) (VROMFUNC)(void); + typedef u64 __attribute__((regparm(2))) (VROMLONGFUNC)(int); + + #define call_vrom_func(rom,func) \ +- (((VROMFUNC *)(rom->func))()) ++ (((VROMFUNC *)(ktva_ktla(rom.func)))()) + + #define call_vrom_long_func(rom,func,arg) \ +- (((VROMLONGFUNC *)(rom->func)) (arg)) ++({\ ++ u64 __reloc = ((VROMLONGFUNC *)(ktva_ktla(rom.func))) (arg);\ ++ struct vmi_relocation_info *const __rel = (struct vmi_relocation_info *)&__reloc;\ ++ __rel->eip = (unsigned char *)ktva_ktla((unsigned long)__rel->eip);\ ++ __reloc;\ ++}) + +-static struct vrom_header *vmi_rom; ++static struct vrom_header vmi_rom __attribute((__section__(".vmi.rom"), __aligned__(PAGE_SIZE))); + static int disable_pge; + static int disable_pse; + static int disable_sep; +@@ -76,10 +81,10 @@ static struct { + void (*set_initial_ap_state)(int, int); + void (*halt)(void); + void (*set_lazy_mode)(int mode); +-} vmi_ops; ++} __no_const vmi_ops __read_only; + + /* Cached VMI operations */ +-struct vmi_timer_ops vmi_timer_ops; ++struct vmi_timer_ops vmi_timer_ops __read_only; + + /* + * VMI patching routines. +@@ -94,7 +99,7 @@ struct vmi_timer_ops vmi_timer_ops; + static inline void patch_offset(void *insnbuf, + unsigned long ip, unsigned long dest) + { +- *(unsigned long *)(insnbuf+1) = dest-ip-5; ++ *(unsigned long *)(insnbuf+1) = dest-ip-5; + } + + static unsigned patch_internal(int call, unsigned len, void *insnbuf, +@@ -102,6 +107,7 @@ static unsigned patch_internal(int call, unsigned len, void *insnbuf, + { + u64 reloc; + struct vmi_relocation_info *const rel = (struct vmi_relocation_info *)&reloc; ++ + reloc = call_vrom_long_func(vmi_rom, get_reloc, call); + switch(rel->type) { + case VMI_RELOCATION_CALL_REL: +@@ -404,13 +410,13 @@ static void vmi_set_pud(pud_t *pudp, pud_t pudval) + + static void vmi_pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) + { +- const pte_t pte = { .pte = 0 }; ++ const pte_t pte = __pte(0ULL); + vmi_ops.set_pte(pte, ptep, vmi_flags_addr(mm, addr, VMI_PAGE_PT, 0)); + } + + static void vmi_pmd_clear(pmd_t *pmd) + { +- const pte_t pte = { .pte = 0 }; ++ const pte_t pte = __pte(0ULL); + vmi_ops.set_pte(pte, (pte_t *)pmd, VMI_PAGE_PD); + } + #endif +@@ -438,10 +444,10 @@ vmi_startup_ipi_hook(int phys_apicid, unsigned long start_eip, + ap.ss = __KERNEL_DS; + ap.esp = (unsigned long) start_esp; + +- ap.ds = __USER_DS; +- ap.es = __USER_DS; ++ ap.ds = __KERNEL_DS; ++ ap.es = __KERNEL_DS; + ap.fs = __KERNEL_PERCPU; +- ap.gs = __KERNEL_STACK_CANARY; ++ savesegment(gs, ap.gs); + + ap.eflags = 0; + +@@ -486,6 +492,18 @@ static void vmi_leave_lazy_mmu(void) + paravirt_leave_lazy_mmu(); + } + ++#ifdef CONFIG_PAX_KERNEXEC ++static unsigned long vmi_pax_open_kernel(void) ++{ ++ return 0; ++} ++ ++static unsigned long vmi_pax_close_kernel(void) ++{ ++ return 0; ++} ++#endif ++ + static inline int __init check_vmi_rom(struct vrom_header *rom) + { + struct pci_header *pci; +@@ -498,6 +516,10 @@ static inline int __init check_vmi_rom(struct vrom_header *rom) + return 0; + if (rom->vrom_signature != VMI_SIGNATURE) + return 0; ++ if (rom->rom_length * 512 > sizeof(*rom)) { ++ printk(KERN_WARNING "PAX: VMI: ROM size too big: %x\n", rom->rom_length * 512); ++ return 0; ++ } + if (rom->api_version_maj != VMI_API_REV_MAJOR || + rom->api_version_min+1 < VMI_API_REV_MINOR+1) { + printk(KERN_WARNING "VMI: Found mismatched rom version %d.%d\n", +@@ -562,7 +584,7 @@ static inline int __init probe_vmi_rom(void) + struct vrom_header *romstart; + romstart = (struct vrom_header *)isa_bus_to_virt(base); + if (check_vmi_rom(romstart)) { +- vmi_rom = romstart; ++ vmi_rom = *romstart; + return 1; + } + } +@@ -836,6 +858,11 @@ static inline int __init activate_vmi(void) + + para_fill(pv_irq_ops.safe_halt, Halt); + ++#ifdef CONFIG_PAX_KERNEXEC ++ pv_mmu_ops.pax_open_kernel = vmi_pax_open_kernel; ++ pv_mmu_ops.pax_close_kernel = vmi_pax_close_kernel; ++#endif ++ + /* + * Alternative instruction rewriting doesn't happen soon enough + * to convert VMI_IRET to a call instead of a jump; so we have +@@ -853,16 +880,16 @@ static inline int __init activate_vmi(void) + + void __init vmi_init(void) + { +- if (!vmi_rom) ++ if (!vmi_rom.rom_signature) + probe_vmi_rom(); + else +- check_vmi_rom(vmi_rom); ++ check_vmi_rom(&vmi_rom); + + /* In case probing for or validating the ROM failed, basil */ +- if (!vmi_rom) ++ if (!vmi_rom.rom_signature) + return; + +- reserve_top_address(-vmi_rom->virtual_top); ++ reserve_top_address(-vmi_rom.virtual_top); + + #ifdef CONFIG_X86_IO_APIC + /* This is virtual hardware; timer routing is wired correctly */ +@@ -874,7 +901,7 @@ void __init vmi_activate(void) + { + unsigned long flags; + +- if (!vmi_rom) ++ if (!vmi_rom.rom_signature) + return; + + local_irq_save(flags); +diff --git a/arch/x86/kernel/vmlinux.lds.S b/arch/x86/kernel/vmlinux.lds.S +index 3c68fe2..12c8280 100644 +--- a/arch/x86/kernel/vmlinux.lds.S ++++ b/arch/x86/kernel/vmlinux.lds.S +@@ -26,6 +26,13 @@ + #include + #include + #include ++#include ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++#define __KERNEL_TEXT_OFFSET (LOAD_OFFSET + ____LOAD_PHYSICAL_ADDR) ++#else ++#define __KERNEL_TEXT_OFFSET 0 ++#endif + + #undef i386 /* in case the preprocessor is a 32bit one */ + +@@ -34,40 +41,53 @@ OUTPUT_FORMAT(CONFIG_OUTPUT_FORMAT, CONFIG_OUTPUT_FORMAT, CONFIG_OUTPUT_FORMAT) + #ifdef CONFIG_X86_32 + OUTPUT_ARCH(i386) + ENTRY(phys_startup_32) +-jiffies = jiffies_64; + #else + OUTPUT_ARCH(i386:x86-64) + ENTRY(phys_startup_64) +-jiffies_64 = jiffies; + #endif + + PHDRS { + text PT_LOAD FLAGS(5); /* R_E */ +- data PT_LOAD FLAGS(7); /* RWE */ ++#ifdef CONFIG_X86_32 ++ module PT_LOAD FLAGS(5); /* R_E */ ++#endif ++#ifdef CONFIG_XEN ++ rodata PT_LOAD FLAGS(5); /* R_E */ ++#else ++ rodata PT_LOAD FLAGS(4); /* R__ */ ++#endif ++ data PT_LOAD FLAGS(6); /* RW_ */ + #ifdef CONFIG_X86_64 + user PT_LOAD FLAGS(5); /* R_E */ ++#endif ++ init.begin PT_LOAD FLAGS(6); /* RW_ */ + #ifdef CONFIG_SMP + percpu PT_LOAD FLAGS(6); /* RW_ */ + #endif ++ text.init PT_LOAD FLAGS(5); /* R_E */ ++ text.exit PT_LOAD FLAGS(5); /* R_E */ + init PT_LOAD FLAGS(7); /* RWE */ +-#endif + note PT_NOTE FLAGS(0); /* ___ */ + } + + SECTIONS + { + #ifdef CONFIG_X86_32 +- . = LOAD_OFFSET + LOAD_PHYSICAL_ADDR; +- phys_startup_32 = startup_32 - LOAD_OFFSET; ++ . = LOAD_OFFSET + ____LOAD_PHYSICAL_ADDR; + #else +- . = __START_KERNEL; +- phys_startup_64 = startup_64 - LOAD_OFFSET; ++ . = __START_KERNEL; + #endif + + /* Text and read-only data */ +- .text : AT(ADDR(.text) - LOAD_OFFSET) { +- _text = .; ++ .text (. - __KERNEL_TEXT_OFFSET): AT(ADDR(.text) - LOAD_OFFSET + __KERNEL_TEXT_OFFSET) { + /* bootstrapping code */ ++#ifdef CONFIG_X86_32 ++ phys_startup_32 = startup_32 - LOAD_OFFSET + __KERNEL_TEXT_OFFSET; ++#else ++ phys_startup_64 = startup_64 - LOAD_OFFSET + __KERNEL_TEXT_OFFSET; ++#endif ++ __LOAD_PHYSICAL_ADDR = . - LOAD_OFFSET + __KERNEL_TEXT_OFFSET; ++ _text = .; + HEAD_TEXT + #ifdef CONFIG_X86_32 + . = ALIGN(PAGE_SIZE); +@@ -82,28 +102,71 @@ SECTIONS + IRQENTRY_TEXT + *(.fixup) + *(.gnu.warning) +- /* End of text section */ +- _etext = .; + } :text = 0x9090 + +- NOTES :text :note ++ . += __KERNEL_TEXT_OFFSET; + +- EXCEPTION_TABLE(16) :text = 0x9090 ++#ifdef CONFIG_X86_32 ++ . = ALIGN(PAGE_SIZE); ++ .vmi.rom : AT(ADDR(.vmi.rom) - LOAD_OFFSET) { ++ *(.vmi.rom) ++ } :module ++ ++ . = ALIGN(PAGE_SIZE); ++ .module.text : AT(ADDR(.module.text) - LOAD_OFFSET) { ++ ++#if defined(CONFIG_PAX_KERNEXEC) && defined(CONFIG_MODULES) ++ MODULES_EXEC_VADDR = .; ++ BYTE(0) ++ . += (CONFIG_PAX_KERNEXEC_MODULE_TEXT * 1024 * 1024); ++ . = ALIGN(HPAGE_SIZE); ++ MODULES_EXEC_END = . - 1; ++#endif ++ ++ } :module ++#endif ++ ++ .text.end : AT(ADDR(.text.end) - LOAD_OFFSET) { ++ /* End of text section */ ++ _etext = . - __KERNEL_TEXT_OFFSET; ++ } ++ ++#ifdef CONFIG_X86_32 ++ . = ALIGN(PAGE_SIZE); ++ .rodata.page_aligned : AT(ADDR(.rodata.page_aligned) - LOAD_OFFSET) { ++ *(.idt) ++ . = ALIGN(PAGE_SIZE); ++ *(.empty_zero_page) ++ *(.swapper_pg_fixmap) ++ *(.swapper_pg_pmd) ++ *(.swapper_pg_dir) ++ *(.trampoline_pg_dir) ++ } :rodata ++#endif ++ ++ . = ALIGN(PAGE_SIZE); ++ NOTES :rodata :note ++ ++ EXCEPTION_TABLE(16) :rodata + + RO_DATA(PAGE_SIZE) + + /* Data */ + .data : AT(ADDR(.data) - LOAD_OFFSET) { ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ . = ALIGN(HPAGE_SIZE); ++#else ++ . = ALIGN(PAGE_SIZE); ++#endif ++ + /* Start of data section */ + _sdata = .; + + /* init_task */ + INIT_TASK_DATA(THREAD_SIZE) + +-#ifdef CONFIG_X86_32 +- /* 32 bit has nosave before _edata */ + NOSAVE_DATA +-#endif + + PAGE_ALIGNED_DATA(PAGE_SIZE) + +@@ -112,6 +175,8 @@ SECTIONS + DATA_DATA + CONSTRUCTORS + ++ jiffies = jiffies_64; ++ + /* rarely changed data like cpu maps */ + READ_MOSTLY_DATA(CONFIG_X86_INTERNODE_CACHE_BYTES) + +@@ -166,12 +231,6 @@ SECTIONS + } + vgetcpu_mode = VVIRT(.vgetcpu_mode); + +- . = ALIGN(CONFIG_X86_L1_CACHE_BYTES); +- .jiffies : AT(VLOAD(.jiffies)) { +- *(.jiffies) +- } +- jiffies = VVIRT(.jiffies); +- + .vsyscall_3 ADDR(.vsyscall_0) + 3072: AT(VLOAD(.vsyscall_3)) { + *(.vsyscall_3) + } +@@ -187,12 +246,19 @@ SECTIONS + #endif /* CONFIG_X86_64 */ + + /* Init code and data - will be freed after init */ +- . = ALIGN(PAGE_SIZE); + .init.begin : AT(ADDR(.init.begin) - LOAD_OFFSET) { ++ BYTE(0) ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ . = ALIGN(HPAGE_SIZE); ++#else ++ . = ALIGN(PAGE_SIZE); ++#endif ++ + __init_begin = .; /* paired with __init_end */ +- } ++ } :init.begin + +-#if defined(CONFIG_X86_64) && defined(CONFIG_SMP) ++#ifdef CONFIG_SMP + /* + * percpu offsets are zero-based on SMP. PERCPU_VADDR() changes the + * output PHDR, so the next output section - .init.text - should +@@ -201,12 +267,27 @@ SECTIONS + PERCPU_VADDR(0, :percpu) + #endif + +- INIT_TEXT_SECTION(PAGE_SIZE) +-#ifdef CONFIG_X86_64 +- :init +-#endif ++ . = ALIGN(PAGE_SIZE); ++ init_begin = .; ++ .init.text (. - __KERNEL_TEXT_OFFSET): AT(init_begin - LOAD_OFFSET) { ++ VMLINUX_SYMBOL(_sinittext) = .; ++ INIT_TEXT ++ VMLINUX_SYMBOL(_einittext) = .; ++ . = ALIGN(PAGE_SIZE); ++ } :text.init + +- INIT_DATA_SECTION(16) ++ /* ++ * .exit.text is discard at runtime, not link time, to deal with ++ * references from .altinstructions and .eh_frame ++ */ ++ .exit.text : AT(ADDR(.exit.text) - LOAD_OFFSET + __KERNEL_TEXT_OFFSET) { ++ EXIT_TEXT ++ . = ALIGN(16); ++ } :text.exit ++ . = init_begin + SIZEOF(.init.text) + SIZEOF(.exit.text); ++ ++ . = ALIGN(PAGE_SIZE); ++ INIT_DATA_SECTION(16) :init + + .x86_cpu_dev.init : AT(ADDR(.x86_cpu_dev.init) - LOAD_OFFSET) { + __x86_cpu_dev_start = .; +@@ -232,19 +313,11 @@ SECTIONS + *(.altinstr_replacement) + } + +- /* +- * .exit.text is discard at runtime, not link time, to deal with +- * references from .altinstructions and .eh_frame +- */ +- .exit.text : AT(ADDR(.exit.text) - LOAD_OFFSET) { +- EXIT_TEXT +- } +- + .exit.data : AT(ADDR(.exit.data) - LOAD_OFFSET) { + EXIT_DATA + } + +-#if !defined(CONFIG_X86_64) || !defined(CONFIG_SMP) ++#ifndef CONFIG_SMP + PERCPU(PAGE_SIZE) + #endif + +@@ -267,12 +340,6 @@ SECTIONS + . = ALIGN(PAGE_SIZE); + } + +-#ifdef CONFIG_X86_64 +- .data_nosave : AT(ADDR(.data_nosave) - LOAD_OFFSET) { +- NOSAVE_DATA +- } +-#endif +- + /* BSS */ + . = ALIGN(PAGE_SIZE); + .bss : AT(ADDR(.bss) - LOAD_OFFSET) { +@@ -288,6 +355,7 @@ SECTIONS + __brk_base = .; + . += 64 * 1024; /* 64k alignment slop space */ + *(.brk_reservation) /* areas brk users have reserved */ ++ . = ALIGN(HPAGE_SIZE); + __brk_limit = .; + } + +@@ -316,13 +384,12 @@ SECTIONS + * for the boot processor. + */ + #define INIT_PER_CPU(x) init_per_cpu__##x = per_cpu__##x + __per_cpu_load +-INIT_PER_CPU(gdt_page); + INIT_PER_CPU(irq_stack_union); + + /* + * Build-time check on the image size: + */ +-. = ASSERT((_end - _text <= KERNEL_IMAGE_SIZE), ++. = ASSERT((_end - _text - __KERNEL_TEXT_OFFSET <= KERNEL_IMAGE_SIZE), + "kernel image bigger than KERNEL_IMAGE_SIZE"); + + #ifdef CONFIG_SMP +diff --git a/arch/x86/kernel/vsyscall_64.c b/arch/x86/kernel/vsyscall_64.c +index 62f39d7..3bc46a1 100644 +--- a/arch/x86/kernel/vsyscall_64.c ++++ b/arch/x86/kernel/vsyscall_64.c +@@ -80,6 +80,7 @@ void update_vsyscall(struct timespec *wall_time, struct clocksource *clock, + + write_seqlock_irqsave(&vsyscall_gtod_data.lock, flags); + /* copy vsyscall data */ ++ strlcpy(vsyscall_gtod_data.clock.name, clock->name, sizeof vsyscall_gtod_data.clock.name); + vsyscall_gtod_data.clock.vread = clock->vread; + vsyscall_gtod_data.clock.cycle_last = clock->cycle_last; + vsyscall_gtod_data.clock.mask = clock->mask; +@@ -203,7 +204,7 @@ vgetcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache) + We do this here because otherwise user space would do it on + its own in a likely inferior way (no access to jiffies). + If you don't like it pass NULL. */ +- if (tcache && tcache->blob[0] == (j = __jiffies)) { ++ if (tcache && tcache->blob[0] == (j = jiffies)) { + p = tcache->blob[1]; + } else if (__vgetcpu_mode == VGETCPU_RDTSCP) { + /* Load per CPU data from RDTSCP */ +diff --git a/arch/x86/kernel/x8664_ksyms_64.c b/arch/x86/kernel/x8664_ksyms_64.c +index 3909e3b..5433a97 100644 +--- a/arch/x86/kernel/x8664_ksyms_64.c ++++ b/arch/x86/kernel/x8664_ksyms_64.c +@@ -30,8 +30,6 @@ EXPORT_SYMBOL(__put_user_8); + + EXPORT_SYMBOL(copy_user_generic); + EXPORT_SYMBOL(__copy_user_nocache); +-EXPORT_SYMBOL(copy_from_user); +-EXPORT_SYMBOL(copy_to_user); + EXPORT_SYMBOL(__copy_from_user_inatomic); + + EXPORT_SYMBOL(copy_page); +diff --git a/arch/x86/kernel/xsave.c b/arch/x86/kernel/xsave.c +index c5ee17e..d63218f 100644 +--- a/arch/x86/kernel/xsave.c ++++ b/arch/x86/kernel/xsave.c +@@ -54,7 +54,7 @@ int check_for_xstate(struct i387_fxsave_struct __user *buf, + fx_sw_user->xstate_size > fx_sw_user->extended_size) + return -1; + +- err = __get_user(magic2, (__u32 *) (((void *)fpstate) + ++ err = __get_user(magic2, (__u32 __user *) (((void __user *)fpstate) + + fx_sw_user->extended_size - + FP_XSTATE_MAGIC2_SIZE)); + /* +@@ -196,7 +196,7 @@ fx_only: + * the other extended state. + */ + xrstor_state(init_xstate_buf, pcntxt_mask & ~XSTATE_FPSSE); +- return fxrstor_checking((__force struct i387_fxsave_struct *)buf); ++ return fxrstor_checking((struct i387_fxsave_struct __force_kernel *)buf); + } + + /* +@@ -228,7 +228,7 @@ int restore_i387_xstate(void __user *buf) + if (task_thread_info(tsk)->status & TS_XSAVE) + err = restore_user_xstate(buf); + else +- err = fxrstor_checking((__force struct i387_fxsave_struct *) ++ err = fxrstor_checking((struct i387_fxsave_struct __user *) + buf); + if (unlikely(err)) { + /* +diff --git a/arch/x86/kvm/emulate.c b/arch/x86/kvm/emulate.c +index 1350e43..a94b011 100644 +--- a/arch/x86/kvm/emulate.c ++++ b/arch/x86/kvm/emulate.c +@@ -81,8 +81,8 @@ + #define Src2CL (1<<29) + #define Src2ImmByte (2<<29) + #define Src2One (3<<29) +-#define Src2Imm16 (4<<29) +-#define Src2Mask (7<<29) ++#define Src2Imm16 (4U<<29) ++#define Src2Mask (7U<<29) + + enum { + Group1_80, Group1_81, Group1_82, Group1_83, +@@ -411,6 +411,7 @@ static u32 group2_table[] = { + + #define ____emulate_2op(_op, _src, _dst, _eflags, _x, _y, _suffix) \ + do { \ ++ unsigned long _tmp; \ + __asm__ __volatile__ ( \ + _PRE_EFLAGS("0", "4", "2") \ + _op _suffix " %"_x"3,%1; " \ +@@ -424,8 +425,6 @@ static u32 group2_table[] = { + /* Raw emulation: instruction has two explicit operands. */ + #define __emulate_2op_nobyte(_op,_src,_dst,_eflags,_wx,_wy,_lx,_ly,_qx,_qy) \ + do { \ +- unsigned long _tmp; \ +- \ + switch ((_dst).bytes) { \ + case 2: \ + ____emulate_2op(_op,_src,_dst,_eflags,_wx,_wy,"w"); \ +@@ -441,7 +440,6 @@ static u32 group2_table[] = { + + #define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \ + do { \ +- unsigned long _tmp; \ + switch ((_dst).bytes) { \ + case 1: \ + ____emulate_2op(_op,_src,_dst,_eflags,_bx,_by,"b"); \ +diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c +index 8dfeaaa..4daa395 100644 +--- a/arch/x86/kvm/lapic.c ++++ b/arch/x86/kvm/lapic.c +@@ -52,7 +52,7 @@ + #define APIC_BUS_CYCLE_NS 1 + + /* #define apic_debug(fmt,arg...) printk(KERN_WARNING fmt,##arg) */ +-#define apic_debug(fmt, arg...) ++#define apic_debug(fmt, arg...) do {} while (0) + + #define APIC_LVT_NUM 6 + /* 14 is the version for Xeon and Pentium 8.4.8*/ +diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h +index 3bc2707..dd157e2 100644 +--- a/arch/x86/kvm/paging_tmpl.h ++++ b/arch/x86/kvm/paging_tmpl.h +@@ -416,6 +416,8 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, + int level = PT_PAGE_TABLE_LEVEL; + unsigned long mmu_seq; + ++ pax_track_stack(); ++ + pgprintk("%s: addr %lx err %x\n", __func__, addr, error_code); + kvm_mmu_audit(vcpu, "pre page fault"); + +@@ -461,6 +463,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, + kvm_mmu_free_some_pages(vcpu); + sptep = FNAME(fetch)(vcpu, addr, &walker, user_fault, write_fault, + level, &write_pt, pfn); ++ (void)sptep; + pgprintk("%s: shadow pte %p %llx ptwrite %d\n", __func__, + sptep, *sptep, write_pt); + +diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c +index 7c6e63e..c5d92c1 100644 +--- a/arch/x86/kvm/svm.c ++++ b/arch/x86/kvm/svm.c +@@ -2486,7 +2486,11 @@ static void reload_tss(struct kvm_vcpu *vcpu) + int cpu = raw_smp_processor_id(); + + struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu); ++ ++ pax_open_kernel(); + svm_data->tss_desc->type = 9; /* available 32/64-bit TSS */ ++ pax_close_kernel(); ++ + load_TR_desc(); + } + +@@ -2947,7 +2951,7 @@ static bool svm_gb_page_enable(void) + return true; + } + +-static struct kvm_x86_ops svm_x86_ops = { ++static const struct kvm_x86_ops svm_x86_ops = { + .cpu_has_kvm_support = has_svm, + .disabled_by_bios = is_disabled, + .hardware_setup = svm_hardware_setup, +diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c +index e6d925f..e7a4af8 100644 +--- a/arch/x86/kvm/vmx.c ++++ b/arch/x86/kvm/vmx.c +@@ -570,7 +570,11 @@ static void reload_tss(void) + + kvm_get_gdt(&gdt); + descs = (void *)gdt.base; ++ ++ pax_open_kernel(); + descs[GDT_ENTRY_TSS].type = 9; /* available TSS */ ++ pax_close_kernel(); ++ + load_TR_desc(); + } + +@@ -1410,8 +1414,11 @@ static __init int hardware_setup(void) + if (!cpu_has_vmx_flexpriority()) + flexpriority_enabled = 0; + +- if (!cpu_has_vmx_tpr_shadow()) +- kvm_x86_ops->update_cr8_intercept = NULL; ++ if (!cpu_has_vmx_tpr_shadow()) { ++ pax_open_kernel(); ++ *(void **)&kvm_x86_ops->update_cr8_intercept = NULL; ++ pax_close_kernel(); ++ } + + if (enable_ept && !cpu_has_vmx_ept_2m_page()) + kvm_disable_largepages(); +@@ -2362,7 +2369,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) + vmcs_writel(HOST_IDTR_BASE, dt.base); /* 22.2.4 */ + + asm("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return)); +- vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */ ++ vmcs_writel(HOST_RIP, ktla_ktva(kvm_vmx_return)); /* 22.2.5 */ + vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0); + vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0); + vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, 0); +@@ -3718,6 +3725,12 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) + "jmp .Lkvm_vmx_return \n\t" + ".Llaunched: " __ex(ASM_VMX_VMRESUME) "\n\t" + ".Lkvm_vmx_return: " ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ "ljmp %[cs],$.Lkvm_vmx_return2\n\t" ++ ".Lkvm_vmx_return2: " ++#endif ++ + /* Save guest registers, load host registers, keep flags */ + "xchg %0, (%%"R"sp) \n\t" + "mov %%"R"ax, %c[rax](%0) \n\t" +@@ -3764,8 +3777,13 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) + [r15]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R15])), + #endif + [cr2]"i"(offsetof(struct vcpu_vmx, vcpu.arch.cr2)) ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ ,[cs]"i"(__KERNEL_CS) ++#endif ++ + : "cc", "memory" +- , R"bx", R"di", R"si" ++ , R"ax", R"bx", R"di", R"si" + #ifdef CONFIG_X86_64 + , "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15" + #endif +@@ -3782,7 +3800,16 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) + if (vmx->rmode.irq.pending) + fixup_rmode_irq(vmx); + +- asm("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); ++ asm("mov %0, %%ds; mov %0, %%es; mov %0, %%ss" : : "r"(__KERNEL_DS)); ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ loadsegment(fs, __KERNEL_PERCPU); ++#endif ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ __set_fs(current_thread_info()->addr_limit); ++#endif ++ + vmx->launched = 1; + + vmx_complete_interrupts(vmx); +@@ -3957,7 +3984,7 @@ static bool vmx_gb_page_enable(void) + return false; + } + +-static struct kvm_x86_ops vmx_x86_ops = { ++static const struct kvm_x86_ops vmx_x86_ops = { + .cpu_has_kvm_support = cpu_has_kvm_support, + .disabled_by_bios = vmx_disabled_by_bios, + .hardware_setup = hardware_setup, +diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c +index df1cefb..5e882ad 100644 +--- a/arch/x86/kvm/x86.c ++++ b/arch/x86/kvm/x86.c +@@ -82,7 +82,7 @@ static void update_cr8_intercept(struct kvm_vcpu *vcpu); + static int kvm_dev_ioctl_get_supported_cpuid(struct kvm_cpuid2 *cpuid, + struct kvm_cpuid_entry2 __user *entries); + +-struct kvm_x86_ops *kvm_x86_ops; ++const struct kvm_x86_ops *kvm_x86_ops; + EXPORT_SYMBOL_GPL(kvm_x86_ops); + + int ignore_msrs = 0; +@@ -1430,15 +1430,20 @@ static int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu, + struct kvm_cpuid2 *cpuid, + struct kvm_cpuid_entry2 __user *entries) + { +- int r; ++ int r, i; + + r = -E2BIG; + if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) + goto out; + r = -EFAULT; +- if (copy_from_user(&vcpu->arch.cpuid_entries, entries, +- cpuid->nent * sizeof(struct kvm_cpuid_entry2))) ++ if (!access_ok(VERIFY_READ, entries, cpuid->nent * sizeof(struct kvm_cpuid_entry2))) + goto out; ++ for (i = 0; i < cpuid->nent; ++i) { ++ struct kvm_cpuid_entry2 cpuid_entry; ++ if (__copy_from_user(&cpuid_entry, entries + i, sizeof(cpuid_entry))) ++ goto out; ++ vcpu->arch.cpuid_entries[i] = cpuid_entry; ++ } + vcpu->arch.cpuid_nent = cpuid->nent; + kvm_apic_set_version(vcpu); + return 0; +@@ -1451,16 +1456,20 @@ static int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu, + struct kvm_cpuid2 *cpuid, + struct kvm_cpuid_entry2 __user *entries) + { +- int r; ++ int r, i; + + vcpu_load(vcpu); + r = -E2BIG; + if (cpuid->nent < vcpu->arch.cpuid_nent) + goto out; + r = -EFAULT; +- if (copy_to_user(entries, &vcpu->arch.cpuid_entries, +- vcpu->arch.cpuid_nent * sizeof(struct kvm_cpuid_entry2))) ++ if (!access_ok(VERIFY_WRITE, entries, vcpu->arch.cpuid_nent * sizeof(struct kvm_cpuid_entry2))) + goto out; ++ for (i = 0; i < vcpu->arch.cpuid_nent; ++i) { ++ struct kvm_cpuid_entry2 cpuid_entry = vcpu->arch.cpuid_entries[i]; ++ if (__copy_to_user(entries + i, &cpuid_entry, sizeof(cpuid_entry))) ++ goto out; ++ } + return 0; + + out: +@@ -1678,7 +1687,7 @@ static int kvm_vcpu_ioctl_set_lapic(struct kvm_vcpu *vcpu, + static int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, + struct kvm_interrupt *irq) + { +- if (irq->irq < 0 || irq->irq >= 256) ++ if (irq->irq >= 256) + return -EINVAL; + if (irqchip_in_kernel(vcpu->kvm)) + return -ENXIO; +@@ -3260,10 +3269,10 @@ static struct notifier_block kvmclock_cpufreq_notifier_block = { + .notifier_call = kvmclock_cpufreq_notifier + }; + +-int kvm_arch_init(void *opaque) ++int kvm_arch_init(const void *opaque) + { + int r, cpu; +- struct kvm_x86_ops *ops = (struct kvm_x86_ops *)opaque; ++ const struct kvm_x86_ops *ops = (const struct kvm_x86_ops *)opaque; + + if (kvm_x86_ops) { + printk(KERN_ERR "kvm: already loaded the other module\n"); +diff --git a/arch/x86/lguest/boot.c b/arch/x86/lguest/boot.c +index 7e59dc1..b88c98f 100644 +--- a/arch/x86/lguest/boot.c ++++ b/arch/x86/lguest/boot.c +@@ -1172,9 +1172,10 @@ static __init int early_put_chars(u32 vtermno, const char *buf, int count) + * Rebooting also tells the Host we're finished, but the RESTART flag tells the + * Launcher to reboot us. + */ +-static void lguest_restart(char *reason) ++static __noreturn void lguest_restart(char *reason) + { + kvm_hypercall2(LHCALL_SHUTDOWN, __pa(reason), LGUEST_SHUTDOWN_RESTART); ++ BUG(); + } + + /*G:050 +diff --git a/arch/x86/lib/atomic64_32.c b/arch/x86/lib/atomic64_32.c +index 824fa0b..c619e96 100644 +--- a/arch/x86/lib/atomic64_32.c ++++ b/arch/x86/lib/atomic64_32.c +@@ -25,6 +25,12 @@ u64 atomic64_cmpxchg(atomic64_t *ptr, u64 old_val, u64 new_val) + } + EXPORT_SYMBOL(atomic64_cmpxchg); + ++u64 atomic64_cmpxchg_unchecked(atomic64_unchecked_t *ptr, u64 old_val, u64 new_val) ++{ ++ return cmpxchg8b(&ptr->counter, old_val, new_val); ++} ++EXPORT_SYMBOL(atomic64_cmpxchg_unchecked); ++ + /** + * atomic64_xchg - xchg atomic64 variable + * @ptr: pointer to type atomic64_t +@@ -56,6 +62,36 @@ u64 atomic64_xchg(atomic64_t *ptr, u64 new_val) + EXPORT_SYMBOL(atomic64_xchg); + + /** ++ * atomic64_xchg_unchecked - xchg atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * @new_val: value to assign ++ * ++ * Atomically xchgs the value of @ptr to @new_val and returns ++ * the old value. ++ */ ++u64 atomic64_xchg_unchecked(atomic64_unchecked_t *ptr, u64 new_val) ++{ ++ /* ++ * Try first with a (possibly incorrect) assumption about ++ * what we have there. We'll do two loops most likely, ++ * but we'll get an ownership MESI transaction straight away ++ * instead of a read transaction followed by a ++ * flush-for-ownership transaction: ++ */ ++ u64 old_val, real_val = 0; ++ ++ do { ++ old_val = real_val; ++ ++ real_val = atomic64_cmpxchg_unchecked(ptr, old_val, new_val); ++ ++ } while (real_val != old_val); ++ ++ return old_val; ++} ++EXPORT_SYMBOL(atomic64_xchg_unchecked); ++ ++/** + * atomic64_set - set atomic64 variable + * @ptr: pointer to type atomic64_t + * @new_val: value to assign +@@ -69,7 +105,19 @@ void atomic64_set(atomic64_t *ptr, u64 new_val) + EXPORT_SYMBOL(atomic64_set); + + /** +-EXPORT_SYMBOL(atomic64_read); ++ * atomic64_unchecked_set - set atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * @new_val: value to assign ++ * ++ * Atomically sets the value of @ptr to @new_val. ++ */ ++void atomic64_set_unchecked(atomic64_unchecked_t *ptr, u64 new_val) ++{ ++ atomic64_xchg_unchecked(ptr, new_val); ++} ++EXPORT_SYMBOL(atomic64_set_unchecked); ++ ++/** + * atomic64_add_return - add and return + * @delta: integer value to add + * @ptr: pointer to type atomic64_t +@@ -99,24 +147,72 @@ noinline u64 atomic64_add_return(u64 delta, atomic64_t *ptr) + } + EXPORT_SYMBOL(atomic64_add_return); + ++/** ++ * atomic64_add_return_unchecked - add and return ++ * @delta: integer value to add ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically adds @delta to @ptr and returns @delta + *@ptr ++ */ ++noinline u64 atomic64_add_return_unchecked(u64 delta, atomic64_unchecked_t *ptr) ++{ ++ /* ++ * Try first with a (possibly incorrect) assumption about ++ * what we have there. We'll do two loops most likely, ++ * but we'll get an ownership MESI transaction straight away ++ * instead of a read transaction followed by a ++ * flush-for-ownership transaction: ++ */ ++ u64 old_val, new_val, real_val = 0; ++ ++ do { ++ old_val = real_val; ++ new_val = old_val + delta; ++ ++ real_val = atomic64_cmpxchg_unchecked(ptr, old_val, new_val); ++ ++ } while (real_val != old_val); ++ ++ return new_val; ++} ++EXPORT_SYMBOL(atomic64_add_return_unchecked); ++ + u64 atomic64_sub_return(u64 delta, atomic64_t *ptr) + { + return atomic64_add_return(-delta, ptr); + } + EXPORT_SYMBOL(atomic64_sub_return); + ++u64 atomic64_sub_return_unchecked(u64 delta, atomic64_unchecked_t *ptr) ++{ ++ return atomic64_add_return_unchecked(-delta, ptr); ++} ++EXPORT_SYMBOL(atomic64_sub_return_unchecked); ++ + u64 atomic64_inc_return(atomic64_t *ptr) + { + return atomic64_add_return(1, ptr); + } + EXPORT_SYMBOL(atomic64_inc_return); + ++u64 atomic64_inc_return_unchecked(atomic64_unchecked_t *ptr) ++{ ++ return atomic64_add_return_unchecked(1, ptr); ++} ++EXPORT_SYMBOL(atomic64_inc_return_unchecked); ++ + u64 atomic64_dec_return(atomic64_t *ptr) + { + return atomic64_sub_return(1, ptr); + } + EXPORT_SYMBOL(atomic64_dec_return); + ++u64 atomic64_dec_return_unchecked(atomic64_unchecked_t *ptr) ++{ ++ return atomic64_sub_return_unchecked(1, ptr); ++} ++EXPORT_SYMBOL(atomic64_dec_return_unchecked); ++ + /** + * atomic64_add - add integer to atomic64 variable + * @delta: integer value to add +@@ -131,6 +227,19 @@ void atomic64_add(u64 delta, atomic64_t *ptr) + EXPORT_SYMBOL(atomic64_add); + + /** ++ * atomic64_add_unchecked - add integer to atomic64 variable ++ * @delta: integer value to add ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically adds @delta to @ptr. ++ */ ++void atomic64_add_unchecked(u64 delta, atomic64_unchecked_t *ptr) ++{ ++ atomic64_add_return_unchecked(delta, ptr); ++} ++EXPORT_SYMBOL(atomic64_add_unchecked); ++ ++/** + * atomic64_sub - subtract the atomic64 variable + * @delta: integer value to subtract + * @ptr: pointer to type atomic64_t +@@ -144,6 +253,19 @@ void atomic64_sub(u64 delta, atomic64_t *ptr) + EXPORT_SYMBOL(atomic64_sub); + + /** ++ * atomic64_sub_unchecked - subtract the atomic64 variable ++ * @delta: integer value to subtract ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically subtracts @delta from @ptr. ++ */ ++void atomic64_sub_unchecked(u64 delta, atomic64_unchecked_t *ptr) ++{ ++ atomic64_add_unchecked(-delta, ptr); ++} ++EXPORT_SYMBOL(atomic64_sub_unchecked); ++ ++/** + * atomic64_sub_and_test - subtract value from variable and test result + * @delta: integer value to subtract + * @ptr: pointer to type atomic64_t +@@ -173,6 +295,18 @@ void atomic64_inc(atomic64_t *ptr) + EXPORT_SYMBOL(atomic64_inc); + + /** ++ * atomic64_inc_unchecked - increment atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically increments @ptr by 1. ++ */ ++void atomic64_inc_unchecked(atomic64_unchecked_t *ptr) ++{ ++ atomic64_add_unchecked(1, ptr); ++} ++EXPORT_SYMBOL(atomic64_inc_unchecked); ++ ++/** + * atomic64_dec - decrement atomic64 variable + * @ptr: pointer to type atomic64_t + * +@@ -185,6 +319,18 @@ void atomic64_dec(atomic64_t *ptr) + EXPORT_SYMBOL(atomic64_dec); + + /** ++ * atomic64_dec_unchecked - decrement atomic64 variable ++ * @ptr: pointer to type atomic64_unchecked_t ++ * ++ * Atomically decrements @ptr by 1. ++ */ ++void atomic64_dec_unchecked(atomic64_unchecked_t *ptr) ++{ ++ atomic64_sub_unchecked(1, ptr); ++} ++EXPORT_SYMBOL(atomic64_dec_unchecked); ++ ++/** + * atomic64_dec_and_test - decrement and test + * @ptr: pointer to type atomic64_t + * +diff --git a/arch/x86/lib/checksum_32.S b/arch/x86/lib/checksum_32.S +index adbccd0..98f96c8 100644 +--- a/arch/x86/lib/checksum_32.S ++++ b/arch/x86/lib/checksum_32.S +@@ -28,7 +28,8 @@ + #include + #include + #include +- ++#include ++ + /* + * computes a partial checksum, e.g. for TCP/UDP fragments + */ +@@ -304,9 +305,28 @@ unsigned int csum_partial_copy_generic (const char *src, char *dst, + + #define ARGBASE 16 + #define FP 12 +- +-ENTRY(csum_partial_copy_generic) ++ ++ENTRY(csum_partial_copy_generic_to_user) + CFI_STARTPROC ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pushl %gs ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %es ++ CFI_ADJUST_CFA_OFFSET -4 ++ jmp csum_partial_copy_generic ++#endif ++ ++ENTRY(csum_partial_copy_generic_from_user) ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pushl %gs ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %ds ++ CFI_ADJUST_CFA_OFFSET -4 ++#endif ++ ++ENTRY(csum_partial_copy_generic) + subl $4,%esp + CFI_ADJUST_CFA_OFFSET 4 + pushl %edi +@@ -331,7 +351,7 @@ ENTRY(csum_partial_copy_generic) + jmp 4f + SRC(1: movw (%esi), %bx ) + addl $2, %esi +-DST( movw %bx, (%edi) ) ++DST( movw %bx, %es:(%edi) ) + addl $2, %edi + addw %bx, %ax + adcl $0, %eax +@@ -343,30 +363,30 @@ DST( movw %bx, (%edi) ) + SRC(1: movl (%esi), %ebx ) + SRC( movl 4(%esi), %edx ) + adcl %ebx, %eax +-DST( movl %ebx, (%edi) ) ++DST( movl %ebx, %es:(%edi) ) + adcl %edx, %eax +-DST( movl %edx, 4(%edi) ) ++DST( movl %edx, %es:4(%edi) ) + + SRC( movl 8(%esi), %ebx ) + SRC( movl 12(%esi), %edx ) + adcl %ebx, %eax +-DST( movl %ebx, 8(%edi) ) ++DST( movl %ebx, %es:8(%edi) ) + adcl %edx, %eax +-DST( movl %edx, 12(%edi) ) ++DST( movl %edx, %es:12(%edi) ) + + SRC( movl 16(%esi), %ebx ) + SRC( movl 20(%esi), %edx ) + adcl %ebx, %eax +-DST( movl %ebx, 16(%edi) ) ++DST( movl %ebx, %es:16(%edi) ) + adcl %edx, %eax +-DST( movl %edx, 20(%edi) ) ++DST( movl %edx, %es:20(%edi) ) + + SRC( movl 24(%esi), %ebx ) + SRC( movl 28(%esi), %edx ) + adcl %ebx, %eax +-DST( movl %ebx, 24(%edi) ) ++DST( movl %ebx, %es:24(%edi) ) + adcl %edx, %eax +-DST( movl %edx, 28(%edi) ) ++DST( movl %edx, %es:28(%edi) ) + + lea 32(%esi), %esi + lea 32(%edi), %edi +@@ -380,7 +400,7 @@ DST( movl %edx, 28(%edi) ) + shrl $2, %edx # This clears CF + SRC(3: movl (%esi), %ebx ) + adcl %ebx, %eax +-DST( movl %ebx, (%edi) ) ++DST( movl %ebx, %es:(%edi) ) + lea 4(%esi), %esi + lea 4(%edi), %edi + dec %edx +@@ -392,12 +412,12 @@ DST( movl %ebx, (%edi) ) + jb 5f + SRC( movw (%esi), %cx ) + leal 2(%esi), %esi +-DST( movw %cx, (%edi) ) ++DST( movw %cx, %es:(%edi) ) + leal 2(%edi), %edi + je 6f + shll $16,%ecx + SRC(5: movb (%esi), %cl ) +-DST( movb %cl, (%edi) ) ++DST( movb %cl, %es:(%edi) ) + 6: addl %ecx, %eax + adcl $0, %eax + 7: +@@ -408,7 +428,7 @@ DST( movb %cl, (%edi) ) + + 6001: + movl ARGBASE+20(%esp), %ebx # src_err_ptr +- movl $-EFAULT, (%ebx) ++ movl $-EFAULT, %ss:(%ebx) + + # zero the complete destination - computing the rest + # is too much work +@@ -421,11 +441,19 @@ DST( movb %cl, (%edi) ) + + 6002: + movl ARGBASE+24(%esp), %ebx # dst_err_ptr +- movl $-EFAULT,(%ebx) ++ movl $-EFAULT,%ss:(%ebx) + jmp 5000b + + .previous + ++ pushl %ss ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %ds ++ CFI_ADJUST_CFA_OFFSET -4 ++ pushl %ss ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %es ++ CFI_ADJUST_CFA_OFFSET -4 + popl %ebx + CFI_ADJUST_CFA_OFFSET -4 + CFI_RESTORE ebx +@@ -439,26 +467,47 @@ DST( movb %cl, (%edi) ) + CFI_ADJUST_CFA_OFFSET -4 + ret + CFI_ENDPROC +-ENDPROC(csum_partial_copy_generic) ++ENDPROC(csum_partial_copy_generic_to_user) + + #else + + /* Version for PentiumII/PPro */ + + #define ROUND1(x) \ ++ nop; nop; nop; \ + SRC(movl x(%esi), %ebx ) ; \ + addl %ebx, %eax ; \ +- DST(movl %ebx, x(%edi) ) ; ++ DST(movl %ebx, %es:x(%edi)) ; + + #define ROUND(x) \ ++ nop; nop; nop; \ + SRC(movl x(%esi), %ebx ) ; \ + adcl %ebx, %eax ; \ +- DST(movl %ebx, x(%edi) ) ; ++ DST(movl %ebx, %es:x(%edi)) ; + + #define ARGBASE 12 +- +-ENTRY(csum_partial_copy_generic) ++ ++ENTRY(csum_partial_copy_generic_to_user) + CFI_STARTPROC ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pushl %gs ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %es ++ CFI_ADJUST_CFA_OFFSET -4 ++ jmp csum_partial_copy_generic ++#endif ++ ++ENTRY(csum_partial_copy_generic_from_user) ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pushl %gs ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %ds ++ CFI_ADJUST_CFA_OFFSET -4 ++#endif ++ ++ENTRY(csum_partial_copy_generic) + pushl %ebx + CFI_ADJUST_CFA_OFFSET 4 + CFI_REL_OFFSET ebx, 0 +@@ -482,7 +531,7 @@ ENTRY(csum_partial_copy_generic) + subl %ebx, %edi + lea -1(%esi),%edx + andl $-32,%edx +- lea 3f(%ebx,%ebx), %ebx ++ lea 3f(%ebx,%ebx,2), %ebx + testl %esi, %esi + jmp *%ebx + 1: addl $64,%esi +@@ -503,19 +552,19 @@ ENTRY(csum_partial_copy_generic) + jb 5f + SRC( movw (%esi), %dx ) + leal 2(%esi), %esi +-DST( movw %dx, (%edi) ) ++DST( movw %dx, %es:(%edi) ) + leal 2(%edi), %edi + je 6f + shll $16,%edx + 5: + SRC( movb (%esi), %dl ) +-DST( movb %dl, (%edi) ) ++DST( movb %dl, %es:(%edi) ) + 6: addl %edx, %eax + adcl $0, %eax + 7: + .section .fixup, "ax" + 6001: movl ARGBASE+20(%esp), %ebx # src_err_ptr +- movl $-EFAULT, (%ebx) ++ movl $-EFAULT, %ss:(%ebx) + # zero the complete destination (computing the rest is too much work) + movl ARGBASE+8(%esp),%edi # dst + movl ARGBASE+12(%esp),%ecx # len +@@ -523,10 +572,21 @@ DST( movb %dl, (%edi) ) + rep; stosb + jmp 7b + 6002: movl ARGBASE+24(%esp), %ebx # dst_err_ptr +- movl $-EFAULT, (%ebx) ++ movl $-EFAULT, %ss:(%ebx) + jmp 7b + .previous + ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ pushl %ss ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %ds ++ CFI_ADJUST_CFA_OFFSET -4 ++ pushl %ss ++ CFI_ADJUST_CFA_OFFSET 4 ++ popl %es ++ CFI_ADJUST_CFA_OFFSET -4 ++#endif ++ + popl %esi + CFI_ADJUST_CFA_OFFSET -4 + CFI_RESTORE esi +@@ -538,7 +598,7 @@ DST( movb %dl, (%edi) ) + CFI_RESTORE ebx + ret + CFI_ENDPROC +-ENDPROC(csum_partial_copy_generic) ++ENDPROC(csum_partial_copy_generic_to_user) + + #undef ROUND + #undef ROUND1 +diff --git a/arch/x86/lib/clear_page_64.S b/arch/x86/lib/clear_page_64.S +index ebeafcc..1e3a402 100644 +--- a/arch/x86/lib/clear_page_64.S ++++ b/arch/x86/lib/clear_page_64.S +@@ -1,5 +1,6 @@ + #include + #include ++#include + + /* + * Zero a page. +@@ -10,6 +11,7 @@ ENTRY(clear_page_c) + movl $4096/8,%ecx + xorl %eax,%eax + rep stosq ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(clear_page_c) +@@ -33,6 +35,7 @@ ENTRY(clear_page) + leaq 64(%rdi),%rdi + jnz .Lloop + nop ++ pax_force_retaddr + ret + CFI_ENDPROC + .Lclear_page_end: +@@ -43,7 +46,7 @@ ENDPROC(clear_page) + + #include + +- .section .altinstr_replacement,"ax" ++ .section .altinstr_replacement,"a" + 1: .byte 0xeb /* jmp */ + .byte (clear_page_c - clear_page) - (2f - 1b) /* offset */ + 2: +diff --git a/arch/x86/lib/copy_page_64.S b/arch/x86/lib/copy_page_64.S +index 727a5d4..333818a 100644 +--- a/arch/x86/lib/copy_page_64.S ++++ b/arch/x86/lib/copy_page_64.S +@@ -2,12 +2,14 @@ + + #include + #include ++#include + + ALIGN + copy_page_c: + CFI_STARTPROC + movl $4096/8,%ecx + rep movsq ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(copy_page_c) +@@ -38,7 +40,7 @@ ENTRY(copy_page) + movq 16 (%rsi), %rdx + movq 24 (%rsi), %r8 + movq 32 (%rsi), %r9 +- movq 40 (%rsi), %r10 ++ movq 40 (%rsi), %r13 + movq 48 (%rsi), %r11 + movq 56 (%rsi), %r12 + +@@ -49,7 +51,7 @@ ENTRY(copy_page) + movq %rdx, 16 (%rdi) + movq %r8, 24 (%rdi) + movq %r9, 32 (%rdi) +- movq %r10, 40 (%rdi) ++ movq %r13, 40 (%rdi) + movq %r11, 48 (%rdi) + movq %r12, 56 (%rdi) + +@@ -68,7 +70,7 @@ ENTRY(copy_page) + movq 16 (%rsi), %rdx + movq 24 (%rsi), %r8 + movq 32 (%rsi), %r9 +- movq 40 (%rsi), %r10 ++ movq 40 (%rsi), %r13 + movq 48 (%rsi), %r11 + movq 56 (%rsi), %r12 + +@@ -77,7 +79,7 @@ ENTRY(copy_page) + movq %rdx, 16 (%rdi) + movq %r8, 24 (%rdi) + movq %r9, 32 (%rdi) +- movq %r10, 40 (%rdi) ++ movq %r13, 40 (%rdi) + movq %r11, 48 (%rdi) + movq %r12, 56 (%rdi) + +@@ -94,6 +96,7 @@ ENTRY(copy_page) + CFI_RESTORE r13 + addq $3*8,%rsp + CFI_ADJUST_CFA_OFFSET -3*8 ++ pax_force_retaddr + ret + .Lcopy_page_end: + CFI_ENDPROC +@@ -104,7 +107,7 @@ ENDPROC(copy_page) + + #include + +- .section .altinstr_replacement,"ax" ++ .section .altinstr_replacement,"a" + 1: .byte 0xeb /* jmp */ + .byte (copy_page_c - copy_page) - (2f - 1b) /* offset */ + 2: +diff --git a/arch/x86/lib/copy_user_64.S b/arch/x86/lib/copy_user_64.S +index af8debd..40c75f3 100644 +--- a/arch/x86/lib/copy_user_64.S ++++ b/arch/x86/lib/copy_user_64.S +@@ -15,13 +15,15 @@ + #include + #include + #include ++#include ++#include + + .macro ALTERNATIVE_JUMP feature,orig,alt + 0: + .byte 0xe9 /* 32bit jump */ + .long \orig-1f /* by default jump to orig */ + 1: +- .section .altinstr_replacement,"ax" ++ .section .altinstr_replacement,"a" + 2: .byte 0xe9 /* near jump with 32bit immediate */ + .long \alt-1b /* offset */ /* or alternatively to alt */ + .previous +@@ -64,55 +66,26 @@ + #endif + .endm + +-/* Standard copy_to_user with segment limit checking */ +-ENTRY(copy_to_user) +- CFI_STARTPROC +- GET_THREAD_INFO(%rax) +- movq %rdi,%rcx +- addq %rdx,%rcx +- jc bad_to_user +- cmpq TI_addr_limit(%rax),%rcx +- ja bad_to_user +- ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string +- CFI_ENDPROC +-ENDPROC(copy_to_user) +- +-/* Standard copy_from_user with segment limit checking */ +-ENTRY(copy_from_user) +- CFI_STARTPROC +- GET_THREAD_INFO(%rax) +- movq %rsi,%rcx +- addq %rdx,%rcx +- jc bad_from_user +- cmpq TI_addr_limit(%rax),%rcx +- ja bad_from_user +- ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string +- CFI_ENDPROC +-ENDPROC(copy_from_user) +- + ENTRY(copy_user_generic) + CFI_STARTPROC + ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string + CFI_ENDPROC + ENDPROC(copy_user_generic) + +-ENTRY(__copy_from_user_inatomic) +- CFI_STARTPROC +- ALTERNATIVE_JUMP X86_FEATURE_REP_GOOD,copy_user_generic_unrolled,copy_user_generic_string +- CFI_ENDPROC +-ENDPROC(__copy_from_user_inatomic) +- + .section .fixup,"ax" + /* must zero dest */ + ENTRY(bad_from_user) + bad_from_user: + CFI_STARTPROC ++ testl %edx,%edx ++ js bad_to_user + movl %edx,%ecx + xorl %eax,%eax + rep + stosb + bad_to_user: + movl %edx,%eax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(bad_from_user) +@@ -142,19 +115,19 @@ ENTRY(copy_user_generic_unrolled) + jz 17f + 1: movq (%rsi),%r8 + 2: movq 1*8(%rsi),%r9 +-3: movq 2*8(%rsi),%r10 ++3: movq 2*8(%rsi),%rax + 4: movq 3*8(%rsi),%r11 + 5: movq %r8,(%rdi) + 6: movq %r9,1*8(%rdi) +-7: movq %r10,2*8(%rdi) ++7: movq %rax,2*8(%rdi) + 8: movq %r11,3*8(%rdi) + 9: movq 4*8(%rsi),%r8 + 10: movq 5*8(%rsi),%r9 +-11: movq 6*8(%rsi),%r10 ++11: movq 6*8(%rsi),%rax + 12: movq 7*8(%rsi),%r11 + 13: movq %r8,4*8(%rdi) + 14: movq %r9,5*8(%rdi) +-15: movq %r10,6*8(%rdi) ++15: movq %rax,6*8(%rdi) + 16: movq %r11,7*8(%rdi) + leaq 64(%rsi),%rsi + leaq 64(%rdi),%rdi +@@ -180,6 +153,7 @@ ENTRY(copy_user_generic_unrolled) + decl %ecx + jnz 21b + 23: xor %eax,%eax ++ pax_force_retaddr + ret + + .section .fixup,"ax" +@@ -252,6 +226,7 @@ ENTRY(copy_user_generic_string) + 3: rep + movsb + 4: xorl %eax,%eax ++ pax_force_retaddr + ret + + .section .fixup,"ax" +diff --git a/arch/x86/lib/copy_user_nocache_64.S b/arch/x86/lib/copy_user_nocache_64.S +index cb0c112..e3a6895 100644 +--- a/arch/x86/lib/copy_user_nocache_64.S ++++ b/arch/x86/lib/copy_user_nocache_64.S +@@ -8,12 +8,14 @@ + + #include + #include ++#include + + #define FIX_ALIGNMENT 1 + + #include + #include + #include ++#include + + .macro ALIGN_DESTINATION + #ifdef FIX_ALIGNMENT +@@ -50,6 +52,15 @@ + */ + ENTRY(__copy_user_nocache) + CFI_STARTPROC ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ mov $PAX_USER_SHADOW_BASE,%rcx ++ cmp %rcx,%rsi ++ jae 1f ++ add %rcx,%rsi ++1: ++#endif ++ + cmpl $8,%edx + jb 20f /* less then 8 bytes, go to byte copy loop */ + ALIGN_DESTINATION +@@ -59,19 +70,19 @@ ENTRY(__copy_user_nocache) + jz 17f + 1: movq (%rsi),%r8 + 2: movq 1*8(%rsi),%r9 +-3: movq 2*8(%rsi),%r10 ++3: movq 2*8(%rsi),%rax + 4: movq 3*8(%rsi),%r11 + 5: movnti %r8,(%rdi) + 6: movnti %r9,1*8(%rdi) +-7: movnti %r10,2*8(%rdi) ++7: movnti %rax,2*8(%rdi) + 8: movnti %r11,3*8(%rdi) + 9: movq 4*8(%rsi),%r8 + 10: movq 5*8(%rsi),%r9 +-11: movq 6*8(%rsi),%r10 ++11: movq 6*8(%rsi),%rax + 12: movq 7*8(%rsi),%r11 + 13: movnti %r8,4*8(%rdi) + 14: movnti %r9,5*8(%rdi) +-15: movnti %r10,6*8(%rdi) ++15: movnti %rax,6*8(%rdi) + 16: movnti %r11,7*8(%rdi) + leaq 64(%rsi),%rsi + leaq 64(%rdi),%rdi +@@ -98,6 +109,7 @@ ENTRY(__copy_user_nocache) + jnz 21b + 23: xorl %eax,%eax + sfence ++ pax_force_retaddr + ret + + .section .fixup,"ax" +diff --git a/arch/x86/lib/csum-copy_64.S b/arch/x86/lib/csum-copy_64.S +index f0dba36..48cb4d6 100644 +--- a/arch/x86/lib/csum-copy_64.S ++++ b/arch/x86/lib/csum-copy_64.S +@@ -8,6 +8,7 @@ + #include + #include + #include ++#include + + /* + * Checksum copy with exception handling. +@@ -228,6 +229,7 @@ ENTRY(csum_partial_copy_generic) + CFI_RESTORE rbp + addq $7*8,%rsp + CFI_ADJUST_CFA_OFFSET -7*8 ++ pax_force_retaddr 0, 1 + ret + CFI_RESTORE_STATE + +diff --git a/arch/x86/lib/csum-wrappers_64.c b/arch/x86/lib/csum-wrappers_64.c +index 459b58a..9570bc7 100644 +--- a/arch/x86/lib/csum-wrappers_64.c ++++ b/arch/x86/lib/csum-wrappers_64.c +@@ -52,7 +52,13 @@ csum_partial_copy_from_user(const void __user *src, void *dst, + len -= 2; + } + } +- isum = csum_partial_copy_generic((__force const void *)src, ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++#endif ++ ++ isum = csum_partial_copy_generic((const void __force_kernel *)src, + dst, len, isum, errp, NULL); + if (unlikely(*errp)) + goto out_err; +@@ -105,7 +111,13 @@ csum_partial_copy_to_user(const void *src, void __user *dst, + } + + *errp = 0; +- return csum_partial_copy_generic(src, (void __force *)dst, ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)dst < PAX_USER_SHADOW_BASE) ++ dst += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return csum_partial_copy_generic(src, (void __force_kernel *)dst, + len, isum, NULL, errp); + } + EXPORT_SYMBOL(csum_partial_copy_to_user); +diff --git a/arch/x86/lib/getuser.S b/arch/x86/lib/getuser.S +index 51f1504..ddac4c1 100644 +--- a/arch/x86/lib/getuser.S ++++ b/arch/x86/lib/getuser.S +@@ -33,15 +33,38 @@ + #include + #include + #include ++#include ++#include ++#include ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define __copyuser_seg gs; ++#else ++#define __copyuser_seg ++#endif + + .text + ENTRY(__get_user_1) + CFI_STARTPROC ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) + GET_THREAD_INFO(%_ASM_DX) + cmp TI_addr_limit(%_ASM_DX),%_ASM_AX + jae bad_get_user +-1: movzb (%_ASM_AX),%edx ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_DX ++ cmp %_ASM_DX,%_ASM_AX ++ jae 1234f ++ add %_ASM_DX,%_ASM_AX ++1234: ++#endif ++ ++#endif ++ ++1: __copyuser_seg movzb (%_ASM_AX),%edx + xor %eax,%eax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(__get_user_1) +@@ -49,12 +72,26 @@ ENDPROC(__get_user_1) + ENTRY(__get_user_2) + CFI_STARTPROC + add $1,%_ASM_AX ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) + jc bad_get_user + GET_THREAD_INFO(%_ASM_DX) + cmp TI_addr_limit(%_ASM_DX),%_ASM_AX + jae bad_get_user +-2: movzwl -1(%_ASM_AX),%edx ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_DX ++ cmp %_ASM_DX,%_ASM_AX ++ jae 1234f ++ add %_ASM_DX,%_ASM_AX ++1234: ++#endif ++ ++#endif ++ ++2: __copyuser_seg movzwl -1(%_ASM_AX),%edx + xor %eax,%eax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(__get_user_2) +@@ -62,12 +99,26 @@ ENDPROC(__get_user_2) + ENTRY(__get_user_4) + CFI_STARTPROC + add $3,%_ASM_AX ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) + jc bad_get_user + GET_THREAD_INFO(%_ASM_DX) + cmp TI_addr_limit(%_ASM_DX),%_ASM_AX + jae bad_get_user +-3: mov -3(%_ASM_AX),%edx ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_DX ++ cmp %_ASM_DX,%_ASM_AX ++ jae 1234f ++ add %_ASM_DX,%_ASM_AX ++1234: ++#endif ++ ++#endif ++ ++3: __copyuser_seg mov -3(%_ASM_AX),%edx + xor %eax,%eax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(__get_user_4) +@@ -80,8 +131,18 @@ ENTRY(__get_user_8) + GET_THREAD_INFO(%_ASM_DX) + cmp TI_addr_limit(%_ASM_DX),%_ASM_AX + jae bad_get_user ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ mov $PAX_USER_SHADOW_BASE,%_ASM_DX ++ cmp %_ASM_DX,%_ASM_AX ++ jae 1234f ++ add %_ASM_DX,%_ASM_AX ++1234: ++#endif ++ + 4: movq -7(%_ASM_AX),%_ASM_DX + xor %eax,%eax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(__get_user_8) +@@ -91,6 +152,7 @@ bad_get_user: + CFI_STARTPROC + xor %edx,%edx + mov $(-EFAULT),%_ASM_AX ++ pax_force_retaddr + ret + CFI_ENDPROC + END(bad_get_user) +diff --git a/arch/x86/lib/iomap_copy_64.S b/arch/x86/lib/iomap_copy_64.S +index 05a95e7..326f2fa 100644 +--- a/arch/x86/lib/iomap_copy_64.S ++++ b/arch/x86/lib/iomap_copy_64.S +@@ -17,6 +17,7 @@ + + #include + #include ++#include + + /* + * override generic version in lib/iomap_copy.c +@@ -25,6 +26,7 @@ ENTRY(__iowrite32_copy) + CFI_STARTPROC + movl %edx,%ecx + rep movsd ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(__iowrite32_copy) +diff --git a/arch/x86/lib/memcpy_64.S b/arch/x86/lib/memcpy_64.S +index ad5441e..610e351 100644 +--- a/arch/x86/lib/memcpy_64.S ++++ b/arch/x86/lib/memcpy_64.S +@@ -4,6 +4,7 @@ + + #include + #include ++#include + + /* + * memcpy - Copy a memory block. +@@ -34,6 +35,7 @@ memcpy_c: + rep movsq + movl %edx, %ecx + rep movsb ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(memcpy_c) +@@ -118,6 +120,7 @@ ENTRY(memcpy) + jnz .Lloop_1 + + .Lend: ++ pax_force_retaddr 0, 1 + ret + CFI_ENDPROC + ENDPROC(memcpy) +@@ -128,7 +131,7 @@ ENDPROC(__memcpy) + * It is also a lot simpler. Use this when possible: + */ + +- .section .altinstr_replacement, "ax" ++ .section .altinstr_replacement, "a" + 1: .byte 0xeb /* jmp */ + .byte (memcpy_c - memcpy) - (2f - 1b) /* offset */ + 2: +diff --git a/arch/x86/lib/memset_64.S b/arch/x86/lib/memset_64.S +index 2c59481..7e9ba4e 100644 +--- a/arch/x86/lib/memset_64.S ++++ b/arch/x86/lib/memset_64.S +@@ -2,6 +2,7 @@ + + #include + #include ++#include + + /* + * ISO C memset - set a memory block to a byte value. +@@ -28,6 +29,7 @@ memset_c: + movl %r8d,%ecx + rep stosb + movq %r9,%rax ++ pax_force_retaddr + ret + CFI_ENDPROC + ENDPROC(memset_c) +@@ -35,13 +37,13 @@ ENDPROC(memset_c) + ENTRY(memset) + ENTRY(__memset) + CFI_STARTPROC +- movq %rdi,%r10 + movq %rdx,%r11 + + /* expand byte value */ + movzbl %sil,%ecx + movabs $0x0101010101010101,%rax + mul %rcx /* with rax, clobbers rdx */ ++ movq %rdi,%rdx + + /* align dst */ + movl %edi,%r9d +@@ -95,7 +97,8 @@ ENTRY(__memset) + jnz .Lloop_1 + + .Lende: +- movq %r10,%rax ++ movq %rdx,%rax ++ pax_force_retaddr + ret + + CFI_RESTORE_STATE +@@ -118,7 +121,7 @@ ENDPROC(__memset) + + #include + +- .section .altinstr_replacement,"ax" ++ .section .altinstr_replacement,"a" + 1: .byte 0xeb /* jmp */ + .byte (memset_c - memset) - (2f - 1b) /* offset */ + 2: +diff --git a/arch/x86/lib/mmx_32.c b/arch/x86/lib/mmx_32.c +index c9f2d9b..e7fd2c0 100644 +--- a/arch/x86/lib/mmx_32.c ++++ b/arch/x86/lib/mmx_32.c +@@ -29,6 +29,7 @@ void *_mmx_memcpy(void *to, const void *from, size_t len) + { + void *p; + int i; ++ unsigned long cr0; + + if (unlikely(in_interrupt())) + return __memcpy(to, from, len); +@@ -39,44 +40,72 @@ void *_mmx_memcpy(void *to, const void *from, size_t len) + kernel_fpu_begin(); + + __asm__ __volatile__ ( +- "1: prefetch (%0)\n" /* This set is 28 bytes */ +- " prefetch 64(%0)\n" +- " prefetch 128(%0)\n" +- " prefetch 192(%0)\n" +- " prefetch 256(%0)\n" ++ "1: prefetch (%1)\n" /* This set is 28 bytes */ ++ " prefetch 64(%1)\n" ++ " prefetch 128(%1)\n" ++ " prefetch 192(%1)\n" ++ " prefetch 256(%1)\n" + "2: \n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ "3: \n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" + _ASM_EXTABLE(1b, 3b) +- : : "r" (from)); ++ : "=&r" (cr0) : "r" (from) : "ax"); + + for ( ; i > 5; i--) { + __asm__ __volatile__ ( +- "1: prefetch 320(%0)\n" +- "2: movq (%0), %%mm0\n" +- " movq 8(%0), %%mm1\n" +- " movq 16(%0), %%mm2\n" +- " movq 24(%0), %%mm3\n" +- " movq %%mm0, (%1)\n" +- " movq %%mm1, 8(%1)\n" +- " movq %%mm2, 16(%1)\n" +- " movq %%mm3, 24(%1)\n" +- " movq 32(%0), %%mm0\n" +- " movq 40(%0), %%mm1\n" +- " movq 48(%0), %%mm2\n" +- " movq 56(%0), %%mm3\n" +- " movq %%mm0, 32(%1)\n" +- " movq %%mm1, 40(%1)\n" +- " movq %%mm2, 48(%1)\n" +- " movq %%mm3, 56(%1)\n" ++ "1: prefetch 320(%1)\n" ++ "2: movq (%1), %%mm0\n" ++ " movq 8(%1), %%mm1\n" ++ " movq 16(%1), %%mm2\n" ++ " movq 24(%1), %%mm3\n" ++ " movq %%mm0, (%2)\n" ++ " movq %%mm1, 8(%2)\n" ++ " movq %%mm2, 16(%2)\n" ++ " movq %%mm3, 24(%2)\n" ++ " movq 32(%1), %%mm0\n" ++ " movq 40(%1), %%mm1\n" ++ " movq 48(%1), %%mm2\n" ++ " movq 56(%1), %%mm3\n" ++ " movq %%mm0, 32(%2)\n" ++ " movq %%mm1, 40(%2)\n" ++ " movq %%mm2, 48(%2)\n" ++ " movq %%mm3, 56(%2)\n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ "3:\n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" + _ASM_EXTABLE(1b, 3b) +- : : "r" (from), "r" (to) : "memory"); ++ : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax"); + + from += 64; + to += 64; +@@ -158,6 +187,7 @@ static void fast_clear_page(void *page) + static void fast_copy_page(void *to, void *from) + { + int i; ++ unsigned long cr0; + + kernel_fpu_begin(); + +@@ -166,42 +196,70 @@ static void fast_copy_page(void *to, void *from) + * but that is for later. -AV + */ + __asm__ __volatile__( +- "1: prefetch (%0)\n" +- " prefetch 64(%0)\n" +- " prefetch 128(%0)\n" +- " prefetch 192(%0)\n" +- " prefetch 256(%0)\n" ++ "1: prefetch (%1)\n" ++ " prefetch 64(%1)\n" ++ " prefetch 128(%1)\n" ++ " prefetch 192(%1)\n" ++ " prefetch 256(%1)\n" + "2: \n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ "3: \n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" +- _ASM_EXTABLE(1b, 3b) : : "r" (from)); ++ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from) : "ax"); + + for (i = 0; i < (4096-320)/64; i++) { + __asm__ __volatile__ ( +- "1: prefetch 320(%0)\n" +- "2: movq (%0), %%mm0\n" +- " movntq %%mm0, (%1)\n" +- " movq 8(%0), %%mm1\n" +- " movntq %%mm1, 8(%1)\n" +- " movq 16(%0), %%mm2\n" +- " movntq %%mm2, 16(%1)\n" +- " movq 24(%0), %%mm3\n" +- " movntq %%mm3, 24(%1)\n" +- " movq 32(%0), %%mm4\n" +- " movntq %%mm4, 32(%1)\n" +- " movq 40(%0), %%mm5\n" +- " movntq %%mm5, 40(%1)\n" +- " movq 48(%0), %%mm6\n" +- " movntq %%mm6, 48(%1)\n" +- " movq 56(%0), %%mm7\n" +- " movntq %%mm7, 56(%1)\n" ++ "1: prefetch 320(%1)\n" ++ "2: movq (%1), %%mm0\n" ++ " movntq %%mm0, (%2)\n" ++ " movq 8(%1), %%mm1\n" ++ " movntq %%mm1, 8(%2)\n" ++ " movq 16(%1), %%mm2\n" ++ " movntq %%mm2, 16(%2)\n" ++ " movq 24(%1), %%mm3\n" ++ " movntq %%mm3, 24(%2)\n" ++ " movq 32(%1), %%mm4\n" ++ " movntq %%mm4, 32(%2)\n" ++ " movq 40(%1), %%mm5\n" ++ " movntq %%mm5, 40(%2)\n" ++ " movq 48(%1), %%mm6\n" ++ " movntq %%mm6, 48(%2)\n" ++ " movq 56(%1), %%mm7\n" ++ " movntq %%mm7, 56(%2)\n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ "3:\n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" +- _ASM_EXTABLE(1b, 3b) : : "r" (from), "r" (to) : "memory"); ++ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax"); + + from += 64; + to += 64; +@@ -280,47 +338,76 @@ static void fast_clear_page(void *page) + static void fast_copy_page(void *to, void *from) + { + int i; ++ unsigned long cr0; + + kernel_fpu_begin(); + + __asm__ __volatile__ ( +- "1: prefetch (%0)\n" +- " prefetch 64(%0)\n" +- " prefetch 128(%0)\n" +- " prefetch 192(%0)\n" +- " prefetch 256(%0)\n" ++ "1: prefetch (%1)\n" ++ " prefetch 64(%1)\n" ++ " prefetch 128(%1)\n" ++ " prefetch 192(%1)\n" ++ " prefetch 256(%1)\n" + "2: \n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ "3: \n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x1AEB, 1b\n" /* jmp on 26 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" +- _ASM_EXTABLE(1b, 3b) : : "r" (from)); ++ _ASM_EXTABLE(1b, 3b) : "=&r" (cr0) : "r" (from) : "ax"); + + for (i = 0; i < 4096/64; i++) { + __asm__ __volatile__ ( +- "1: prefetch 320(%0)\n" +- "2: movq (%0), %%mm0\n" +- " movq 8(%0), %%mm1\n" +- " movq 16(%0), %%mm2\n" +- " movq 24(%0), %%mm3\n" +- " movq %%mm0, (%1)\n" +- " movq %%mm1, 8(%1)\n" +- " movq %%mm2, 16(%1)\n" +- " movq %%mm3, 24(%1)\n" +- " movq 32(%0), %%mm0\n" +- " movq 40(%0), %%mm1\n" +- " movq 48(%0), %%mm2\n" +- " movq 56(%0), %%mm3\n" +- " movq %%mm0, 32(%1)\n" +- " movq %%mm1, 40(%1)\n" +- " movq %%mm2, 48(%1)\n" +- " movq %%mm3, 56(%1)\n" ++ "1: prefetch 320(%1)\n" ++ "2: movq (%1), %%mm0\n" ++ " movq 8(%1), %%mm1\n" ++ " movq 16(%1), %%mm2\n" ++ " movq 24(%1), %%mm3\n" ++ " movq %%mm0, (%2)\n" ++ " movq %%mm1, 8(%2)\n" ++ " movq %%mm2, 16(%2)\n" ++ " movq %%mm3, 24(%2)\n" ++ " movq 32(%1), %%mm0\n" ++ " movq 40(%1), %%mm1\n" ++ " movq 48(%1), %%mm2\n" ++ " movq 56(%1), %%mm3\n" ++ " movq %%mm0, 32(%2)\n" ++ " movq %%mm1, 40(%2)\n" ++ " movq %%mm2, 48(%2)\n" ++ " movq %%mm3, 56(%2)\n" + ".section .fixup, \"ax\"\n" +- "3: movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ "3:\n" ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %%cr0, %0\n" ++ " movl %0, %%eax\n" ++ " andl $0xFFFEFFFF, %%eax\n" ++ " movl %%eax, %%cr0\n" ++#endif ++ ++ " movw $0x05EB, 1b\n" /* jmp on 5 bytes */ ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ " movl %0, %%cr0\n" ++#endif ++ + " jmp 2b\n" + ".previous\n" + _ASM_EXTABLE(1b, 3b) +- : : "r" (from), "r" (to) : "memory"); ++ : "=&r" (cr0) : "r" (from), "r" (to) : "memory", "ax"); + + from += 64; + to += 64; +diff --git a/arch/x86/lib/msr-reg.S b/arch/x86/lib/msr-reg.S +index 69fa106..adda88b 100644 +--- a/arch/x86/lib/msr-reg.S ++++ b/arch/x86/lib/msr-reg.S +@@ -3,6 +3,7 @@ + #include + #include + #include ++#include + + #ifdef CONFIG_X86_64 + /* +@@ -16,7 +17,7 @@ ENTRY(native_\op\()_safe_regs) + CFI_STARTPROC + pushq_cfi %rbx + pushq_cfi %rbp +- movq %rdi, %r10 /* Save pointer */ ++ movq %rdi, %r9 /* Save pointer */ + xorl %r11d, %r11d /* Return value */ + movl (%rdi), %eax + movl 4(%rdi), %ecx +@@ -27,16 +28,17 @@ ENTRY(native_\op\()_safe_regs) + movl 28(%rdi), %edi + CFI_REMEMBER_STATE + 1: \op +-2: movl %eax, (%r10) ++2: movl %eax, (%r9) + movl %r11d, %eax /* Return value */ +- movl %ecx, 4(%r10) +- movl %edx, 8(%r10) +- movl %ebx, 12(%r10) +- movl %ebp, 20(%r10) +- movl %esi, 24(%r10) +- movl %edi, 28(%r10) ++ movl %ecx, 4(%r9) ++ movl %edx, 8(%r9) ++ movl %ebx, 12(%r9) ++ movl %ebp, 20(%r9) ++ movl %esi, 24(%r9) ++ movl %edi, 28(%r9) + popq_cfi %rbp + popq_cfi %rbx ++ pax_force_retaddr + ret + 3: + CFI_RESTORE_STATE +diff --git a/arch/x86/lib/putuser.S b/arch/x86/lib/putuser.S +index 36b0d15..d381858 100644 +--- a/arch/x86/lib/putuser.S ++++ b/arch/x86/lib/putuser.S +@@ -15,7 +15,9 @@ + #include + #include + #include +- ++#include ++#include ++#include + + /* + * __put_user_X +@@ -29,52 +31,119 @@ + * as they get called from within inline assembly. + */ + +-#define ENTER CFI_STARTPROC ; \ +- GET_THREAD_INFO(%_ASM_BX) +-#define EXIT ret ; \ ++#define ENTER CFI_STARTPROC ++#define EXIT pax_force_retaddr; ret ; \ + CFI_ENDPROC + ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define _DEST %_ASM_CX,%_ASM_BX ++#else ++#define _DEST %_ASM_CX ++#endif ++ ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#define __copyuser_seg gs; ++#else ++#define __copyuser_seg ++#endif ++ + .text + ENTRY(__put_user_1) + ENTER ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) ++ GET_THREAD_INFO(%_ASM_BX) + cmp TI_addr_limit(%_ASM_BX),%_ASM_CX + jae bad_put_user +-1: movb %al,(%_ASM_CX) ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_BX ++ cmp %_ASM_BX,%_ASM_CX ++ jb 1234f ++ xor %ebx,%ebx ++1234: ++#endif ++ ++#endif ++ ++1: __copyuser_seg movb %al,(_DEST) + xor %eax,%eax + EXIT + ENDPROC(__put_user_1) + + ENTRY(__put_user_2) + ENTER ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) ++ GET_THREAD_INFO(%_ASM_BX) + mov TI_addr_limit(%_ASM_BX),%_ASM_BX + sub $1,%_ASM_BX + cmp %_ASM_BX,%_ASM_CX + jae bad_put_user +-2: movw %ax,(%_ASM_CX) ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_BX ++ cmp %_ASM_BX,%_ASM_CX ++ jb 1234f ++ xor %ebx,%ebx ++1234: ++#endif ++ ++#endif ++ ++2: __copyuser_seg movw %ax,(_DEST) + xor %eax,%eax + EXIT + ENDPROC(__put_user_2) + + ENTRY(__put_user_4) + ENTER ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) ++ GET_THREAD_INFO(%_ASM_BX) + mov TI_addr_limit(%_ASM_BX),%_ASM_BX + sub $3,%_ASM_BX + cmp %_ASM_BX,%_ASM_CX + jae bad_put_user +-3: movl %eax,(%_ASM_CX) ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_BX ++ cmp %_ASM_BX,%_ASM_CX ++ jb 1234f ++ xor %ebx,%ebx ++1234: ++#endif ++ ++#endif ++ ++3: __copyuser_seg movl %eax,(_DEST) + xor %eax,%eax + EXIT + ENDPROC(__put_user_4) + + ENTRY(__put_user_8) + ENTER ++ ++#if !defined(CONFIG_X86_32) || !defined(CONFIG_PAX_MEMORY_UDEREF) ++ GET_THREAD_INFO(%_ASM_BX) + mov TI_addr_limit(%_ASM_BX),%_ASM_BX + sub $7,%_ASM_BX + cmp %_ASM_BX,%_ASM_CX + jae bad_put_user +-4: mov %_ASM_AX,(%_ASM_CX) ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ mov $PAX_USER_SHADOW_BASE,%_ASM_BX ++ cmp %_ASM_BX,%_ASM_CX ++ jb 1234f ++ xor %ebx,%ebx ++1234: ++#endif ++ ++#endif ++ ++4: __copyuser_seg mov %_ASM_AX,(_DEST) + #ifdef CONFIG_X86_32 +-5: movl %edx,4(%_ASM_CX) ++5: __copyuser_seg movl %edx,4(_DEST) + #endif + xor %eax,%eax + EXIT +diff --git a/arch/x86/lib/rwlock_64.S b/arch/x86/lib/rwlock_64.S +index 05ea55f..6345b9a 100644 +--- a/arch/x86/lib/rwlock_64.S ++++ b/arch/x86/lib/rwlock_64.S +@@ -2,6 +2,7 @@ + + #include + #include ++#include + #include + #include + +@@ -10,13 +11,34 @@ ENTRY(__write_lock_failed) + CFI_STARTPROC + LOCK_PREFIX + addl $RW_LOCK_BIAS,(%rdi) ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ jno 1234f ++ LOCK_PREFIX ++ subl $RW_LOCK_BIAS,(%rdi) ++ int $4 ++1234: ++ _ASM_EXTABLE(1234b, 1234b) ++#endif ++ + 1: rep + nop + cmpl $RW_LOCK_BIAS,(%rdi) + jne 1b + LOCK_PREFIX + subl $RW_LOCK_BIAS,(%rdi) ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ jno 1234f ++ LOCK_PREFIX ++ addl $RW_LOCK_BIAS,(%rdi) ++ int $4 ++1234: ++ _ASM_EXTABLE(1234b, 1234b) ++#endif ++ + jnz __write_lock_failed ++ pax_force_retaddr + ret + CFI_ENDPROC + END(__write_lock_failed) +@@ -26,13 +48,34 @@ ENTRY(__read_lock_failed) + CFI_STARTPROC + LOCK_PREFIX + incl (%rdi) ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ jno 1234f ++ LOCK_PREFIX ++ decl (%rdi) ++ int $4 ++1234: ++ _ASM_EXTABLE(1234b, 1234b) ++#endif ++ + 1: rep + nop + cmpl $1,(%rdi) + js 1b + LOCK_PREFIX + decl (%rdi) ++ ++#ifdef CONFIG_PAX_REFCOUNT ++ jno 1234f ++ LOCK_PREFIX ++ incl (%rdi) ++ int $4 ++1234: ++ _ASM_EXTABLE(1234b, 1234b) ++#endif ++ + js __read_lock_failed ++ pax_force_retaddr + ret + CFI_ENDPROC + END(__read_lock_failed) +diff --git a/arch/x86/lib/rwsem_64.S b/arch/x86/lib/rwsem_64.S +index 15acecf..f768b10 100644 +--- a/arch/x86/lib/rwsem_64.S ++++ b/arch/x86/lib/rwsem_64.S +@@ -48,6 +48,7 @@ ENTRY(call_rwsem_down_read_failed) + call rwsem_down_read_failed + popq %rdx + restore_common_regs ++ pax_force_retaddr + ret + ENDPROC(call_rwsem_down_read_failed) + +@@ -56,6 +57,7 @@ ENTRY(call_rwsem_down_write_failed) + movq %rax,%rdi + call rwsem_down_write_failed + restore_common_regs ++ pax_force_retaddr + ret + ENDPROC(call_rwsem_down_write_failed) + +@@ -66,7 +68,8 @@ ENTRY(call_rwsem_wake) + movq %rax,%rdi + call rwsem_wake + restore_common_regs +-1: ret ++1: pax_force_retaddr ++ ret + ENDPROC(call_rwsem_wake) + + /* Fix up special calling conventions */ +@@ -77,5 +80,6 @@ ENTRY(call_rwsem_downgrade_wake) + call rwsem_downgrade_wake + popq %rdx + restore_common_regs ++ pax_force_retaddr + ret + ENDPROC(call_rwsem_downgrade_wake) +diff --git a/arch/x86/lib/thunk_64.S b/arch/x86/lib/thunk_64.S +index bf9a7d5..fb06ab5 100644 +--- a/arch/x86/lib/thunk_64.S ++++ b/arch/x86/lib/thunk_64.S +@@ -10,7 +10,8 @@ + #include + #include + #include +- ++ #include ++ + /* rdi: arg1 ... normal C conventions. rax is saved/restored. */ + .macro thunk name,func + .globl \name +@@ -70,6 +71,7 @@ + SAVE_ARGS + restore: + RESTORE_ARGS ++ pax_force_retaddr + ret + CFI_ENDPROC + +@@ -77,5 +79,6 @@ restore: + SAVE_ARGS + restore_norax: + RESTORE_ARGS 1 ++ pax_force_retaddr + ret + CFI_ENDPROC +diff --git a/arch/x86/lib/usercopy_32.c b/arch/x86/lib/usercopy_32.c +index 1f118d4..ec4a953 100644 +--- a/arch/x86/lib/usercopy_32.c ++++ b/arch/x86/lib/usercopy_32.c +@@ -43,7 +43,7 @@ do { \ + __asm__ __volatile__( \ + " testl %1,%1\n" \ + " jz 2f\n" \ +- "0: lodsb\n" \ ++ "0: "__copyuser_seg"lodsb\n" \ + " stosb\n" \ + " testb %%al,%%al\n" \ + " jz 1f\n" \ +@@ -128,10 +128,12 @@ do { \ + int __d0; \ + might_fault(); \ + __asm__ __volatile__( \ ++ __COPYUSER_SET_ES \ + "0: rep; stosl\n" \ + " movl %2,%0\n" \ + "1: rep; stosb\n" \ + "2:\n" \ ++ __COPYUSER_RESTORE_ES \ + ".section .fixup,\"ax\"\n" \ + "3: lea 0(%2,%0,4),%0\n" \ + " jmp 2b\n" \ +@@ -200,6 +202,7 @@ long strnlen_user(const char __user *s, long n) + might_fault(); + + __asm__ __volatile__( ++ __COPYUSER_SET_ES + " testl %0, %0\n" + " jz 3f\n" + " andl %0,%%ecx\n" +@@ -208,6 +211,7 @@ long strnlen_user(const char __user *s, long n) + " subl %%ecx,%0\n" + " addl %0,%%eax\n" + "1:\n" ++ __COPYUSER_RESTORE_ES + ".section .fixup,\"ax\"\n" + "2: xorl %%eax,%%eax\n" + " jmp 1b\n" +@@ -227,7 +231,7 @@ EXPORT_SYMBOL(strnlen_user); + + #ifdef CONFIG_X86_INTEL_USERCOPY + static unsigned long +-__copy_user_intel(void __user *to, const void *from, unsigned long size) ++__generic_copy_to_user_intel(void __user *to, const void *from, unsigned long size) + { + int d0, d1; + __asm__ __volatile__( +@@ -239,36 +243,36 @@ __copy_user_intel(void __user *to, const void *from, unsigned long size) + " .align 2,0x90\n" + "3: movl 0(%4), %%eax\n" + "4: movl 4(%4), %%edx\n" +- "5: movl %%eax, 0(%3)\n" +- "6: movl %%edx, 4(%3)\n" ++ "5: "__copyuser_seg" movl %%eax, 0(%3)\n" ++ "6: "__copyuser_seg" movl %%edx, 4(%3)\n" + "7: movl 8(%4), %%eax\n" + "8: movl 12(%4),%%edx\n" +- "9: movl %%eax, 8(%3)\n" +- "10: movl %%edx, 12(%3)\n" ++ "9: "__copyuser_seg" movl %%eax, 8(%3)\n" ++ "10: "__copyuser_seg" movl %%edx, 12(%3)\n" + "11: movl 16(%4), %%eax\n" + "12: movl 20(%4), %%edx\n" +- "13: movl %%eax, 16(%3)\n" +- "14: movl %%edx, 20(%3)\n" ++ "13: "__copyuser_seg" movl %%eax, 16(%3)\n" ++ "14: "__copyuser_seg" movl %%edx, 20(%3)\n" + "15: movl 24(%4), %%eax\n" + "16: movl 28(%4), %%edx\n" +- "17: movl %%eax, 24(%3)\n" +- "18: movl %%edx, 28(%3)\n" ++ "17: "__copyuser_seg" movl %%eax, 24(%3)\n" ++ "18: "__copyuser_seg" movl %%edx, 28(%3)\n" + "19: movl 32(%4), %%eax\n" + "20: movl 36(%4), %%edx\n" +- "21: movl %%eax, 32(%3)\n" +- "22: movl %%edx, 36(%3)\n" ++ "21: "__copyuser_seg" movl %%eax, 32(%3)\n" ++ "22: "__copyuser_seg" movl %%edx, 36(%3)\n" + "23: movl 40(%4), %%eax\n" + "24: movl 44(%4), %%edx\n" +- "25: movl %%eax, 40(%3)\n" +- "26: movl %%edx, 44(%3)\n" ++ "25: "__copyuser_seg" movl %%eax, 40(%3)\n" ++ "26: "__copyuser_seg" movl %%edx, 44(%3)\n" + "27: movl 48(%4), %%eax\n" + "28: movl 52(%4), %%edx\n" +- "29: movl %%eax, 48(%3)\n" +- "30: movl %%edx, 52(%3)\n" ++ "29: "__copyuser_seg" movl %%eax, 48(%3)\n" ++ "30: "__copyuser_seg" movl %%edx, 52(%3)\n" + "31: movl 56(%4), %%eax\n" + "32: movl 60(%4), %%edx\n" +- "33: movl %%eax, 56(%3)\n" +- "34: movl %%edx, 60(%3)\n" ++ "33: "__copyuser_seg" movl %%eax, 56(%3)\n" ++ "34: "__copyuser_seg" movl %%edx, 60(%3)\n" + " addl $-64, %0\n" + " addl $64, %4\n" + " addl $64, %3\n" +@@ -278,10 +282,119 @@ __copy_user_intel(void __user *to, const void *from, unsigned long size) + " shrl $2, %0\n" + " andl $3, %%eax\n" + " cld\n" ++ __COPYUSER_SET_ES + "99: rep; movsl\n" + "36: movl %%eax, %0\n" + "37: rep; movsb\n" + "100:\n" ++ __COPYUSER_RESTORE_ES ++ ".section .fixup,\"ax\"\n" ++ "101: lea 0(%%eax,%0,4),%0\n" ++ " jmp 100b\n" ++ ".previous\n" ++ ".section __ex_table,\"a\"\n" ++ " .align 4\n" ++ " .long 1b,100b\n" ++ " .long 2b,100b\n" ++ " .long 3b,100b\n" ++ " .long 4b,100b\n" ++ " .long 5b,100b\n" ++ " .long 6b,100b\n" ++ " .long 7b,100b\n" ++ " .long 8b,100b\n" ++ " .long 9b,100b\n" ++ " .long 10b,100b\n" ++ " .long 11b,100b\n" ++ " .long 12b,100b\n" ++ " .long 13b,100b\n" ++ " .long 14b,100b\n" ++ " .long 15b,100b\n" ++ " .long 16b,100b\n" ++ " .long 17b,100b\n" ++ " .long 18b,100b\n" ++ " .long 19b,100b\n" ++ " .long 20b,100b\n" ++ " .long 21b,100b\n" ++ " .long 22b,100b\n" ++ " .long 23b,100b\n" ++ " .long 24b,100b\n" ++ " .long 25b,100b\n" ++ " .long 26b,100b\n" ++ " .long 27b,100b\n" ++ " .long 28b,100b\n" ++ " .long 29b,100b\n" ++ " .long 30b,100b\n" ++ " .long 31b,100b\n" ++ " .long 32b,100b\n" ++ " .long 33b,100b\n" ++ " .long 34b,100b\n" ++ " .long 35b,100b\n" ++ " .long 36b,100b\n" ++ " .long 37b,100b\n" ++ " .long 99b,101b\n" ++ ".previous" ++ : "=&c"(size), "=&D" (d0), "=&S" (d1) ++ : "1"(to), "2"(from), "0"(size) ++ : "eax", "edx", "memory"); ++ return size; ++} ++ ++static unsigned long ++__generic_copy_from_user_intel(void *to, const void __user *from, unsigned long size) ++{ ++ int d0, d1; ++ __asm__ __volatile__( ++ " .align 2,0x90\n" ++ "1: "__copyuser_seg" movl 32(%4), %%eax\n" ++ " cmpl $67, %0\n" ++ " jbe 3f\n" ++ "2: "__copyuser_seg" movl 64(%4), %%eax\n" ++ " .align 2,0x90\n" ++ "3: "__copyuser_seg" movl 0(%4), %%eax\n" ++ "4: "__copyuser_seg" movl 4(%4), %%edx\n" ++ "5: movl %%eax, 0(%3)\n" ++ "6: movl %%edx, 4(%3)\n" ++ "7: "__copyuser_seg" movl 8(%4), %%eax\n" ++ "8: "__copyuser_seg" movl 12(%4),%%edx\n" ++ "9: movl %%eax, 8(%3)\n" ++ "10: movl %%edx, 12(%3)\n" ++ "11: "__copyuser_seg" movl 16(%4), %%eax\n" ++ "12: "__copyuser_seg" movl 20(%4), %%edx\n" ++ "13: movl %%eax, 16(%3)\n" ++ "14: movl %%edx, 20(%3)\n" ++ "15: "__copyuser_seg" movl 24(%4), %%eax\n" ++ "16: "__copyuser_seg" movl 28(%4), %%edx\n" ++ "17: movl %%eax, 24(%3)\n" ++ "18: movl %%edx, 28(%3)\n" ++ "19: "__copyuser_seg" movl 32(%4), %%eax\n" ++ "20: "__copyuser_seg" movl 36(%4), %%edx\n" ++ "21: movl %%eax, 32(%3)\n" ++ "22: movl %%edx, 36(%3)\n" ++ "23: "__copyuser_seg" movl 40(%4), %%eax\n" ++ "24: "__copyuser_seg" movl 44(%4), %%edx\n" ++ "25: movl %%eax, 40(%3)\n" ++ "26: movl %%edx, 44(%3)\n" ++ "27: "__copyuser_seg" movl 48(%4), %%eax\n" ++ "28: "__copyuser_seg" movl 52(%4), %%edx\n" ++ "29: movl %%eax, 48(%3)\n" ++ "30: movl %%edx, 52(%3)\n" ++ "31: "__copyuser_seg" movl 56(%4), %%eax\n" ++ "32: "__copyuser_seg" movl 60(%4), %%edx\n" ++ "33: movl %%eax, 56(%3)\n" ++ "34: movl %%edx, 60(%3)\n" ++ " addl $-64, %0\n" ++ " addl $64, %4\n" ++ " addl $64, %3\n" ++ " cmpl $63, %0\n" ++ " ja 1b\n" ++ "35: movl %0, %%eax\n" ++ " shrl $2, %0\n" ++ " andl $3, %%eax\n" ++ " cld\n" ++ "99: rep; "__copyuser_seg" movsl\n" ++ "36: movl %%eax, %0\n" ++ "37: rep; "__copyuser_seg" movsb\n" ++ "100:\n" + ".section .fixup,\"ax\"\n" + "101: lea 0(%%eax,%0,4),%0\n" + " jmp 100b\n" +@@ -339,41 +452,41 @@ __copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size) + int d0, d1; + __asm__ __volatile__( + " .align 2,0x90\n" +- "0: movl 32(%4), %%eax\n" ++ "0: "__copyuser_seg" movl 32(%4), %%eax\n" + " cmpl $67, %0\n" + " jbe 2f\n" +- "1: movl 64(%4), %%eax\n" ++ "1: "__copyuser_seg" movl 64(%4), %%eax\n" + " .align 2,0x90\n" +- "2: movl 0(%4), %%eax\n" +- "21: movl 4(%4), %%edx\n" ++ "2: "__copyuser_seg" movl 0(%4), %%eax\n" ++ "21: "__copyuser_seg" movl 4(%4), %%edx\n" + " movl %%eax, 0(%3)\n" + " movl %%edx, 4(%3)\n" +- "3: movl 8(%4), %%eax\n" +- "31: movl 12(%4),%%edx\n" ++ "3: "__copyuser_seg" movl 8(%4), %%eax\n" ++ "31: "__copyuser_seg" movl 12(%4),%%edx\n" + " movl %%eax, 8(%3)\n" + " movl %%edx, 12(%3)\n" +- "4: movl 16(%4), %%eax\n" +- "41: movl 20(%4), %%edx\n" ++ "4: "__copyuser_seg" movl 16(%4), %%eax\n" ++ "41: "__copyuser_seg" movl 20(%4), %%edx\n" + " movl %%eax, 16(%3)\n" + " movl %%edx, 20(%3)\n" +- "10: movl 24(%4), %%eax\n" +- "51: movl 28(%4), %%edx\n" ++ "10: "__copyuser_seg" movl 24(%4), %%eax\n" ++ "51: "__copyuser_seg" movl 28(%4), %%edx\n" + " movl %%eax, 24(%3)\n" + " movl %%edx, 28(%3)\n" +- "11: movl 32(%4), %%eax\n" +- "61: movl 36(%4), %%edx\n" ++ "11: "__copyuser_seg" movl 32(%4), %%eax\n" ++ "61: "__copyuser_seg" movl 36(%4), %%edx\n" + " movl %%eax, 32(%3)\n" + " movl %%edx, 36(%3)\n" +- "12: movl 40(%4), %%eax\n" +- "71: movl 44(%4), %%edx\n" ++ "12: "__copyuser_seg" movl 40(%4), %%eax\n" ++ "71: "__copyuser_seg" movl 44(%4), %%edx\n" + " movl %%eax, 40(%3)\n" + " movl %%edx, 44(%3)\n" +- "13: movl 48(%4), %%eax\n" +- "81: movl 52(%4), %%edx\n" ++ "13: "__copyuser_seg" movl 48(%4), %%eax\n" ++ "81: "__copyuser_seg" movl 52(%4), %%edx\n" + " movl %%eax, 48(%3)\n" + " movl %%edx, 52(%3)\n" +- "14: movl 56(%4), %%eax\n" +- "91: movl 60(%4), %%edx\n" ++ "14: "__copyuser_seg" movl 56(%4), %%eax\n" ++ "91: "__copyuser_seg" movl 60(%4), %%edx\n" + " movl %%eax, 56(%3)\n" + " movl %%edx, 60(%3)\n" + " addl $-64, %0\n" +@@ -385,9 +498,9 @@ __copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size) + " shrl $2, %0\n" + " andl $3, %%eax\n" + " cld\n" +- "6: rep; movsl\n" ++ "6: rep; "__copyuser_seg" movsl\n" + " movl %%eax,%0\n" +- "7: rep; movsb\n" ++ "7: rep; "__copyuser_seg" movsb\n" + "8:\n" + ".section .fixup,\"ax\"\n" + "9: lea 0(%%eax,%0,4),%0\n" +@@ -440,41 +553,41 @@ static unsigned long __copy_user_zeroing_intel_nocache(void *to, + + __asm__ __volatile__( + " .align 2,0x90\n" +- "0: movl 32(%4), %%eax\n" ++ "0: "__copyuser_seg" movl 32(%4), %%eax\n" + " cmpl $67, %0\n" + " jbe 2f\n" +- "1: movl 64(%4), %%eax\n" ++ "1: "__copyuser_seg" movl 64(%4), %%eax\n" + " .align 2,0x90\n" +- "2: movl 0(%4), %%eax\n" +- "21: movl 4(%4), %%edx\n" ++ "2: "__copyuser_seg" movl 0(%4), %%eax\n" ++ "21: "__copyuser_seg" movl 4(%4), %%edx\n" + " movnti %%eax, 0(%3)\n" + " movnti %%edx, 4(%3)\n" +- "3: movl 8(%4), %%eax\n" +- "31: movl 12(%4),%%edx\n" ++ "3: "__copyuser_seg" movl 8(%4), %%eax\n" ++ "31: "__copyuser_seg" movl 12(%4),%%edx\n" + " movnti %%eax, 8(%3)\n" + " movnti %%edx, 12(%3)\n" +- "4: movl 16(%4), %%eax\n" +- "41: movl 20(%4), %%edx\n" ++ "4: "__copyuser_seg" movl 16(%4), %%eax\n" ++ "41: "__copyuser_seg" movl 20(%4), %%edx\n" + " movnti %%eax, 16(%3)\n" + " movnti %%edx, 20(%3)\n" +- "10: movl 24(%4), %%eax\n" +- "51: movl 28(%4), %%edx\n" ++ "10: "__copyuser_seg" movl 24(%4), %%eax\n" ++ "51: "__copyuser_seg" movl 28(%4), %%edx\n" + " movnti %%eax, 24(%3)\n" + " movnti %%edx, 28(%3)\n" +- "11: movl 32(%4), %%eax\n" +- "61: movl 36(%4), %%edx\n" ++ "11: "__copyuser_seg" movl 32(%4), %%eax\n" ++ "61: "__copyuser_seg" movl 36(%4), %%edx\n" + " movnti %%eax, 32(%3)\n" + " movnti %%edx, 36(%3)\n" +- "12: movl 40(%4), %%eax\n" +- "71: movl 44(%4), %%edx\n" ++ "12: "__copyuser_seg" movl 40(%4), %%eax\n" ++ "71: "__copyuser_seg" movl 44(%4), %%edx\n" + " movnti %%eax, 40(%3)\n" + " movnti %%edx, 44(%3)\n" +- "13: movl 48(%4), %%eax\n" +- "81: movl 52(%4), %%edx\n" ++ "13: "__copyuser_seg" movl 48(%4), %%eax\n" ++ "81: "__copyuser_seg" movl 52(%4), %%edx\n" + " movnti %%eax, 48(%3)\n" + " movnti %%edx, 52(%3)\n" +- "14: movl 56(%4), %%eax\n" +- "91: movl 60(%4), %%edx\n" ++ "14: "__copyuser_seg" movl 56(%4), %%eax\n" ++ "91: "__copyuser_seg" movl 60(%4), %%edx\n" + " movnti %%eax, 56(%3)\n" + " movnti %%edx, 60(%3)\n" + " addl $-64, %0\n" +@@ -487,9 +600,9 @@ static unsigned long __copy_user_zeroing_intel_nocache(void *to, + " shrl $2, %0\n" + " andl $3, %%eax\n" + " cld\n" +- "6: rep; movsl\n" ++ "6: rep; "__copyuser_seg" movsl\n" + " movl %%eax,%0\n" +- "7: rep; movsb\n" ++ "7: rep; "__copyuser_seg" movsb\n" + "8:\n" + ".section .fixup,\"ax\"\n" + "9: lea 0(%%eax,%0,4),%0\n" +@@ -537,41 +650,41 @@ static unsigned long __copy_user_intel_nocache(void *to, + + __asm__ __volatile__( + " .align 2,0x90\n" +- "0: movl 32(%4), %%eax\n" ++ "0: "__copyuser_seg" movl 32(%4), %%eax\n" + " cmpl $67, %0\n" + " jbe 2f\n" +- "1: movl 64(%4), %%eax\n" ++ "1: "__copyuser_seg" movl 64(%4), %%eax\n" + " .align 2,0x90\n" +- "2: movl 0(%4), %%eax\n" +- "21: movl 4(%4), %%edx\n" ++ "2: "__copyuser_seg" movl 0(%4), %%eax\n" ++ "21: "__copyuser_seg" movl 4(%4), %%edx\n" + " movnti %%eax, 0(%3)\n" + " movnti %%edx, 4(%3)\n" +- "3: movl 8(%4), %%eax\n" +- "31: movl 12(%4),%%edx\n" ++ "3: "__copyuser_seg" movl 8(%4), %%eax\n" ++ "31: "__copyuser_seg" movl 12(%4),%%edx\n" + " movnti %%eax, 8(%3)\n" + " movnti %%edx, 12(%3)\n" +- "4: movl 16(%4), %%eax\n" +- "41: movl 20(%4), %%edx\n" ++ "4: "__copyuser_seg" movl 16(%4), %%eax\n" ++ "41: "__copyuser_seg" movl 20(%4), %%edx\n" + " movnti %%eax, 16(%3)\n" + " movnti %%edx, 20(%3)\n" +- "10: movl 24(%4), %%eax\n" +- "51: movl 28(%4), %%edx\n" ++ "10: "__copyuser_seg" movl 24(%4), %%eax\n" ++ "51: "__copyuser_seg" movl 28(%4), %%edx\n" + " movnti %%eax, 24(%3)\n" + " movnti %%edx, 28(%3)\n" +- "11: movl 32(%4), %%eax\n" +- "61: movl 36(%4), %%edx\n" ++ "11: "__copyuser_seg" movl 32(%4), %%eax\n" ++ "61: "__copyuser_seg" movl 36(%4), %%edx\n" + " movnti %%eax, 32(%3)\n" + " movnti %%edx, 36(%3)\n" +- "12: movl 40(%4), %%eax\n" +- "71: movl 44(%4), %%edx\n" ++ "12: "__copyuser_seg" movl 40(%4), %%eax\n" ++ "71: "__copyuser_seg" movl 44(%4), %%edx\n" + " movnti %%eax, 40(%3)\n" + " movnti %%edx, 44(%3)\n" +- "13: movl 48(%4), %%eax\n" +- "81: movl 52(%4), %%edx\n" ++ "13: "__copyuser_seg" movl 48(%4), %%eax\n" ++ "81: "__copyuser_seg" movl 52(%4), %%edx\n" + " movnti %%eax, 48(%3)\n" + " movnti %%edx, 52(%3)\n" +- "14: movl 56(%4), %%eax\n" +- "91: movl 60(%4), %%edx\n" ++ "14: "__copyuser_seg" movl 56(%4), %%eax\n" ++ "91: "__copyuser_seg" movl 60(%4), %%edx\n" + " movnti %%eax, 56(%3)\n" + " movnti %%edx, 60(%3)\n" + " addl $-64, %0\n" +@@ -584,9 +697,9 @@ static unsigned long __copy_user_intel_nocache(void *to, + " shrl $2, %0\n" + " andl $3, %%eax\n" + " cld\n" +- "6: rep; movsl\n" ++ "6: rep; "__copyuser_seg" movsl\n" + " movl %%eax,%0\n" +- "7: rep; movsb\n" ++ "7: rep; "__copyuser_seg" movsb\n" + "8:\n" + ".section .fixup,\"ax\"\n" + "9: lea 0(%%eax,%0,4),%0\n" +@@ -629,32 +742,36 @@ static unsigned long __copy_user_intel_nocache(void *to, + */ + unsigned long __copy_user_zeroing_intel(void *to, const void __user *from, + unsigned long size); +-unsigned long __copy_user_intel(void __user *to, const void *from, ++unsigned long __generic_copy_to_user_intel(void __user *to, const void *from, ++ unsigned long size); ++unsigned long __generic_copy_from_user_intel(void *to, const void __user *from, + unsigned long size); + unsigned long __copy_user_zeroing_intel_nocache(void *to, + const void __user *from, unsigned long size); + #endif /* CONFIG_X86_INTEL_USERCOPY */ + + /* Generic arbitrary sized copy. */ +-#define __copy_user(to, from, size) \ ++#define __copy_user(to, from, size, prefix, set, restore) \ + do { \ + int __d0, __d1, __d2; \ + __asm__ __volatile__( \ ++ set \ + " cmp $7,%0\n" \ + " jbe 1f\n" \ + " movl %1,%0\n" \ + " negl %0\n" \ + " andl $7,%0\n" \ + " subl %0,%3\n" \ +- "4: rep; movsb\n" \ ++ "4: rep; "prefix"movsb\n" \ + " movl %3,%0\n" \ + " shrl $2,%0\n" \ + " andl $3,%3\n" \ + " .align 2,0x90\n" \ +- "0: rep; movsl\n" \ ++ "0: rep; "prefix"movsl\n" \ + " movl %3,%0\n" \ +- "1: rep; movsb\n" \ ++ "1: rep; "prefix"movsb\n" \ + "2:\n" \ ++ restore \ + ".section .fixup,\"ax\"\n" \ + "5: addl %3,%0\n" \ + " jmp 2b\n" \ +@@ -682,14 +799,14 @@ do { \ + " negl %0\n" \ + " andl $7,%0\n" \ + " subl %0,%3\n" \ +- "4: rep; movsb\n" \ ++ "4: rep; "__copyuser_seg"movsb\n" \ + " movl %3,%0\n" \ + " shrl $2,%0\n" \ + " andl $3,%3\n" \ + " .align 2,0x90\n" \ +- "0: rep; movsl\n" \ ++ "0: rep; "__copyuser_seg"movsl\n" \ + " movl %3,%0\n" \ +- "1: rep; movsb\n" \ ++ "1: rep; "__copyuser_seg"movsb\n" \ + "2:\n" \ + ".section .fixup,\"ax\"\n" \ + "5: addl %3,%0\n" \ +@@ -775,9 +892,9 @@ survive: + } + #endif + if (movsl_is_ok(to, from, n)) +- __copy_user(to, from, n); ++ __copy_user(to, from, n, "", __COPYUSER_SET_ES, __COPYUSER_RESTORE_ES); + else +- n = __copy_user_intel(to, from, n); ++ n = __generic_copy_to_user_intel(to, from, n); + return n; + } + EXPORT_SYMBOL(__copy_to_user_ll); +@@ -797,10 +914,9 @@ unsigned long __copy_from_user_ll_nozero(void *to, const void __user *from, + unsigned long n) + { + if (movsl_is_ok(to, from, n)) +- __copy_user(to, from, n); ++ __copy_user(to, from, n, __copyuser_seg, "", ""); + else +- n = __copy_user_intel((void __user *)to, +- (const void *)from, n); ++ n = __generic_copy_from_user_intel(to, from, n); + return n; + } + EXPORT_SYMBOL(__copy_from_user_ll_nozero); +@@ -827,59 +943,38 @@ unsigned long __copy_from_user_ll_nocache_nozero(void *to, const void __user *fr + if (n > 64 && cpu_has_xmm2) + n = __copy_user_intel_nocache(to, from, n); + else +- __copy_user(to, from, n); ++ __copy_user(to, from, n, __copyuser_seg, "", ""); + #else +- __copy_user(to, from, n); ++ __copy_user(to, from, n, __copyuser_seg, "", ""); + #endif + return n; + } + EXPORT_SYMBOL(__copy_from_user_ll_nocache_nozero); + +-/** +- * copy_to_user: - Copy a block of data into user space. +- * @to: Destination address, in user space. +- * @from: Source address, in kernel space. +- * @n: Number of bytes to copy. +- * +- * Context: User context only. This function may sleep. +- * +- * Copy data from kernel space to user space. +- * +- * Returns number of bytes that could not be copied. +- * On success, this will be zero. +- */ +-unsigned long +-copy_to_user(void __user *to, const void *from, unsigned long n) ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++void __set_fs(mm_segment_t x) + { +- if (access_ok(VERIFY_WRITE, to, n)) +- n = __copy_to_user(to, from, n); +- return n; ++ switch (x.seg) { ++ case 0: ++ loadsegment(gs, 0); ++ break; ++ case TASK_SIZE_MAX: ++ loadsegment(gs, __USER_DS); ++ break; ++ case -1UL: ++ loadsegment(gs, __KERNEL_DS); ++ break; ++ default: ++ BUG(); ++ } ++ return; + } +-EXPORT_SYMBOL(copy_to_user); ++EXPORT_SYMBOL(__set_fs); + +-/** +- * copy_from_user: - Copy a block of data from user space. +- * @to: Destination address, in kernel space. +- * @from: Source address, in user space. +- * @n: Number of bytes to copy. +- * +- * Context: User context only. This function may sleep. +- * +- * Copy data from user space to kernel space. +- * +- * Returns number of bytes that could not be copied. +- * On success, this will be zero. +- * +- * If some data could not be copied, this function will pad the copied +- * data to the requested size using zero bytes. +- */ +-unsigned long +-copy_from_user(void *to, const void __user *from, unsigned long n) ++void set_fs(mm_segment_t x) + { +- if (access_ok(VERIFY_READ, from, n)) +- n = __copy_from_user(to, from, n); +- else +- memset(to, 0, n); +- return n; ++ current_thread_info()->addr_limit = x; ++ __set_fs(x); + } +-EXPORT_SYMBOL(copy_from_user); ++EXPORT_SYMBOL(set_fs); ++#endif +diff --git a/arch/x86/lib/usercopy_64.c b/arch/x86/lib/usercopy_64.c +index b7c2849..8633ad8 100644 +--- a/arch/x86/lib/usercopy_64.c ++++ b/arch/x86/lib/usercopy_64.c +@@ -42,6 +42,12 @@ long + __strncpy_from_user(char *dst, const char __user *src, long count) + { + long res; ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)src < PAX_USER_SHADOW_BASE) ++ src += PAX_USER_SHADOW_BASE; ++#endif ++ + __do_strncpy_from_user(dst, src, count, res); + return res; + } +@@ -65,6 +71,12 @@ unsigned long __clear_user(void __user *addr, unsigned long size) + { + long __d0; + might_fault(); ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)addr < PAX_USER_SHADOW_BASE) ++ addr += PAX_USER_SHADOW_BASE; ++#endif ++ + /* no memory constraint because it doesn't change any memory gcc knows + about */ + asm volatile( +@@ -149,12 +161,20 @@ long strlen_user(const char __user *s) + } + EXPORT_SYMBOL(strlen_user); + +-unsigned long copy_in_user(void __user *to, const void __user *from, unsigned len) ++unsigned long copy_in_user(void __user *to, const void __user *from, unsigned long len) + { +- if (access_ok(VERIFY_WRITE, to, len) && access_ok(VERIFY_READ, from, len)) { +- return copy_user_generic((__force void *)to, (__force void *)from, len); +- } +- return len; ++ if (access_ok(VERIFY_WRITE, to, len) && access_ok(VERIFY_READ, from, len)) { ++ ++#ifdef CONFIG_PAX_MEMORY_UDEREF ++ if ((unsigned long)to < PAX_USER_SHADOW_BASE) ++ to += PAX_USER_SHADOW_BASE; ++ if ((unsigned long)from < PAX_USER_SHADOW_BASE) ++ from += PAX_USER_SHADOW_BASE; ++#endif ++ ++ return copy_user_generic((void __force_kernel *)to, (void __force_kernel *)from, len); ++ } ++ return len; + } + EXPORT_SYMBOL(copy_in_user); + +@@ -164,7 +184,7 @@ EXPORT_SYMBOL(copy_in_user); + * it is not necessary to optimize tail handling. + */ + unsigned long +-copy_user_handle_tail(char *to, char *from, unsigned len, unsigned zerorest) ++copy_user_handle_tail(char __user *to, char __user *from, unsigned long len, unsigned zerorest) + { + char c; + unsigned zero_len; +diff --git a/arch/x86/mm/extable.c b/arch/x86/mm/extable.c +index 61b41ca..5fef66a 100644 +--- a/arch/x86/mm/extable.c ++++ b/arch/x86/mm/extable.c +@@ -1,14 +1,71 @@ + #include + #include ++#include + #include ++#include + ++/* ++ * The exception table needs to be sorted so that the binary ++ * search that we use to find entries in it works properly. ++ * This is used both for the kernel exception table and for ++ * the exception tables of modules that get loaded. ++ */ ++static int cmp_ex(const void *a, const void *b) ++{ ++ const struct exception_table_entry *x = a, *y = b; ++ ++ /* avoid overflow */ ++ if (x->insn > y->insn) ++ return 1; ++ if (x->insn < y->insn) ++ return -1; ++ return 0; ++} ++ ++static void swap_ex(void *a, void *b, int size) ++{ ++ struct exception_table_entry t, *x = a, *y = b; ++ ++ t = *x; ++ ++ pax_open_kernel(); ++ *x = *y; ++ *y = t; ++ pax_close_kernel(); ++} ++ ++void sort_extable(struct exception_table_entry *start, ++ struct exception_table_entry *finish) ++{ ++ sort(start, finish - start, sizeof(struct exception_table_entry), ++ cmp_ex, swap_ex); ++} ++ ++#ifdef CONFIG_MODULES ++/* ++ * If the exception table is sorted, any referring to the module init ++ * will be at the beginning or the end. ++ */ ++void trim_init_extable(struct module *m) ++{ ++ /*trim the beginning*/ ++ while (m->num_exentries && within_module_init(m->extable[0].insn, m)) { ++ m->extable++; ++ m->num_exentries--; ++ } ++ /*trim the end*/ ++ while (m->num_exentries && ++ within_module_init(m->extable[m->num_exentries-1].insn, m)) ++ m->num_exentries--; ++} ++#endif /* CONFIG_MODULES */ + + int fixup_exception(struct pt_regs *regs) + { + const struct exception_table_entry *fixup; + + #ifdef CONFIG_PNPBIOS +- if (unlikely(SEGMENT_IS_PNP_CODE(regs->cs))) { ++ if (unlikely(!v8086_mode(regs) && SEGMENT_IS_PNP_CODE(regs->cs))) { + extern u32 pnp_bios_fault_eip, pnp_bios_fault_esp; + extern u32 pnp_bios_is_utter_crap; + pnp_bios_is_utter_crap = 1; +diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c +index 8ac0d76..ca501e2 100644 +--- a/arch/x86/mm/fault.c ++++ b/arch/x86/mm/fault.c +@@ -11,10 +11,19 @@ + #include /* __kprobes, ... */ + #include /* kmmio_handler, ... */ + #include /* perf_sw_event */ ++#include ++#include + + #include /* dotraplinkage, ... */ + #include /* pgd_*(), ... */ + #include /* kmemcheck_*(), ... */ ++#include ++#include ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++#include ++#include "../kernel/dumpstack.h" ++#endif + + /* + * Page fault error code bits: +@@ -51,7 +60,7 @@ static inline int notify_page_fault(struct pt_regs *regs) + int ret = 0; + + /* kprobe_running() needs smp_processor_id() */ +- if (kprobes_built_in() && !user_mode_vm(regs)) { ++ if (kprobes_built_in() && !user_mode(regs)) { + preempt_disable(); + if (kprobe_running() && kprobe_fault_handler(regs, 14)) + ret = 1; +@@ -112,7 +121,10 @@ check_prefetch_opcode(struct pt_regs *regs, unsigned char *instr, + return !instr_lo || (instr_lo>>1) == 1; + case 0x00: + /* Prefetch instruction is 0x0F0D or 0x0F18 */ +- if (probe_kernel_address(instr, opcode)) ++ if (user_mode(regs)) { ++ if (__copy_from_user_inatomic(&opcode, (unsigned char __force_user *)(instr), 1)) ++ return 0; ++ } else if (probe_kernel_address(instr, opcode)) + return 0; + + *prefetch = (instr_lo == 0xF) && +@@ -146,7 +158,10 @@ is_prefetch(struct pt_regs *regs, unsigned long error_code, unsigned long addr) + while (instr < max_instr) { + unsigned char opcode; + +- if (probe_kernel_address(instr, opcode)) ++ if (user_mode(regs)) { ++ if (__copy_from_user_inatomic(&opcode, (unsigned char __force_user *)(instr), 1)) ++ break; ++ } else if (probe_kernel_address(instr, opcode)) + break; + + instr++; +@@ -172,6 +187,34 @@ force_sig_info_fault(int si_signo, int si_code, unsigned long address, + force_sig_info(si_signo, &info, tsk); + } + ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++static bool pax_is_fetch_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address); ++#endif ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++static int pax_handle_fetch_fault(struct pt_regs *regs); ++#endif ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++static inline pmd_t * pax_get_pmd(struct mm_struct *mm, unsigned long address) ++{ ++ pgd_t *pgd; ++ pud_t *pud; ++ pmd_t *pmd; ++ ++ pgd = pgd_offset(mm, address); ++ if (!pgd_present(*pgd)) ++ return NULL; ++ pud = pud_offset(pgd, address); ++ if (!pud_present(*pud)) ++ return NULL; ++ pmd = pmd_offset(pud, address); ++ if (!pmd_present(*pmd)) ++ return NULL; ++ return pmd; ++} ++#endif ++ + DEFINE_SPINLOCK(pgd_lock); + LIST_HEAD(pgd_list); + +@@ -224,11 +267,24 @@ void vmalloc_sync_all(void) + address += PMD_SIZE) { + + unsigned long flags; ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ unsigned long cpu; ++#else + struct page *page; ++#endif + + spin_lock_irqsave(&pgd_lock, flags); ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ for (cpu = 0; cpu < nr_cpu_ids; ++cpu) { ++ pgd_t *pgd = get_cpu_pgd(cpu); ++#else + list_for_each_entry(page, &pgd_list, lru) { +- if (!vmalloc_sync_one(page_address(page), address)) ++ pgd_t *pgd = page_address(page); ++#endif ++ ++ if (!vmalloc_sync_one(pgd, address)) + break; + } + spin_unlock_irqrestore(&pgd_lock, flags); +@@ -258,6 +314,11 @@ static noinline int vmalloc_fault(unsigned long address) + * an interrupt in the middle of a task switch.. + */ + pgd_paddr = read_cr3(); ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ BUG_ON(__pa(get_cpu_pgd(smp_processor_id())) != (pgd_paddr & PHYSICAL_PAGE_MASK)); ++#endif ++ + pmd_k = vmalloc_sync_one(__va(pgd_paddr), address); + if (!pmd_k) + return -1; +@@ -332,15 +393,27 @@ void vmalloc_sync_all(void) + + const pgd_t *pgd_ref = pgd_offset_k(address); + unsigned long flags; ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ unsigned long cpu; ++#else + struct page *page; ++#endif + + if (pgd_none(*pgd_ref)) + continue; + + spin_lock_irqsave(&pgd_lock, flags); ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ for (cpu = 0; cpu < nr_cpu_ids; ++cpu) { ++ pgd_t *pgd = pgd_offset_cpu(cpu, address); ++#else + list_for_each_entry(page, &pgd_list, lru) { + pgd_t *pgd; + pgd = (pgd_t *)page_address(page) + pgd_index(address); ++#endif ++ + if (pgd_none(*pgd)) + set_pgd(pgd, *pgd_ref); + else +@@ -373,7 +446,14 @@ static noinline int vmalloc_fault(unsigned long address) + * happen within a race in page table update. In the later + * case just flush: + */ ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ BUG_ON(__pa(get_cpu_pgd(smp_processor_id())) != (read_cr3() & PHYSICAL_PAGE_MASK)); ++ pgd = pgd_offset_cpu(smp_processor_id(), address); ++#else + pgd = pgd_offset(current->active_mm, address); ++#endif ++ + pgd_ref = pgd_offset_k(address); + if (pgd_none(*pgd_ref)) + return -1; +@@ -535,7 +615,7 @@ static int is_errata93(struct pt_regs *regs, unsigned long address) + static int is_errata100(struct pt_regs *regs, unsigned long address) + { + #ifdef CONFIG_X86_64 +- if ((regs->cs == __USER32_CS || (regs->cs & (1<<2))) && (address >> 32)) ++ if ((regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT)) && (address >> 32)) + return 1; + #endif + return 0; +@@ -562,7 +642,7 @@ static int is_f00f_bug(struct pt_regs *regs, unsigned long address) + } + + static const char nx_warning[] = KERN_CRIT +-"kernel tried to execute NX-protected page - exploit attempt? (uid: %d)\n"; ++"kernel tried to execute NX-protected page - exploit attempt? (uid: %d, task: %s, pid: %d)\n"; + + static void + show_fault_oops(struct pt_regs *regs, unsigned long error_code, +@@ -571,15 +651,26 @@ show_fault_oops(struct pt_regs *regs, unsigned long error_code, + if (!oops_may_print()) + return; + +- if (error_code & PF_INSTR) { ++ if (nx_enabled && (error_code & PF_INSTR)) { + unsigned int level; + + pte_t *pte = lookup_address(address, &level); + + if (pte && pte_present(*pte) && !pte_exec(*pte)) +- printk(nx_warning, current_uid()); ++ printk(nx_warning, current_uid(), current->comm, task_pid_nr(current)); + } + ++#ifdef CONFIG_PAX_KERNEXEC ++ if (init_mm.start_code <= address && address < init_mm.end_code) { ++ if (current->signal->curr_ip) ++ printk(KERN_ERR "PAX: From %pI4: %s:%d, uid/euid: %u/%u, attempted to modify kernel code\n", ++ ¤t->signal->curr_ip, current->comm, task_pid_nr(current), current_uid(), current_euid()); ++ else ++ printk(KERN_ERR "PAX: %s:%d, uid/euid: %u/%u, attempted to modify kernel code\n", ++ current->comm, task_pid_nr(current), current_uid(), current_euid()); ++ } ++#endif ++ + printk(KERN_ALERT "BUG: unable to handle kernel "); + if (address < PAGE_SIZE) + printk(KERN_CONT "NULL pointer dereference"); +@@ -705,6 +796,23 @@ __bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code, + { + struct task_struct *tsk = current; + ++#ifdef CONFIG_X86_64 ++ struct mm_struct *mm = tsk->mm; ++ ++ if (mm && (error_code & PF_INSTR) && mm->context.vdso) { ++ if (regs->ip == (unsigned long)vgettimeofday) { ++ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, fallback_gettimeofday); ++ return; ++ } else if (regs->ip == (unsigned long)vtime) { ++ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, fallback_time); ++ return; ++ } else if (regs->ip == (unsigned long)vgetcpu) { ++ regs->ip = (unsigned long)VDSO64_SYMBOL(mm->context.vdso, getcpu); ++ return; ++ } ++ } ++#endif ++ + /* User mode accesses just cause a SIGSEGV */ + if (error_code & PF_USER) { + /* +@@ -722,6 +830,21 @@ __bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code, + if (is_errata100(regs, address)) + return; + ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++ if (pax_is_fetch_fault(regs, error_code, address)) { ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++ switch (pax_handle_fetch_fault(regs)) { ++ case 2: ++ return; ++ } ++#endif ++ ++ pax_report_fault(regs, (void *)regs->ip, (void *)regs->sp); ++ do_group_exit(SIGKILL); ++ } ++#endif ++ + if (unlikely(show_unhandled_signals)) + show_signal_msg(regs, error_code, address, tsk); + +@@ -818,7 +941,7 @@ do_sigbus(struct pt_regs *regs, unsigned long error_code, unsigned long address, + if (fault & VM_FAULT_HWPOISON) { + printk(KERN_ERR + "MCE: Killing %s:%d due to hardware memory corruption fault at %lx\n", +- tsk->comm, tsk->pid, address); ++ tsk->comm, task_pid_nr(tsk), address); + code = BUS_MCEERR_AR; + } + #endif +@@ -857,6 +980,99 @@ static int spurious_fault_check(unsigned long error_code, pte_t *pte) + return 1; + } + ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) ++static int pax_handle_pageexec_fault(struct pt_regs *regs, struct mm_struct *mm, unsigned long address, unsigned long error_code) ++{ ++ pte_t *pte; ++ pmd_t *pmd; ++ spinlock_t *ptl; ++ unsigned char pte_mask; ++ ++ if (nx_enabled || (error_code & (PF_PROT|PF_USER)) != (PF_PROT|PF_USER) || v8086_mode(regs) || ++ !(mm->pax_flags & MF_PAX_PAGEEXEC)) ++ return 0; ++ ++ /* PaX: it's our fault, let's handle it if we can */ ++ ++ /* PaX: take a look at read faults before acquiring any locks */ ++ if (unlikely(!(error_code & PF_WRITE) && (regs->ip == address))) { ++ /* instruction fetch attempt from a protected page in user mode */ ++ up_read(&mm->mmap_sem); ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++ switch (pax_handle_fetch_fault(regs)) { ++ case 2: ++ return 1; ++ } ++#endif ++ ++ pax_report_fault(regs, (void *)regs->ip, (void *)regs->sp); ++ do_group_exit(SIGKILL); ++ } ++ ++ pmd = pax_get_pmd(mm, address); ++ if (unlikely(!pmd)) ++ return 0; ++ ++ pte = pte_offset_map_lock(mm, pmd, address, &ptl); ++ if (unlikely(!(pte_val(*pte) & _PAGE_PRESENT) || pte_user(*pte))) { ++ pte_unmap_unlock(pte, ptl); ++ return 0; ++ } ++ ++ if (unlikely((error_code & PF_WRITE) && !pte_write(*pte))) { ++ /* write attempt to a protected page in user mode */ ++ pte_unmap_unlock(pte, ptl); ++ return 0; ++ } ++ ++#ifdef CONFIG_SMP ++ if (likely(address > get_limit(regs->cs) && cpu_isset(smp_processor_id(), mm->context.cpu_user_cs_mask))) ++#else ++ if (likely(address > get_limit(regs->cs))) ++#endif ++ { ++ set_pte(pte, pte_mkread(*pte)); ++ __flush_tlb_one(address); ++ pte_unmap_unlock(pte, ptl); ++ up_read(&mm->mmap_sem); ++ return 1; ++ } ++ ++ pte_mask = _PAGE_ACCESSED | _PAGE_USER | ((error_code & PF_WRITE) << (_PAGE_BIT_DIRTY-1)); ++ ++ /* ++ * PaX: fill DTLB with user rights and retry ++ */ ++ __asm__ __volatile__ ( ++ "orb %2,(%1)\n" ++#if defined(CONFIG_M586) || defined(CONFIG_M586TSC) ++/* ++ * PaX: let this uncommented 'invlpg' remind us on the behaviour of Intel's ++ * (and AMD's) TLBs. namely, they do not cache PTEs that would raise *any* ++ * page fault when examined during a TLB load attempt. this is true not only ++ * for PTEs holding a non-present entry but also present entries that will ++ * raise a page fault (such as those set up by PaX, or the copy-on-write ++ * mechanism). in effect it means that we do *not* need to flush the TLBs ++ * for our target pages since their PTEs are simply not in the TLBs at all. ++ ++ * the best thing in omitting it is that we gain around 15-20% speed in the ++ * fast path of the page fault handler and can get rid of tracing since we ++ * can no longer flush unintended entries. ++ */ ++ "invlpg (%0)\n" ++#endif ++ __copyuser_seg"testb $0,(%0)\n" ++ "xorb %3,(%1)\n" ++ : ++ : "r" (address), "r" (pte), "q" (pte_mask), "i" (_PAGE_USER) ++ : "memory", "cc"); ++ pte_unmap_unlock(pte, ptl); ++ up_read(&mm->mmap_sem); ++ return 1; ++} ++#endif ++ + /* + * Handle a spurious fault caused by a stale TLB entry. + * +@@ -923,6 +1139,9 @@ int show_unhandled_signals = 1; + static inline int + access_error(unsigned long error_code, int write, struct vm_area_struct *vma) + { ++ if (nx_enabled && (error_code & PF_INSTR) && !(vma->vm_flags & VM_EXEC)) ++ return 1; ++ + if (write) { + /* write, present and write, not present: */ + if (unlikely(!(vma->vm_flags & VM_WRITE))) +@@ -956,16 +1175,30 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) + { + struct vm_area_struct *vma; + struct task_struct *tsk; +- unsigned long address; + struct mm_struct *mm; + int write; + int fault; + +- tsk = current; +- mm = tsk->mm; +- + /* Get the faulting address: */ +- address = read_cr2(); ++ unsigned long address = read_cr2(); ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ if (!user_mode(regs) && address < 2 * PAX_USER_SHADOW_BASE) { ++ if (!search_exception_tables(regs->ip)) { ++ bad_area_nosemaphore(regs, error_code, address); ++ return; ++ } ++ if (address < PAX_USER_SHADOW_BASE) { ++ printk(KERN_ERR "PAX: please report this to pageexec@freemail.hu\n"); ++ printk(KERN_ERR "PAX: faulting IP: %pA\n", (void *)regs->ip); ++ show_trace_log_lvl(NULL, NULL, (void *)regs->sp, regs->bp, KERN_ERR); ++ } else ++ address -= PAX_USER_SHADOW_BASE; ++ } ++#endif ++ ++ tsk = current; ++ mm = tsk->mm; + + /* + * Detect and handle instructions that would cause a page fault for +@@ -1026,7 +1259,7 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) + * User-mode registers count as a user access even for any + * potential system fault or CPU buglet: + */ +- if (user_mode_vm(regs)) { ++ if (user_mode(regs)) { + local_irq_enable(); + error_code |= PF_USER; + } else { +@@ -1080,6 +1313,11 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) + might_sleep(); + } + ++#if defined(CONFIG_X86_32) && defined(CONFIG_PAX_PAGEEXEC) ++ if (pax_handle_pageexec_fault(regs, mm, address, error_code)) ++ return; ++#endif ++ + vma = find_vma(mm, address); + if (unlikely(!vma)) { + bad_area(regs, error_code, address); +@@ -1091,18 +1329,24 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) + bad_area(regs, error_code, address); + return; + } +- if (error_code & PF_USER) { +- /* +- * Accessing the stack below %sp is always a bug. +- * The large cushion allows instructions like enter +- * and pusha to work. ("enter $65535, $31" pushes +- * 32 pointers and then decrements %sp by 65535.) +- */ +- if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < regs->sp)) { +- bad_area(regs, error_code, address); +- return; +- } ++ /* ++ * Accessing the stack below %sp is always a bug. ++ * The large cushion allows instructions like enter ++ * and pusha to work. ("enter $65535, $31" pushes ++ * 32 pointers and then decrements %sp by 65535.) ++ */ ++ if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < task_pt_regs(tsk)->sp)) { ++ bad_area(regs, error_code, address); ++ return; + } ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (unlikely((mm->pax_flags & MF_PAX_SEGMEXEC) && vma->vm_end - SEGMEXEC_TASK_SIZE - 1 < address - SEGMEXEC_TASK_SIZE - 1)) { ++ bad_area(regs, error_code, address); ++ return; ++ } ++#endif ++ + if (unlikely(expand_stack(vma, address))) { + bad_area(regs, error_code, address); + return; +@@ -1146,3 +1390,292 @@ good_area: + + up_read(&mm->mmap_sem); + } ++ ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++static bool pax_is_fetch_fault(struct pt_regs *regs, unsigned long error_code, unsigned long address) ++{ ++ struct mm_struct *mm = current->mm; ++ unsigned long ip = regs->ip; ++ ++ if (v8086_mode(regs)) ++ ip = ((regs->cs & 0xffff) << 4) + (ip & 0xffff); ++ ++#ifdef CONFIG_PAX_PAGEEXEC ++ if (mm->pax_flags & MF_PAX_PAGEEXEC) { ++ if ((__supported_pte_mask & _PAGE_NX) && (error_code & PF_INSTR)) ++ return true; ++ if (!(error_code & (PF_PROT | PF_WRITE)) && ip == address) ++ return true; ++ return false; ++ } ++#endif ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) { ++ if (!(error_code & (PF_PROT | PF_WRITE)) && (ip + SEGMEXEC_TASK_SIZE == address)) ++ return true; ++ return false; ++ } ++#endif ++ ++ return false; ++} ++#endif ++ ++#ifdef CONFIG_PAX_EMUTRAMP ++static int pax_handle_fetch_fault_32(struct pt_regs *regs) ++{ ++ int err; ++ ++ do { /* PaX: libffi trampoline emulation */ ++ unsigned char mov, jmp; ++ unsigned int addr1, addr2; ++ ++#ifdef CONFIG_X86_64 ++ if ((regs->ip + 9) >> 32) ++ break; ++#endif ++ ++ err = get_user(mov, (unsigned char __user *)regs->ip); ++ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 1)); ++ err |= get_user(jmp, (unsigned char __user *)(regs->ip + 5)); ++ err |= get_user(addr2, (unsigned int __user *)(regs->ip + 6)); ++ ++ if (err) ++ break; ++ ++ if (mov == 0xB8 && jmp == 0xE9) { ++ regs->ax = addr1; ++ regs->ip = (unsigned int)(regs->ip + addr2 + 10); ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: gcc trampoline emulation #1 */ ++ unsigned char mov1, mov2; ++ unsigned short jmp; ++ unsigned int addr1, addr2; ++ ++#ifdef CONFIG_X86_64 ++ if ((regs->ip + 11) >> 32) ++ break; ++#endif ++ ++ err = get_user(mov1, (unsigned char __user *)regs->ip); ++ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 1)); ++ err |= get_user(mov2, (unsigned char __user *)(regs->ip + 5)); ++ err |= get_user(addr2, (unsigned int __user *)(regs->ip + 6)); ++ err |= get_user(jmp, (unsigned short __user *)(regs->ip + 10)); ++ ++ if (err) ++ break; ++ ++ if (mov1 == 0xB9 && mov2 == 0xB8 && jmp == 0xE0FF) { ++ regs->cx = addr1; ++ regs->ax = addr2; ++ regs->ip = addr2; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: gcc trampoline emulation #2 */ ++ unsigned char mov, jmp; ++ unsigned int addr1, addr2; ++ ++#ifdef CONFIG_X86_64 ++ if ((regs->ip + 9) >> 32) ++ break; ++#endif ++ ++ err = get_user(mov, (unsigned char __user *)regs->ip); ++ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 1)); ++ err |= get_user(jmp, (unsigned char __user *)(regs->ip + 5)); ++ err |= get_user(addr2, (unsigned int __user *)(regs->ip + 6)); ++ ++ if (err) ++ break; ++ ++ if (mov == 0xB9 && jmp == 0xE9) { ++ regs->cx = addr1; ++ regs->ip = (unsigned int)(regs->ip + addr2 + 10); ++ return 2; ++ } ++ } while (0); ++ ++ return 1; /* PaX in action */ ++} ++ ++#ifdef CONFIG_X86_64 ++static int pax_handle_fetch_fault_64(struct pt_regs *regs) ++{ ++ int err; ++ ++ do { /* PaX: libffi trampoline emulation */ ++ unsigned short mov1, mov2, jmp1; ++ unsigned char stcclc, jmp2; ++ unsigned long addr1, addr2; ++ ++ err = get_user(mov1, (unsigned short __user *)regs->ip); ++ err |= get_user(addr1, (unsigned long __user *)(regs->ip + 2)); ++ err |= get_user(mov2, (unsigned short __user *)(regs->ip + 10)); ++ err |= get_user(addr2, (unsigned long __user *)(regs->ip + 12)); ++ err |= get_user(stcclc, (unsigned char __user *)(regs->ip + 20)); ++ err |= get_user(jmp1, (unsigned short __user *)(regs->ip + 21)); ++ err |= get_user(jmp2, (unsigned char __user *)(regs->ip + 23)); ++ ++ if (err) ++ break; ++ ++ if (mov1 == 0xBB49 && mov2 == 0xBA49 && (stcclc == 0xF8 || stcclc == 0xF9) && jmp1 == 0xFF49 && jmp2 == 0xE3) { ++ regs->r11 = addr1; ++ regs->r10 = addr2; ++ if (stcclc == 0xF8) ++ regs->flags &= ~X86_EFLAGS_CF; ++ else ++ regs->flags |= X86_EFLAGS_CF; ++ regs->ip = addr1; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: gcc trampoline emulation #1 */ ++ unsigned short mov1, mov2, jmp1; ++ unsigned char jmp2; ++ unsigned int addr1; ++ unsigned long addr2; ++ ++ err = get_user(mov1, (unsigned short __user *)regs->ip); ++ err |= get_user(addr1, (unsigned int __user *)(regs->ip + 2)); ++ err |= get_user(mov2, (unsigned short __user *)(regs->ip + 6)); ++ err |= get_user(addr2, (unsigned long __user *)(regs->ip + 8)); ++ err |= get_user(jmp1, (unsigned short __user *)(regs->ip + 16)); ++ err |= get_user(jmp2, (unsigned char __user *)(regs->ip + 18)); ++ ++ if (err) ++ break; ++ ++ if (mov1 == 0xBB41 && mov2 == 0xBA49 && jmp1 == 0xFF49 && jmp2 == 0xE3) { ++ regs->r11 = addr1; ++ regs->r10 = addr2; ++ regs->ip = addr1; ++ return 2; ++ } ++ } while (0); ++ ++ do { /* PaX: gcc trampoline emulation #2 */ ++ unsigned short mov1, mov2, jmp1; ++ unsigned char jmp2; ++ unsigned long addr1, addr2; ++ ++ err = get_user(mov1, (unsigned short __user *)regs->ip); ++ err |= get_user(addr1, (unsigned long __user *)(regs->ip + 2)); ++ err |= get_user(mov2, (unsigned short __user *)(regs->ip + 10)); ++ err |= get_user(addr2, (unsigned long __user *)(regs->ip + 12)); ++ err |= get_user(jmp1, (unsigned short __user *)(regs->ip + 20)); ++ err |= get_user(jmp2, (unsigned char __user *)(regs->ip + 22)); ++ ++ if (err) ++ break; ++ ++ if (mov1 == 0xBB49 && mov2 == 0xBA49 && jmp1 == 0xFF49 && jmp2 == 0xE3) { ++ regs->r11 = addr1; ++ regs->r10 = addr2; ++ regs->ip = addr1; ++ return 2; ++ } ++ } while (0); ++ ++ return 1; /* PaX in action */ ++} ++#endif ++ ++/* ++ * PaX: decide what to do with offenders (regs->ip = fault address) ++ * ++ * returns 1 when task should be killed ++ * 2 when gcc trampoline was detected ++ */ ++static int pax_handle_fetch_fault(struct pt_regs *regs) ++{ ++ if (v8086_mode(regs)) ++ return 1; ++ ++ if (!(current->mm->pax_flags & MF_PAX_EMUTRAMP)) ++ return 1; ++ ++#ifdef CONFIG_X86_32 ++ return pax_handle_fetch_fault_32(regs); ++#else ++ if (regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT)) ++ return pax_handle_fetch_fault_32(regs); ++ else ++ return pax_handle_fetch_fault_64(regs); ++#endif ++} ++#endif ++ ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++void pax_report_insns(struct pt_regs *regs, void *pc, void *sp) ++{ ++ long i; ++ ++ printk(KERN_ERR "PAX: bytes at PC: "); ++ for (i = 0; i < 20; i++) { ++ unsigned char c; ++ if (get_user(c, (unsigned char __force_user *)pc+i)) ++ printk(KERN_CONT "?? "); ++ else ++ printk(KERN_CONT "%02x ", c); ++ } ++ printk("\n"); ++ ++ printk(KERN_ERR "PAX: bytes at SP-%lu: ", (unsigned long)sizeof(long)); ++ for (i = -1; i < 80 / (long)sizeof(long); i++) { ++ unsigned long c; ++ if (get_user(c, (unsigned long __force_user *)sp+i)) { ++#ifdef CONFIG_X86_32 ++ printk(KERN_CONT "???????? "); ++#else ++ if ((regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT))) ++ printk(KERN_CONT "???????? ???????? "); ++ else ++ printk(KERN_CONT "???????????????? "); ++#endif ++ } else { ++#ifdef CONFIG_X86_64 ++ if ((regs->cs == __USER32_CS || (regs->cs & SEGMENT_LDT))) { ++ printk(KERN_CONT "%08x ", (unsigned int)c); ++ printk(KERN_CONT "%08x ", (unsigned int)(c >> 32)); ++ } else ++#endif ++ printk(KERN_CONT "%0*lx ", 2 * (int)sizeof(long), c); ++ } ++ } ++ printk("\n"); ++} ++#endif ++ ++/** ++ * probe_kernel_write(): safely attempt to write to a location ++ * @dst: address to write to ++ * @src: pointer to the data that shall be written ++ * @size: size of the data chunk ++ * ++ * Safely write to address @dst from the buffer at @src. If a kernel fault ++ * happens, handle that and return -EFAULT. ++ */ ++long notrace probe_kernel_write(void *dst, const void *src, size_t size) ++{ ++ long ret; ++ mm_segment_t old_fs = get_fs(); ++ ++ set_fs(KERNEL_DS); ++ pagefault_disable(); ++ pax_open_kernel(); ++ ret = __copy_to_user_inatomic((void __force_user *)dst, src, size); ++ pax_close_kernel(); ++ pagefault_enable(); ++ set_fs(old_fs); ++ ++ return ret ? -EFAULT : 0; ++} +diff --git a/arch/x86/mm/gup.c b/arch/x86/mm/gup.c +index 71da1bc..7a16bf4 100644 +--- a/arch/x86/mm/gup.c ++++ b/arch/x86/mm/gup.c +@@ -237,7 +237,7 @@ int __get_user_pages_fast(unsigned long start, int nr_pages, int write, + addr = start; + len = (unsigned long) nr_pages << PAGE_SHIFT; + end = start + len; +- if (unlikely(!access_ok(write ? VERIFY_WRITE : VERIFY_READ, ++ if (unlikely(!__access_ok(write ? VERIFY_WRITE : VERIFY_READ, + (void __user *)start, len))) + return 0; + +diff --git a/arch/x86/mm/highmem_32.c b/arch/x86/mm/highmem_32.c +index 63a6ba6..79abd7a 100644 +--- a/arch/x86/mm/highmem_32.c ++++ b/arch/x86/mm/highmem_32.c +@@ -43,7 +43,10 @@ void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot) + idx = type + KM_TYPE_NR*smp_processor_id(); + vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); + BUG_ON(!pte_none(*(kmap_pte-idx))); ++ ++ pax_open_kernel(); + set_pte(kmap_pte-idx, mk_pte(page, prot)); ++ pax_close_kernel(); + + return (void *)vaddr; + } +diff --git a/arch/x86/mm/hugetlbpage.c b/arch/x86/mm/hugetlbpage.c +index f46c3407..6ff9a26 100644 +--- a/arch/x86/mm/hugetlbpage.c ++++ b/arch/x86/mm/hugetlbpage.c +@@ -267,13 +267,20 @@ static unsigned long hugetlb_get_unmapped_area_bottomup(struct file *file, + struct hstate *h = hstate_file(file); + struct mm_struct *mm = current->mm; + struct vm_area_struct *vma; +- unsigned long start_addr; ++ unsigned long start_addr, pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ pax_task_size -= PAGE_SIZE; + + if (len > mm->cached_hole_size) { +- start_addr = mm->free_area_cache; ++ start_addr = mm->free_area_cache; + } else { +- start_addr = TASK_UNMAPPED_BASE; +- mm->cached_hole_size = 0; ++ start_addr = mm->mmap_base; ++ mm->cached_hole_size = 0; + } + + full_search: +@@ -281,26 +288,27 @@ full_search: + + for (vma = find_vma(mm, addr); ; vma = vma->vm_next) { + /* At this point: (!vma || addr < vma->vm_end). */ +- if (TASK_SIZE - len < addr) { ++ if (pax_task_size - len < addr) { + /* + * Start a new search - just in case we missed + * some holes. + */ +- if (start_addr != TASK_UNMAPPED_BASE) { +- start_addr = TASK_UNMAPPED_BASE; ++ if (start_addr != mm->mmap_base) { ++ start_addr = mm->mmap_base; + mm->cached_hole_size = 0; + goto full_search; + } + return -ENOMEM; + } +- if (!vma || addr + len <= vma->vm_start) { +- mm->free_area_cache = addr + len; +- return addr; +- } ++ if (check_heap_stack_gap(vma, addr, len)) ++ break; + if (addr + mm->cached_hole_size < vma->vm_start) + mm->cached_hole_size = vma->vm_start - addr; + addr = ALIGN(vma->vm_end, huge_page_size(h)); + } ++ ++ mm->free_area_cache = addr + len; ++ return addr; + } + + static unsigned long hugetlb_get_unmapped_area_topdown(struct file *file, +@@ -309,10 +317,9 @@ static unsigned long hugetlb_get_unmapped_area_topdown(struct file *file, + { + struct hstate *h = hstate_file(file); + struct mm_struct *mm = current->mm; +- struct vm_area_struct *vma, *prev_vma; +- unsigned long base = mm->mmap_base, addr = addr0; ++ struct vm_area_struct *vma; ++ unsigned long base = mm->mmap_base, addr; + unsigned long largest_hole = mm->cached_hole_size; +- int first_time = 1; + + /* don't allow allocations above current base */ + if (mm->free_area_cache > base) +@@ -322,64 +329,63 @@ static unsigned long hugetlb_get_unmapped_area_topdown(struct file *file, + largest_hole = 0; + mm->free_area_cache = base; + } +-try_again: ++ + /* make sure it can fit in the remaining address space */ + if (mm->free_area_cache < len) + goto fail; + + /* either no address requested or cant fit in requested address hole */ +- addr = (mm->free_area_cache - len) & huge_page_mask(h); ++ addr = (mm->free_area_cache - len); + do { ++ addr &= huge_page_mask(h); ++ vma = find_vma(mm, addr); + /* + * Lookup failure means no vma is above this address, + * i.e. return with success: +- */ +- if (!(vma = find_vma_prev(mm, addr, &prev_vma))) +- return addr; +- +- /* + * new region fits between prev_vma->vm_end and + * vma->vm_start, use it: + */ +- if (addr + len <= vma->vm_start && +- (!prev_vma || (addr >= prev_vma->vm_end))) { ++ if (check_heap_stack_gap(vma, addr, len)) { + /* remember the address as a hint for next time */ +- mm->cached_hole_size = largest_hole; +- return (mm->free_area_cache = addr); +- } else { +- /* pull free_area_cache down to the first hole */ +- if (mm->free_area_cache == vma->vm_end) { +- mm->free_area_cache = vma->vm_start; +- mm->cached_hole_size = largest_hole; +- } ++ mm->cached_hole_size = largest_hole; ++ return (mm->free_area_cache = addr); ++ } ++ /* pull free_area_cache down to the first hole */ ++ if (mm->free_area_cache == vma->vm_end) { ++ mm->free_area_cache = vma->vm_start; ++ mm->cached_hole_size = largest_hole; + } + + /* remember the largest hole we saw so far */ + if (addr + largest_hole < vma->vm_start) +- largest_hole = vma->vm_start - addr; ++ largest_hole = vma->vm_start - addr; + + /* try just below the current vma->vm_start */ +- addr = (vma->vm_start - len) & huge_page_mask(h); +- } while (len <= vma->vm_start); ++ addr = skip_heap_stack_gap(vma, len); ++ } while (!IS_ERR_VALUE(addr)); + + fail: + /* +- * if hint left us with no space for the requested +- * mapping then try again: +- */ +- if (first_time) { +- mm->free_area_cache = base; +- largest_hole = 0; +- first_time = 0; +- goto try_again; +- } +- /* + * A failed mmap() very likely causes application failure, + * so fall back to the bottom-up function here. This scenario + * can happen with large stack limits and large mmap() + * allocations. + */ +- mm->free_area_cache = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ mm->mmap_base = SEGMEXEC_TASK_UNMAPPED_BASE; ++ else ++#endif ++ ++ mm->mmap_base = TASK_UNMAPPED_BASE; ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ ++ mm->free_area_cache = mm->mmap_base; + mm->cached_hole_size = ~0UL; + addr = hugetlb_get_unmapped_area_bottomup(file, addr0, + len, pgoff, flags); +@@ -387,6 +393,7 @@ fail: + /* + * Restore the topdown base: + */ ++ mm->mmap_base = base; + mm->free_area_cache = base; + mm->cached_hole_size = ~0UL; + +@@ -400,10 +407,19 @@ hugetlb_get_unmapped_area(struct file *file, unsigned long addr, + struct hstate *h = hstate_file(file); + struct mm_struct *mm = current->mm; + struct vm_area_struct *vma; ++ unsigned long pax_task_size = TASK_SIZE; + + if (len & ~huge_page_mask(h)) + return -EINVAL; +- if (len > TASK_SIZE) ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif ++ ++ pax_task_size -= PAGE_SIZE; ++ ++ if (len > pax_task_size) + return -ENOMEM; + + if (flags & MAP_FIXED) { +@@ -415,8 +431,7 @@ hugetlb_get_unmapped_area(struct file *file, unsigned long addr, + if (addr) { + addr = ALIGN(addr, huge_page_size(h)); + vma = find_vma(mm, addr); +- if (TASK_SIZE - len >= addr && +- (!vma || addr + len <= vma->vm_start)) ++ if (pax_task_size - len >= addr && check_heap_stack_gap(vma, addr, len)) + return addr; + } + if (mm->get_unmapped_area == arch_get_unmapped_area) +diff --git a/arch/x86/mm/init.c b/arch/x86/mm/init.c +index 73ffd55..f61c2a7 100644 +--- a/arch/x86/mm/init.c ++++ b/arch/x86/mm/init.c +@@ -69,11 +69,7 @@ static void __init find_early_table_space(unsigned long end, int use_pse, + * cause a hotspot and fill up ZONE_DMA. The page tables + * need roughly 0.5KB per GB. + */ +-#ifdef CONFIG_X86_32 +- start = 0x7000; +-#else +- start = 0x8000; +-#endif ++ start = 0x100000; + e820_table_start = find_e820_area(start, max_pfn_mapped<> PAGE_SHIFT) == pagenr) ++ return 1; ++ /* allow ISA/video mem */ ++ if ((ISA_START_ADDRESS >> PAGE_SHIFT) <= pagenr && pagenr < (ISA_END_ADDRESS >> PAGE_SHIFT)) ++ return 1; ++ /* throw out everything else below 1MB */ ++ if (pagenr <= 256) ++ return 0; ++#else + if (pagenr <= 256) + return 1; ++#endif ++ + if (iomem_is_exclusive(pagenr << PAGE_SHIFT)) + return 0; + if (!page_is_ram(pagenr)) +@@ -379,6 +392,86 @@ void free_init_pages(char *what, unsigned long begin, unsigned long end) + + void free_initmem(void) + { ++ ++#ifdef CONFIG_PAX_KERNEXEC ++#ifdef CONFIG_X86_32 ++ /* PaX: limit KERNEL_CS to actual size */ ++ unsigned long addr, limit; ++ struct desc_struct d; ++ int cpu; ++ ++ limit = paravirt_enabled() ? ktva_ktla(0xffffffff) : (unsigned long)&_etext; ++ limit = (limit - 1UL) >> PAGE_SHIFT; ++ ++ memset(__LOAD_PHYSICAL_ADDR + PAGE_OFFSET, POISON_FREE_INITMEM, PAGE_SIZE); ++ for (cpu = 0; cpu < nr_cpu_ids; cpu++) { ++ pack_descriptor(&d, get_desc_base(&get_cpu_gdt_table(cpu)[GDT_ENTRY_KERNEL_CS]), limit, 0x9B, 0xC); ++ write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_KERNEL_CS, &d, DESCTYPE_S); ++ } ++ ++ /* PaX: make KERNEL_CS read-only */ ++ addr = PFN_ALIGN(ktla_ktva((unsigned long)&_text)); ++ if (!paravirt_enabled()) ++ set_memory_ro(addr, (PFN_ALIGN(_sdata) - addr) >> PAGE_SHIFT); ++/* ++ for (addr = ktla_ktva((unsigned long)&_text); addr < (unsigned long)&_sdata; addr += PMD_SIZE) { ++ pgd = pgd_offset_k(addr); ++ pud = pud_offset(pgd, addr); ++ pmd = pmd_offset(pud, addr); ++ set_pmd(pmd, __pmd(pmd_val(*pmd) & ~_PAGE_RW)); ++ } ++*/ ++#ifdef CONFIG_X86_PAE ++ set_memory_nx(PFN_ALIGN(__init_begin), (PFN_ALIGN(__init_end) - PFN_ALIGN(__init_begin)) >> PAGE_SHIFT); ++/* ++ for (addr = (unsigned long)&__init_begin; addr < (unsigned long)&__init_end; addr += PMD_SIZE) { ++ pgd = pgd_offset_k(addr); ++ pud = pud_offset(pgd, addr); ++ pmd = pmd_offset(pud, addr); ++ set_pmd(pmd, __pmd(pmd_val(*pmd) | (_PAGE_NX & __supported_pte_mask))); ++ } ++*/ ++#endif ++ ++#ifdef CONFIG_MODULES ++ set_memory_4k((unsigned long)MODULES_EXEC_VADDR, (MODULES_EXEC_END - MODULES_EXEC_VADDR) >> PAGE_SHIFT); ++#endif ++ ++#else ++ pgd_t *pgd; ++ pud_t *pud; ++ pmd_t *pmd; ++ unsigned long addr, end; ++ ++ /* PaX: make kernel code/rodata read-only, rest non-executable */ ++ for (addr = __START_KERNEL_map; addr < __START_KERNEL_map + KERNEL_IMAGE_SIZE; addr += PMD_SIZE) { ++ pgd = pgd_offset_k(addr); ++ pud = pud_offset(pgd, addr); ++ pmd = pmd_offset(pud, addr); ++ if (!pmd_present(*pmd)) ++ continue; ++ if ((unsigned long)_text <= addr && addr < (unsigned long)_sdata) ++ set_pmd(pmd, __pmd(pmd_val(*pmd) & ~_PAGE_RW)); ++ else ++ set_pmd(pmd, __pmd(pmd_val(*pmd) | (_PAGE_NX & __supported_pte_mask))); ++ } ++ ++ addr = (unsigned long)__va(__pa(__START_KERNEL_map)); ++ end = addr + KERNEL_IMAGE_SIZE; ++ for (; addr < end; addr += PMD_SIZE) { ++ pgd = pgd_offset_k(addr); ++ pud = pud_offset(pgd, addr); ++ pmd = pmd_offset(pud, addr); ++ if (!pmd_present(*pmd)) ++ continue; ++ if ((unsigned long)__va(__pa(_text)) <= addr && addr < (unsigned long)__va(__pa(_sdata))) ++ set_pmd(pmd, __pmd(pmd_val(*pmd) & ~_PAGE_RW)); ++ } ++#endif ++ ++ flush_tlb_all(); ++#endif ++ + free_init_pages("unused kernel memory", + (unsigned long)(&__init_begin), + (unsigned long)(&__init_end)); +diff --git a/arch/x86/mm/init_32.c b/arch/x86/mm/init_32.c +index 30938c1..bda3d5d 100644 +--- a/arch/x86/mm/init_32.c ++++ b/arch/x86/mm/init_32.c +@@ -72,36 +72,6 @@ static __init void *alloc_low_page(void) + } + + /* +- * Creates a middle page table and puts a pointer to it in the +- * given global directory entry. This only returns the gd entry +- * in non-PAE compilation mode, since the middle layer is folded. +- */ +-static pmd_t * __init one_md_table_init(pgd_t *pgd) +-{ +- pud_t *pud; +- pmd_t *pmd_table; +- +-#ifdef CONFIG_X86_PAE +- if (!(pgd_val(*pgd) & _PAGE_PRESENT)) { +- if (after_bootmem) +- pmd_table = (pmd_t *)alloc_bootmem_pages(PAGE_SIZE); +- else +- pmd_table = (pmd_t *)alloc_low_page(); +- paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT); +- set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); +- pud = pud_offset(pgd, 0); +- BUG_ON(pmd_table != pmd_offset(pud, 0)); +- +- return pmd_table; +- } +-#endif +- pud = pud_offset(pgd, 0); +- pmd_table = pmd_offset(pud, 0); +- +- return pmd_table; +-} +- +-/* + * Create a page table and place a pointer to it in a middle page + * directory entry: + */ +@@ -121,13 +91,28 @@ static pte_t * __init one_page_table_init(pmd_t *pmd) + page_table = (pte_t *)alloc_low_page(); + + paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); ++#if defined(CONFIG_PAX_PAGEEXEC) || defined(CONFIG_PAX_SEGMEXEC) ++ set_pmd(pmd, __pmd(__pa(page_table) | _KERNPG_TABLE)); ++#else + set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); ++#endif + BUG_ON(page_table != pte_offset_kernel(pmd, 0)); + } + + return pte_offset_kernel(pmd, 0); + } + ++static pmd_t * __init one_md_table_init(pgd_t *pgd) ++{ ++ pud_t *pud; ++ pmd_t *pmd_table; ++ ++ pud = pud_offset(pgd, 0); ++ pmd_table = pmd_offset(pud, 0); ++ ++ return pmd_table; ++} ++ + pmd_t * __init populate_extra_pmd(unsigned long vaddr) + { + int pgd_idx = pgd_index(vaddr); +@@ -201,6 +186,7 @@ page_table_range_init(unsigned long start, unsigned long end, pgd_t *pgd_base) + int pgd_idx, pmd_idx; + unsigned long vaddr; + pgd_t *pgd; ++ pud_t *pud; + pmd_t *pmd; + pte_t *pte = NULL; + +@@ -210,8 +196,13 @@ page_table_range_init(unsigned long start, unsigned long end, pgd_t *pgd_base) + pgd = pgd_base + pgd_idx; + + for ( ; (pgd_idx < PTRS_PER_PGD) && (vaddr != end); pgd++, pgd_idx++) { +- pmd = one_md_table_init(pgd); +- pmd = pmd + pmd_index(vaddr); ++ pud = pud_offset(pgd, vaddr); ++ pmd = pmd_offset(pud, vaddr); ++ ++#ifdef CONFIG_X86_PAE ++ paravirt_alloc_pmd(&init_mm, __pa(pmd) >> PAGE_SHIFT); ++#endif ++ + for (; (pmd_idx < PTRS_PER_PMD) && (vaddr != end); + pmd++, pmd_idx++) { + pte = page_table_kmap_check(one_page_table_init(pmd), +@@ -223,11 +214,20 @@ page_table_range_init(unsigned long start, unsigned long end, pgd_t *pgd_base) + } + } + +-static inline int is_kernel_text(unsigned long addr) ++static inline int is_kernel_text(unsigned long start, unsigned long end) + { +- if (addr >= PAGE_OFFSET && addr <= (unsigned long)__init_end) +- return 1; +- return 0; ++ if ((start > ktla_ktva((unsigned long)_etext) || ++ end <= ktla_ktva((unsigned long)_stext)) && ++ (start > ktla_ktva((unsigned long)_einittext) || ++ end <= ktla_ktva((unsigned long)_sinittext)) && ++ ++#ifdef CONFIG_ACPI_SLEEP ++ (start > (unsigned long)__va(acpi_wakeup_address) + 0x4000 || end <= (unsigned long)__va(acpi_wakeup_address)) && ++#endif ++ ++ (start > (unsigned long)__va(0xfffff) || end <= (unsigned long)__va(0xc0000))) ++ return 0; ++ return 1; + } + + /* +@@ -243,9 +243,10 @@ kernel_physical_mapping_init(unsigned long start, + int use_pse = page_size_mask == (1<> PAGE_SHIFT); ++#endif + + if (pfn >= end_pfn) + continue; +@@ -291,14 +297,13 @@ repeat: + #endif + for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; + pmd++, pmd_idx++) { +- unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; ++ unsigned long address = pfn * PAGE_SIZE + PAGE_OFFSET; + + /* + * Map with big pages if possible, otherwise + * create normal page tables: + */ + if (use_pse) { +- unsigned int addr2; + pgprot_t prot = PAGE_KERNEL_LARGE; + /* + * first pass will use the same initial +@@ -308,11 +313,7 @@ repeat: + __pgprot(PTE_IDENT_ATTR | + _PAGE_PSE); + +- addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + +- PAGE_OFFSET + PAGE_SIZE-1; +- +- if (is_kernel_text(addr) || +- is_kernel_text(addr2)) ++ if (is_kernel_text(address, address + PMD_SIZE)) + prot = PAGE_KERNEL_LARGE_EXEC; + + pages_2m++; +@@ -329,7 +330,7 @@ repeat: + pte_ofs = pte_index((pfn<> 10, + +- (unsigned long)&_etext, (unsigned long)&_edata, +- ((unsigned long)&_edata - (unsigned long)&_etext) >> 10, ++ (unsigned long)&_sdata, (unsigned long)&_edata, ++ ((unsigned long)&_edata - (unsigned long)&_sdata) >> 10, + +- (unsigned long)&_text, (unsigned long)&_etext, ++ ktla_ktva((unsigned long)&_text), ktla_ktva((unsigned long)&_etext), + ((unsigned long)&_etext - (unsigned long)&_text) >> 10); + + /* +@@ -1007,6 +1012,7 @@ void set_kernel_text_rw(void) + if (!kernel_set_to_readonly) + return; + ++ start = ktla_ktva(start); + pr_debug("Set kernel text: %lx - %lx for read write\n", + start, start+size); + +@@ -1021,6 +1027,7 @@ void set_kernel_text_ro(void) + if (!kernel_set_to_readonly) + return; + ++ start = ktla_ktva(start); + pr_debug("Set kernel text: %lx - %lx for read only\n", + start, start+size); + +@@ -1032,6 +1039,7 @@ void mark_rodata_ro(void) + unsigned long start = PFN_ALIGN(_text); + unsigned long size = PFN_ALIGN(_etext) - start; + ++ start = ktla_ktva(start); + set_pages_ro(virt_to_page(start), size >> PAGE_SHIFT); + printk(KERN_INFO "Write protecting the kernel text: %luk\n", + size >> 10); +diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c +index 7d095ad..25d2549 100644 +--- a/arch/x86/mm/init_64.c ++++ b/arch/x86/mm/init_64.c +@@ -164,7 +164,9 @@ void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) + pmd = fill_pmd(pud, vaddr); + pte = fill_pte(pmd, vaddr); + ++ pax_open_kernel(); + set_pte(pte, new_pte); ++ pax_close_kernel(); + + /* + * It's enough to flush this one mapping. +@@ -223,14 +225,12 @@ static void __init __init_extra_mapping(unsigned long phys, unsigned long size, + pgd = pgd_offset_k((unsigned long)__va(phys)); + if (pgd_none(*pgd)) { + pud = (pud_t *) spp_getpage(); +- set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE | +- _PAGE_USER)); ++ set_pgd(pgd, __pgd(__pa(pud) | _PAGE_TABLE)); + } + pud = pud_offset(pgd, (unsigned long)__va(phys)); + if (pud_none(*pud)) { + pmd = (pmd_t *) spp_getpage(); +- set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE | +- _PAGE_USER)); ++ set_pud(pud, __pud(__pa(pmd) | _PAGE_TABLE)); + } + pmd = pmd_offset(pud, phys); + BUG_ON(!pmd_none(*pmd)); +@@ -675,6 +675,12 @@ void __init mem_init(void) + + pci_iommu_alloc(); + ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ clone_pgd_range(get_cpu_pgd(0) + KERNEL_PGD_BOUNDARY, ++ swapper_pg_dir + KERNEL_PGD_BOUNDARY, ++ KERNEL_PGD_PTRS); ++#endif ++ + /* clear_bss() already clear the empty_zero_page */ + + reservedpages = 0; +@@ -861,8 +867,8 @@ int kern_addr_valid(unsigned long addr) + static struct vm_area_struct gate_vma = { + .vm_start = VSYSCALL_START, + .vm_end = VSYSCALL_START + (VSYSCALL_MAPPED_PAGES * PAGE_SIZE), +- .vm_page_prot = PAGE_READONLY_EXEC, +- .vm_flags = VM_READ | VM_EXEC ++ .vm_page_prot = PAGE_READONLY, ++ .vm_flags = VM_READ + }; + + struct vm_area_struct *get_gate_vma(struct task_struct *tsk) +@@ -896,7 +902,7 @@ int in_gate_area_no_task(unsigned long addr) + + const char *arch_vma_name(struct vm_area_struct *vma) + { +- if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) ++ if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso) + return "[vdso]"; + if (vma == &gate_vma) + return "[vsyscall]"; +diff --git a/arch/x86/mm/iomap_32.c b/arch/x86/mm/iomap_32.c +index 84e236c..69bd3f6 100644 +--- a/arch/x86/mm/iomap_32.c ++++ b/arch/x86/mm/iomap_32.c +@@ -65,7 +65,11 @@ void *kmap_atomic_prot_pfn(unsigned long pfn, enum km_type type, pgprot_t prot) + debug_kmap_atomic(type); + idx = type + KM_TYPE_NR * smp_processor_id(); + vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); ++ ++ pax_open_kernel(); + set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); ++ pax_close_kernel(); ++ + arch_flush_lazy_mmu_mode(); + + return (void *)vaddr; +diff --git a/arch/x86/mm/ioremap.c b/arch/x86/mm/ioremap.c +index 2feb9bd..ab91e7b 100644 +--- a/arch/x86/mm/ioremap.c ++++ b/arch/x86/mm/ioremap.c +@@ -41,8 +41,8 @@ int page_is_ram(unsigned long pagenr) + * Second special case: Some BIOSen report the PC BIOS + * area (640->1Mb) as ram even though it is not. + */ +- if (pagenr >= (BIOS_BEGIN >> PAGE_SHIFT) && +- pagenr < (BIOS_END >> PAGE_SHIFT)) ++ if (pagenr >= (ISA_START_ADDRESS >> PAGE_SHIFT) && ++ pagenr < (ISA_END_ADDRESS >> PAGE_SHIFT)) + return 0; + + for (i = 0; i < e820.nr_map; i++) { +@@ -137,13 +137,10 @@ static void __iomem *__ioremap_caller(resource_size_t phys_addr, + /* + * Don't allow anybody to remap normal RAM that we're using.. + */ +- for (pfn = phys_addr >> PAGE_SHIFT; +- (pfn << PAGE_SHIFT) < (last_addr & PAGE_MASK); +- pfn++) { +- ++ for (pfn = phys_addr >> PAGE_SHIFT; ((resource_size_t)pfn << PAGE_SHIFT) < (last_addr & PAGE_MASK); pfn++) { + int is_ram = page_is_ram(pfn); + +- if (is_ram && pfn_valid(pfn) && !PageReserved(pfn_to_page(pfn))) ++ if (is_ram && pfn_valid(pfn) && (pfn >= 0x100 || !PageReserved(pfn_to_page(pfn)))) + return NULL; + WARN_ON_ONCE(is_ram); + } +@@ -378,6 +375,9 @@ void *xlate_dev_mem_ptr(unsigned long phys) + + /* If page is RAM, we can use __va. Otherwise ioremap and unmap. */ + if (page_is_ram(start >> PAGE_SHIFT)) ++#ifdef CONFIG_HIGHMEM ++ if ((start >> PAGE_SHIFT) < max_low_pfn) ++#endif + return __va(phys); + + addr = (void __force *)ioremap_default(start, PAGE_SIZE); +@@ -407,7 +407,7 @@ static int __init early_ioremap_debug_setup(char *str) + early_param("early_ioremap_debug", early_ioremap_debug_setup); + + static __initdata int after_paging_init; +-static pte_t bm_pte[PAGE_SIZE/sizeof(pte_t)] __page_aligned_bss; ++static pte_t bm_pte[PAGE_SIZE/sizeof(pte_t)] __read_only __aligned(PAGE_SIZE); + + static inline pmd_t * __init early_ioremap_pmd(unsigned long addr) + { +@@ -439,8 +439,7 @@ void __init early_ioremap_init(void) + slot_virt[i] = __fix_to_virt(FIX_BTMAP_BEGIN - NR_FIX_BTMAPS*i); + + pmd = early_ioremap_pmd(fix_to_virt(FIX_BTMAP_BEGIN)); +- memset(bm_pte, 0, sizeof(bm_pte)); +- pmd_populate_kernel(&init_mm, pmd, bm_pte); ++ pmd_populate_user(&init_mm, pmd, bm_pte); + + /* + * The boot-ioremap range spans multiple pmds, for which +diff --git a/arch/x86/mm/kmemcheck/kmemcheck.c b/arch/x86/mm/kmemcheck/kmemcheck.c +index 8cc1833..1abbc5b 100644 +--- a/arch/x86/mm/kmemcheck/kmemcheck.c ++++ b/arch/x86/mm/kmemcheck/kmemcheck.c +@@ -622,9 +622,9 @@ bool kmemcheck_fault(struct pt_regs *regs, unsigned long address, + * memory (e.g. tracked pages)? For now, we need this to avoid + * invoking kmemcheck for PnP BIOS calls. + */ +- if (regs->flags & X86_VM_MASK) ++ if (v8086_mode(regs)) + return false; +- if (regs->cs != __KERNEL_CS) ++ if (regs->cs != __KERNEL_CS && regs->cs != __KERNEXEC_KERNEL_CS) + return false; + + pte = kmemcheck_pte_lookup(address); +diff --git a/arch/x86/mm/mmap.c b/arch/x86/mm/mmap.c +index c9e57af..07a321b 100644 +--- a/arch/x86/mm/mmap.c ++++ b/arch/x86/mm/mmap.c +@@ -49,7 +49,7 @@ static unsigned int stack_maxrandom_size(void) + * Leave an at least ~128 MB hole with possible stack randomization. + */ + #define MIN_GAP (128*1024*1024UL + stack_maxrandom_size()) +-#define MAX_GAP (TASK_SIZE/6*5) ++#define MAX_GAP (pax_task_size/6*5) + + /* + * True on X86_32 or when emulating IA32 on X86_64 +@@ -94,27 +94,40 @@ static unsigned long mmap_rnd(void) + return rnd << PAGE_SHIFT; + } + +-static unsigned long mmap_base(void) ++static unsigned long mmap_base(struct mm_struct *mm) + { + unsigned long gap = current->signal->rlim[RLIMIT_STACK].rlim_cur; ++ unsigned long pax_task_size = TASK_SIZE; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ pax_task_size = SEGMEXEC_TASK_SIZE; ++#endif + + if (gap < MIN_GAP) + gap = MIN_GAP; + else if (gap > MAX_GAP) + gap = MAX_GAP; + +- return PAGE_ALIGN(TASK_SIZE - gap - mmap_rnd()); ++ return PAGE_ALIGN(pax_task_size - gap - mmap_rnd()); + } + + /* + * Bottom-up (legacy) layout on X86_32 did not support randomization, X86_64 + * does, but not when emulating X86_32 + */ +-static unsigned long mmap_legacy_base(void) ++static unsigned long mmap_legacy_base(struct mm_struct *mm) + { +- if (mmap_is_ia32()) ++ if (mmap_is_ia32()) { ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (mm->pax_flags & MF_PAX_SEGMEXEC) ++ return SEGMEXEC_TASK_UNMAPPED_BASE; ++ else ++#endif ++ + return TASK_UNMAPPED_BASE; +- else ++ } else + return TASK_UNMAPPED_BASE + mmap_rnd(); + } + +@@ -125,11 +138,23 @@ static unsigned long mmap_legacy_base(void) + void arch_pick_mmap_layout(struct mm_struct *mm) + { + if (mmap_is_legacy()) { +- mm->mmap_base = mmap_legacy_base(); ++ mm->mmap_base = mmap_legacy_base(mm); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base += mm->delta_mmap; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area; + mm->unmap_area = arch_unmap_area; + } else { +- mm->mmap_base = mmap_base(); ++ mm->mmap_base = mmap_base(mm); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (mm->pax_flags & MF_PAX_RANDMMAP) ++ mm->mmap_base -= mm->delta_mmap + mm->delta_stack; ++#endif ++ + mm->get_unmapped_area = arch_get_unmapped_area_topdown; + mm->unmap_area = arch_unmap_area_topdown; + } +diff --git a/arch/x86/mm/mmio-mod.c b/arch/x86/mm/mmio-mod.c +index 132772a..b961f11 100644 +--- a/arch/x86/mm/mmio-mod.c ++++ b/arch/x86/mm/mmio-mod.c +@@ -193,7 +193,7 @@ static void pre(struct kmmio_probe *p, struct pt_regs *regs, + break; + default: + { +- unsigned char *ip = (unsigned char *)instptr; ++ unsigned char *ip = (unsigned char *)ktla_ktva(instptr); + my_trace->opcode = MMIO_UNKNOWN_OP; + my_trace->width = 0; + my_trace->value = (*ip) << 16 | *(ip + 1) << 8 | +@@ -233,7 +233,7 @@ static void post(struct kmmio_probe *p, unsigned long condition, + static void ioremap_trace_core(resource_size_t offset, unsigned long size, + void __iomem *addr) + { +- static atomic_t next_id; ++ static atomic_unchecked_t next_id; + struct remap_trace *trace = kmalloc(sizeof(*trace), GFP_KERNEL); + /* These are page-unaligned. */ + struct mmiotrace_map map = { +@@ -257,7 +257,7 @@ static void ioremap_trace_core(resource_size_t offset, unsigned long size, + .private = trace + }, + .phys = offset, +- .id = atomic_inc_return(&next_id) ++ .id = atomic_inc_return_unchecked(&next_id) + }; + map.map_id = trace->id; + +diff --git a/arch/x86/mm/numa_32.c b/arch/x86/mm/numa_32.c +index d253006..e56dd6a 100644 +--- a/arch/x86/mm/numa_32.c ++++ b/arch/x86/mm/numa_32.c +@@ -98,7 +98,6 @@ unsigned long node_memmap_size_bytes(int nid, unsigned long start_pfn, + } + #endif + +-extern unsigned long find_max_low_pfn(void); + extern unsigned long highend_pfn, highstart_pfn; + + #define LARGE_PAGE_BYTES (PTRS_PER_PTE * PAGE_SIZE) +diff --git a/arch/x86/mm/pageattr-test.c b/arch/x86/mm/pageattr-test.c +index e1d1069..2251ff3 100644 +--- a/arch/x86/mm/pageattr-test.c ++++ b/arch/x86/mm/pageattr-test.c +@@ -36,7 +36,7 @@ enum { + + static int pte_testbit(pte_t pte) + { +- return pte_flags(pte) & _PAGE_UNUSED1; ++ return pte_flags(pte) & _PAGE_CPA_TEST; + } + + struct split_state { +diff --git a/arch/x86/mm/pageattr.c b/arch/x86/mm/pageattr.c +index dd38bfb..b72c63e 100644 +--- a/arch/x86/mm/pageattr.c ++++ b/arch/x86/mm/pageattr.c +@@ -261,16 +261,17 @@ static inline pgprot_t static_protections(pgprot_t prot, unsigned long address, + * PCI BIOS based config access (CONFIG_PCI_GOBIOS) support. + */ + if (within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) +- pgprot_val(forbidden) |= _PAGE_NX; ++ pgprot_val(forbidden) |= _PAGE_NX & __supported_pte_mask; + + /* + * The kernel text needs to be executable for obvious reasons + * Does not cover __inittext since that is gone later on. On + * 64bit we do not enforce !NX on the low mapping + */ +- if (within(address, (unsigned long)_text, (unsigned long)_etext)) +- pgprot_val(forbidden) |= _PAGE_NX; ++ if (within(address, ktla_ktva((unsigned long)_text), ktla_ktva((unsigned long)_etext))) ++ pgprot_val(forbidden) |= _PAGE_NX & __supported_pte_mask; + ++#ifdef CONFIG_DEBUG_RODATA + /* + * The .rodata section needs to be read-only. Using the pfn + * catches all aliases. +@@ -278,6 +279,14 @@ static inline pgprot_t static_protections(pgprot_t prot, unsigned long address, + if (within(pfn, __pa((unsigned long)__start_rodata) >> PAGE_SHIFT, + __pa((unsigned long)__end_rodata) >> PAGE_SHIFT)) + pgprot_val(forbidden) |= _PAGE_RW; ++#endif ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ if (within(pfn, __pa((unsigned long)&_text), __pa((unsigned long)&_sdata))) { ++ pgprot_val(forbidden) |= _PAGE_RW; ++ pgprot_val(forbidden) |= _PAGE_NX & __supported_pte_mask; ++ } ++#endif + + prot = __pgprot(pgprot_val(prot) & ~pgprot_val(forbidden)); + +@@ -331,23 +340,37 @@ EXPORT_SYMBOL_GPL(lookup_address); + static void __set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte) + { + /* change init_mm */ ++ pax_open_kernel(); + set_pte_atomic(kpte, pte); ++ + #ifdef CONFIG_X86_32 + if (!SHARED_KERNEL_PMD) { ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ unsigned long cpu; ++#else + struct page *page; ++#endif + ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ for (cpu = 0; cpu < nr_cpu_ids; ++cpu) { ++ pgd_t *pgd = get_cpu_pgd(cpu); ++#else + list_for_each_entry(page, &pgd_list, lru) { +- pgd_t *pgd; ++ pgd_t *pgd = (pgd_t *)page_address(page); ++#endif ++ + pud_t *pud; + pmd_t *pmd; + +- pgd = (pgd_t *)page_address(page) + pgd_index(address); ++ pgd += pgd_index(address); + pud = pud_offset(pgd, address); + pmd = pmd_offset(pud, address); + set_pte_atomic((pte_t *)pmd, pte); + } + } + #endif ++ pax_close_kernel(); + } + + static int +diff --git a/arch/x86/mm/pat.c b/arch/x86/mm/pat.c +index e78cd0e..de0a817 100644 +--- a/arch/x86/mm/pat.c ++++ b/arch/x86/mm/pat.c +@@ -258,7 +258,7 @@ chk_conflict(struct memtype *new, struct memtype *entry, unsigned long *type) + + conflict: + printk(KERN_INFO "%s:%d conflicting memory types " +- "%Lx-%Lx %s<->%s\n", current->comm, current->pid, new->start, ++ "%Lx-%Lx %s<->%s\n", current->comm, task_pid_nr(current), new->start, + new->end, cattr_name(new->type), cattr_name(entry->type)); + return -EBUSY; + } +@@ -559,7 +559,7 @@ unlock_ret: + + if (err) { + printk(KERN_INFO "%s:%d freeing invalid memtype %Lx-%Lx\n", +- current->comm, current->pid, start, end); ++ current->comm, task_pid_nr(current), start, end); + } + + dprintk("free_memtype request 0x%Lx-0x%Lx\n", start, end); +@@ -689,8 +689,8 @@ static inline int range_is_allowed(unsigned long pfn, unsigned long size) + while (cursor < to) { + if (!devmem_is_allowed(pfn)) { + printk(KERN_INFO +- "Program %s tried to access /dev/mem between %Lx->%Lx.\n", +- current->comm, from, to); ++ "Program %s tried to access /dev/mem between %Lx->%Lx (%Lx).\n", ++ current->comm, from, to, cursor); + return 0; + } + cursor += PAGE_SIZE; +@@ -755,7 +755,7 @@ int kernel_map_sync_memtype(u64 base, unsigned long size, unsigned long flags) + printk(KERN_INFO + "%s:%d ioremap_change_attr failed %s " + "for %Lx-%Lx\n", +- current->comm, current->pid, ++ current->comm, task_pid_nr(current), + cattr_name(flags), + base, (unsigned long long)(base + size)); + return -EINVAL; +@@ -813,7 +813,7 @@ static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t *vma_prot, + free_memtype(paddr, paddr + size); + printk(KERN_ERR "%s:%d map pfn expected mapping type %s" + " for %Lx-%Lx, got %s\n", +- current->comm, current->pid, ++ current->comm, task_pid_nr(current), + cattr_name(want_flags), + (unsigned long long)paddr, + (unsigned long long)(paddr + size), +diff --git a/arch/x86/mm/pf_in.c b/arch/x86/mm/pf_in.c +index df3d5c8..c2223e1 100644 +--- a/arch/x86/mm/pf_in.c ++++ b/arch/x86/mm/pf_in.c +@@ -148,7 +148,7 @@ enum reason_type get_ins_type(unsigned long ins_addr) + int i; + enum reason_type rv = OTHERS; + +- p = (unsigned char *)ins_addr; ++ p = (unsigned char *)ktla_ktva(ins_addr); + p += skip_prefix(p, &prf); + p += get_opcode(p, &opcode); + +@@ -168,7 +168,7 @@ static unsigned int get_ins_reg_width(unsigned long ins_addr) + struct prefix_bits prf; + int i; + +- p = (unsigned char *)ins_addr; ++ p = (unsigned char *)ktla_ktva(ins_addr); + p += skip_prefix(p, &prf); + p += get_opcode(p, &opcode); + +@@ -191,7 +191,7 @@ unsigned int get_ins_mem_width(unsigned long ins_addr) + struct prefix_bits prf; + int i; + +- p = (unsigned char *)ins_addr; ++ p = (unsigned char *)ktla_ktva(ins_addr); + p += skip_prefix(p, &prf); + p += get_opcode(p, &opcode); + +@@ -417,7 +417,7 @@ unsigned long get_ins_reg_val(unsigned long ins_addr, struct pt_regs *regs) + int i; + unsigned long rv; + +- p = (unsigned char *)ins_addr; ++ p = (unsigned char *)ktla_ktva(ins_addr); + p += skip_prefix(p, &prf); + p += get_opcode(p, &opcode); + for (i = 0; i < ARRAY_SIZE(reg_rop); i++) +@@ -472,7 +472,7 @@ unsigned long get_ins_imm_val(unsigned long ins_addr) + int i; + unsigned long rv; + +- p = (unsigned char *)ins_addr; ++ p = (unsigned char *)ktla_ktva(ins_addr); + p += skip_prefix(p, &prf); + p += get_opcode(p, &opcode); + for (i = 0; i < ARRAY_SIZE(imm_wop); i++) +diff --git a/arch/x86/mm/pgtable.c b/arch/x86/mm/pgtable.c +index e0e6fad..c56b495 100644 +--- a/arch/x86/mm/pgtable.c ++++ b/arch/x86/mm/pgtable.c +@@ -83,9 +83,52 @@ static inline void pgd_list_del(pgd_t *pgd) + list_del(&page->lru); + } + +-#define UNSHARED_PTRS_PER_PGD \ +- (SHARED_KERNEL_PMD ? KERNEL_PGD_BOUNDARY : PTRS_PER_PGD) ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++pgdval_t clone_pgd_mask __read_only = ~_PAGE_PRESENT; + ++void __shadow_user_pgds(pgd_t *dst, const pgd_t *src, int count) ++{ ++ while (count--) ++ *dst++ = __pgd((pgd_val(*src++) | (_PAGE_NX & __supported_pte_mask)) & ~_PAGE_USER); ++} ++#endif ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++void __clone_user_pgds(pgd_t *dst, const pgd_t *src, int count) ++{ ++ while (count--) ++ ++#if defined(CONFIG_X86_64) && defined(CONFIG_PAX_MEMORY_UDEREF) ++ *dst++ = __pgd(pgd_val(*src++) & clone_pgd_mask); ++#else ++ *dst++ = *src++; ++#endif ++ ++} ++#endif ++ ++#ifdef CONFIG_X86_64 ++#define pxd_t pud_t ++#define pyd_t pgd_t ++#define paravirt_release_pxd(pfn) paravirt_release_pud(pfn) ++#define pxd_free(mm, pud) pud_free((mm), (pud)) ++#define pyd_populate(mm, pgd, pud) pgd_populate((mm), (pgd), (pud)) ++#define pyd_offset(mm, address) pgd_offset((mm), (address)) ++#define PYD_SIZE PGDIR_SIZE ++#else ++#define pxd_t pmd_t ++#define pyd_t pud_t ++#define paravirt_release_pxd(pfn) paravirt_release_pmd(pfn) ++#define pxd_free(mm, pud) pmd_free((mm), (pud)) ++#define pyd_populate(mm, pgd, pud) pud_populate((mm), (pgd), (pud)) ++#define pyd_offset(mm, address) pud_offset((mm), (address)) ++#define PYD_SIZE PUD_SIZE ++#endif ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++static inline void pgd_ctor(pgd_t *pgd) {} ++static inline void pgd_dtor(pgd_t *pgd) {} ++#else + static void pgd_ctor(pgd_t *pgd) + { + /* If the pgd points to a shared pagetable level (either the +@@ -119,6 +162,7 @@ static void pgd_dtor(pgd_t *pgd) + pgd_list_del(pgd); + spin_unlock_irqrestore(&pgd_lock, flags); + } ++#endif + + /* + * List of all pgd's needed for non-PAE so it can invalidate entries +@@ -131,7 +175,7 @@ static void pgd_dtor(pgd_t *pgd) + * -- wli + */ + +-#ifdef CONFIG_X86_PAE ++#if defined(CONFIG_X86_32) && defined(CONFIG_X86_PAE) + /* + * In PAE mode, we need to do a cr3 reload (=tlb flush) when + * updating the top-level pagetable entries to guarantee the +@@ -143,7 +187,7 @@ static void pgd_dtor(pgd_t *pgd) + * not shared between pagetables (!SHARED_KERNEL_PMDS), we allocate + * and initialize the kernel pmds here. + */ +-#define PREALLOCATED_PMDS UNSHARED_PTRS_PER_PGD ++#define PREALLOCATED_PXDS (SHARED_KERNEL_PMD ? KERNEL_PGD_BOUNDARY : PTRS_PER_PGD) + + void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd) + { +@@ -161,36 +205,38 @@ void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd) + */ + flush_tlb_mm(mm); + } ++#elif defined(CONFIG_X86_64) && defined(CONFIG_PAX_PER_CPU_PGD) ++#define PREALLOCATED_PXDS USER_PGD_PTRS + #else /* !CONFIG_X86_PAE */ + + /* No need to prepopulate any pagetable entries in non-PAE modes. */ +-#define PREALLOCATED_PMDS 0 ++#define PREALLOCATED_PXDS 0 + + #endif /* CONFIG_X86_PAE */ + +-static void free_pmds(pmd_t *pmds[]) ++static void free_pxds(pxd_t *pxds[]) + { + int i; + +- for(i = 0; i < PREALLOCATED_PMDS; i++) +- if (pmds[i]) +- free_page((unsigned long)pmds[i]); ++ for(i = 0; i < PREALLOCATED_PXDS; i++) ++ if (pxds[i]) ++ free_page((unsigned long)pxds[i]); + } + +-static int preallocate_pmds(pmd_t *pmds[]) ++static int preallocate_pxds(pxd_t *pxds[]) + { + int i; + bool failed = false; + +- for(i = 0; i < PREALLOCATED_PMDS; i++) { +- pmd_t *pmd = (pmd_t *)__get_free_page(PGALLOC_GFP); +- if (pmd == NULL) ++ for(i = 0; i < PREALLOCATED_PXDS; i++) { ++ pxd_t *pxd = (pxd_t *)__get_free_page(PGALLOC_GFP); ++ if (pxd == NULL) + failed = true; +- pmds[i] = pmd; ++ pxds[i] = pxd; + } + + if (failed) { +- free_pmds(pmds); ++ free_pxds(pxds); + return -ENOMEM; + } + +@@ -203,51 +249,56 @@ static int preallocate_pmds(pmd_t *pmds[]) + * preallocate which never got a corresponding vma will need to be + * freed manually. + */ +-static void pgd_mop_up_pmds(struct mm_struct *mm, pgd_t *pgdp) ++static void pgd_mop_up_pxds(struct mm_struct *mm, pgd_t *pgdp) + { + int i; + +- for(i = 0; i < PREALLOCATED_PMDS; i++) { ++ for(i = 0; i < PREALLOCATED_PXDS; i++) { + pgd_t pgd = pgdp[i]; + + if (pgd_val(pgd) != 0) { +- pmd_t *pmd = (pmd_t *)pgd_page_vaddr(pgd); ++ pxd_t *pxd = (pxd_t *)pgd_page_vaddr(pgd); + +- pgdp[i] = native_make_pgd(0); ++ set_pgd(pgdp + i, native_make_pgd(0)); + +- paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT); +- pmd_free(mm, pmd); ++ paravirt_release_pxd(pgd_val(pgd) >> PAGE_SHIFT); ++ pxd_free(mm, pxd); + } + } + } + +-static void pgd_prepopulate_pmd(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmds[]) ++static void pgd_prepopulate_pxd(struct mm_struct *mm, pgd_t *pgd, pxd_t *pxds[]) + { +- pud_t *pud; ++ pyd_t *pyd; + unsigned long addr; + int i; + +- if (PREALLOCATED_PMDS == 0) /* Work around gcc-3.4.x bug */ ++ if (PREALLOCATED_PXDS == 0) /* Work around gcc-3.4.x bug */ + return; + +- pud = pud_offset(pgd, 0); ++#ifdef CONFIG_X86_64 ++ pyd = pyd_offset(mm, 0L); ++#else ++ pyd = pyd_offset(pgd, 0L); ++#endif + +- for (addr = i = 0; i < PREALLOCATED_PMDS; +- i++, pud++, addr += PUD_SIZE) { +- pmd_t *pmd = pmds[i]; ++ for (addr = i = 0; i < PREALLOCATED_PXDS; ++ i++, pyd++, addr += PYD_SIZE) { ++ pxd_t *pxd = pxds[i]; + + if (i >= KERNEL_PGD_BOUNDARY) +- memcpy(pmd, (pmd_t *)pgd_page_vaddr(swapper_pg_dir[i]), +- sizeof(pmd_t) * PTRS_PER_PMD); ++ memcpy(pxd, (pxd_t *)pgd_page_vaddr(swapper_pg_dir[i]), ++ sizeof(pxd_t) * PTRS_PER_PMD); + +- pud_populate(mm, pud, pmd); ++ pyd_populate(mm, pyd, pxd); + } + } + + pgd_t *pgd_alloc(struct mm_struct *mm) + { + pgd_t *pgd; +- pmd_t *pmds[PREALLOCATED_PMDS]; ++ pxd_t *pxds[PREALLOCATED_PXDS]; ++ + unsigned long flags; + + pgd = (pgd_t *)__get_free_page(PGALLOC_GFP); +@@ -257,11 +308,11 @@ pgd_t *pgd_alloc(struct mm_struct *mm) + + mm->pgd = pgd; + +- if (preallocate_pmds(pmds) != 0) ++ if (preallocate_pxds(pxds) != 0) + goto out_free_pgd; + + if (paravirt_pgd_alloc(mm) != 0) +- goto out_free_pmds; ++ goto out_free_pxds; + + /* + * Make sure that pre-populating the pmds is atomic with +@@ -271,14 +322,14 @@ pgd_t *pgd_alloc(struct mm_struct *mm) + spin_lock_irqsave(&pgd_lock, flags); + + pgd_ctor(pgd); +- pgd_prepopulate_pmd(mm, pgd, pmds); ++ pgd_prepopulate_pxd(mm, pgd, pxds); + + spin_unlock_irqrestore(&pgd_lock, flags); + + return pgd; + +-out_free_pmds: +- free_pmds(pmds); ++out_free_pxds: ++ free_pxds(pxds); + out_free_pgd: + free_page((unsigned long)pgd); + out: +@@ -287,7 +338,7 @@ out: + + void pgd_free(struct mm_struct *mm, pgd_t *pgd) + { +- pgd_mop_up_pmds(mm, pgd); ++ pgd_mop_up_pxds(mm, pgd); + pgd_dtor(pgd); + paravirt_pgd_free(mm, pgd); + free_page((unsigned long)pgd); +diff --git a/arch/x86/mm/pgtable_32.c b/arch/x86/mm/pgtable_32.c +index 46c8834..fcab43d 100644 +--- a/arch/x86/mm/pgtable_32.c ++++ b/arch/x86/mm/pgtable_32.c +@@ -49,10 +49,13 @@ void set_pte_vaddr(unsigned long vaddr, pte_t pteval) + return; + } + pte = pte_offset_kernel(pmd, vaddr); ++ ++ pax_open_kernel(); + if (pte_val(pteval)) + set_pte_at(&init_mm, vaddr, pte, pteval); + else + pte_clear(&init_mm, vaddr, pte); ++ pax_close_kernel(); + + /* + * It's enough to flush this one mapping. +diff --git a/arch/x86/mm/setup_nx.c b/arch/x86/mm/setup_nx.c +index 513d8ed..978c161 100644 +--- a/arch/x86/mm/setup_nx.c ++++ b/arch/x86/mm/setup_nx.c +@@ -4,11 +4,10 @@ + + #include + ++#if defined(CONFIG_X86_32) && defined(CONFIG_X86_PAE) + int nx_enabled; + +-#if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE) +-static int disable_nx __cpuinitdata; +- ++#ifndef CONFIG_PAX_PAGEEXEC + /* + * noexec = on|off + * +@@ -22,32 +21,26 @@ static int __init noexec_setup(char *str) + if (!str) + return -EINVAL; + if (!strncmp(str, "on", 2)) { +- __supported_pte_mask |= _PAGE_NX; +- disable_nx = 0; ++ nx_enabled = 1; + } else if (!strncmp(str, "off", 3)) { +- disable_nx = 1; +- __supported_pte_mask &= ~_PAGE_NX; ++ nx_enabled = 0; + } + return 0; + } + early_param("noexec", noexec_setup); + #endif ++#endif + + #ifdef CONFIG_X86_PAE + void __init set_nx(void) + { +- unsigned int v[4], l, h; ++ if (!nx_enabled && cpu_has_nx) { ++ unsigned l, h; + +- if (cpu_has_pae && (cpuid_eax(0x80000000) > 0x80000001)) { +- cpuid(0x80000001, &v[0], &v[1], &v[2], &v[3]); +- +- if ((v[3] & (1 << 20)) && !disable_nx) { +- rdmsr(MSR_EFER, l, h); +- l |= EFER_NX; +- wrmsr(MSR_EFER, l, h); +- nx_enabled = 1; +- __supported_pte_mask |= _PAGE_NX; +- } ++ __supported_pte_mask &= ~_PAGE_NX; ++ rdmsr(MSR_EFER, l, h); ++ l &= ~EFER_NX; ++ wrmsr(MSR_EFER, l, h); + } + } + #else +@@ -62,7 +55,7 @@ void __cpuinit check_efer(void) + unsigned long efer; + + rdmsrl(MSR_EFER, efer); +- if (!(efer & EFER_NX) || disable_nx) ++ if (!(efer & EFER_NX) || !nx_enabled) + __supported_pte_mask &= ~_PAGE_NX; + } + #endif +diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c +index 36fe08e..b123d3a 100644 +--- a/arch/x86/mm/tlb.c ++++ b/arch/x86/mm/tlb.c +@@ -61,7 +61,11 @@ void leave_mm(int cpu) + BUG(); + cpumask_clear_cpu(cpu, + mm_cpumask(percpu_read(cpu_tlbstate.active_mm))); ++ ++#ifndef CONFIG_PAX_PER_CPU_PGD + load_cr3(swapper_pg_dir); ++#endif ++ + } + EXPORT_SYMBOL_GPL(leave_mm); + +diff --git a/arch/x86/oprofile/backtrace.c b/arch/x86/oprofile/backtrace.c +index 829edf0..672adb3 100644 +--- a/arch/x86/oprofile/backtrace.c ++++ b/arch/x86/oprofile/backtrace.c +@@ -115,7 +115,7 @@ x86_backtrace(struct pt_regs * const regs, unsigned int depth) + { + struct frame_head *head = (struct frame_head *)frame_pointer(regs); + +- if (!user_mode_vm(regs)) { ++ if (!user_mode(regs)) { + unsigned long stack = kernel_stack_pointer(regs); + if (depth) + dump_trace(NULL, regs, (unsigned long *)stack, 0, +diff --git a/arch/x86/oprofile/op_model_p4.c b/arch/x86/oprofile/op_model_p4.c +index e6a160a..36deff6 100644 +--- a/arch/x86/oprofile/op_model_p4.c ++++ b/arch/x86/oprofile/op_model_p4.c +@@ -50,7 +50,7 @@ static inline void setup_num_counters(void) + #endif + } + +-static int inline addr_increment(void) ++static inline int addr_increment(void) + { + #ifdef CONFIG_SMP + return smp_num_siblings == 2 ? 2 : 1; +diff --git a/arch/x86/pci/common.c b/arch/x86/pci/common.c +index 1331fcf..03901b2 100644 +--- a/arch/x86/pci/common.c ++++ b/arch/x86/pci/common.c +@@ -31,8 +31,8 @@ int noioapicreroute = 1; + int pcibios_last_bus = -1; + unsigned long pirq_table_addr; + struct pci_bus *pci_root_bus; +-struct pci_raw_ops *raw_pci_ops; +-struct pci_raw_ops *raw_pci_ext_ops; ++const struct pci_raw_ops *raw_pci_ops; ++const struct pci_raw_ops *raw_pci_ext_ops; + + int raw_pci_read(unsigned int domain, unsigned int bus, unsigned int devfn, + int reg, int len, u32 *val) +diff --git a/arch/x86/pci/direct.c b/arch/x86/pci/direct.c +index 347d882..4baf6b6 100644 +--- a/arch/x86/pci/direct.c ++++ b/arch/x86/pci/direct.c +@@ -79,7 +79,7 @@ static int pci_conf1_write(unsigned int seg, unsigned int bus, + + #undef PCI_CONF1_ADDRESS + +-struct pci_raw_ops pci_direct_conf1 = { ++const struct pci_raw_ops pci_direct_conf1 = { + .read = pci_conf1_read, + .write = pci_conf1_write, + }; +@@ -173,7 +173,7 @@ static int pci_conf2_write(unsigned int seg, unsigned int bus, + + #undef PCI_CONF2_ADDRESS + +-struct pci_raw_ops pci_direct_conf2 = { ++const struct pci_raw_ops pci_direct_conf2 = { + .read = pci_conf2_read, + .write = pci_conf2_write, + }; +@@ -189,7 +189,7 @@ struct pci_raw_ops pci_direct_conf2 = { + * This should be close to trivial, but it isn't, because there are buggy + * chipsets (yes, you guessed it, by Intel and Compaq) that have no class ID. + */ +-static int __init pci_sanity_check(struct pci_raw_ops *o) ++static int __init pci_sanity_check(const struct pci_raw_ops *o) + { + u32 x = 0; + int year, devfn; +diff --git a/arch/x86/pci/mmconfig_32.c b/arch/x86/pci/mmconfig_32.c +index f10a7e9..0425342 100644 +--- a/arch/x86/pci/mmconfig_32.c ++++ b/arch/x86/pci/mmconfig_32.c +@@ -125,7 +125,7 @@ static int pci_mmcfg_write(unsigned int seg, unsigned int bus, + return 0; + } + +-static struct pci_raw_ops pci_mmcfg = { ++static const struct pci_raw_ops pci_mmcfg = { + .read = pci_mmcfg_read, + .write = pci_mmcfg_write, + }; +diff --git a/arch/x86/pci/mmconfig_64.c b/arch/x86/pci/mmconfig_64.c +index 94349f8..41600a7 100644 +--- a/arch/x86/pci/mmconfig_64.c ++++ b/arch/x86/pci/mmconfig_64.c +@@ -104,7 +104,7 @@ static int pci_mmcfg_write(unsigned int seg, unsigned int bus, + return 0; + } + +-static struct pci_raw_ops pci_mmcfg = { ++static const struct pci_raw_ops pci_mmcfg = { + .read = pci_mmcfg_read, + .write = pci_mmcfg_write, + }; +diff --git a/arch/x86/pci/numaq_32.c b/arch/x86/pci/numaq_32.c +index 8eb295e..86bd657 100644 +--- a/arch/x86/pci/numaq_32.c ++++ b/arch/x86/pci/numaq_32.c +@@ -112,7 +112,7 @@ static int pci_conf1_mq_write(unsigned int seg, unsigned int bus, + + #undef PCI_CONF1_MQ_ADDRESS + +-static struct pci_raw_ops pci_direct_conf1_mq = { ++static const struct pci_raw_ops pci_direct_conf1_mq = { + .read = pci_conf1_mq_read, + .write = pci_conf1_mq_write + }; +diff --git a/arch/x86/pci/olpc.c b/arch/x86/pci/olpc.c +index b889d82..5a58a0a 100644 +--- a/arch/x86/pci/olpc.c ++++ b/arch/x86/pci/olpc.c +@@ -297,7 +297,7 @@ static int pci_olpc_write(unsigned int seg, unsigned int bus, + return 0; + } + +-static struct pci_raw_ops pci_olpc_conf = { ++static const struct pci_raw_ops pci_olpc_conf = { + .read = pci_olpc_read, + .write = pci_olpc_write, + }; +diff --git a/arch/x86/pci/pcbios.c b/arch/x86/pci/pcbios.c +index 1c975cc..b8e16c2 100644 +--- a/arch/x86/pci/pcbios.c ++++ b/arch/x86/pci/pcbios.c +@@ -56,50 +56,93 @@ union bios32 { + static struct { + unsigned long address; + unsigned short segment; +-} bios32_indirect = { 0, __KERNEL_CS }; ++} bios32_indirect __read_only = { 0, __PCIBIOS_CS }; + + /* + * Returns the entry point for the given service, NULL on error + */ + +-static unsigned long bios32_service(unsigned long service) ++static unsigned long __devinit bios32_service(unsigned long service) + { + unsigned char return_code; /* %al */ + unsigned long address; /* %ebx */ + unsigned long length; /* %ecx */ + unsigned long entry; /* %edx */ + unsigned long flags; ++ struct desc_struct d, *gdt; + + local_irq_save(flags); +- __asm__("lcall *(%%edi); cld" ++ ++ gdt = get_cpu_gdt_table(smp_processor_id()); ++ ++ pack_descriptor(&d, 0UL, 0xFFFFFUL, 0x9B, 0xC); ++ write_gdt_entry(gdt, GDT_ENTRY_PCIBIOS_CS, &d, DESCTYPE_S); ++ pack_descriptor(&d, 0UL, 0xFFFFFUL, 0x93, 0xC); ++ write_gdt_entry(gdt, GDT_ENTRY_PCIBIOS_DS, &d, DESCTYPE_S); ++ ++ __asm__("movw %w7, %%ds; lcall *(%%edi); push %%ss; pop %%ds; cld" + : "=a" (return_code), + "=b" (address), + "=c" (length), + "=d" (entry) + : "0" (service), + "1" (0), +- "D" (&bios32_indirect)); ++ "D" (&bios32_indirect), ++ "r"(__PCIBIOS_DS) ++ : "memory"); ++ ++ pax_open_kernel(); ++ gdt[GDT_ENTRY_PCIBIOS_CS].a = 0; ++ gdt[GDT_ENTRY_PCIBIOS_CS].b = 0; ++ gdt[GDT_ENTRY_PCIBIOS_DS].a = 0; ++ gdt[GDT_ENTRY_PCIBIOS_DS].b = 0; ++ pax_close_kernel(); ++ + local_irq_restore(flags); + + switch (return_code) { +- case 0: +- return address + entry; +- case 0x80: /* Not present */ +- printk(KERN_WARNING "bios32_service(0x%lx): not present\n", service); +- return 0; +- default: /* Shouldn't happen */ +- printk(KERN_WARNING "bios32_service(0x%lx): returned 0x%x -- BIOS bug!\n", +- service, return_code); ++ case 0: { ++ int cpu; ++ unsigned char flags; ++ ++ printk(KERN_INFO "bios32_service: base:%08lx length:%08lx entry:%08lx\n", address, length, entry); ++ if (address >= 0xFFFF0 || length > 0x100000 - address || length <= entry) { ++ printk(KERN_WARNING "bios32_service: not valid\n"); + return 0; ++ } ++ address = address + PAGE_OFFSET; ++ length += 16UL; /* some BIOSs underreport this... */ ++ flags = 4; ++ if (length >= 64*1024*1024) { ++ length >>= PAGE_SHIFT; ++ flags |= 8; ++ } ++ ++ for (cpu = 0; cpu < nr_cpu_ids; cpu++) { ++ gdt = get_cpu_gdt_table(cpu); ++ pack_descriptor(&d, address, length, 0x9b, flags); ++ write_gdt_entry(gdt, GDT_ENTRY_PCIBIOS_CS, &d, DESCTYPE_S); ++ pack_descriptor(&d, address, length, 0x93, flags); ++ write_gdt_entry(gdt, GDT_ENTRY_PCIBIOS_DS, &d, DESCTYPE_S); ++ } ++ return entry; ++ } ++ case 0x80: /* Not present */ ++ printk(KERN_WARNING "bios32_service(0x%lx): not present\n", service); ++ return 0; ++ default: /* Shouldn't happen */ ++ printk(KERN_WARNING "bios32_service(0x%lx): returned 0x%x -- BIOS bug!\n", ++ service, return_code); ++ return 0; + } + } + + static struct { + unsigned long address; + unsigned short segment; +-} pci_indirect = { 0, __KERNEL_CS }; ++} pci_indirect __read_only = { 0, __PCIBIOS_CS }; + +-static int pci_bios_present; ++static int pci_bios_present __read_only; + + static int __devinit check_pcibios(void) + { +@@ -108,11 +151,13 @@ static int __devinit check_pcibios(void) + unsigned long flags, pcibios_entry; + + if ((pcibios_entry = bios32_service(PCI_SERVICE))) { +- pci_indirect.address = pcibios_entry + PAGE_OFFSET; ++ pci_indirect.address = pcibios_entry; + + local_irq_save(flags); +- __asm__( +- "lcall *(%%edi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%edi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -121,7 +166,8 @@ static int __devinit check_pcibios(void) + "=b" (ebx), + "=c" (ecx) + : "1" (PCIBIOS_PCI_BIOS_PRESENT), +- "D" (&pci_indirect) ++ "D" (&pci_indirect), ++ "r" (__PCIBIOS_DS) + : "memory"); + local_irq_restore(flags); + +@@ -165,7 +211,10 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + + switch (len) { + case 1: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -174,7 +223,8 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + : "1" (PCIBIOS_READ_CONFIG_BYTE), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + /* + * Zero-extend the result beyond 8 bits, do not trust the + * BIOS having done it: +@@ -182,7 +232,10 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + *value &= 0xff; + break; + case 2: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -191,7 +244,8 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + : "1" (PCIBIOS_READ_CONFIG_WORD), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + /* + * Zero-extend the result beyond 16 bits, do not trust the + * BIOS having done it: +@@ -199,7 +253,10 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + *value &= 0xffff; + break; + case 4: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -208,7 +265,8 @@ static int pci_bios_read(unsigned int seg, unsigned int bus, + : "1" (PCIBIOS_READ_CONFIG_DWORD), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + break; + } + +@@ -231,7 +289,10 @@ static int pci_bios_write(unsigned int seg, unsigned int bus, + + switch (len) { + case 1: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -240,10 +301,14 @@ static int pci_bios_write(unsigned int seg, unsigned int bus, + "c" (value), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + break; + case 2: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -252,10 +317,14 @@ static int pci_bios_write(unsigned int seg, unsigned int bus, + "c" (value), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + break; + case 4: +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w6, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n\t" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -264,7 +333,8 @@ static int pci_bios_write(unsigned int seg, unsigned int bus, + "c" (value), + "b" (bx), + "D" ((long)reg), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + break; + } + +@@ -278,7 +348,7 @@ static int pci_bios_write(unsigned int seg, unsigned int bus, + * Function table for BIOS32 access + */ + +-static struct pci_raw_ops pci_bios_access = { ++static const struct pci_raw_ops pci_bios_access = { + .read = pci_bios_read, + .write = pci_bios_write + }; +@@ -287,7 +357,7 @@ static struct pci_raw_ops pci_bios_access = { + * Try to find PCI BIOS. + */ + +-static struct pci_raw_ops * __devinit pci_find_bios(void) ++static const struct pci_raw_ops * __devinit pci_find_bios(void) + { + union bios32 *check; + unsigned char sum; +@@ -368,10 +438,13 @@ struct irq_routing_table * pcibios_get_irq_routing_table(void) + + DBG("PCI: Fetching IRQ routing table... "); + __asm__("push %%es\n\t" ++ "movw %w8, %%ds\n\t" + "push %%ds\n\t" + "pop %%es\n\t" +- "lcall *(%%esi); cld\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" + "pop %%es\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -382,7 +455,8 @@ struct irq_routing_table * pcibios_get_irq_routing_table(void) + "1" (0), + "D" ((long) &opt), + "S" (&pci_indirect), +- "m" (opt) ++ "m" (opt), ++ "r" (__PCIBIOS_DS) + : "memory"); + DBG("OK ret=%d, size=%d, map=%x\n", ret, opt.size, map); + if (ret & 0xff00) +@@ -406,7 +480,10 @@ int pcibios_set_irq_routing(struct pci_dev *dev, int pin, int irq) + { + int ret; + +- __asm__("lcall *(%%esi); cld\n\t" ++ __asm__("movw %w5, %%ds\n\t" ++ "lcall *%%ss:(%%esi); cld\n\t" ++ "push %%ss\n\t" ++ "pop %%ds\n" + "jc 1f\n\t" + "xor %%ah, %%ah\n" + "1:" +@@ -414,7 +491,8 @@ int pcibios_set_irq_routing(struct pci_dev *dev, int pin, int irq) + : "0" (PCIBIOS_SET_PCI_HW_INT), + "b" ((dev->bus->number << 8) | dev->devfn), + "c" ((irq << 8) | (pin + 10)), +- "S" (&pci_indirect)); ++ "S" (&pci_indirect), ++ "r" (__PCIBIOS_DS)); + return !(ret & 0xff00); + } + EXPORT_SYMBOL(pcibios_set_irq_routing); +diff --git a/arch/x86/power/cpu.c b/arch/x86/power/cpu.c +index fa0f651..9d8f3d9 100644 +--- a/arch/x86/power/cpu.c ++++ b/arch/x86/power/cpu.c +@@ -129,7 +129,7 @@ static void do_fpu_end(void) + static void fix_processor_context(void) + { + int cpu = smp_processor_id(); +- struct tss_struct *t = &per_cpu(init_tss, cpu); ++ struct tss_struct *t = init_tss + cpu; + + set_tss_desc(cpu, t); /* + * This just modifies memory; should not be +@@ -139,7 +139,9 @@ static void fix_processor_context(void) + */ + + #ifdef CONFIG_X86_64 ++ pax_open_kernel(); + get_cpu_gdt_table(cpu)[GDT_ENTRY_TSS].type = 9; ++ pax_close_kernel(); + + syscall_init(); /* This sets MSR_*STAR and related */ + #endif +diff --git a/arch/x86/vdso/Makefile b/arch/x86/vdso/Makefile +index dd78ef6..f9d928d 100644 +--- a/arch/x86/vdso/Makefile ++++ b/arch/x86/vdso/Makefile +@@ -122,7 +122,7 @@ quiet_cmd_vdso = VDSO $@ + $(VDSO_LDFLAGS) $(VDSO_LDFLAGS_$(filter %.lds,$(^F))) \ + -Wl,-T,$(filter %.lds,$^) $(filter %.o,$^) + +-VDSO_LDFLAGS = -fPIC -shared $(call cc-ldoption, -Wl$(comma)--hash-style=sysv) ++VDSO_LDFLAGS = -fPIC -shared -Wl,--no-undefined $(call cc-ldoption, -Wl$(comma)--hash-style=sysv) + GCOV_PROFILE := n + + # +diff --git a/arch/x86/vdso/vclock_gettime.c b/arch/x86/vdso/vclock_gettime.c +index ee55754..0013b2e 100644 +--- a/arch/x86/vdso/vclock_gettime.c ++++ b/arch/x86/vdso/vclock_gettime.c +@@ -22,24 +22,48 @@ + #include + #include + #include ++#include + #include "vextern.h" + + #define gtod vdso_vsyscall_gtod_data + ++notrace noinline long __vdso_fallback_time(long *t) ++{ ++ long secs; ++ asm volatile("syscall" ++ : "=a" (secs) ++ : "0" (__NR_time),"D" (t) : "r11", "cx", "memory"); ++ return secs; ++} ++ + notrace static long vdso_fallback_gettime(long clock, struct timespec *ts) + { + long ret; + asm("syscall" : "=a" (ret) : +- "0" (__NR_clock_gettime),"D" (clock), "S" (ts) : "memory"); ++ "0" (__NR_clock_gettime),"D" (clock), "S" (ts) : "r11", "cx", "memory"); + return ret; + } + ++notrace static inline cycle_t __vdso_vread_hpet(void) ++{ ++ return readl((const void __iomem *)fix_to_virt(VSYSCALL_HPET) + 0xf0); ++} ++ ++notrace static inline cycle_t __vdso_vread_tsc(void) ++{ ++ cycle_t ret = (cycle_t)vget_cycles(); ++ ++ return ret >= gtod->clock.cycle_last ? ret : gtod->clock.cycle_last; ++} ++ + notrace static inline long vgetns(void) + { + long v; +- cycles_t (*vread)(void); +- vread = gtod->clock.vread; +- v = (vread() - gtod->clock.cycle_last) & gtod->clock.mask; ++ if (gtod->clock.name[0] == 't' && gtod->clock.name[1] == 's' && gtod->clock.name[2] == 'c' && !gtod->clock.name[3]) ++ v = __vdso_vread_tsc(); ++ else ++ v = __vdso_vread_hpet(); ++ v = (v - gtod->clock.cycle_last) & gtod->clock.mask; + return (v * gtod->clock.mult) >> gtod->clock.shift; + } + +@@ -113,7 +137,9 @@ notrace static noinline int do_monotonic_coarse(struct timespec *ts) + + notrace int __vdso_clock_gettime(clockid_t clock, struct timespec *ts) + { +- if (likely(gtod->sysctl_enabled)) ++ if (likely(gtod->sysctl_enabled && ++ ((gtod->clock.name[0] == 'h' && gtod->clock.name[1] == 'p' && gtod->clock.name[2] == 'e' && gtod->clock.name[3] == 't' && !gtod->clock.name[4]) || ++ (gtod->clock.name[0] == 't' && gtod->clock.name[1] == 's' && gtod->clock.name[2] == 'c' && !gtod->clock.name[3])))) + switch (clock) { + case CLOCK_REALTIME: + if (likely(gtod->clock.vread)) +@@ -133,10 +159,20 @@ notrace int __vdso_clock_gettime(clockid_t clock, struct timespec *ts) + int clock_gettime(clockid_t, struct timespec *) + __attribute__((weak, alias("__vdso_clock_gettime"))); + ++notrace noinline int __vdso_fallback_gettimeofday(struct timeval *tv, struct timezone *tz) ++{ ++ long ret; ++ asm("syscall" : "=a" (ret) : ++ "0" (__NR_gettimeofday), "D" (tv), "S" (tz) : "r11", "cx", "memory"); ++ return ret; ++} ++ + notrace int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) + { +- long ret; +- if (likely(gtod->sysctl_enabled && gtod->clock.vread)) { ++ if (likely(gtod->sysctl_enabled && ++ ((gtod->clock.name[0] == 'h' && gtod->clock.name[1] == 'p' && gtod->clock.name[2] == 'e' && gtod->clock.name[3] == 't' && !gtod->clock.name[4]) || ++ (gtod->clock.name[0] == 't' && gtod->clock.name[1] == 's' && gtod->clock.name[2] == 'c' && !gtod->clock.name[3])))) ++ { + if (likely(tv != NULL)) { + BUILD_BUG_ON(offsetof(struct timeval, tv_usec) != + offsetof(struct timespec, tv_nsec) || +@@ -151,9 +187,7 @@ notrace int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) + } + return 0; + } +- asm("syscall" : "=a" (ret) : +- "0" (__NR_gettimeofday), "D" (tv), "S" (tz) : "memory"); +- return ret; ++ return __vdso_fallback_gettimeofday(tv, tz); + } + int gettimeofday(struct timeval *, struct timezone *) + __attribute__((weak, alias("__vdso_gettimeofday"))); +diff --git a/arch/x86/vdso/vdso.lds.S b/arch/x86/vdso/vdso.lds.S +index 4e5dd3b..00ba15e 100644 +--- a/arch/x86/vdso/vdso.lds.S ++++ b/arch/x86/vdso/vdso.lds.S +@@ -35,3 +35,9 @@ VDSO64_PRELINK = VDSO_PRELINK; + #define VEXTERN(x) VDSO64_ ## x = vdso_ ## x; + #include "vextern.h" + #undef VEXTERN ++ ++#define VEXTERN(x) VDSO64_ ## x = __vdso_ ## x; ++VEXTERN(fallback_gettimeofday) ++VEXTERN(fallback_time) ++VEXTERN(getcpu) ++#undef VEXTERN +diff --git a/arch/x86/vdso/vdso32-setup.c b/arch/x86/vdso/vdso32-setup.c +index 58bc00f..d53fb48 100644 +--- a/arch/x86/vdso/vdso32-setup.c ++++ b/arch/x86/vdso/vdso32-setup.c +@@ -25,6 +25,7 @@ + #include + #include + #include ++#include + + enum { + VDSO_DISABLED = 0, +@@ -226,7 +227,7 @@ static inline void map_compat_vdso(int map) + void enable_sep_cpu(void) + { + int cpu = get_cpu(); +- struct tss_struct *tss = &per_cpu(init_tss, cpu); ++ struct tss_struct *tss = init_tss + cpu; + + if (!boot_cpu_has(X86_FEATURE_SEP)) { + put_cpu(); +@@ -249,7 +250,7 @@ static int __init gate_vma_init(void) + gate_vma.vm_start = FIXADDR_USER_START; + gate_vma.vm_end = FIXADDR_USER_END; + gate_vma.vm_flags = VM_READ | VM_MAYREAD | VM_EXEC | VM_MAYEXEC; +- gate_vma.vm_page_prot = __P101; ++ gate_vma.vm_page_prot = vm_get_page_prot(gate_vma.vm_flags); + /* + * Make sure the vDSO gets into every core dump. + * Dumping its contents makes post-mortem fully interpretable later +@@ -331,14 +332,14 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + if (compat) + addr = VDSO_HIGH_BASE; + else { +- addr = get_unmapped_area(NULL, 0, PAGE_SIZE, 0, 0); ++ addr = get_unmapped_area(NULL, 0, PAGE_SIZE, 0, MAP_EXECUTABLE); + if (IS_ERR_VALUE(addr)) { + ret = addr; + goto up_fail; + } + } + +- current->mm->context.vdso = (void *)addr; ++ current->mm->context.vdso = addr; + + if (compat_uses_vma || !compat) { + /* +@@ -361,11 +362,11 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + } + + current_thread_info()->sysenter_return = +- VDSO32_SYMBOL(addr, SYSENTER_RETURN); ++ (__force void __user *)VDSO32_SYMBOL(addr, SYSENTER_RETURN); + + up_fail: + if (ret) +- current->mm->context.vdso = NULL; ++ current->mm->context.vdso = 0; + + up_write(&mm->mmap_sem); + +@@ -413,8 +414,14 @@ __initcall(ia32_binfmt_init); + + const char *arch_vma_name(struct vm_area_struct *vma) + { +- if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) ++ if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso) + return "[vdso]"; ++ ++#ifdef CONFIG_PAX_SEGMEXEC ++ if (vma->vm_mm && vma->vm_mirror && vma->vm_mirror->vm_start == vma->vm_mm->context.vdso) ++ return "[vdso]"; ++#endif ++ + return NULL; + } + +@@ -423,7 +430,7 @@ struct vm_area_struct *get_gate_vma(struct task_struct *tsk) + struct mm_struct *mm = tsk->mm; + + /* Check to see if this task was created in compat vdso mode */ +- if (mm && mm->context.vdso == (void *)VDSO_HIGH_BASE) ++ if (mm && mm->context.vdso == VDSO_HIGH_BASE) + return &gate_vma; + return NULL; + } +diff --git a/arch/x86/vdso/vextern.h b/arch/x86/vdso/vextern.h +index 1683ba2..48d07f3 100644 +--- a/arch/x86/vdso/vextern.h ++++ b/arch/x86/vdso/vextern.h +@@ -11,6 +11,5 @@ + put into vextern.h and be referenced as a pointer with vdso prefix. + The main kernel later fills in the values. */ + +-VEXTERN(jiffies) + VEXTERN(vgetcpu_mode) + VEXTERN(vsyscall_gtod_data) +diff --git a/arch/x86/vdso/vma.c b/arch/x86/vdso/vma.c +index 21e1aeb..2c0b3c4 100644 +--- a/arch/x86/vdso/vma.c ++++ b/arch/x86/vdso/vma.c +@@ -17,8 +17,6 @@ + #include "vextern.h" /* Just for VMAGIC. */ + #undef VEXTERN + +-unsigned int __read_mostly vdso_enabled = 1; +- + extern char vdso_start[], vdso_end[]; + extern unsigned short vdso_sync_cpuid; + +@@ -27,10 +25,8 @@ static unsigned vdso_size; + + static inline void *var_ref(void *p, char *name) + { +- if (*(void **)p != (void *)VMAGIC) { +- printk("VDSO: variable %s broken\n", name); +- vdso_enabled = 0; +- } ++ if (*(void **)p != (void *)VMAGIC) ++ panic("VDSO: variable %s broken\n", name); + return p; + } + +@@ -57,21 +53,18 @@ static int __init init_vdso_vars(void) + if (!vbase) + goto oom; + +- if (memcmp(vbase, "\177ELF", 4)) { +- printk("VDSO: I'm broken; not ELF\n"); +- vdso_enabled = 0; +- } ++ if (memcmp(vbase, ELFMAG, SELFMAG)) ++ panic("VDSO: I'm broken; not ELF\n"); + + #define VEXTERN(x) \ + *(typeof(__ ## x) **) var_ref(VDSO64_SYMBOL(vbase, x), #x) = &__ ## x; + #include "vextern.h" + #undef VEXTERN ++ vunmap(vbase); + return 0; + + oom: +- printk("Cannot allocate vdso\n"); +- vdso_enabled = 0; +- return -ENOMEM; ++ panic("Cannot allocate vdso\n"); + } + __initcall(init_vdso_vars); + +@@ -102,13 +95,15 @@ static unsigned long vdso_addr(unsigned long start, unsigned len) + int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + { + struct mm_struct *mm = current->mm; +- unsigned long addr; ++ unsigned long addr = 0; + int ret; + +- if (!vdso_enabled) +- return 0; +- + down_write(&mm->mmap_sem); ++ ++#ifdef CONFIG_PAX_RANDMMAP ++ if (!(mm->pax_flags & MF_PAX_RANDMMAP)) ++#endif ++ + addr = vdso_addr(mm->start_stack, vdso_size); + addr = get_unmapped_area(NULL, addr, vdso_size, 0, 0); + if (IS_ERR_VALUE(addr)) { +@@ -116,7 +111,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + goto up_fail; + } + +- current->mm->context.vdso = (void *)addr; ++ current->mm->context.vdso = addr; + + ret = install_special_mapping(mm, addr, vdso_size, + VM_READ|VM_EXEC| +@@ -124,7 +119,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp) + VM_ALWAYSDUMP, + vdso_pages); + if (ret) { +- current->mm->context.vdso = NULL; ++ current->mm->context.vdso = 0; + goto up_fail; + } + +@@ -132,10 +127,3 @@ up_fail: + up_write(&mm->mmap_sem); + return ret; + } +- +-static __init int vdso_setup(char *s) +-{ +- vdso_enabled = simple_strtoul(s, NULL, 0); +- return 0; +-} +-__setup("vdso=", vdso_setup); +diff --git a/arch/x86/xen/enlighten.c b/arch/x86/xen/enlighten.c +index 0087b00..eecb34f 100644 +--- a/arch/x86/xen/enlighten.c ++++ b/arch/x86/xen/enlighten.c +@@ -71,8 +71,6 @@ EXPORT_SYMBOL_GPL(xen_start_info); + + struct shared_info xen_dummy_shared_info; + +-void *xen_initial_gdt; +- + /* + * Point at some empty memory to start with. We map the real shared_info + * page as soon as fixmap is up and running. +@@ -548,7 +546,7 @@ static void xen_write_idt_entry(gate_desc *dt, int entrynum, const gate_desc *g) + + preempt_disable(); + +- start = __get_cpu_var(idt_desc).address; ++ start = (unsigned long)__get_cpu_var(idt_desc).address; + end = start + __get_cpu_var(idt_desc).size + 1; + + xen_mc_flush(); +@@ -993,7 +991,7 @@ static const struct pv_apic_ops xen_apic_ops __initdata = { + #endif + }; + +-static void xen_reboot(int reason) ++static __noreturn void xen_reboot(int reason) + { + struct sched_shutdown r = { .reason = reason }; + +@@ -1001,17 +999,17 @@ static void xen_reboot(int reason) + BUG(); + } + +-static void xen_restart(char *msg) ++static __noreturn void xen_restart(char *msg) + { + xen_reboot(SHUTDOWN_reboot); + } + +-static void xen_emergency_restart(void) ++static __noreturn void xen_emergency_restart(void) + { + xen_reboot(SHUTDOWN_reboot); + } + +-static void xen_machine_halt(void) ++static __noreturn void xen_machine_halt(void) + { + xen_reboot(SHUTDOWN_poweroff); + } +@@ -1095,9 +1093,20 @@ asmlinkage void __init xen_start_kernel(void) + */ + __userpte_alloc_gfp &= ~__GFP_HIGHMEM; + +-#ifdef CONFIG_X86_64 + /* Work out if we support NX */ +- check_efer(); ++#if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE) ++ if ((cpuid_eax(0x80000000) & 0xffff0000) == 0x80000000 && ++ (cpuid_edx(0x80000001) & (1U << (X86_FEATURE_NX & 31)))) { ++ unsigned l, h; ++ ++#ifdef CONFIG_X86_PAE ++ nx_enabled = 1; ++#endif ++ __supported_pte_mask |= _PAGE_NX; ++ rdmsr(MSR_EFER, l, h); ++ l |= EFER_NX; ++ wrmsr(MSR_EFER, l, h); ++ } + #endif + + xen_setup_features(); +@@ -1129,13 +1138,6 @@ asmlinkage void __init xen_start_kernel(void) + + machine_ops = xen_machine_ops; + +- /* +- * The only reliable way to retain the initial address of the +- * percpu gdt_page is to remember it here, so we can go and +- * mark it RW later, when the initial percpu area is freed. +- */ +- xen_initial_gdt = &per_cpu(gdt_page, 0); +- + xen_smp_init(); + + pgd = (pgd_t *)xen_start_info->pt_base; +diff --git a/arch/x86/xen/mmu.c b/arch/x86/xen/mmu.c +index 3f90a2c..2c2ad84 100644 +--- a/arch/x86/xen/mmu.c ++++ b/arch/x86/xen/mmu.c +@@ -1719,6 +1719,9 @@ __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, + convert_pfn_mfn(init_level4_pgt); + convert_pfn_mfn(level3_ident_pgt); + convert_pfn_mfn(level3_kernel_pgt); ++ convert_pfn_mfn(level3_vmalloc_start_pgt); ++ convert_pfn_mfn(level3_vmalloc_end_pgt); ++ convert_pfn_mfn(level3_vmemmap_pgt); + + l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd); + l2 = m2v(l3[pud_index(__START_KERNEL_map)].pud); +@@ -1737,7 +1740,11 @@ __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, + set_page_prot(init_level4_pgt, PAGE_KERNEL_RO); + set_page_prot(level3_ident_pgt, PAGE_KERNEL_RO); + set_page_prot(level3_kernel_pgt, PAGE_KERNEL_RO); ++ set_page_prot(level3_vmalloc_start_pgt, PAGE_KERNEL_RO); ++ set_page_prot(level3_vmalloc_end_pgt, PAGE_KERNEL_RO); ++ set_page_prot(level3_vmemmap_pgt, PAGE_KERNEL_RO); + set_page_prot(level3_user_vsyscall, PAGE_KERNEL_RO); ++ set_page_prot(level2_vmemmap_pgt, PAGE_KERNEL_RO); + set_page_prot(level2_kernel_pgt, PAGE_KERNEL_RO); + set_page_prot(level2_fixmap_pgt, PAGE_KERNEL_RO); + +@@ -1860,6 +1867,7 @@ static __init void xen_post_allocator_init(void) + pv_mmu_ops.set_pud = xen_set_pud; + #if PAGETABLE_LEVELS == 4 + pv_mmu_ops.set_pgd = xen_set_pgd; ++ pv_mmu_ops.set_pgd_batched = xen_set_pgd; + #endif + + /* This will work as long as patching hasn't happened yet +@@ -1946,6 +1954,7 @@ static const struct pv_mmu_ops xen_mmu_ops __initdata = { + .pud_val = PV_CALLEE_SAVE(xen_pud_val), + .make_pud = PV_CALLEE_SAVE(xen_make_pud), + .set_pgd = xen_set_pgd_hyper, ++ .set_pgd_batched = xen_set_pgd_hyper, + + .alloc_pud = xen_alloc_pmd_init, + .release_pud = xen_release_pmd_init, +diff --git a/arch/x86/xen/smp.c b/arch/x86/xen/smp.c +index a96204a..fca9b8e 100644 +--- a/arch/x86/xen/smp.c ++++ b/arch/x86/xen/smp.c +@@ -168,11 +168,6 @@ static void __init xen_smp_prepare_boot_cpu(void) + { + BUG_ON(smp_processor_id() != 0); + native_smp_prepare_boot_cpu(); +- +- /* We've switched to the "real" per-cpu gdt, so make sure the +- old memory can be recycled */ +- make_lowmem_page_readwrite(xen_initial_gdt); +- + xen_setup_vcpu_info_placement(); + } + +@@ -241,12 +236,12 @@ cpu_initialize_context(unsigned int cpu, struct task_struct *idle) + gdt = get_cpu_gdt_table(cpu); + + ctxt->flags = VGCF_IN_KERNEL; +- ctxt->user_regs.ds = __USER_DS; +- ctxt->user_regs.es = __USER_DS; ++ ctxt->user_regs.ds = __KERNEL_DS; ++ ctxt->user_regs.es = __KERNEL_DS; + ctxt->user_regs.ss = __KERNEL_DS; + #ifdef CONFIG_X86_32 + ctxt->user_regs.fs = __KERNEL_PERCPU; +- ctxt->user_regs.gs = __KERNEL_STACK_CANARY; ++ savesegment(gs, ctxt->user_regs.gs); + #else + ctxt->gs_base_kernel = per_cpu_offset(cpu); + #endif +@@ -297,13 +292,12 @@ static int __cpuinit xen_cpu_up(unsigned int cpu) + int rc; + + per_cpu(current_task, cpu) = idle; ++ per_cpu(current_tinfo, cpu) = &idle->tinfo; + #ifdef CONFIG_X86_32 + irq_ctx_init(cpu); + #else + clear_tsk_thread_flag(idle, TIF_FORK); +- per_cpu(kernel_stack, cpu) = +- (unsigned long)task_stack_page(idle) - +- KERNEL_STACK_OFFSET + THREAD_SIZE; ++ per_cpu(kernel_stack, cpu) = (unsigned long)task_stack_page(idle) - 16 + THREAD_SIZE; + #endif + xen_setup_runstate_info(cpu); + xen_setup_timer(cpu); +diff --git a/arch/x86/xen/xen-asm_32.S b/arch/x86/xen/xen-asm_32.S +index 9a95a9c..4f39e774 100644 +--- a/arch/x86/xen/xen-asm_32.S ++++ b/arch/x86/xen/xen-asm_32.S +@@ -83,14 +83,14 @@ ENTRY(xen_iret) + ESP_OFFSET=4 # bytes pushed onto stack + + /* +- * Store vcpu_info pointer for easy access. Do it this way to +- * avoid having to reload %fs ++ * Store vcpu_info pointer for easy access. + */ + #ifdef CONFIG_SMP +- GET_THREAD_INFO(%eax) +- movl TI_cpu(%eax), %eax +- movl __per_cpu_offset(,%eax,4), %eax +- mov per_cpu__xen_vcpu(%eax), %eax ++ push %fs ++ mov $(__KERNEL_PERCPU), %eax ++ mov %eax, %fs ++ mov PER_CPU_VAR(xen_vcpu), %eax ++ pop %fs + #else + movl per_cpu__xen_vcpu, %eax + #endif +diff --git a/arch/x86/xen/xen-head.S b/arch/x86/xen/xen-head.S +index 1a5ff24..a187d40 100644 +--- a/arch/x86/xen/xen-head.S ++++ b/arch/x86/xen/xen-head.S +@@ -19,6 +19,17 @@ ENTRY(startup_xen) + #ifdef CONFIG_X86_32 + mov %esi,xen_start_info + mov $init_thread_union+THREAD_SIZE,%esp ++#ifdef CONFIG_SMP ++ movl $cpu_gdt_table,%edi ++ movl $__per_cpu_load,%eax ++ movw %ax,__KERNEL_PERCPU + 2(%edi) ++ rorl $16,%eax ++ movb %al,__KERNEL_PERCPU + 4(%edi) ++ movb %ah,__KERNEL_PERCPU + 7(%edi) ++ movl $__per_cpu_end - 1,%eax ++ subl $__per_cpu_start,%eax ++ movw %ax,__KERNEL_PERCPU + 0(%edi) ++#endif + #else + mov %rsi,xen_start_info + mov $init_thread_union+THREAD_SIZE,%rsp +diff --git a/arch/x86/xen/xen-ops.h b/arch/x86/xen/xen-ops.h +index f9153a3..51eab3d 100644 +--- a/arch/x86/xen/xen-ops.h ++++ b/arch/x86/xen/xen-ops.h +@@ -10,8 +10,6 @@ + extern const char xen_hypervisor_callback[]; + extern const char xen_failsafe_callback[]; + +-extern void *xen_initial_gdt; +- + struct trap_info; + void xen_copy_trap_info(struct trap_info *traps); + +diff --git a/arch/xtensa/variants/dc232b/include/variant/core.h b/arch/xtensa/variants/dc232b/include/variant/core.h +index 525bd3d..ef888b1 100644 +--- a/arch/xtensa/variants/dc232b/include/variant/core.h ++++ b/arch/xtensa/variants/dc232b/include/variant/core.h +@@ -119,9 +119,9 @@ + ----------------------------------------------------------------------*/ + + #define XCHAL_ICACHE_LINESIZE 32 /* I-cache line size in bytes */ +-#define XCHAL_DCACHE_LINESIZE 32 /* D-cache line size in bytes */ + #define XCHAL_ICACHE_LINEWIDTH 5 /* log2(I line size in bytes) */ + #define XCHAL_DCACHE_LINEWIDTH 5 /* log2(D line size in bytes) */ ++#define XCHAL_DCACHE_LINESIZE (_AC(1,UL) << XCHAL_DCACHE_LINEWIDTH) /* D-cache line size in bytes */ + + #define XCHAL_ICACHE_SIZE 16384 /* I-cache size in bytes or 0 */ + #define XCHAL_DCACHE_SIZE 16384 /* D-cache size in bytes or 0 */ +diff --git a/arch/xtensa/variants/fsf/include/variant/core.h b/arch/xtensa/variants/fsf/include/variant/core.h +index 2f33760..835e50a 100644 +--- a/arch/xtensa/variants/fsf/include/variant/core.h ++++ b/arch/xtensa/variants/fsf/include/variant/core.h +@@ -11,6 +11,7 @@ + #ifndef _XTENSA_CORE_H + #define _XTENSA_CORE_H + ++#include + + /**************************************************************************** + Parameters Useful for Any Code, USER or PRIVILEGED +@@ -112,9 +113,9 @@ + ----------------------------------------------------------------------*/ + + #define XCHAL_ICACHE_LINESIZE 16 /* I-cache line size in bytes */ +-#define XCHAL_DCACHE_LINESIZE 16 /* D-cache line size in bytes */ + #define XCHAL_ICACHE_LINEWIDTH 4 /* log2(I line size in bytes) */ + #define XCHAL_DCACHE_LINEWIDTH 4 /* log2(D line size in bytes) */ ++#define XCHAL_DCACHE_LINESIZE (_AC(1,UL) << XCHAL_DCACHE_LINEWIDTH) /* D-cache line size in bytes */ + + #define XCHAL_ICACHE_SIZE 8192 /* I-cache size in bytes or 0 */ + #define XCHAL_DCACHE_SIZE 8192 /* D-cache size in bytes or 0 */ +diff --git a/arch/xtensa/variants/s6000/include/variant/core.h b/arch/xtensa/variants/s6000/include/variant/core.h +index af00795..2bb8105 100644 +--- a/arch/xtensa/variants/s6000/include/variant/core.h ++++ b/arch/xtensa/variants/s6000/include/variant/core.h +@@ -11,6 +11,7 @@ + #ifndef _XTENSA_CORE_CONFIGURATION_H + #define _XTENSA_CORE_CONFIGURATION_H + ++#include + + /**************************************************************************** + Parameters Useful for Any Code, USER or PRIVILEGED +@@ -118,9 +119,9 @@ + ----------------------------------------------------------------------*/ + + #define XCHAL_ICACHE_LINESIZE 16 /* I-cache line size in bytes */ +-#define XCHAL_DCACHE_LINESIZE 16 /* D-cache line size in bytes */ + #define XCHAL_ICACHE_LINEWIDTH 4 /* log2(I line size in bytes) */ + #define XCHAL_DCACHE_LINEWIDTH 4 /* log2(D line size in bytes) */ ++#define XCHAL_DCACHE_LINESIZE (_AC(1,UL) << XCHAL_DCACHE_LINEWIDTH) /* D-cache line size in bytes */ + + #define XCHAL_ICACHE_SIZE 32768 /* I-cache size in bytes or 0 */ + #define XCHAL_DCACHE_SIZE 32768 /* D-cache size in bytes or 0 */ +diff --git a/block/blk-integrity.c b/block/blk-integrity.c +index 15c6308..96e83c2 100644 +--- a/block/blk-integrity.c ++++ b/block/blk-integrity.c +@@ -278,7 +278,7 @@ static struct attribute *integrity_attrs[] = { + NULL, + }; + +-static struct sysfs_ops integrity_ops = { ++static const struct sysfs_ops integrity_ops = { + .show = &integrity_attr_show, + .store = &integrity_attr_store, + }; +diff --git a/block/blk-ioc.c b/block/blk-ioc.c +index d4ed600..cbdabb0 100644 +--- a/block/blk-ioc.c ++++ b/block/blk-ioc.c +@@ -66,22 +66,22 @@ static void cfq_exit(struct io_context *ioc) + } + + /* Called by the exitting task */ +-void exit_io_context(void) ++void exit_io_context(struct task_struct *task) + { + struct io_context *ioc; + +- task_lock(current); +- ioc = current->io_context; +- current->io_context = NULL; +- task_unlock(current); ++ task_lock(task); ++ ioc = task->io_context; ++ task->io_context = NULL; ++ task_unlock(task); + + if (atomic_dec_and_test(&ioc->nr_tasks)) { + if (ioc->aic && ioc->aic->exit) + ioc->aic->exit(ioc->aic); + cfq_exit(ioc); + +- put_io_context(ioc); + } ++ put_io_context(ioc); + } + + struct io_context *alloc_io_context(gfp_t gfp_flags, int node) +diff --git a/block/blk-iopoll.c b/block/blk-iopoll.c +index ca56420..f2fc409 100644 +--- a/block/blk-iopoll.c ++++ b/block/blk-iopoll.c +@@ -77,7 +77,7 @@ void blk_iopoll_complete(struct blk_iopoll *iopoll) + } + EXPORT_SYMBOL(blk_iopoll_complete); + +-static void blk_iopoll_softirq(struct softirq_action *h) ++static void blk_iopoll_softirq(void) + { + struct list_head *list = &__get_cpu_var(blk_cpu_iopoll); + int rearm = 0, budget = blk_iopoll_budget; +diff --git a/block/blk-map.c b/block/blk-map.c +index 30a7e51..0aeec6a 100644 +--- a/block/blk-map.c ++++ b/block/blk-map.c +@@ -54,7 +54,7 @@ static int __blk_rq_map_user(struct request_queue *q, struct request *rq, + * direct dma. else, set up kernel bounce buffers + */ + uaddr = (unsigned long) ubuf; +- if (blk_rq_aligned(q, ubuf, len) && !map_data) ++ if (blk_rq_aligned(q, (__force void *)ubuf, len) && !map_data) + bio = bio_map_user(q, NULL, uaddr, len, reading, gfp_mask); + else + bio = bio_copy_user(q, map_data, uaddr, len, reading, gfp_mask); +@@ -201,12 +201,13 @@ int blk_rq_map_user_iov(struct request_queue *q, struct request *rq, + for (i = 0; i < iov_count; i++) { + unsigned long uaddr = (unsigned long)iov[i].iov_base; + ++ if (!iov[i].iov_len) ++ return -EINVAL; ++ + if (uaddr & queue_dma_alignment(q)) { + unaligned = 1; + break; + } +- if (!iov[i].iov_len) +- return -EINVAL; + } + + if (unaligned || (q->dma_pad_mask & len) || map_data) +@@ -299,7 +300,7 @@ int blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf, + if (!len || !kbuf) + return -EINVAL; + +- do_copy = !blk_rq_aligned(q, kbuf, len) || object_is_on_stack(kbuf); ++ do_copy = !blk_rq_aligned(q, kbuf, len) || object_starts_on_stack(kbuf); + if (do_copy) + bio = bio_copy_kern(q, kbuf, len, gfp_mask, reading); + else +diff --git a/block/blk-softirq.c b/block/blk-softirq.c +index ee9c216..58d410a 100644 +--- a/block/blk-softirq.c ++++ b/block/blk-softirq.c +@@ -17,7 +17,7 @@ static DEFINE_PER_CPU(struct list_head, blk_cpu_done); + * Softirq action handler - move entries to local list and loop over them + * while passing them to the queue registered handler. + */ +-static void blk_done_softirq(struct softirq_action *h) ++static void blk_done_softirq(void) + { + struct list_head *cpu_list, local_list; + +diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c +index bb9c5ea..5330d48 100644 +--- a/block/blk-sysfs.c ++++ b/block/blk-sysfs.c +@@ -414,7 +414,7 @@ static void blk_release_queue(struct kobject *kobj) + kmem_cache_free(blk_requestq_cachep, q); + } + +-static struct sysfs_ops queue_sysfs_ops = { ++static const struct sysfs_ops queue_sysfs_ops = { + .show = queue_attr_show, + .store = queue_attr_store, + }; +diff --git a/block/bsg.c b/block/bsg.c +index 7154a7a..08ac2f0 100644 +--- a/block/bsg.c ++++ b/block/bsg.c +@@ -175,16 +175,24 @@ static int blk_fill_sgv4_hdr_rq(struct request_queue *q, struct request *rq, + struct sg_io_v4 *hdr, struct bsg_device *bd, + fmode_t has_write_perm) + { ++ unsigned char tmpcmd[sizeof(rq->__cmd)]; ++ unsigned char *cmdptr; ++ + if (hdr->request_len > BLK_MAX_CDB) { + rq->cmd = kzalloc(hdr->request_len, GFP_KERNEL); + if (!rq->cmd) + return -ENOMEM; +- } ++ cmdptr = rq->cmd; ++ } else ++ cmdptr = tmpcmd; + +- if (copy_from_user(rq->cmd, (void *)(unsigned long)hdr->request, ++ if (copy_from_user(cmdptr, (void __user *)(unsigned long)hdr->request, + hdr->request_len)) + return -EFAULT; + ++ if (cmdptr != rq->cmd) ++ memcpy(rq->cmd, cmdptr, hdr->request_len); ++ + if (hdr->subprotocol == BSG_SUB_PROTOCOL_SCSI_CMD) { + if (blk_verify_command(rq->cmd, has_write_perm)) + return -EPERM; +@@ -282,7 +290,7 @@ bsg_map_hdr(struct bsg_device *bd, struct sg_io_v4 *hdr, fmode_t has_write_perm, + rq->next_rq = next_rq; + next_rq->cmd_type = rq->cmd_type; + +- dxferp = (void*)(unsigned long)hdr->din_xferp; ++ dxferp = (void __user *)(unsigned long)hdr->din_xferp; + ret = blk_rq_map_user(q, next_rq, NULL, dxferp, + hdr->din_xfer_len, GFP_KERNEL); + if (ret) +@@ -291,10 +299,10 @@ bsg_map_hdr(struct bsg_device *bd, struct sg_io_v4 *hdr, fmode_t has_write_perm, + + if (hdr->dout_xfer_len) { + dxfer_len = hdr->dout_xfer_len; +- dxferp = (void*)(unsigned long)hdr->dout_xferp; ++ dxferp = (void __user *)(unsigned long)hdr->dout_xferp; + } else if (hdr->din_xfer_len) { + dxfer_len = hdr->din_xfer_len; +- dxferp = (void*)(unsigned long)hdr->din_xferp; ++ dxferp = (void __user *)(unsigned long)hdr->din_xferp; + } else + dxfer_len = 0; + +@@ -436,7 +444,7 @@ static int blk_complete_sgv4_hdr_rq(struct request *rq, struct sg_io_v4 *hdr, + int len = min_t(unsigned int, hdr->max_response_len, + rq->sense_len); + +- ret = copy_to_user((void*)(unsigned long)hdr->response, ++ ret = copy_to_user((void __user *)(unsigned long)hdr->response, + rq->sense, len); + if (!ret) + hdr->response_len = len; +diff --git a/block/compat_ioctl.c b/block/compat_ioctl.c +index 9bd086c..ca1fc22 100644 +--- a/block/compat_ioctl.c ++++ b/block/compat_ioctl.c +@@ -354,7 +354,7 @@ static int compat_fd_ioctl(struct block_device *bdev, fmode_t mode, + err |= __get_user(f->spec1, &uf->spec1); + err |= __get_user(f->fmt_gap, &uf->fmt_gap); + err |= __get_user(name, &uf->name); +- f->name = compat_ptr(name); ++ f->name = (void __force_kernel *)compat_ptr(name); + if (err) { + err = -EFAULT; + goto out; +diff --git a/block/elevator.c b/block/elevator.c +index a847046..75a1746 100644 +--- a/block/elevator.c ++++ b/block/elevator.c +@@ -889,7 +889,7 @@ elv_attr_store(struct kobject *kobj, struct attribute *attr, + return error; + } + +-static struct sysfs_ops elv_sysfs_ops = { ++static const struct sysfs_ops elv_sysfs_ops = { + .show = elv_attr_show, + .store = elv_attr_store, + }; +diff --git a/block/scsi_ioctl.c b/block/scsi_ioctl.c +index 2be0a97..bded3fd 100644 +--- a/block/scsi_ioctl.c ++++ b/block/scsi_ioctl.c +@@ -221,8 +221,20 @@ EXPORT_SYMBOL(blk_verify_command); + static int blk_fill_sghdr_rq(struct request_queue *q, struct request *rq, + struct sg_io_hdr *hdr, fmode_t mode) + { +- if (copy_from_user(rq->cmd, hdr->cmdp, hdr->cmd_len)) ++ unsigned char tmpcmd[sizeof(rq->__cmd)]; ++ unsigned char *cmdptr; ++ ++ if (rq->cmd != rq->__cmd) ++ cmdptr = rq->cmd; ++ else ++ cmdptr = tmpcmd; ++ ++ if (copy_from_user(cmdptr, hdr->cmdp, hdr->cmd_len)) + return -EFAULT; ++ ++ if (cmdptr != rq->cmd) ++ memcpy(rq->cmd, cmdptr, hdr->cmd_len); ++ + if (blk_verify_command(rq->cmd, mode & FMODE_WRITE)) + return -EPERM; + +@@ -431,6 +443,8 @@ int sg_scsi_ioctl(struct request_queue *q, struct gendisk *disk, fmode_t mode, + int err; + unsigned int in_len, out_len, bytes, opcode, cmdlen; + char *buffer = NULL, sense[SCSI_SENSE_BUFFERSIZE]; ++ unsigned char tmpcmd[sizeof(rq->__cmd)]; ++ unsigned char *cmdptr; + + if (!sic) + return -EINVAL; +@@ -464,9 +478,18 @@ int sg_scsi_ioctl(struct request_queue *q, struct gendisk *disk, fmode_t mode, + */ + err = -EFAULT; + rq->cmd_len = cmdlen; +- if (copy_from_user(rq->cmd, sic->data, cmdlen)) ++ ++ if (rq->cmd != rq->__cmd) ++ cmdptr = rq->cmd; ++ else ++ cmdptr = tmpcmd; ++ ++ if (copy_from_user(cmdptr, sic->data, cmdlen)) + goto error; + ++ if (rq->cmd != cmdptr) ++ memcpy(rq->cmd, cmdptr, cmdlen); ++ + if (in_len && copy_from_user(buffer, sic->data + cmdlen, in_len)) + goto error; + +diff --git a/crypto/cryptd.c b/crypto/cryptd.c +index 3533582..f143117 100644 +--- a/crypto/cryptd.c ++++ b/crypto/cryptd.c +@@ -50,7 +50,7 @@ struct cryptd_blkcipher_ctx { + + struct cryptd_blkcipher_request_ctx { + crypto_completion_t complete; +-}; ++} __no_const; + + struct cryptd_hash_ctx { + struct crypto_shash *child; +diff --git a/crypto/gf128mul.c b/crypto/gf128mul.c +index a90d260..7a9765e 100644 +--- a/crypto/gf128mul.c ++++ b/crypto/gf128mul.c +@@ -182,7 +182,7 @@ void gf128mul_lle(be128 *r, const be128 *b) + for (i = 0; i < 7; ++i) + gf128mul_x_lle(&p[i + 1], &p[i]); + +- memset(r, 0, sizeof(r)); ++ memset(r, 0, sizeof(*r)); + for (i = 0;;) { + u8 ch = ((u8 *)b)[15 - i]; + +@@ -220,7 +220,7 @@ void gf128mul_bbe(be128 *r, const be128 *b) + for (i = 0; i < 7; ++i) + gf128mul_x_bbe(&p[i + 1], &p[i]); + +- memset(r, 0, sizeof(r)); ++ memset(r, 0, sizeof(*r)); + for (i = 0;;) { + u8 ch = ((u8 *)b)[i]; + +diff --git a/crypto/serpent.c b/crypto/serpent.c +index b651a55..023297d 100644 +--- a/crypto/serpent.c ++++ b/crypto/serpent.c +@@ -21,6 +21,7 @@ + #include + #include + #include ++#include + + /* Key is padded to the maximum of 256 bits before round key generation. + * Any key length <= 256 bits (32 bytes) is allowed by the algorithm. +@@ -224,6 +225,8 @@ static int serpent_setkey(struct crypto_tfm *tfm, const u8 *key, + u32 r0,r1,r2,r3,r4; + int i; + ++ pax_track_stack(); ++ + /* Copy key, add padding */ + + for (i = 0; i < keylen; ++i) +diff --git a/drivers/acpi/acpi_pad.c b/drivers/acpi/acpi_pad.c +index 0d2cdb8..d8de48d 100644 +--- a/drivers/acpi/acpi_pad.c ++++ b/drivers/acpi/acpi_pad.c +@@ -30,7 +30,7 @@ + #include + #include + +-#define ACPI_PROCESSOR_AGGREGATOR_CLASS "processor_aggregator" ++#define ACPI_PROCESSOR_AGGREGATOR_CLASS "acpi_pad" + #define ACPI_PROCESSOR_AGGREGATOR_DEVICE_NAME "Processor Aggregator" + #define ACPI_PROCESSOR_AGGREGATOR_NOTIFY 0x80 + static DEFINE_MUTEX(isolated_cpus_lock); +diff --git a/drivers/acpi/battery.c b/drivers/acpi/battery.c +index 3f4602b..2e41d36 100644 +--- a/drivers/acpi/battery.c ++++ b/drivers/acpi/battery.c +@@ -763,7 +763,7 @@ DECLARE_FILE_FUNCTIONS(alarm); + } + + static struct battery_file { +- struct file_operations ops; ++ const struct file_operations ops; + mode_t mode; + const char *name; + } acpi_battery_file[] = { +diff --git a/drivers/acpi/dock.c b/drivers/acpi/dock.c +index 7338b6a..82f0257 100644 +--- a/drivers/acpi/dock.c ++++ b/drivers/acpi/dock.c +@@ -77,7 +77,7 @@ struct dock_dependent_device { + struct list_head list; + struct list_head hotplug_list; + acpi_handle handle; +- struct acpi_dock_ops *ops; ++ const struct acpi_dock_ops *ops; + void *context; + }; + +@@ -605,7 +605,7 @@ EXPORT_SYMBOL_GPL(unregister_dock_notifier); + * the dock driver after _DCK is executed. + */ + int +-register_hotplug_dock_device(acpi_handle handle, struct acpi_dock_ops *ops, ++register_hotplug_dock_device(acpi_handle handle, const struct acpi_dock_ops *ops, + void *context) + { + struct dock_dependent_device *dd; +diff --git a/drivers/acpi/osl.c b/drivers/acpi/osl.c +index 7c1c59e..2993595 100644 +--- a/drivers/acpi/osl.c ++++ b/drivers/acpi/osl.c +@@ -523,6 +523,8 @@ acpi_os_read_memory(acpi_physical_address phys_addr, u32 * value, u32 width) + void __iomem *virt_addr; + + virt_addr = ioremap(phys_addr, width); ++ if (!virt_addr) ++ return AE_NO_MEMORY; + if (!value) + value = &dummy; + +@@ -551,6 +553,8 @@ acpi_os_write_memory(acpi_physical_address phys_addr, u32 value, u32 width) + void __iomem *virt_addr; + + virt_addr = ioremap(phys_addr, width); ++ if (!virt_addr) ++ return AE_NO_MEMORY; + + switch (width) { + case 8: +diff --git a/drivers/acpi/power_meter.c b/drivers/acpi/power_meter.c +index c216062..eec10d2 100644 +--- a/drivers/acpi/power_meter.c ++++ b/drivers/acpi/power_meter.c +@@ -315,8 +315,6 @@ static ssize_t set_trip(struct device *dev, struct device_attribute *devattr, + return res; + + temp /= 1000; +- if (temp < 0) +- return -EINVAL; + + mutex_lock(&resource->lock); + resource->trip[attr->index - 7] = temp; +diff --git a/drivers/acpi/proc.c b/drivers/acpi/proc.c +index d0d25e2..961643d 100644 +--- a/drivers/acpi/proc.c ++++ b/drivers/acpi/proc.c +@@ -391,20 +391,15 @@ acpi_system_write_wakeup_device(struct file *file, + size_t count, loff_t * ppos) + { + struct list_head *node, *next; +- char strbuf[5]; +- char str[5] = ""; +- unsigned int len = count; ++ char strbuf[5] = {0}; + struct acpi_device *found_dev = NULL; + +- if (len > 4) +- len = 4; +- if (len < 0) +- return -EFAULT; ++ if (count > 4) ++ count = 4; + +- if (copy_from_user(strbuf, buffer, len)) ++ if (copy_from_user(strbuf, buffer, count)) + return -EFAULT; +- strbuf[len] = '\0'; +- sscanf(strbuf, "%s", str); ++ strbuf[count] = '\0'; + + mutex_lock(&acpi_device_lock); + list_for_each_safe(node, next, &acpi_wakeup_device_list) { +@@ -413,7 +408,7 @@ acpi_system_write_wakeup_device(struct file *file, + if (!dev->wakeup.flags.valid) + continue; + +- if (!strncmp(dev->pnp.bus_id, str, 4)) { ++ if (!strncmp(dev->pnp.bus_id, strbuf, 4)) { + dev->wakeup.state.enabled = + dev->wakeup.state.enabled ? 0 : 1; + found_dev = dev; +diff --git a/drivers/acpi/processor_core.c b/drivers/acpi/processor_core.c +index 7102474..de8ad22 100644 +--- a/drivers/acpi/processor_core.c ++++ b/drivers/acpi/processor_core.c +@@ -790,7 +790,7 @@ static int __cpuinit acpi_processor_add(struct acpi_device *device) + return 0; + } + +- BUG_ON((pr->id >= nr_cpu_ids) || (pr->id < 0)); ++ BUG_ON(pr->id >= nr_cpu_ids); + + /* + * Buggy BIOS check +diff --git a/drivers/acpi/sbshc.c b/drivers/acpi/sbshc.c +index d933980..5761f13 100644 +--- a/drivers/acpi/sbshc.c ++++ b/drivers/acpi/sbshc.c +@@ -17,7 +17,7 @@ + + #define PREFIX "ACPI: " + +-#define ACPI_SMB_HC_CLASS "smbus_host_controller" ++#define ACPI_SMB_HC_CLASS "smbus_host_ctl" + #define ACPI_SMB_HC_DEVICE_NAME "ACPI SMBus HC" + + struct acpi_smb_hc { +diff --git a/drivers/acpi/sleep.c b/drivers/acpi/sleep.c +index 0458094..6978e7b 100644 +--- a/drivers/acpi/sleep.c ++++ b/drivers/acpi/sleep.c +@@ -283,7 +283,7 @@ static int acpi_suspend_state_valid(suspend_state_t pm_state) + } + } + +-static struct platform_suspend_ops acpi_suspend_ops = { ++static const struct platform_suspend_ops acpi_suspend_ops = { + .valid = acpi_suspend_state_valid, + .begin = acpi_suspend_begin, + .prepare_late = acpi_pm_prepare, +@@ -311,7 +311,7 @@ static int acpi_suspend_begin_old(suspend_state_t pm_state) + * The following callbacks are used if the pre-ACPI 2.0 suspend ordering has + * been requested. + */ +-static struct platform_suspend_ops acpi_suspend_ops_old = { ++static const struct platform_suspend_ops acpi_suspend_ops_old = { + .valid = acpi_suspend_state_valid, + .begin = acpi_suspend_begin_old, + .prepare_late = acpi_pm_disable_gpes, +@@ -460,7 +460,7 @@ static void acpi_pm_enable_gpes(void) + acpi_enable_all_runtime_gpes(); + } + +-static struct platform_hibernation_ops acpi_hibernation_ops = { ++static const struct platform_hibernation_ops acpi_hibernation_ops = { + .begin = acpi_hibernation_begin, + .end = acpi_pm_end, + .pre_snapshot = acpi_hibernation_pre_snapshot, +@@ -513,7 +513,7 @@ static int acpi_hibernation_pre_snapshot_old(void) + * The following callbacks are used if the pre-ACPI 2.0 suspend ordering has + * been requested. + */ +-static struct platform_hibernation_ops acpi_hibernation_ops_old = { ++static const struct platform_hibernation_ops acpi_hibernation_ops_old = { + .begin = acpi_hibernation_begin_old, + .end = acpi_pm_end, + .pre_snapshot = acpi_hibernation_pre_snapshot_old, +diff --git a/drivers/acpi/video.c b/drivers/acpi/video.c +index 05dff63..b662ab7 100644 +--- a/drivers/acpi/video.c ++++ b/drivers/acpi/video.c +@@ -359,7 +359,7 @@ static int acpi_video_set_brightness(struct backlight_device *bd) + vd->brightness->levels[request_level]); + } + +-static struct backlight_ops acpi_backlight_ops = { ++static const struct backlight_ops acpi_backlight_ops = { + .get_brightness = acpi_video_get_brightness, + .update_status = acpi_video_set_brightness, + }; +diff --git a/drivers/ata/ahci.c b/drivers/ata/ahci.c +index 6787aab..23ffb0e 100644 +--- a/drivers/ata/ahci.c ++++ b/drivers/ata/ahci.c +@@ -387,7 +387,7 @@ static struct scsi_host_template ahci_sht = { + .sdev_attrs = ahci_sdev_attrs, + }; + +-static struct ata_port_operations ahci_ops = { ++static const struct ata_port_operations ahci_ops = { + .inherits = &sata_pmp_port_ops, + + .qc_defer = sata_pmp_qc_defer_cmd_switch, +@@ -424,17 +424,17 @@ static struct ata_port_operations ahci_ops = { + .port_stop = ahci_port_stop, + }; + +-static struct ata_port_operations ahci_vt8251_ops = { ++static const struct ata_port_operations ahci_vt8251_ops = { + .inherits = &ahci_ops, + .hardreset = ahci_vt8251_hardreset, + }; + +-static struct ata_port_operations ahci_p5wdh_ops = { ++static const struct ata_port_operations ahci_p5wdh_ops = { + .inherits = &ahci_ops, + .hardreset = ahci_p5wdh_hardreset, + }; + +-static struct ata_port_operations ahci_sb600_ops = { ++static const struct ata_port_operations ahci_sb600_ops = { + .inherits = &ahci_ops, + .softreset = ahci_sb600_softreset, + .pmp_softreset = ahci_sb600_softreset, +diff --git a/drivers/ata/ata_generic.c b/drivers/ata/ata_generic.c +index 99e7196..4968c77 100644 +--- a/drivers/ata/ata_generic.c ++++ b/drivers/ata/ata_generic.c +@@ -104,7 +104,7 @@ static struct scsi_host_template generic_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations generic_port_ops = { ++static const struct ata_port_operations generic_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_unknown, + .set_mode = generic_set_mode, +diff --git a/drivers/ata/ata_piix.c b/drivers/ata/ata_piix.c +index c33591d..000c121 100644 +--- a/drivers/ata/ata_piix.c ++++ b/drivers/ata/ata_piix.c +@@ -318,7 +318,7 @@ static struct scsi_host_template piix_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations piix_pata_ops = { ++static const struct ata_port_operations piix_pata_ops = { + .inherits = &ata_bmdma32_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = piix_set_piomode, +@@ -326,22 +326,22 @@ static struct ata_port_operations piix_pata_ops = { + .prereset = piix_pata_prereset, + }; + +-static struct ata_port_operations piix_vmw_ops = { ++static const struct ata_port_operations piix_vmw_ops = { + .inherits = &piix_pata_ops, + .bmdma_status = piix_vmw_bmdma_status, + }; + +-static struct ata_port_operations ich_pata_ops = { ++static const struct ata_port_operations ich_pata_ops = { + .inherits = &piix_pata_ops, + .cable_detect = ich_pata_cable_detect, + .set_dmamode = ich_set_dmamode, + }; + +-static struct ata_port_operations piix_sata_ops = { ++static const struct ata_port_operations piix_sata_ops = { + .inherits = &ata_bmdma_port_ops, + }; + +-static struct ata_port_operations piix_sidpr_sata_ops = { ++static const struct ata_port_operations piix_sidpr_sata_ops = { + .inherits = &piix_sata_ops, + .hardreset = sata_std_hardreset, + .scr_read = piix_sidpr_scr_read, +diff --git a/drivers/ata/libata-acpi.c b/drivers/ata/libata-acpi.c +index b0882cd..c295d65 100644 +--- a/drivers/ata/libata-acpi.c ++++ b/drivers/ata/libata-acpi.c +@@ -223,12 +223,12 @@ static void ata_acpi_dev_uevent(acpi_handle handle, u32 event, void *data) + ata_acpi_uevent(dev->link->ap, dev, event); + } + +-static struct acpi_dock_ops ata_acpi_dev_dock_ops = { ++static const struct acpi_dock_ops ata_acpi_dev_dock_ops = { + .handler = ata_acpi_dev_notify_dock, + .uevent = ata_acpi_dev_uevent, + }; + +-static struct acpi_dock_ops ata_acpi_ap_dock_ops = { ++static const struct acpi_dock_ops ata_acpi_ap_dock_ops = { + .handler = ata_acpi_ap_notify_dock, + .uevent = ata_acpi_ap_uevent, + }; +diff --git a/drivers/ata/libata-core.c b/drivers/ata/libata-core.c +index d4f7f99..94f603e 100644 +--- a/drivers/ata/libata-core.c ++++ b/drivers/ata/libata-core.c +@@ -4954,7 +4954,7 @@ void ata_qc_free(struct ata_queued_cmd *qc) + struct ata_port *ap; + unsigned int tag; + +- WARN_ON_ONCE(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ ++ BUG_ON(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ + ap = qc->ap; + + qc->flags = 0; +@@ -4970,7 +4970,7 @@ void __ata_qc_complete(struct ata_queued_cmd *qc) + struct ata_port *ap; + struct ata_link *link; + +- WARN_ON_ONCE(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ ++ BUG_ON(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ + WARN_ON_ONCE(!(qc->flags & ATA_QCFLAG_ACTIVE)); + ap = qc->ap; + link = qc->dev->link; +@@ -5987,7 +5987,7 @@ static void ata_host_stop(struct device *gendev, void *res) + * LOCKING: + * None. + */ +-static void ata_finalize_port_ops(struct ata_port_operations *ops) ++static void ata_finalize_port_ops(const struct ata_port_operations *ops) + { + static DEFINE_SPINLOCK(lock); + const struct ata_port_operations *cur; +@@ -5999,6 +5999,7 @@ static void ata_finalize_port_ops(struct ata_port_operations *ops) + return; + + spin_lock(&lock); ++ pax_open_kernel(); + + for (cur = ops->inherits; cur; cur = cur->inherits) { + void **inherit = (void **)cur; +@@ -6012,8 +6013,9 @@ static void ata_finalize_port_ops(struct ata_port_operations *ops) + if (IS_ERR(*pp)) + *pp = NULL; + +- ops->inherits = NULL; ++ *(struct ata_port_operations **)&ops->inherits = NULL; + ++ pax_close_kernel(); + spin_unlock(&lock); + } + +@@ -6110,7 +6112,7 @@ int ata_host_start(struct ata_host *host) + */ + /* KILLME - the only user left is ipr */ + void ata_host_init(struct ata_host *host, struct device *dev, +- unsigned long flags, struct ata_port_operations *ops) ++ unsigned long flags, const struct ata_port_operations *ops) + { + spin_lock_init(&host->lock); + host->dev = dev; +@@ -6773,7 +6775,7 @@ static void ata_dummy_error_handler(struct ata_port *ap) + /* truly dummy */ + } + +-struct ata_port_operations ata_dummy_port_ops = { ++const struct ata_port_operations ata_dummy_port_ops = { + .qc_prep = ata_noop_qc_prep, + .qc_issue = ata_dummy_qc_issue, + .error_handler = ata_dummy_error_handler, +diff --git a/drivers/ata/libata-eh.c b/drivers/ata/libata-eh.c +index e5bdb9b..45a8e72 100644 +--- a/drivers/ata/libata-eh.c ++++ b/drivers/ata/libata-eh.c +@@ -2423,6 +2423,8 @@ void ata_eh_report(struct ata_port *ap) + { + struct ata_link *link; + ++ pax_track_stack(); ++ + ata_for_each_link(link, ap, HOST_FIRST) + ata_eh_link_report(link); + } +@@ -3594,7 +3596,7 @@ void ata_do_eh(struct ata_port *ap, ata_prereset_fn_t prereset, + */ + void ata_std_error_handler(struct ata_port *ap) + { +- struct ata_port_operations *ops = ap->ops; ++ const struct ata_port_operations *ops = ap->ops; + ata_reset_fn_t hardreset = ops->hardreset; + + /* ignore built-in hardreset if SCR access is not available */ +diff --git a/drivers/ata/libata-pmp.c b/drivers/ata/libata-pmp.c +index 51f0ffb..19ce3e3 100644 +--- a/drivers/ata/libata-pmp.c ++++ b/drivers/ata/libata-pmp.c +@@ -841,7 +841,7 @@ static int sata_pmp_handle_link_fail(struct ata_link *link, int *link_tries) + */ + static int sata_pmp_eh_recover(struct ata_port *ap) + { +- struct ata_port_operations *ops = ap->ops; ++ const struct ata_port_operations *ops = ap->ops; + int pmp_tries, link_tries[SATA_PMP_MAX_PORTS]; + struct ata_link *pmp_link = &ap->link; + struct ata_device *pmp_dev = pmp_link->device; +diff --git a/drivers/ata/pata_acpi.c b/drivers/ata/pata_acpi.c +index d8f35fe..288180a 100644 +--- a/drivers/ata/pata_acpi.c ++++ b/drivers/ata/pata_acpi.c +@@ -215,7 +215,7 @@ static struct scsi_host_template pacpi_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pacpi_ops = { ++static const struct ata_port_operations pacpi_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_issue = pacpi_qc_issue, + .cable_detect = pacpi_cable_detect, +diff --git a/drivers/ata/pata_ali.c b/drivers/ata/pata_ali.c +index 9434114..1f2f364 100644 +--- a/drivers/ata/pata_ali.c ++++ b/drivers/ata/pata_ali.c +@@ -365,7 +365,7 @@ static struct scsi_host_template ali_sht = { + * Port operations for PIO only ALi + */ + +-static struct ata_port_operations ali_early_port_ops = { ++static const struct ata_port_operations ali_early_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = ali_set_piomode, +@@ -382,7 +382,7 @@ static const struct ata_port_operations ali_dma_base_ops = { + * Port operations for DMA capable ALi without cable + * detect + */ +-static struct ata_port_operations ali_20_port_ops = { ++static const struct ata_port_operations ali_20_port_ops = { + .inherits = &ali_dma_base_ops, + .cable_detect = ata_cable_40wire, + .mode_filter = ali_20_filter, +@@ -393,7 +393,7 @@ static struct ata_port_operations ali_20_port_ops = { + /* + * Port operations for DMA capable ALi with cable detect + */ +-static struct ata_port_operations ali_c2_port_ops = { ++static const struct ata_port_operations ali_c2_port_ops = { + .inherits = &ali_dma_base_ops, + .check_atapi_dma = ali_check_atapi_dma, + .cable_detect = ali_c2_cable_detect, +@@ -404,7 +404,7 @@ static struct ata_port_operations ali_c2_port_ops = { + /* + * Port operations for DMA capable ALi with cable detect + */ +-static struct ata_port_operations ali_c4_port_ops = { ++static const struct ata_port_operations ali_c4_port_ops = { + .inherits = &ali_dma_base_ops, + .check_atapi_dma = ali_check_atapi_dma, + .cable_detect = ali_c2_cable_detect, +@@ -414,7 +414,7 @@ static struct ata_port_operations ali_c4_port_ops = { + /* + * Port operations for DMA capable ALi with cable detect and LBA48 + */ +-static struct ata_port_operations ali_c5_port_ops = { ++static const struct ata_port_operations ali_c5_port_ops = { + .inherits = &ali_dma_base_ops, + .check_atapi_dma = ali_check_atapi_dma, + .dev_config = ali_warn_atapi_dma, +diff --git a/drivers/ata/pata_amd.c b/drivers/ata/pata_amd.c +index 567f3f7..c8ee0da 100644 +--- a/drivers/ata/pata_amd.c ++++ b/drivers/ata/pata_amd.c +@@ -397,28 +397,28 @@ static const struct ata_port_operations amd_base_port_ops = { + .prereset = amd_pre_reset, + }; + +-static struct ata_port_operations amd33_port_ops = { ++static const struct ata_port_operations amd33_port_ops = { + .inherits = &amd_base_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = amd33_set_piomode, + .set_dmamode = amd33_set_dmamode, + }; + +-static struct ata_port_operations amd66_port_ops = { ++static const struct ata_port_operations amd66_port_ops = { + .inherits = &amd_base_port_ops, + .cable_detect = ata_cable_unknown, + .set_piomode = amd66_set_piomode, + .set_dmamode = amd66_set_dmamode, + }; + +-static struct ata_port_operations amd100_port_ops = { ++static const struct ata_port_operations amd100_port_ops = { + .inherits = &amd_base_port_ops, + .cable_detect = ata_cable_unknown, + .set_piomode = amd100_set_piomode, + .set_dmamode = amd100_set_dmamode, + }; + +-static struct ata_port_operations amd133_port_ops = { ++static const struct ata_port_operations amd133_port_ops = { + .inherits = &amd_base_port_ops, + .cable_detect = amd_cable_detect, + .set_piomode = amd133_set_piomode, +@@ -433,13 +433,13 @@ static const struct ata_port_operations nv_base_port_ops = { + .host_stop = nv_host_stop, + }; + +-static struct ata_port_operations nv100_port_ops = { ++static const struct ata_port_operations nv100_port_ops = { + .inherits = &nv_base_port_ops, + .set_piomode = nv100_set_piomode, + .set_dmamode = nv100_set_dmamode, + }; + +-static struct ata_port_operations nv133_port_ops = { ++static const struct ata_port_operations nv133_port_ops = { + .inherits = &nv_base_port_ops, + .set_piomode = nv133_set_piomode, + .set_dmamode = nv133_set_dmamode, +diff --git a/drivers/ata/pata_artop.c b/drivers/ata/pata_artop.c +index d332cfd..4b7eaae 100644 +--- a/drivers/ata/pata_artop.c ++++ b/drivers/ata/pata_artop.c +@@ -311,7 +311,7 @@ static struct scsi_host_template artop_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations artop6210_ops = { ++static const struct ata_port_operations artop6210_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = artop6210_set_piomode, +@@ -320,7 +320,7 @@ static struct ata_port_operations artop6210_ops = { + .qc_defer = artop6210_qc_defer, + }; + +-static struct ata_port_operations artop6260_ops = { ++static const struct ata_port_operations artop6260_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = artop6260_cable_detect, + .set_piomode = artop6260_set_piomode, +diff --git a/drivers/ata/pata_at32.c b/drivers/ata/pata_at32.c +index 5c129f9..7bb7ccb 100644 +--- a/drivers/ata/pata_at32.c ++++ b/drivers/ata/pata_at32.c +@@ -172,7 +172,7 @@ static struct scsi_host_template at32_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations at32_port_ops = { ++static const struct ata_port_operations at32_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = pata_at32_set_piomode, +diff --git a/drivers/ata/pata_at91.c b/drivers/ata/pata_at91.c +index 41c94b1..829006d 100644 +--- a/drivers/ata/pata_at91.c ++++ b/drivers/ata/pata_at91.c +@@ -195,7 +195,7 @@ static struct scsi_host_template pata_at91_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pata_at91_port_ops = { ++static const struct ata_port_operations pata_at91_port_ops = { + .inherits = &ata_sff_port_ops, + + .sff_data_xfer = pata_at91_data_xfer_noirq, +diff --git a/drivers/ata/pata_atiixp.c b/drivers/ata/pata_atiixp.c +index ae4454d..d391eb4 100644 +--- a/drivers/ata/pata_atiixp.c ++++ b/drivers/ata/pata_atiixp.c +@@ -205,7 +205,7 @@ static struct scsi_host_template atiixp_sht = { + .sg_tablesize = LIBATA_DUMB_MAX_PRD, + }; + +-static struct ata_port_operations atiixp_port_ops = { ++static const struct ata_port_operations atiixp_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .qc_prep = ata_sff_dumb_qc_prep, +diff --git a/drivers/ata/pata_atp867x.c b/drivers/ata/pata_atp867x.c +index 6fe7ded..2a425dc 100644 +--- a/drivers/ata/pata_atp867x.c ++++ b/drivers/ata/pata_atp867x.c +@@ -274,7 +274,7 @@ static struct scsi_host_template atp867x_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations atp867x_ops = { ++static const struct ata_port_operations atp867x_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = atp867x_cable_detect, + .set_piomode = atp867x_set_piomode, +diff --git a/drivers/ata/pata_bf54x.c b/drivers/ata/pata_bf54x.c +index c4b47a3..b27a367 100644 +--- a/drivers/ata/pata_bf54x.c ++++ b/drivers/ata/pata_bf54x.c +@@ -1464,7 +1464,7 @@ static struct scsi_host_template bfin_sht = { + .dma_boundary = ATA_DMA_BOUNDARY, + }; + +-static struct ata_port_operations bfin_pata_ops = { ++static const struct ata_port_operations bfin_pata_ops = { + .inherits = &ata_sff_port_ops, + + .set_piomode = bfin_set_piomode, +diff --git a/drivers/ata/pata_cmd640.c b/drivers/ata/pata_cmd640.c +index 5acf9fa..84248be 100644 +--- a/drivers/ata/pata_cmd640.c ++++ b/drivers/ata/pata_cmd640.c +@@ -168,7 +168,7 @@ static struct scsi_host_template cmd640_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations cmd640_port_ops = { ++static const struct ata_port_operations cmd640_port_ops = { + .inherits = &ata_bmdma_port_ops, + /* In theory xfer_noirq is not needed once we kill the prefetcher */ + .sff_data_xfer = ata_sff_data_xfer_noirq, +diff --git a/drivers/ata/pata_cmd64x.c b/drivers/ata/pata_cmd64x.c +index ccd2694..c869c3d 100644 +--- a/drivers/ata/pata_cmd64x.c ++++ b/drivers/ata/pata_cmd64x.c +@@ -271,18 +271,18 @@ static const struct ata_port_operations cmd64x_base_ops = { + .set_dmamode = cmd64x_set_dmamode, + }; + +-static struct ata_port_operations cmd64x_port_ops = { ++static const struct ata_port_operations cmd64x_port_ops = { + .inherits = &cmd64x_base_ops, + .cable_detect = ata_cable_40wire, + }; + +-static struct ata_port_operations cmd646r1_port_ops = { ++static const struct ata_port_operations cmd646r1_port_ops = { + .inherits = &cmd64x_base_ops, + .bmdma_stop = cmd646r1_bmdma_stop, + .cable_detect = ata_cable_40wire, + }; + +-static struct ata_port_operations cmd648_port_ops = { ++static const struct ata_port_operations cmd648_port_ops = { + .inherits = &cmd64x_base_ops, + .bmdma_stop = cmd648_bmdma_stop, + .cable_detect = cmd648_cable_detect, +diff --git a/drivers/ata/pata_cs5520.c b/drivers/ata/pata_cs5520.c +index 0df83cf..d7595b0 100644 +--- a/drivers/ata/pata_cs5520.c ++++ b/drivers/ata/pata_cs5520.c +@@ -144,7 +144,7 @@ static struct scsi_host_template cs5520_sht = { + .sg_tablesize = LIBATA_DUMB_MAX_PRD, + }; + +-static struct ata_port_operations cs5520_port_ops = { ++static const struct ata_port_operations cs5520_port_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_prep = ata_sff_dumb_qc_prep, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_cs5530.c b/drivers/ata/pata_cs5530.c +index c974b05..6d26b11 100644 +--- a/drivers/ata/pata_cs5530.c ++++ b/drivers/ata/pata_cs5530.c +@@ -164,7 +164,7 @@ static struct scsi_host_template cs5530_sht = { + .sg_tablesize = LIBATA_DUMB_MAX_PRD, + }; + +-static struct ata_port_operations cs5530_port_ops = { ++static const struct ata_port_operations cs5530_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .qc_prep = ata_sff_dumb_qc_prep, +diff --git a/drivers/ata/pata_cs5535.c b/drivers/ata/pata_cs5535.c +index 403f561..aacd26b 100644 +--- a/drivers/ata/pata_cs5535.c ++++ b/drivers/ata/pata_cs5535.c +@@ -160,7 +160,7 @@ static struct scsi_host_template cs5535_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations cs5535_port_ops = { ++static const struct ata_port_operations cs5535_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = cs5535_cable_detect, + .set_piomode = cs5535_set_piomode, +diff --git a/drivers/ata/pata_cs5536.c b/drivers/ata/pata_cs5536.c +index 6da4cb4..de24a25 100644 +--- a/drivers/ata/pata_cs5536.c ++++ b/drivers/ata/pata_cs5536.c +@@ -223,7 +223,7 @@ static struct scsi_host_template cs5536_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations cs5536_port_ops = { ++static const struct ata_port_operations cs5536_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = cs5536_cable_detect, + .set_piomode = cs5536_set_piomode, +diff --git a/drivers/ata/pata_cypress.c b/drivers/ata/pata_cypress.c +index 8fb040b..b16a9c9 100644 +--- a/drivers/ata/pata_cypress.c ++++ b/drivers/ata/pata_cypress.c +@@ -113,7 +113,7 @@ static struct scsi_host_template cy82c693_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations cy82c693_port_ops = { ++static const struct ata_port_operations cy82c693_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = cy82c693_set_piomode, +diff --git a/drivers/ata/pata_efar.c b/drivers/ata/pata_efar.c +index 2a6412f..555ee11 100644 +--- a/drivers/ata/pata_efar.c ++++ b/drivers/ata/pata_efar.c +@@ -222,7 +222,7 @@ static struct scsi_host_template efar_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations efar_ops = { ++static const struct ata_port_operations efar_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = efar_cable_detect, + .set_piomode = efar_set_piomode, +diff --git a/drivers/ata/pata_hpt366.c b/drivers/ata/pata_hpt366.c +index b9d8836..0b92030 100644 +--- a/drivers/ata/pata_hpt366.c ++++ b/drivers/ata/pata_hpt366.c +@@ -282,7 +282,7 @@ static struct scsi_host_template hpt36x_sht = { + * Configuration for HPT366/68 + */ + +-static struct ata_port_operations hpt366_port_ops = { ++static const struct ata_port_operations hpt366_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = hpt36x_cable_detect, + .mode_filter = hpt366_filter, +diff --git a/drivers/ata/pata_hpt37x.c b/drivers/ata/pata_hpt37x.c +index 5af7f19..00c4980 100644 +--- a/drivers/ata/pata_hpt37x.c ++++ b/drivers/ata/pata_hpt37x.c +@@ -576,7 +576,7 @@ static struct scsi_host_template hpt37x_sht = { + * Configuration for HPT370 + */ + +-static struct ata_port_operations hpt370_port_ops = { ++static const struct ata_port_operations hpt370_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .bmdma_stop = hpt370_bmdma_stop, +@@ -591,7 +591,7 @@ static struct ata_port_operations hpt370_port_ops = { + * Configuration for HPT370A. Close to 370 but less filters + */ + +-static struct ata_port_operations hpt370a_port_ops = { ++static const struct ata_port_operations hpt370a_port_ops = { + .inherits = &hpt370_port_ops, + .mode_filter = hpt370a_filter, + }; +@@ -601,7 +601,7 @@ static struct ata_port_operations hpt370a_port_ops = { + * and DMA mode setting functionality. + */ + +-static struct ata_port_operations hpt372_port_ops = { ++static const struct ata_port_operations hpt372_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .bmdma_stop = hpt37x_bmdma_stop, +@@ -616,7 +616,7 @@ static struct ata_port_operations hpt372_port_ops = { + * but we have a different cable detection procedure for function 1. + */ + +-static struct ata_port_operations hpt374_fn1_port_ops = { ++static const struct ata_port_operations hpt374_fn1_port_ops = { + .inherits = &hpt372_port_ops, + .prereset = hpt374_fn1_pre_reset, + }; +diff --git a/drivers/ata/pata_hpt3x2n.c b/drivers/ata/pata_hpt3x2n.c +index 100f227..2e39382 100644 +--- a/drivers/ata/pata_hpt3x2n.c ++++ b/drivers/ata/pata_hpt3x2n.c +@@ -337,7 +337,7 @@ static struct scsi_host_template hpt3x2n_sht = { + * Configuration for HPT3x2n. + */ + +-static struct ata_port_operations hpt3x2n_port_ops = { ++static const struct ata_port_operations hpt3x2n_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .bmdma_stop = hpt3x2n_bmdma_stop, +diff --git a/drivers/ata/pata_hpt3x3.c b/drivers/ata/pata_hpt3x3.c +index 7e31025..6fca8f4 100644 +--- a/drivers/ata/pata_hpt3x3.c ++++ b/drivers/ata/pata_hpt3x3.c +@@ -141,7 +141,7 @@ static struct scsi_host_template hpt3x3_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations hpt3x3_port_ops = { ++static const struct ata_port_operations hpt3x3_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = hpt3x3_set_piomode, +diff --git a/drivers/ata/pata_icside.c b/drivers/ata/pata_icside.c +index b663b7f..9a26c2a 100644 +--- a/drivers/ata/pata_icside.c ++++ b/drivers/ata/pata_icside.c +@@ -319,7 +319,7 @@ static void pata_icside_postreset(struct ata_link *link, unsigned int *classes) + } + } + +-static struct ata_port_operations pata_icside_port_ops = { ++static const struct ata_port_operations pata_icside_port_ops = { + .inherits = &ata_sff_port_ops, + /* no need to build any PRD tables for DMA */ + .qc_prep = ata_noop_qc_prep, +diff --git a/drivers/ata/pata_isapnp.c b/drivers/ata/pata_isapnp.c +index 4bceb88..457dfb6 100644 +--- a/drivers/ata/pata_isapnp.c ++++ b/drivers/ata/pata_isapnp.c +@@ -23,12 +23,12 @@ static struct scsi_host_template isapnp_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations isapnp_port_ops = { ++static const struct ata_port_operations isapnp_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + }; + +-static struct ata_port_operations isapnp_noalt_port_ops = { ++static const struct ata_port_operations isapnp_noalt_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + /* No altstatus so we don't want to use the lost interrupt poll */ +diff --git a/drivers/ata/pata_it8213.c b/drivers/ata/pata_it8213.c +index f156da8..24976e2 100644 +--- a/drivers/ata/pata_it8213.c ++++ b/drivers/ata/pata_it8213.c +@@ -234,7 +234,7 @@ static struct scsi_host_template it8213_sht = { + }; + + +-static struct ata_port_operations it8213_ops = { ++static const struct ata_port_operations it8213_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = it8213_cable_detect, + .set_piomode = it8213_set_piomode, +diff --git a/drivers/ata/pata_it821x.c b/drivers/ata/pata_it821x.c +index 188bc2f..ca9e785 100644 +--- a/drivers/ata/pata_it821x.c ++++ b/drivers/ata/pata_it821x.c +@@ -800,7 +800,7 @@ static struct scsi_host_template it821x_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations it821x_smart_port_ops = { ++static const struct ata_port_operations it821x_smart_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .check_atapi_dma= it821x_check_atapi_dma, +@@ -814,7 +814,7 @@ static struct ata_port_operations it821x_smart_port_ops = { + .port_start = it821x_port_start, + }; + +-static struct ata_port_operations it821x_passthru_port_ops = { ++static const struct ata_port_operations it821x_passthru_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .check_atapi_dma= it821x_check_atapi_dma, +@@ -830,7 +830,7 @@ static struct ata_port_operations it821x_passthru_port_ops = { + .port_start = it821x_port_start, + }; + +-static struct ata_port_operations it821x_rdc_port_ops = { ++static const struct ata_port_operations it821x_rdc_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .check_atapi_dma= it821x_check_atapi_dma, +diff --git a/drivers/ata/pata_ixp4xx_cf.c b/drivers/ata/pata_ixp4xx_cf.c +index ba54b08..4b952b7 100644 +--- a/drivers/ata/pata_ixp4xx_cf.c ++++ b/drivers/ata/pata_ixp4xx_cf.c +@@ -89,7 +89,7 @@ static struct scsi_host_template ixp4xx_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations ixp4xx_port_ops = { ++static const struct ata_port_operations ixp4xx_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ixp4xx_mmio_data_xfer, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_jmicron.c b/drivers/ata/pata_jmicron.c +index 3a1474a..434b0ff 100644 +--- a/drivers/ata/pata_jmicron.c ++++ b/drivers/ata/pata_jmicron.c +@@ -111,7 +111,7 @@ static struct scsi_host_template jmicron_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations jmicron_ops = { ++static const struct ata_port_operations jmicron_ops = { + .inherits = &ata_bmdma_port_ops, + .prereset = jmicron_pre_reset, + }; +diff --git a/drivers/ata/pata_legacy.c b/drivers/ata/pata_legacy.c +index 6932e56..220e71d 100644 +--- a/drivers/ata/pata_legacy.c ++++ b/drivers/ata/pata_legacy.c +@@ -106,7 +106,7 @@ struct legacy_probe { + + struct legacy_controller { + const char *name; +- struct ata_port_operations *ops; ++ const struct ata_port_operations *ops; + unsigned int pio_mask; + unsigned int flags; + unsigned int pflags; +@@ -223,12 +223,12 @@ static const struct ata_port_operations legacy_base_port_ops = { + * pio_mask as well. + */ + +-static struct ata_port_operations simple_port_ops = { ++static const struct ata_port_operations simple_port_ops = { + .inherits = &legacy_base_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + }; + +-static struct ata_port_operations legacy_port_ops = { ++static const struct ata_port_operations legacy_port_ops = { + .inherits = &legacy_base_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + .set_mode = legacy_set_mode, +@@ -324,7 +324,7 @@ static unsigned int pdc_data_xfer_vlb(struct ata_device *dev, + return buflen; + } + +-static struct ata_port_operations pdc20230_port_ops = { ++static const struct ata_port_operations pdc20230_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = pdc20230_set_piomode, + .sff_data_xfer = pdc_data_xfer_vlb, +@@ -357,7 +357,7 @@ static void ht6560a_set_piomode(struct ata_port *ap, struct ata_device *adev) + ioread8(ap->ioaddr.status_addr); + } + +-static struct ata_port_operations ht6560a_port_ops = { ++static const struct ata_port_operations ht6560a_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = ht6560a_set_piomode, + }; +@@ -400,7 +400,7 @@ static void ht6560b_set_piomode(struct ata_port *ap, struct ata_device *adev) + ioread8(ap->ioaddr.status_addr); + } + +-static struct ata_port_operations ht6560b_port_ops = { ++static const struct ata_port_operations ht6560b_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = ht6560b_set_piomode, + }; +@@ -499,7 +499,7 @@ static void opti82c611a_set_piomode(struct ata_port *ap, + } + + +-static struct ata_port_operations opti82c611a_port_ops = { ++static const struct ata_port_operations opti82c611a_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = opti82c611a_set_piomode, + }; +@@ -609,7 +609,7 @@ static unsigned int opti82c46x_qc_issue(struct ata_queued_cmd *qc) + return ata_sff_qc_issue(qc); + } + +-static struct ata_port_operations opti82c46x_port_ops = { ++static const struct ata_port_operations opti82c46x_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = opti82c46x_set_piomode, + .qc_issue = opti82c46x_qc_issue, +@@ -771,20 +771,20 @@ static int qdi_port(struct platform_device *dev, + return 0; + } + +-static struct ata_port_operations qdi6500_port_ops = { ++static const struct ata_port_operations qdi6500_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = qdi6500_set_piomode, + .qc_issue = qdi_qc_issue, + .sff_data_xfer = vlb32_data_xfer, + }; + +-static struct ata_port_operations qdi6580_port_ops = { ++static const struct ata_port_operations qdi6580_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = qdi6580_set_piomode, + .sff_data_xfer = vlb32_data_xfer, + }; + +-static struct ata_port_operations qdi6580dp_port_ops = { ++static const struct ata_port_operations qdi6580dp_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = qdi6580dp_set_piomode, + .sff_data_xfer = vlb32_data_xfer, +@@ -855,7 +855,7 @@ static int winbond_port(struct platform_device *dev, + return 0; + } + +-static struct ata_port_operations winbond_port_ops = { ++static const struct ata_port_operations winbond_port_ops = { + .inherits = &legacy_base_port_ops, + .set_piomode = winbond_set_piomode, + .sff_data_xfer = vlb32_data_xfer, +@@ -978,7 +978,7 @@ static __init int legacy_init_one(struct legacy_probe *probe) + int pio_modes = controller->pio_mask; + unsigned long io = probe->port; + u32 mask = (1 << probe->slot); +- struct ata_port_operations *ops = controller->ops; ++ const struct ata_port_operations *ops = controller->ops; + struct legacy_data *ld = &legacy_data[probe->slot]; + struct ata_host *host = NULL; + struct ata_port *ap; +diff --git a/drivers/ata/pata_marvell.c b/drivers/ata/pata_marvell.c +index 2096fb7..4d090fc 100644 +--- a/drivers/ata/pata_marvell.c ++++ b/drivers/ata/pata_marvell.c +@@ -100,7 +100,7 @@ static struct scsi_host_template marvell_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations marvell_ops = { ++static const struct ata_port_operations marvell_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = marvell_cable_detect, + .prereset = marvell_pre_reset, +diff --git a/drivers/ata/pata_mpc52xx.c b/drivers/ata/pata_mpc52xx.c +index 99d41be..7d56aa8 100644 +--- a/drivers/ata/pata_mpc52xx.c ++++ b/drivers/ata/pata_mpc52xx.c +@@ -609,7 +609,7 @@ static struct scsi_host_template mpc52xx_ata_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations mpc52xx_ata_port_ops = { ++static const struct ata_port_operations mpc52xx_ata_port_ops = { + .inherits = &ata_bmdma_port_ops, + .sff_dev_select = mpc52xx_ata_dev_select, + .set_piomode = mpc52xx_ata_set_piomode, +diff --git a/drivers/ata/pata_mpiix.c b/drivers/ata/pata_mpiix.c +index b21f002..0a27e7f 100644 +--- a/drivers/ata/pata_mpiix.c ++++ b/drivers/ata/pata_mpiix.c +@@ -140,7 +140,7 @@ static struct scsi_host_template mpiix_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations mpiix_port_ops = { ++static const struct ata_port_operations mpiix_port_ops = { + .inherits = &ata_sff_port_ops, + .qc_issue = mpiix_qc_issue, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_netcell.c b/drivers/ata/pata_netcell.c +index f0d52f7..89c3be3 100644 +--- a/drivers/ata/pata_netcell.c ++++ b/drivers/ata/pata_netcell.c +@@ -34,7 +34,7 @@ static struct scsi_host_template netcell_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations netcell_ops = { ++static const struct ata_port_operations netcell_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_80wire, + .read_id = netcell_read_id, +diff --git a/drivers/ata/pata_ninja32.c b/drivers/ata/pata_ninja32.c +index dd53a66..a3f4317 100644 +--- a/drivers/ata/pata_ninja32.c ++++ b/drivers/ata/pata_ninja32.c +@@ -81,7 +81,7 @@ static struct scsi_host_template ninja32_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations ninja32_port_ops = { ++static const struct ata_port_operations ninja32_port_ops = { + .inherits = &ata_bmdma_port_ops, + .sff_dev_select = ninja32_dev_select, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_ns87410.c b/drivers/ata/pata_ns87410.c +index ca53fac..9aa93ef 100644 +--- a/drivers/ata/pata_ns87410.c ++++ b/drivers/ata/pata_ns87410.c +@@ -132,7 +132,7 @@ static struct scsi_host_template ns87410_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations ns87410_port_ops = { ++static const struct ata_port_operations ns87410_port_ops = { + .inherits = &ata_sff_port_ops, + .qc_issue = ns87410_qc_issue, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_ns87415.c b/drivers/ata/pata_ns87415.c +index 773b159..55f454e 100644 +--- a/drivers/ata/pata_ns87415.c ++++ b/drivers/ata/pata_ns87415.c +@@ -299,7 +299,7 @@ static u8 ns87560_bmdma_status(struct ata_port *ap) + } + #endif /* 87560 SuperIO Support */ + +-static struct ata_port_operations ns87415_pata_ops = { ++static const struct ata_port_operations ns87415_pata_ops = { + .inherits = &ata_bmdma_port_ops, + + .check_atapi_dma = ns87415_check_atapi_dma, +@@ -313,7 +313,7 @@ static struct ata_port_operations ns87415_pata_ops = { + }; + + #if defined(CONFIG_SUPERIO) +-static struct ata_port_operations ns87560_pata_ops = { ++static const struct ata_port_operations ns87560_pata_ops = { + .inherits = &ns87415_pata_ops, + .sff_tf_read = ns87560_tf_read, + .sff_check_status = ns87560_check_status, +diff --git a/drivers/ata/pata_octeon_cf.c b/drivers/ata/pata_octeon_cf.c +index d6f6956..639295b 100644 +--- a/drivers/ata/pata_octeon_cf.c ++++ b/drivers/ata/pata_octeon_cf.c +@@ -801,6 +801,7 @@ static unsigned int octeon_cf_qc_issue(struct ata_queued_cmd *qc) + return 0; + } + ++/* cannot be const */ + static struct ata_port_operations octeon_cf_ops = { + .inherits = &ata_sff_port_ops, + .check_atapi_dma = octeon_cf_check_atapi_dma, +diff --git a/drivers/ata/pata_oldpiix.c b/drivers/ata/pata_oldpiix.c +index 84ac503..adee1cd 100644 +--- a/drivers/ata/pata_oldpiix.c ++++ b/drivers/ata/pata_oldpiix.c +@@ -208,7 +208,7 @@ static struct scsi_host_template oldpiix_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations oldpiix_pata_ops = { ++static const struct ata_port_operations oldpiix_pata_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_issue = oldpiix_qc_issue, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_opti.c b/drivers/ata/pata_opti.c +index 99eddda..3a4c0aa 100644 +--- a/drivers/ata/pata_opti.c ++++ b/drivers/ata/pata_opti.c +@@ -152,7 +152,7 @@ static struct scsi_host_template opti_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations opti_port_ops = { ++static const struct ata_port_operations opti_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = opti_set_piomode, +diff --git a/drivers/ata/pata_optidma.c b/drivers/ata/pata_optidma.c +index 86885a4..8e9968d 100644 +--- a/drivers/ata/pata_optidma.c ++++ b/drivers/ata/pata_optidma.c +@@ -337,7 +337,7 @@ static struct scsi_host_template optidma_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations optidma_port_ops = { ++static const struct ata_port_operations optidma_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_40wire, + .set_piomode = optidma_set_pio_mode, +@@ -346,7 +346,7 @@ static struct ata_port_operations optidma_port_ops = { + .prereset = optidma_pre_reset, + }; + +-static struct ata_port_operations optiplus_port_ops = { ++static const struct ata_port_operations optiplus_port_ops = { + .inherits = &optidma_port_ops, + .set_piomode = optiplus_set_pio_mode, + .set_dmamode = optiplus_set_dma_mode, +diff --git a/drivers/ata/pata_palmld.c b/drivers/ata/pata_palmld.c +index 11fb4cc..1a14022 100644 +--- a/drivers/ata/pata_palmld.c ++++ b/drivers/ata/pata_palmld.c +@@ -37,7 +37,7 @@ static struct scsi_host_template palmld_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations palmld_port_ops = { ++static const struct ata_port_operations palmld_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pata_pcmcia.c b/drivers/ata/pata_pcmcia.c +index dc99e26..7f4b1e4 100644 +--- a/drivers/ata/pata_pcmcia.c ++++ b/drivers/ata/pata_pcmcia.c +@@ -162,14 +162,14 @@ static struct scsi_host_template pcmcia_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pcmcia_port_ops = { ++static const struct ata_port_operations pcmcia_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + .cable_detect = ata_cable_40wire, + .set_mode = pcmcia_set_mode, + }; + +-static struct ata_port_operations pcmcia_8bit_port_ops = { ++static const struct ata_port_operations pcmcia_8bit_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ata_data_xfer_8bit, + .cable_detect = ata_cable_40wire, +@@ -256,7 +256,7 @@ static int pcmcia_init_one(struct pcmcia_device *pdev) + unsigned long io_base, ctl_base; + void __iomem *io_addr, *ctl_addr; + int n_ports = 1; +- struct ata_port_operations *ops = &pcmcia_port_ops; ++ const struct ata_port_operations *ops = &pcmcia_port_ops; + + info = kzalloc(sizeof(*info), GFP_KERNEL); + if (info == NULL) +diff --git a/drivers/ata/pata_pdc2027x.c b/drivers/ata/pata_pdc2027x.c +index ca5cad0..3a1f125 100644 +--- a/drivers/ata/pata_pdc2027x.c ++++ b/drivers/ata/pata_pdc2027x.c +@@ -132,14 +132,14 @@ static struct scsi_host_template pdc2027x_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pdc2027x_pata100_ops = { ++static const struct ata_port_operations pdc2027x_pata100_ops = { + .inherits = &ata_bmdma_port_ops, + .check_atapi_dma = pdc2027x_check_atapi_dma, + .cable_detect = pdc2027x_cable_detect, + .prereset = pdc2027x_prereset, + }; + +-static struct ata_port_operations pdc2027x_pata133_ops = { ++static const struct ata_port_operations pdc2027x_pata133_ops = { + .inherits = &pdc2027x_pata100_ops, + .mode_filter = pdc2027x_mode_filter, + .set_piomode = pdc2027x_set_piomode, +diff --git a/drivers/ata/pata_pdc202xx_old.c b/drivers/ata/pata_pdc202xx_old.c +index 2911120..4bf62aa 100644 +--- a/drivers/ata/pata_pdc202xx_old.c ++++ b/drivers/ata/pata_pdc202xx_old.c +@@ -274,7 +274,7 @@ static struct scsi_host_template pdc202xx_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pdc2024x_port_ops = { ++static const struct ata_port_operations pdc2024x_port_ops = { + .inherits = &ata_bmdma_port_ops, + + .cable_detect = ata_cable_40wire, +@@ -284,7 +284,7 @@ static struct ata_port_operations pdc2024x_port_ops = { + .sff_exec_command = pdc202xx_exec_command, + }; + +-static struct ata_port_operations pdc2026x_port_ops = { ++static const struct ata_port_operations pdc2026x_port_ops = { + .inherits = &pdc2024x_port_ops, + + .check_atapi_dma = pdc2026x_check_atapi_dma, +diff --git a/drivers/ata/pata_platform.c b/drivers/ata/pata_platform.c +index 3f6ebc6..a18c358 100644 +--- a/drivers/ata/pata_platform.c ++++ b/drivers/ata/pata_platform.c +@@ -48,7 +48,7 @@ static struct scsi_host_template pata_platform_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations pata_platform_port_ops = { ++static const struct ata_port_operations pata_platform_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + .cable_detect = ata_cable_unknown, +diff --git a/drivers/ata/pata_qdi.c b/drivers/ata/pata_qdi.c +index 45879dc..165a9f9 100644 +--- a/drivers/ata/pata_qdi.c ++++ b/drivers/ata/pata_qdi.c +@@ -157,7 +157,7 @@ static struct scsi_host_template qdi_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations qdi6500_port_ops = { ++static const struct ata_port_operations qdi6500_port_ops = { + .inherits = &ata_sff_port_ops, + .qc_issue = qdi_qc_issue, + .sff_data_xfer = qdi_data_xfer, +@@ -165,7 +165,7 @@ static struct ata_port_operations qdi6500_port_ops = { + .set_piomode = qdi6500_set_piomode, + }; + +-static struct ata_port_operations qdi6580_port_ops = { ++static const struct ata_port_operations qdi6580_port_ops = { + .inherits = &qdi6500_port_ops, + .set_piomode = qdi6580_set_piomode, + }; +diff --git a/drivers/ata/pata_radisys.c b/drivers/ata/pata_radisys.c +index 4401b33..716c5cc 100644 +--- a/drivers/ata/pata_radisys.c ++++ b/drivers/ata/pata_radisys.c +@@ -187,7 +187,7 @@ static struct scsi_host_template radisys_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations radisys_pata_ops = { ++static const struct ata_port_operations radisys_pata_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_issue = radisys_qc_issue, + .cable_detect = ata_cable_unknown, +diff --git a/drivers/ata/pata_rb532_cf.c b/drivers/ata/pata_rb532_cf.c +index 45f1e10..fab6bca 100644 +--- a/drivers/ata/pata_rb532_cf.c ++++ b/drivers/ata/pata_rb532_cf.c +@@ -68,7 +68,7 @@ static irqreturn_t rb532_pata_irq_handler(int irq, void *dev_instance) + return IRQ_HANDLED; + } + +-static struct ata_port_operations rb532_pata_port_ops = { ++static const struct ata_port_operations rb532_pata_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = ata_sff_data_xfer32, + }; +diff --git a/drivers/ata/pata_rdc.c b/drivers/ata/pata_rdc.c +index c843a1e..b5853c3 100644 +--- a/drivers/ata/pata_rdc.c ++++ b/drivers/ata/pata_rdc.c +@@ -272,7 +272,7 @@ static void rdc_set_dmamode(struct ata_port *ap, struct ata_device *adev) + pci_write_config_byte(dev, 0x48, udma_enable); + } + +-static struct ata_port_operations rdc_pata_ops = { ++static const struct ata_port_operations rdc_pata_ops = { + .inherits = &ata_bmdma32_port_ops, + .cable_detect = rdc_pata_cable_detect, + .set_piomode = rdc_set_piomode, +diff --git a/drivers/ata/pata_rz1000.c b/drivers/ata/pata_rz1000.c +index a5e4dfe..080c8c9 100644 +--- a/drivers/ata/pata_rz1000.c ++++ b/drivers/ata/pata_rz1000.c +@@ -54,7 +54,7 @@ static struct scsi_host_template rz1000_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations rz1000_port_ops = { ++static const struct ata_port_operations rz1000_port_ops = { + .inherits = &ata_sff_port_ops, + .cable_detect = ata_cable_40wire, + .set_mode = rz1000_set_mode, +diff --git a/drivers/ata/pata_sc1200.c b/drivers/ata/pata_sc1200.c +index 3bbed83..e309daf 100644 +--- a/drivers/ata/pata_sc1200.c ++++ b/drivers/ata/pata_sc1200.c +@@ -207,7 +207,7 @@ static struct scsi_host_template sc1200_sht = { + .sg_tablesize = LIBATA_DUMB_MAX_PRD, + }; + +-static struct ata_port_operations sc1200_port_ops = { ++static const struct ata_port_operations sc1200_port_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_prep = ata_sff_dumb_qc_prep, + .qc_issue = sc1200_qc_issue, +diff --git a/drivers/ata/pata_scc.c b/drivers/ata/pata_scc.c +index 4257d6b..4c1d9d5 100644 +--- a/drivers/ata/pata_scc.c ++++ b/drivers/ata/pata_scc.c +@@ -965,7 +965,7 @@ static struct scsi_host_template scc_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations scc_pata_ops = { ++static const struct ata_port_operations scc_pata_ops = { + .inherits = &ata_bmdma_port_ops, + + .set_piomode = scc_set_piomode, +diff --git a/drivers/ata/pata_sch.c b/drivers/ata/pata_sch.c +index 99cceb4..e2e0a87 100644 +--- a/drivers/ata/pata_sch.c ++++ b/drivers/ata/pata_sch.c +@@ -75,7 +75,7 @@ static struct scsi_host_template sch_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations sch_pata_ops = { ++static const struct ata_port_operations sch_pata_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = ata_cable_unknown, + .set_piomode = sch_set_piomode, +diff --git a/drivers/ata/pata_serverworks.c b/drivers/ata/pata_serverworks.c +index beaed12..39969f1 100644 +--- a/drivers/ata/pata_serverworks.c ++++ b/drivers/ata/pata_serverworks.c +@@ -299,7 +299,7 @@ static struct scsi_host_template serverworks_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations serverworks_osb4_port_ops = { ++static const struct ata_port_operations serverworks_osb4_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = serverworks_cable_detect, + .mode_filter = serverworks_osb4_filter, +@@ -307,7 +307,7 @@ static struct ata_port_operations serverworks_osb4_port_ops = { + .set_dmamode = serverworks_set_dmamode, + }; + +-static struct ata_port_operations serverworks_csb_port_ops = { ++static const struct ata_port_operations serverworks_csb_port_ops = { + .inherits = &serverworks_osb4_port_ops, + .mode_filter = serverworks_csb_filter, + }; +diff --git a/drivers/ata/pata_sil680.c b/drivers/ata/pata_sil680.c +index a2ace48..0463b44 100644 +--- a/drivers/ata/pata_sil680.c ++++ b/drivers/ata/pata_sil680.c +@@ -194,7 +194,7 @@ static struct scsi_host_template sil680_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations sil680_port_ops = { ++static const struct ata_port_operations sil680_port_ops = { + .inherits = &ata_bmdma32_port_ops, + .cable_detect = sil680_cable_detect, + .set_piomode = sil680_set_piomode, +diff --git a/drivers/ata/pata_sis.c b/drivers/ata/pata_sis.c +index 488e77b..b3724d5 100644 +--- a/drivers/ata/pata_sis.c ++++ b/drivers/ata/pata_sis.c +@@ -503,47 +503,47 @@ static struct scsi_host_template sis_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations sis_133_for_sata_ops = { ++static const struct ata_port_operations sis_133_for_sata_ops = { + .inherits = &ata_bmdma_port_ops, + .set_piomode = sis_133_set_piomode, + .set_dmamode = sis_133_set_dmamode, + .cable_detect = sis_133_cable_detect, + }; + +-static struct ata_port_operations sis_base_ops = { ++static const struct ata_port_operations sis_base_ops = { + .inherits = &ata_bmdma_port_ops, + .prereset = sis_pre_reset, + }; + +-static struct ata_port_operations sis_133_ops = { ++static const struct ata_port_operations sis_133_ops = { + .inherits = &sis_base_ops, + .set_piomode = sis_133_set_piomode, + .set_dmamode = sis_133_set_dmamode, + .cable_detect = sis_133_cable_detect, + }; + +-static struct ata_port_operations sis_133_early_ops = { ++static const struct ata_port_operations sis_133_early_ops = { + .inherits = &sis_base_ops, + .set_piomode = sis_100_set_piomode, + .set_dmamode = sis_133_early_set_dmamode, + .cable_detect = sis_66_cable_detect, + }; + +-static struct ata_port_operations sis_100_ops = { ++static const struct ata_port_operations sis_100_ops = { + .inherits = &sis_base_ops, + .set_piomode = sis_100_set_piomode, + .set_dmamode = sis_100_set_dmamode, + .cable_detect = sis_66_cable_detect, + }; + +-static struct ata_port_operations sis_66_ops = { ++static const struct ata_port_operations sis_66_ops = { + .inherits = &sis_base_ops, + .set_piomode = sis_old_set_piomode, + .set_dmamode = sis_66_set_dmamode, + .cable_detect = sis_66_cable_detect, + }; + +-static struct ata_port_operations sis_old_ops = { ++static const struct ata_port_operations sis_old_ops = { + .inherits = &sis_base_ops, + .set_piomode = sis_old_set_piomode, + .set_dmamode = sis_old_set_dmamode, +diff --git a/drivers/ata/pata_sl82c105.c b/drivers/ata/pata_sl82c105.c +index 29f733c..43e9ca0 100644 +--- a/drivers/ata/pata_sl82c105.c ++++ b/drivers/ata/pata_sl82c105.c +@@ -231,7 +231,7 @@ static struct scsi_host_template sl82c105_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations sl82c105_port_ops = { ++static const struct ata_port_operations sl82c105_port_ops = { + .inherits = &ata_bmdma_port_ops, + .qc_defer = sl82c105_qc_defer, + .bmdma_start = sl82c105_bmdma_start, +diff --git a/drivers/ata/pata_triflex.c b/drivers/ata/pata_triflex.c +index f1f13ff..df39e99 100644 +--- a/drivers/ata/pata_triflex.c ++++ b/drivers/ata/pata_triflex.c +@@ -178,7 +178,7 @@ static struct scsi_host_template triflex_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations triflex_port_ops = { ++static const struct ata_port_operations triflex_port_ops = { + .inherits = &ata_bmdma_port_ops, + .bmdma_start = triflex_bmdma_start, + .bmdma_stop = triflex_bmdma_stop, +diff --git a/drivers/ata/pata_via.c b/drivers/ata/pata_via.c +index 1d73b8d..98a4b29 100644 +--- a/drivers/ata/pata_via.c ++++ b/drivers/ata/pata_via.c +@@ -419,7 +419,7 @@ static struct scsi_host_template via_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations via_port_ops = { ++static const struct ata_port_operations via_port_ops = { + .inherits = &ata_bmdma_port_ops, + .cable_detect = via_cable_detect, + .set_piomode = via_set_piomode, +@@ -429,7 +429,7 @@ static struct ata_port_operations via_port_ops = { + .port_start = via_port_start, + }; + +-static struct ata_port_operations via_port_ops_noirq = { ++static const struct ata_port_operations via_port_ops_noirq = { + .inherits = &via_port_ops, + .sff_data_xfer = ata_sff_data_xfer_noirq, + }; +diff --git a/drivers/ata/pata_winbond.c b/drivers/ata/pata_winbond.c +index 6d8619b..ad511c4 100644 +--- a/drivers/ata/pata_winbond.c ++++ b/drivers/ata/pata_winbond.c +@@ -125,7 +125,7 @@ static struct scsi_host_template winbond_sht = { + ATA_PIO_SHT(DRV_NAME), + }; + +-static struct ata_port_operations winbond_port_ops = { ++static const struct ata_port_operations winbond_port_ops = { + .inherits = &ata_sff_port_ops, + .sff_data_xfer = winbond_data_xfer, + .cable_detect = ata_cable_40wire, +diff --git a/drivers/ata/pdc_adma.c b/drivers/ata/pdc_adma.c +index 6c65b07..f996ec7 100644 +--- a/drivers/ata/pdc_adma.c ++++ b/drivers/ata/pdc_adma.c +@@ -145,7 +145,7 @@ static struct scsi_host_template adma_ata_sht = { + .dma_boundary = ADMA_DMA_BOUNDARY, + }; + +-static struct ata_port_operations adma_ata_ops = { ++static const struct ata_port_operations adma_ata_ops = { + .inherits = &ata_sff_port_ops, + + .lost_interrupt = ATA_OP_NULL, +diff --git a/drivers/ata/sata_fsl.c b/drivers/ata/sata_fsl.c +index 172b57e..c49bc1e 100644 +--- a/drivers/ata/sata_fsl.c ++++ b/drivers/ata/sata_fsl.c +@@ -1258,7 +1258,7 @@ static struct scsi_host_template sata_fsl_sht = { + .dma_boundary = ATA_DMA_BOUNDARY, + }; + +-static struct ata_port_operations sata_fsl_ops = { ++static const struct ata_port_operations sata_fsl_ops = { + .inherits = &sata_pmp_port_ops, + + .qc_defer = ata_std_qc_defer, +diff --git a/drivers/ata/sata_inic162x.c b/drivers/ata/sata_inic162x.c +index 4406902..60603ef 100644 +--- a/drivers/ata/sata_inic162x.c ++++ b/drivers/ata/sata_inic162x.c +@@ -721,7 +721,7 @@ static int inic_port_start(struct ata_port *ap) + return 0; + } + +-static struct ata_port_operations inic_port_ops = { ++static const struct ata_port_operations inic_port_ops = { + .inherits = &sata_port_ops, + + .check_atapi_dma = inic_check_atapi_dma, +diff --git a/drivers/ata/sata_mv.c b/drivers/ata/sata_mv.c +index cf41126..8107be6 100644 +--- a/drivers/ata/sata_mv.c ++++ b/drivers/ata/sata_mv.c +@@ -656,7 +656,7 @@ static struct scsi_host_template mv6_sht = { + .dma_boundary = MV_DMA_BOUNDARY, + }; + +-static struct ata_port_operations mv5_ops = { ++static const struct ata_port_operations mv5_ops = { + .inherits = &ata_sff_port_ops, + + .lost_interrupt = ATA_OP_NULL, +@@ -678,7 +678,7 @@ static struct ata_port_operations mv5_ops = { + .port_stop = mv_port_stop, + }; + +-static struct ata_port_operations mv6_ops = { ++static const struct ata_port_operations mv6_ops = { + .inherits = &mv5_ops, + .dev_config = mv6_dev_config, + .scr_read = mv_scr_read, +@@ -698,7 +698,7 @@ static struct ata_port_operations mv6_ops = { + .bmdma_status = mv_bmdma_status, + }; + +-static struct ata_port_operations mv_iie_ops = { ++static const struct ata_port_operations mv_iie_ops = { + .inherits = &mv6_ops, + .dev_config = ATA_OP_NULL, + .qc_prep = mv_qc_prep_iie, +diff --git a/drivers/ata/sata_nv.c b/drivers/ata/sata_nv.c +index ae2297c..d5c9c33 100644 +--- a/drivers/ata/sata_nv.c ++++ b/drivers/ata/sata_nv.c +@@ -464,7 +464,7 @@ static struct scsi_host_template nv_swncq_sht = { + * cases. Define nv_hardreset() which only kicks in for post-boot + * probing and use it for all variants. + */ +-static struct ata_port_operations nv_generic_ops = { ++static const struct ata_port_operations nv_generic_ops = { + .inherits = &ata_bmdma_port_ops, + .lost_interrupt = ATA_OP_NULL, + .scr_read = nv_scr_read, +@@ -472,20 +472,20 @@ static struct ata_port_operations nv_generic_ops = { + .hardreset = nv_hardreset, + }; + +-static struct ata_port_operations nv_nf2_ops = { ++static const struct ata_port_operations nv_nf2_ops = { + .inherits = &nv_generic_ops, + .freeze = nv_nf2_freeze, + .thaw = nv_nf2_thaw, + }; + +-static struct ata_port_operations nv_ck804_ops = { ++static const struct ata_port_operations nv_ck804_ops = { + .inherits = &nv_generic_ops, + .freeze = nv_ck804_freeze, + .thaw = nv_ck804_thaw, + .host_stop = nv_ck804_host_stop, + }; + +-static struct ata_port_operations nv_adma_ops = { ++static const struct ata_port_operations nv_adma_ops = { + .inherits = &nv_ck804_ops, + + .check_atapi_dma = nv_adma_check_atapi_dma, +@@ -509,7 +509,7 @@ static struct ata_port_operations nv_adma_ops = { + .host_stop = nv_adma_host_stop, + }; + +-static struct ata_port_operations nv_swncq_ops = { ++static const struct ata_port_operations nv_swncq_ops = { + .inherits = &nv_generic_ops, + + .qc_defer = ata_std_qc_defer, +diff --git a/drivers/ata/sata_promise.c b/drivers/ata/sata_promise.c +index 07d8d00..6cc70bb 100644 +--- a/drivers/ata/sata_promise.c ++++ b/drivers/ata/sata_promise.c +@@ -195,7 +195,7 @@ static const struct ata_port_operations pdc_common_ops = { + .error_handler = pdc_error_handler, + }; + +-static struct ata_port_operations pdc_sata_ops = { ++static const struct ata_port_operations pdc_sata_ops = { + .inherits = &pdc_common_ops, + .cable_detect = pdc_sata_cable_detect, + .freeze = pdc_sata_freeze, +@@ -208,14 +208,14 @@ static struct ata_port_operations pdc_sata_ops = { + + /* First-generation chips need a more restrictive ->check_atapi_dma op, + and ->freeze/thaw that ignore the hotplug controls. */ +-static struct ata_port_operations pdc_old_sata_ops = { ++static const struct ata_port_operations pdc_old_sata_ops = { + .inherits = &pdc_sata_ops, + .freeze = pdc_freeze, + .thaw = pdc_thaw, + .check_atapi_dma = pdc_old_sata_check_atapi_dma, + }; + +-static struct ata_port_operations pdc_pata_ops = { ++static const struct ata_port_operations pdc_pata_ops = { + .inherits = &pdc_common_ops, + .cable_detect = pdc_pata_cable_detect, + .freeze = pdc_freeze, +diff --git a/drivers/ata/sata_qstor.c b/drivers/ata/sata_qstor.c +index 326c0cf..36ecebe 100644 +--- a/drivers/ata/sata_qstor.c ++++ b/drivers/ata/sata_qstor.c +@@ -132,7 +132,7 @@ static struct scsi_host_template qs_ata_sht = { + .dma_boundary = QS_DMA_BOUNDARY, + }; + +-static struct ata_port_operations qs_ata_ops = { ++static const struct ata_port_operations qs_ata_ops = { + .inherits = &ata_sff_port_ops, + + .check_atapi_dma = qs_check_atapi_dma, +diff --git a/drivers/ata/sata_sil.c b/drivers/ata/sata_sil.c +index 3cb69d5..0871d3c 100644 +--- a/drivers/ata/sata_sil.c ++++ b/drivers/ata/sata_sil.c +@@ -182,7 +182,7 @@ static struct scsi_host_template sil_sht = { + .sg_tablesize = ATA_MAX_PRD + }; + +-static struct ata_port_operations sil_ops = { ++static const struct ata_port_operations sil_ops = { + .inherits = &ata_bmdma32_port_ops, + .dev_config = sil_dev_config, + .set_mode = sil_set_mode, +diff --git a/drivers/ata/sata_sil24.c b/drivers/ata/sata_sil24.c +index e6946fc..eddb794 100644 +--- a/drivers/ata/sata_sil24.c ++++ b/drivers/ata/sata_sil24.c +@@ -388,7 +388,7 @@ static struct scsi_host_template sil24_sht = { + .dma_boundary = ATA_DMA_BOUNDARY, + }; + +-static struct ata_port_operations sil24_ops = { ++static const struct ata_port_operations sil24_ops = { + .inherits = &sata_pmp_port_ops, + + .qc_defer = sil24_qc_defer, +diff --git a/drivers/ata/sata_sis.c b/drivers/ata/sata_sis.c +index f8a91bf..9cb06b6 100644 +--- a/drivers/ata/sata_sis.c ++++ b/drivers/ata/sata_sis.c +@@ -89,7 +89,7 @@ static struct scsi_host_template sis_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations sis_ops = { ++static const struct ata_port_operations sis_ops = { + .inherits = &ata_bmdma_port_ops, + .scr_read = sis_scr_read, + .scr_write = sis_scr_write, +diff --git a/drivers/ata/sata_svw.c b/drivers/ata/sata_svw.c +index 7257f2d..d04c6f5 100644 +--- a/drivers/ata/sata_svw.c ++++ b/drivers/ata/sata_svw.c +@@ -344,7 +344,7 @@ static struct scsi_host_template k2_sata_sht = { + }; + + +-static struct ata_port_operations k2_sata_ops = { ++static const struct ata_port_operations k2_sata_ops = { + .inherits = &ata_bmdma_port_ops, + .sff_tf_load = k2_sata_tf_load, + .sff_tf_read = k2_sata_tf_read, +diff --git a/drivers/ata/sata_sx4.c b/drivers/ata/sata_sx4.c +index bbcf970..cd0df0d 100644 +--- a/drivers/ata/sata_sx4.c ++++ b/drivers/ata/sata_sx4.c +@@ -248,7 +248,7 @@ static struct scsi_host_template pdc_sata_sht = { + }; + + /* TODO: inherit from base port_ops after converting to new EH */ +-static struct ata_port_operations pdc_20621_ops = { ++static const struct ata_port_operations pdc_20621_ops = { + .inherits = &ata_sff_port_ops, + + .check_atapi_dma = pdc_check_atapi_dma, +diff --git a/drivers/ata/sata_uli.c b/drivers/ata/sata_uli.c +index e5bff47..089d859 100644 +--- a/drivers/ata/sata_uli.c ++++ b/drivers/ata/sata_uli.c +@@ -79,7 +79,7 @@ static struct scsi_host_template uli_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations uli_ops = { ++static const struct ata_port_operations uli_ops = { + .inherits = &ata_bmdma_port_ops, + .scr_read = uli_scr_read, + .scr_write = uli_scr_write, +diff --git a/drivers/ata/sata_via.c b/drivers/ata/sata_via.c +index f5dcca7..77b94eb 100644 +--- a/drivers/ata/sata_via.c ++++ b/drivers/ata/sata_via.c +@@ -115,32 +115,32 @@ static struct scsi_host_template svia_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations svia_base_ops = { ++static const struct ata_port_operations svia_base_ops = { + .inherits = &ata_bmdma_port_ops, + .sff_tf_load = svia_tf_load, + }; + +-static struct ata_port_operations vt6420_sata_ops = { ++static const struct ata_port_operations vt6420_sata_ops = { + .inherits = &svia_base_ops, + .freeze = svia_noop_freeze, + .prereset = vt6420_prereset, + .bmdma_start = vt6420_bmdma_start, + }; + +-static struct ata_port_operations vt6421_pata_ops = { ++static const struct ata_port_operations vt6421_pata_ops = { + .inherits = &svia_base_ops, + .cable_detect = vt6421_pata_cable_detect, + .set_piomode = vt6421_set_pio_mode, + .set_dmamode = vt6421_set_dma_mode, + }; + +-static struct ata_port_operations vt6421_sata_ops = { ++static const struct ata_port_operations vt6421_sata_ops = { + .inherits = &svia_base_ops, + .scr_read = svia_scr_read, + .scr_write = svia_scr_write, + }; + +-static struct ata_port_operations vt8251_ops = { ++static const struct ata_port_operations vt8251_ops = { + .inherits = &svia_base_ops, + .hardreset = sata_std_hardreset, + .scr_read = vt8251_scr_read, +diff --git a/drivers/ata/sata_vsc.c b/drivers/ata/sata_vsc.c +index 8b2a278..51e65d3 100644 +--- a/drivers/ata/sata_vsc.c ++++ b/drivers/ata/sata_vsc.c +@@ -306,7 +306,7 @@ static struct scsi_host_template vsc_sata_sht = { + }; + + +-static struct ata_port_operations vsc_sata_ops = { ++static const struct ata_port_operations vsc_sata_ops = { + .inherits = &ata_bmdma_port_ops, + /* The IRQ handling is not quite standard SFF behaviour so we + cannot use the default lost interrupt handler */ +diff --git a/drivers/atm/adummy.c b/drivers/atm/adummy.c +index 5effec6..7e4019a 100644 +--- a/drivers/atm/adummy.c ++++ b/drivers/atm/adummy.c +@@ -77,7 +77,7 @@ adummy_send(struct atm_vcc *vcc, struct sk_buff *skb) + vcc->pop(vcc, skb); + else + dev_kfree_skb_any(skb); +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + + return 0; + } +diff --git a/drivers/atm/ambassador.c b/drivers/atm/ambassador.c +index 66e1813..26a27c6 100644 +--- a/drivers/atm/ambassador.c ++++ b/drivers/atm/ambassador.c +@@ -453,7 +453,7 @@ static void tx_complete (amb_dev * dev, tx_out * tx) { + PRINTD (DBG_FLOW|DBG_TX, "tx_complete %p %p", dev, tx); + + // VC layer stats +- atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); ++ atomic_inc_unchecked(&ATM_SKB(skb)->vcc->stats->tx); + + // free the descriptor + kfree (tx_descr); +@@ -494,7 +494,7 @@ static void rx_complete (amb_dev * dev, rx_out * rx) { + dump_skb ("<<<", vc, skb); + + // VC layer stats +- atomic_inc(&atm_vcc->stats->rx); ++ atomic_inc_unchecked(&atm_vcc->stats->rx); + __net_timestamp(skb); + // end of our responsability + atm_vcc->push (atm_vcc, skb); +@@ -509,7 +509,7 @@ static void rx_complete (amb_dev * dev, rx_out * rx) { + } else { + PRINTK (KERN_INFO, "dropped over-size frame"); + // should we count this? +- atomic_inc(&atm_vcc->stats->rx_drop); ++ atomic_inc_unchecked(&atm_vcc->stats->rx_drop); + } + + } else { +@@ -1341,7 +1341,7 @@ static int amb_send (struct atm_vcc * atm_vcc, struct sk_buff * skb) { + } + + if (check_area (skb->data, skb->len)) { +- atomic_inc(&atm_vcc->stats->tx_err); ++ atomic_inc_unchecked(&atm_vcc->stats->tx_err); + return -ENOMEM; // ? + } + +diff --git a/drivers/atm/atmtcp.c b/drivers/atm/atmtcp.c +index 02ad83d..6daffeb 100644 +--- a/drivers/atm/atmtcp.c ++++ b/drivers/atm/atmtcp.c +@@ -206,7 +206,7 @@ static int atmtcp_v_send(struct atm_vcc *vcc,struct sk_buff *skb) + if (vcc->pop) vcc->pop(vcc,skb); + else dev_kfree_skb(skb); + if (dev_data) return 0; +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + return -ENOLINK; + } + size = skb->len+sizeof(struct atmtcp_hdr); +@@ -214,7 +214,7 @@ static int atmtcp_v_send(struct atm_vcc *vcc,struct sk_buff *skb) + if (!new_skb) { + if (vcc->pop) vcc->pop(vcc,skb); + else dev_kfree_skb(skb); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + return -ENOBUFS; + } + hdr = (void *) skb_put(new_skb,sizeof(struct atmtcp_hdr)); +@@ -225,8 +225,8 @@ static int atmtcp_v_send(struct atm_vcc *vcc,struct sk_buff *skb) + if (vcc->pop) vcc->pop(vcc,skb); + else dev_kfree_skb(skb); + out_vcc->push(out_vcc,new_skb); +- atomic_inc(&vcc->stats->tx); +- atomic_inc(&out_vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->tx); ++ atomic_inc_unchecked(&out_vcc->stats->rx); + return 0; + } + +@@ -300,7 +300,7 @@ static int atmtcp_c_send(struct atm_vcc *vcc,struct sk_buff *skb) + out_vcc = find_vcc(dev, ntohs(hdr->vpi), ntohs(hdr->vci)); + read_unlock(&vcc_sklist_lock); + if (!out_vcc) { +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + goto done; + } + skb_pull(skb,sizeof(struct atmtcp_hdr)); +@@ -312,8 +312,8 @@ static int atmtcp_c_send(struct atm_vcc *vcc,struct sk_buff *skb) + __net_timestamp(new_skb); + skb_copy_from_linear_data(skb, skb_put(new_skb, skb->len), skb->len); + out_vcc->push(out_vcc,new_skb); +- atomic_inc(&vcc->stats->tx); +- atomic_inc(&out_vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->tx); ++ atomic_inc_unchecked(&out_vcc->stats->rx); + done: + if (vcc->pop) vcc->pop(vcc,skb); + else dev_kfree_skb(skb); +diff --git a/drivers/atm/eni.c b/drivers/atm/eni.c +index 0c30261..3da356e 100644 +--- a/drivers/atm/eni.c ++++ b/drivers/atm/eni.c +@@ -525,7 +525,7 @@ static int rx_aal0(struct atm_vcc *vcc) + DPRINTK(DEV_LABEL "(itf %d): trashing empty cell\n", + vcc->dev->number); + length = 0; +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + } + else { + length = ATM_CELL_SIZE-1; /* no HEC */ +@@ -580,7 +580,7 @@ static int rx_aal5(struct atm_vcc *vcc) + size); + } + eff = length = 0; +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + } + else { + size = (descr & MID_RED_COUNT)*(ATM_CELL_PAYLOAD >> 2); +@@ -597,7 +597,7 @@ static int rx_aal5(struct atm_vcc *vcc) + "(VCI=%d,length=%ld,size=%ld (descr 0x%lx))\n", + vcc->dev->number,vcc->vci,length,size << 2,descr); + length = eff = 0; +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + } + } + skb = eff ? atm_alloc_charge(vcc,eff << 2,GFP_ATOMIC) : NULL; +@@ -770,7 +770,7 @@ rx_dequeued++; + vcc->push(vcc,skb); + pushed++; + } +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + wake_up(&eni_dev->rx_wait); + } +@@ -1227,7 +1227,7 @@ static void dequeue_tx(struct atm_dev *dev) + PCI_DMA_TODEVICE); + if (vcc->pop) vcc->pop(vcc,skb); + else dev_kfree_skb_irq(skb); +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + wake_up(&eni_dev->tx_wait); + dma_complete++; + } +@@ -1570,7 +1570,7 @@ tx_complete++; + /*--------------------------------- entries ---------------------------------*/ + + +-static const char *media_name[] __devinitdata = { ++static const char *media_name[] __devinitconst = { + "MMF", "SMF", "MMF", "03?", /* 0- 3 */ + "UTP", "05?", "06?", "07?", /* 4- 7 */ + "TAXI","09?", "10?", "11?", /* 8-11 */ +diff --git a/drivers/atm/firestream.c b/drivers/atm/firestream.c +index cd5049a..a51209f 100644 +--- a/drivers/atm/firestream.c ++++ b/drivers/atm/firestream.c +@@ -748,7 +748,7 @@ static void process_txdone_queue (struct fs_dev *dev, struct queue *q) + } + } + +- atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); ++ atomic_inc_unchecked(&ATM_SKB(skb)->vcc->stats->tx); + + fs_dprintk (FS_DEBUG_TXMEM, "i"); + fs_dprintk (FS_DEBUG_ALLOC, "Free t-skb: %p\n", skb); +@@ -815,7 +815,7 @@ static void process_incoming (struct fs_dev *dev, struct queue *q) + #endif + skb_put (skb, qe->p1 & 0xffff); + ATM_SKB(skb)->vcc = atm_vcc; +- atomic_inc(&atm_vcc->stats->rx); ++ atomic_inc_unchecked(&atm_vcc->stats->rx); + __net_timestamp(skb); + fs_dprintk (FS_DEBUG_ALLOC, "Free rec-skb: %p (pushed)\n", skb); + atm_vcc->push (atm_vcc, skb); +@@ -836,12 +836,12 @@ static void process_incoming (struct fs_dev *dev, struct queue *q) + kfree (pe); + } + if (atm_vcc) +- atomic_inc(&atm_vcc->stats->rx_drop); ++ atomic_inc_unchecked(&atm_vcc->stats->rx_drop); + break; + case 0x1f: /* Reassembly abort: no buffers. */ + /* Silently increment error counter. */ + if (atm_vcc) +- atomic_inc(&atm_vcc->stats->rx_drop); ++ atomic_inc_unchecked(&atm_vcc->stats->rx_drop); + break; + default: /* Hmm. Haven't written the code to handle the others yet... -- REW */ + printk (KERN_WARNING "Don't know what to do with RX status %x: %s.\n", +diff --git a/drivers/atm/fore200e.c b/drivers/atm/fore200e.c +index f766cc4..a34002e 100644 +--- a/drivers/atm/fore200e.c ++++ b/drivers/atm/fore200e.c +@@ -931,9 +931,9 @@ fore200e_tx_irq(struct fore200e* fore200e) + #endif + /* check error condition */ + if (*entry->status & STATUS_ERROR) +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + else +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + } + } + +@@ -1082,7 +1082,7 @@ fore200e_push_rpd(struct fore200e* fore200e, struct atm_vcc* vcc, struct rpd* rp + if (skb == NULL) { + DPRINTK(2, "unable to alloc new skb, rx PDU length = %d\n", pdu_len); + +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + return -ENOMEM; + } + +@@ -1125,14 +1125,14 @@ fore200e_push_rpd(struct fore200e* fore200e, struct atm_vcc* vcc, struct rpd* rp + + dev_kfree_skb_any(skb); + +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + return -ENOMEM; + } + + ASSERT(atomic_read(&sk_atm(vcc)->sk_wmem_alloc) >= 0); + + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + ASSERT(atomic_read(&sk_atm(vcc)->sk_wmem_alloc) >= 0); + +@@ -1210,7 +1210,7 @@ fore200e_rx_irq(struct fore200e* fore200e) + DPRINTK(2, "damaged PDU on %d.%d.%d\n", + fore200e->atm_dev->number, + entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + } + } + +@@ -1655,7 +1655,7 @@ fore200e_send(struct atm_vcc *vcc, struct sk_buff *skb) + goto retry_here; + } + +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + + fore200e->tx_sat++; + DPRINTK(2, "tx queue of device %s is saturated, PDU dropped - heartbeat is %08x\n", +diff --git a/drivers/atm/he.c b/drivers/atm/he.c +index 7066703..2b130de 100644 +--- a/drivers/atm/he.c ++++ b/drivers/atm/he.c +@@ -1769,7 +1769,7 @@ he_service_rbrq(struct he_dev *he_dev, int group) + + if (RBRQ_HBUF_ERR(he_dev->rbrq_head)) { + hprintk("HBUF_ERR! (cid 0x%x)\n", cid); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + goto return_host_buffers; + } + +@@ -1802,7 +1802,7 @@ he_service_rbrq(struct he_dev *he_dev, int group) + RBRQ_LEN_ERR(he_dev->rbrq_head) + ? "LEN_ERR" : "", + vcc->vpi, vcc->vci); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto return_host_buffers; + } + +@@ -1861,7 +1861,7 @@ he_service_rbrq(struct he_dev *he_dev, int group) + vcc->push(vcc, skb); + spin_lock(&he_dev->global_lock); + +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + return_host_buffers: + ++pdus_assembled; +@@ -2206,7 +2206,7 @@ __enqueue_tpd(struct he_dev *he_dev, struct he_tpd *tpd, unsigned cid) + tpd->vcc->pop(tpd->vcc, tpd->skb); + else + dev_kfree_skb_any(tpd->skb); +- atomic_inc(&tpd->vcc->stats->tx_err); ++ atomic_inc_unchecked(&tpd->vcc->stats->tx_err); + } + pci_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status)); + return; +@@ -2618,7 +2618,7 @@ he_send(struct atm_vcc *vcc, struct sk_buff *skb) + vcc->pop(vcc, skb); + else + dev_kfree_skb_any(skb); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + return -EINVAL; + } + +@@ -2629,7 +2629,7 @@ he_send(struct atm_vcc *vcc, struct sk_buff *skb) + vcc->pop(vcc, skb); + else + dev_kfree_skb_any(skb); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + return -EINVAL; + } + #endif +@@ -2641,7 +2641,7 @@ he_send(struct atm_vcc *vcc, struct sk_buff *skb) + vcc->pop(vcc, skb); + else + dev_kfree_skb_any(skb); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + spin_unlock_irqrestore(&he_dev->global_lock, flags); + return -ENOMEM; + } +@@ -2683,7 +2683,7 @@ he_send(struct atm_vcc *vcc, struct sk_buff *skb) + vcc->pop(vcc, skb); + else + dev_kfree_skb_any(skb); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + spin_unlock_irqrestore(&he_dev->global_lock, flags); + return -ENOMEM; + } +@@ -2714,7 +2714,7 @@ he_send(struct atm_vcc *vcc, struct sk_buff *skb) + __enqueue_tpd(he_dev, tpd, cid); + spin_unlock_irqrestore(&he_dev->global_lock, flags); + +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + + return 0; + } +diff --git a/drivers/atm/horizon.c b/drivers/atm/horizon.c +index 4e49021..01b1512 100644 +--- a/drivers/atm/horizon.c ++++ b/drivers/atm/horizon.c +@@ -1033,7 +1033,7 @@ static void rx_schedule (hrz_dev * dev, int irq) { + { + struct atm_vcc * vcc = ATM_SKB(skb)->vcc; + // VC layer stats +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + __net_timestamp(skb); + // end of our responsability + vcc->push (vcc, skb); +@@ -1185,7 +1185,7 @@ static void tx_schedule (hrz_dev * const dev, int irq) { + dev->tx_iovec = NULL; + + // VC layer stats +- atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); ++ atomic_inc_unchecked(&ATM_SKB(skb)->vcc->stats->tx); + + // free the skb + hrz_kfree_skb (skb); +diff --git a/drivers/atm/idt77252.c b/drivers/atm/idt77252.c +index e33ae00..9deb4ab 100644 +--- a/drivers/atm/idt77252.c ++++ b/drivers/atm/idt77252.c +@@ -810,7 +810,7 @@ drain_scq(struct idt77252_dev *card, struct vc_map *vc) + else + dev_kfree_skb(skb); + +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + } + + atomic_dec(&scq->used); +@@ -1073,13 +1073,13 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + if ((sb = dev_alloc_skb(64)) == NULL) { + printk("%s: Can't allocate buffers for aal0.\n", + card->name); +- atomic_add(i, &vcc->stats->rx_drop); ++ atomic_add_unchecked(i, &vcc->stats->rx_drop); + break; + } + if (!atm_charge(vcc, sb->truesize)) { + RXPRINTK("%s: atm_charge() dropped aal0 packets.\n", + card->name); +- atomic_add(i - 1, &vcc->stats->rx_drop); ++ atomic_add_unchecked(i - 1, &vcc->stats->rx_drop); + dev_kfree_skb(sb); + break; + } +@@ -1096,7 +1096,7 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + ATM_SKB(sb)->vcc = vcc; + __net_timestamp(sb); + vcc->push(vcc, sb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + cell += ATM_CELL_PAYLOAD; + } +@@ -1133,13 +1133,13 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + "(CDC: %08x)\n", + card->name, len, rpp->len, readl(SAR_REG_CDC)); + recycle_rx_pool_skb(card, rpp); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + return; + } + if (stat & SAR_RSQE_CRC) { + RXPRINTK("%s: AAL5 CRC error.\n", card->name); + recycle_rx_pool_skb(card, rpp); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + return; + } + if (skb_queue_len(&rpp->queue) > 1) { +@@ -1150,7 +1150,7 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + RXPRINTK("%s: Can't alloc RX skb.\n", + card->name); + recycle_rx_pool_skb(card, rpp); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + return; + } + if (!atm_charge(vcc, skb->truesize)) { +@@ -1169,7 +1169,7 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + __net_timestamp(skb); + + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + return; + } +@@ -1191,7 +1191,7 @@ dequeue_rx(struct idt77252_dev *card, struct rsq_entry *rsqe) + __net_timestamp(skb); + + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + if (skb->truesize > SAR_FB_SIZE_3) + add_rx_skb(card, 3, SAR_FB_SIZE_3, 1); +@@ -1303,14 +1303,14 @@ idt77252_rx_raw(struct idt77252_dev *card) + if (vcc->qos.aal != ATM_AAL0) { + RPRINTK("%s: raw cell for non AAL0 vc %u.%u\n", + card->name, vpi, vci); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + goto drop; + } + + if ((sb = dev_alloc_skb(64)) == NULL) { + printk("%s: Can't allocate buffers for AAL0.\n", + card->name); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto drop; + } + +@@ -1329,7 +1329,7 @@ idt77252_rx_raw(struct idt77252_dev *card) + ATM_SKB(sb)->vcc = vcc; + __net_timestamp(sb); + vcc->push(vcc, sb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + + drop: + skb_pull(queue, 64); +@@ -1954,13 +1954,13 @@ idt77252_send_skb(struct atm_vcc *vcc, struct sk_buff *skb, int oam) + + if (vc == NULL) { + printk("%s: NULL connection in send().\n", card->name); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb(skb); + return -EINVAL; + } + if (!test_bit(VCF_TX, &vc->flags)) { + printk("%s: Trying to transmit on a non-tx VC.\n", card->name); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb(skb); + return -EINVAL; + } +@@ -1972,14 +1972,14 @@ idt77252_send_skb(struct atm_vcc *vcc, struct sk_buff *skb, int oam) + break; + default: + printk("%s: Unsupported AAL: %d\n", card->name, vcc->qos.aal); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb(skb); + return -EINVAL; + } + + if (skb_shinfo(skb)->nr_frags != 0) { + printk("%s: No scatter-gather yet.\n", card->name); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb(skb); + return -EINVAL; + } +@@ -1987,7 +1987,7 @@ idt77252_send_skb(struct atm_vcc *vcc, struct sk_buff *skb, int oam) + + err = queue_skb(card, vc, skb, oam); + if (err) { +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb(skb); + return err; + } +@@ -2010,7 +2010,7 @@ idt77252_send_oam(struct atm_vcc *vcc, void *cell, int flags) + skb = dev_alloc_skb(64); + if (!skb) { + printk("%s: Out of memory in send_oam().\n", card->name); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + return -ENOMEM; + } + atomic_add(skb->truesize, &sk_atm(vcc)->sk_wmem_alloc); +diff --git a/drivers/atm/iphase.c b/drivers/atm/iphase.c +index b2c1b37..faa672b 100644 +--- a/drivers/atm/iphase.c ++++ b/drivers/atm/iphase.c +@@ -1123,7 +1123,7 @@ static int rx_pkt(struct atm_dev *dev) + status = (u_short) (buf_desc_ptr->desc_mode); + if (status & (RX_CER | RX_PTE | RX_OFL)) + { +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + IF_ERR(printk("IA: bad packet, dropping it");) + if (status & RX_CER) { + IF_ERR(printk(" cause: packet CRC error\n");) +@@ -1146,7 +1146,7 @@ static int rx_pkt(struct atm_dev *dev) + len = dma_addr - buf_addr; + if (len > iadev->rx_buf_sz) { + printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto out_free_desc; + } + +@@ -1296,7 +1296,7 @@ static void rx_dle_intr(struct atm_dev *dev) + ia_vcc = INPH_IA_VCC(vcc); + if (ia_vcc == NULL) + { +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + dev_kfree_skb_any(skb); + atm_return(vcc, atm_guess_pdu2truesize(len)); + goto INCR_DLE; +@@ -1308,7 +1308,7 @@ static void rx_dle_intr(struct atm_dev *dev) + if ((length > iadev->rx_buf_sz) || (length > + (skb->len - sizeof(struct cpcs_trailer)))) + { +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + IF_ERR(printk("rx_dle_intr: Bad AAL5 trailer %d (skb len %d)", + length, skb->len);) + dev_kfree_skb_any(skb); +@@ -1324,7 +1324,7 @@ static void rx_dle_intr(struct atm_dev *dev) + + IF_RX(printk("rx_dle_intr: skb push");) + vcc->push(vcc,skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + iadev->rx_pkt_cnt++; + } + INCR_DLE: +@@ -2806,15 +2806,15 @@ static int ia_ioctl(struct atm_dev *dev, unsigned int cmd, void __user *arg) + { + struct k_sonet_stats *stats; + stats = &PRIV(_ia_dev[board])->sonet_stats; +- printk("section_bip: %d\n", atomic_read(&stats->section_bip)); +- printk("line_bip : %d\n", atomic_read(&stats->line_bip)); +- printk("path_bip : %d\n", atomic_read(&stats->path_bip)); +- printk("line_febe : %d\n", atomic_read(&stats->line_febe)); +- printk("path_febe : %d\n", atomic_read(&stats->path_febe)); +- printk("corr_hcs : %d\n", atomic_read(&stats->corr_hcs)); +- printk("uncorr_hcs : %d\n", atomic_read(&stats->uncorr_hcs)); +- printk("tx_cells : %d\n", atomic_read(&stats->tx_cells)); +- printk("rx_cells : %d\n", atomic_read(&stats->rx_cells)); ++ printk("section_bip: %d\n", atomic_read_unchecked(&stats->section_bip)); ++ printk("line_bip : %d\n", atomic_read_unchecked(&stats->line_bip)); ++ printk("path_bip : %d\n", atomic_read_unchecked(&stats->path_bip)); ++ printk("line_febe : %d\n", atomic_read_unchecked(&stats->line_febe)); ++ printk("path_febe : %d\n", atomic_read_unchecked(&stats->path_febe)); ++ printk("corr_hcs : %d\n", atomic_read_unchecked(&stats->corr_hcs)); ++ printk("uncorr_hcs : %d\n", atomic_read_unchecked(&stats->uncorr_hcs)); ++ printk("tx_cells : %d\n", atomic_read_unchecked(&stats->tx_cells)); ++ printk("rx_cells : %d\n", atomic_read_unchecked(&stats->rx_cells)); + } + ia_cmds.status = 0; + break; +@@ -2919,7 +2919,7 @@ static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb) { + if ((desc == 0) || (desc > iadev->num_tx_desc)) + { + IF_ERR(printk(DEV_LABEL "invalid desc for send: %d\n", desc);) +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + if (vcc->pop) + vcc->pop(vcc, skb); + else +@@ -3024,14 +3024,14 @@ static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb) { + ATM_DESC(skb) = vcc->vci; + skb_queue_tail(&iadev->tx_dma_q, skb); + +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + iadev->tx_pkt_cnt++; + /* Increment transaction counter */ + writel(2, iadev->dma+IPHASE5575_TX_COUNTER); + + #if 0 + /* add flow control logic */ +- if (atomic_read(&vcc->stats->tx) % 20 == 0) { ++ if (atomic_read_unchecked(&vcc->stats->tx) % 20 == 0) { + if (iavcc->vc_desc_cnt > 10) { + vcc->tx_quota = vcc->tx_quota * 3 / 4; + printk("Tx1: vcc->tx_quota = %d \n", (u32)vcc->tx_quota ); +diff --git a/drivers/atm/lanai.c b/drivers/atm/lanai.c +index cf97c34..8d30655 100644 +--- a/drivers/atm/lanai.c ++++ b/drivers/atm/lanai.c +@@ -1305,7 +1305,7 @@ static void lanai_send_one_aal5(struct lanai_dev *lanai, + vcc_tx_add_aal5_trailer(lvcc, skb->len, 0, 0); + lanai_endtx(lanai, lvcc); + lanai_free_skb(lvcc->tx.atmvcc, skb); +- atomic_inc(&lvcc->tx.atmvcc->stats->tx); ++ atomic_inc_unchecked(&lvcc->tx.atmvcc->stats->tx); + } + + /* Try to fill the buffer - don't call unless there is backlog */ +@@ -1428,7 +1428,7 @@ static void vcc_rx_aal5(struct lanai_vcc *lvcc, int endptr) + ATM_SKB(skb)->vcc = lvcc->rx.atmvcc; + __net_timestamp(skb); + lvcc->rx.atmvcc->push(lvcc->rx.atmvcc, skb); +- atomic_inc(&lvcc->rx.atmvcc->stats->rx); ++ atomic_inc_unchecked(&lvcc->rx.atmvcc->stats->rx); + out: + lvcc->rx.buf.ptr = end; + cardvcc_write(lvcc, endptr, vcc_rxreadptr); +@@ -1670,7 +1670,7 @@ static int handle_service(struct lanai_dev *lanai, u32 s) + DPRINTK("(itf %d) got RX service entry 0x%X for non-AAL5 " + "vcc %d\n", lanai->number, (unsigned int) s, vci); + lanai->stats.service_rxnotaal5++; +- atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); ++ atomic_inc_unchecked(&lvcc->rx.atmvcc->stats->rx_err); + return 0; + } + if (likely(!(s & (SERVICE_TRASH | SERVICE_STREAM | SERVICE_CRCERR)))) { +@@ -1682,7 +1682,7 @@ static int handle_service(struct lanai_dev *lanai, u32 s) + int bytes; + read_unlock(&vcc_sklist_lock); + DPRINTK("got trashed rx pdu on vci %d\n", vci); +- atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); ++ atomic_inc_unchecked(&lvcc->rx.atmvcc->stats->rx_err); + lvcc->stats.x.aal5.service_trash++; + bytes = (SERVICE_GET_END(s) * 16) - + (((unsigned long) lvcc->rx.buf.ptr) - +@@ -1694,7 +1694,7 @@ static int handle_service(struct lanai_dev *lanai, u32 s) + } + if (s & SERVICE_STREAM) { + read_unlock(&vcc_sklist_lock); +- atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); ++ atomic_inc_unchecked(&lvcc->rx.atmvcc->stats->rx_err); + lvcc->stats.x.aal5.service_stream++; + printk(KERN_ERR DEV_LABEL "(itf %d): Got AAL5 stream " + "PDU on VCI %d!\n", lanai->number, vci); +@@ -1702,7 +1702,7 @@ static int handle_service(struct lanai_dev *lanai, u32 s) + return 0; + } + DPRINTK("got rx crc error on vci %d\n", vci); +- atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); ++ atomic_inc_unchecked(&lvcc->rx.atmvcc->stats->rx_err); + lvcc->stats.x.aal5.service_rxcrc++; + lvcc->rx.buf.ptr = &lvcc->rx.buf.start[SERVICE_GET_END(s) * 4]; + cardvcc_write(lvcc, SERVICE_GET_END(s), vcc_rxreadptr); +diff --git a/drivers/atm/nicstar.c b/drivers/atm/nicstar.c +index 3da804b..d3b0eed 100644 +--- a/drivers/atm/nicstar.c ++++ b/drivers/atm/nicstar.c +@@ -1723,7 +1723,7 @@ static int ns_send(struct atm_vcc *vcc, struct sk_buff *skb) + if ((vc = (vc_map *) vcc->dev_data) == NULL) + { + printk("nicstar%d: vcc->dev_data == NULL on ns_send().\n", card->index); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb_any(skb); + return -EINVAL; + } +@@ -1731,7 +1731,7 @@ static int ns_send(struct atm_vcc *vcc, struct sk_buff *skb) + if (!vc->tx) + { + printk("nicstar%d: Trying to transmit on a non-tx VC.\n", card->index); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb_any(skb); + return -EINVAL; + } +@@ -1739,7 +1739,7 @@ static int ns_send(struct atm_vcc *vcc, struct sk_buff *skb) + if (vcc->qos.aal != ATM_AAL5 && vcc->qos.aal != ATM_AAL0) + { + printk("nicstar%d: Only AAL0 and AAL5 are supported.\n", card->index); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb_any(skb); + return -EINVAL; + } +@@ -1747,7 +1747,7 @@ static int ns_send(struct atm_vcc *vcc, struct sk_buff *skb) + if (skb_shinfo(skb)->nr_frags != 0) + { + printk("nicstar%d: No scatter-gather yet.\n", card->index); +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb_any(skb); + return -EINVAL; + } +@@ -1792,11 +1792,11 @@ static int ns_send(struct atm_vcc *vcc, struct sk_buff *skb) + + if (push_scqe(card, vc, scq, &scqe, skb) != 0) + { +- atomic_inc(&vcc->stats->tx_err); ++ atomic_inc_unchecked(&vcc->stats->tx_err); + dev_kfree_skb_any(skb); + return -EIO; + } +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + + return 0; + } +@@ -2111,14 +2111,14 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + { + printk("nicstar%d: Can't allocate buffers for aal0.\n", + card->index); +- atomic_add(i,&vcc->stats->rx_drop); ++ atomic_add_unchecked(i,&vcc->stats->rx_drop); + break; + } + if (!atm_charge(vcc, sb->truesize)) + { + RXPRINTK("nicstar%d: atm_charge() dropped aal0 packets.\n", + card->index); +- atomic_add(i-1,&vcc->stats->rx_drop); /* already increased by 1 */ ++ atomic_add_unchecked(i-1,&vcc->stats->rx_drop); /* already increased by 1 */ + dev_kfree_skb_any(sb); + break; + } +@@ -2133,7 +2133,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + ATM_SKB(sb)->vcc = vcc; + __net_timestamp(sb); + vcc->push(vcc, sb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + cell += ATM_CELL_PAYLOAD; + } + +@@ -2152,7 +2152,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + if (iovb == NULL) + { + printk("nicstar%d: Out of iovec buffers.\n", card->index); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + recycle_rx_buf(card, skb); + return; + } +@@ -2182,7 +2182,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + else if (NS_SKB(iovb)->iovcnt >= NS_MAX_IOVECS) + { + printk("nicstar%d: received too big AAL5 SDU.\n", card->index); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + recycle_iovec_rx_bufs(card, (struct iovec *) iovb->data, NS_MAX_IOVECS); + NS_SKB(iovb)->iovcnt = 0; + iovb->len = 0; +@@ -2202,7 +2202,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + printk("nicstar%d: Expected a small buffer, and this is not one.\n", + card->index); + which_list(card, skb); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + recycle_rx_buf(card, skb); + vc->rx_iov = NULL; + recycle_iov_buf(card, iovb); +@@ -2216,7 +2216,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + printk("nicstar%d: Expected a large buffer, and this is not one.\n", + card->index); + which_list(card, skb); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + recycle_iovec_rx_bufs(card, (struct iovec *) iovb->data, + NS_SKB(iovb)->iovcnt); + vc->rx_iov = NULL; +@@ -2240,7 +2240,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + printk(" - PDU size mismatch.\n"); + else + printk(".\n"); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + recycle_iovec_rx_bufs(card, (struct iovec *) iovb->data, + NS_SKB(iovb)->iovcnt); + vc->rx_iov = NULL; +@@ -2256,7 +2256,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + if (!atm_charge(vcc, skb->truesize)) + { + push_rxbufs(card, skb); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + } + else + { +@@ -2268,7 +2268,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + ATM_SKB(skb)->vcc = vcc; + __net_timestamp(skb); + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + } + else if (NS_SKB(iovb)->iovcnt == 2) /* One small plus one large buffer */ +@@ -2283,7 +2283,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + if (!atm_charge(vcc, sb->truesize)) + { + push_rxbufs(card, sb); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + } + else + { +@@ -2295,7 +2295,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + ATM_SKB(sb)->vcc = vcc; + __net_timestamp(sb); + vcc->push(vcc, sb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + + push_rxbufs(card, skb); +@@ -2306,7 +2306,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + if (!atm_charge(vcc, skb->truesize)) + { + push_rxbufs(card, skb); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + } + else + { +@@ -2320,7 +2320,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + ATM_SKB(skb)->vcc = vcc; + __net_timestamp(skb); + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + + push_rxbufs(card, sb); +@@ -2342,7 +2342,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + if (hb == NULL) + { + printk("nicstar%d: Out of huge buffers.\n", card->index); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + recycle_iovec_rx_bufs(card, (struct iovec *) iovb->data, + NS_SKB(iovb)->iovcnt); + vc->rx_iov = NULL; +@@ -2393,7 +2393,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + } + else + dev_kfree_skb_any(hb); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + } + else + { +@@ -2427,7 +2427,7 @@ static void dequeue_rx(ns_dev *card, ns_rsqe *rsqe) + #endif /* NS_USE_DESTRUCTORS */ + __net_timestamp(hb); + vcc->push(vcc, hb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + } + +diff --git a/drivers/atm/solos-pci.c b/drivers/atm/solos-pci.c +index 84c93ff..e6ed269 100644 +--- a/drivers/atm/solos-pci.c ++++ b/drivers/atm/solos-pci.c +@@ -708,7 +708,7 @@ void solos_bh(unsigned long card_arg) + } + atm_charge(vcc, skb->truesize); + vcc->push(vcc, skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + break; + + case PKT_STATUS: +@@ -914,6 +914,8 @@ static int print_buffer(struct sk_buff *buf) + char msg[500]; + char item[10]; + ++ pax_track_stack(); ++ + len = buf->len; + for (i = 0; i < len; i++){ + if(i % 8 == 0) +@@ -1023,7 +1025,7 @@ static uint32_t fpga_tx(struct solos_card *card) + vcc = SKB_CB(oldskb)->vcc; + + if (vcc) { +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + solos_pop(vcc, oldskb); + } else + dev_kfree_skb_irq(oldskb); +diff --git a/drivers/atm/suni.c b/drivers/atm/suni.c +index 6dd3f59..ee377f3 100644 +--- a/drivers/atm/suni.c ++++ b/drivers/atm/suni.c +@@ -49,8 +49,8 @@ static DEFINE_SPINLOCK(sunis_lock); + + + #define ADD_LIMITED(s,v) \ +- atomic_add((v),&stats->s); \ +- if (atomic_read(&stats->s) < 0) atomic_set(&stats->s,INT_MAX); ++ atomic_add_unchecked((v),&stats->s); \ ++ if (atomic_read_unchecked(&stats->s) < 0) atomic_set_unchecked(&stats->s,INT_MAX); + + + static void suni_hz(unsigned long from_timer) +diff --git a/drivers/atm/uPD98402.c b/drivers/atm/uPD98402.c +index fc8cb07..4a80e53 100644 +--- a/drivers/atm/uPD98402.c ++++ b/drivers/atm/uPD98402.c +@@ -41,7 +41,7 @@ static int fetch_stats(struct atm_dev *dev,struct sonet_stats __user *arg,int ze + struct sonet_stats tmp; + int error = 0; + +- atomic_add(GET(HECCT),&PRIV(dev)->sonet_stats.uncorr_hcs); ++ atomic_add_unchecked(GET(HECCT),&PRIV(dev)->sonet_stats.uncorr_hcs); + sonet_copy_stats(&PRIV(dev)->sonet_stats,&tmp); + if (arg) error = copy_to_user(arg,&tmp,sizeof(tmp)); + if (zero && !error) { +@@ -160,9 +160,9 @@ static int uPD98402_ioctl(struct atm_dev *dev,unsigned int cmd,void __user *arg) + + + #define ADD_LIMITED(s,v) \ +- { atomic_add(GET(v),&PRIV(dev)->sonet_stats.s); \ +- if (atomic_read(&PRIV(dev)->sonet_stats.s) < 0) \ +- atomic_set(&PRIV(dev)->sonet_stats.s,INT_MAX); } ++ { atomic_add_unchecked(GET(v),&PRIV(dev)->sonet_stats.s); \ ++ if (atomic_read_unchecked(&PRIV(dev)->sonet_stats.s) < 0) \ ++ atomic_set_unchecked(&PRIV(dev)->sonet_stats.s,INT_MAX); } + + + static void stat_event(struct atm_dev *dev) +@@ -193,7 +193,7 @@ static void uPD98402_int(struct atm_dev *dev) + if (reason & uPD98402_INT_PFM) stat_event(dev); + if (reason & uPD98402_INT_PCO) { + (void) GET(PCOCR); /* clear interrupt cause */ +- atomic_add(GET(HECCT), ++ atomic_add_unchecked(GET(HECCT), + &PRIV(dev)->sonet_stats.uncorr_hcs); + } + if ((reason & uPD98402_INT_RFO) && +@@ -221,9 +221,9 @@ static int uPD98402_start(struct atm_dev *dev) + PUT(~(uPD98402_INT_PFM | uPD98402_INT_ALM | uPD98402_INT_RFO | + uPD98402_INT_LOS),PIMR); /* enable them */ + (void) fetch_stats(dev,NULL,1); /* clear kernel counters */ +- atomic_set(&PRIV(dev)->sonet_stats.corr_hcs,-1); +- atomic_set(&PRIV(dev)->sonet_stats.tx_cells,-1); +- atomic_set(&PRIV(dev)->sonet_stats.rx_cells,-1); ++ atomic_set_unchecked(&PRIV(dev)->sonet_stats.corr_hcs,-1); ++ atomic_set_unchecked(&PRIV(dev)->sonet_stats.tx_cells,-1); ++ atomic_set_unchecked(&PRIV(dev)->sonet_stats.rx_cells,-1); + return 0; + } + +diff --git a/drivers/atm/zatm.c b/drivers/atm/zatm.c +index 2e9635b..32927b4 100644 +--- a/drivers/atm/zatm.c ++++ b/drivers/atm/zatm.c +@@ -458,7 +458,7 @@ printk("dummy: 0x%08lx, 0x%08lx\n",dummy[0],dummy[1]); + } + if (!size) { + dev_kfree_skb_irq(skb); +- if (vcc) atomic_inc(&vcc->stats->rx_err); ++ if (vcc) atomic_inc_unchecked(&vcc->stats->rx_err); + continue; + } + if (!atm_charge(vcc,skb->truesize)) { +@@ -468,7 +468,7 @@ printk("dummy: 0x%08lx, 0x%08lx\n",dummy[0],dummy[1]); + skb->len = size; + ATM_SKB(skb)->vcc = vcc; + vcc->push(vcc,skb); +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + } + zout(pos & 0xffff,MTA(mbx)); + #if 0 /* probably a stupid idea */ +@@ -732,7 +732,7 @@ if (*ZATM_PRV_DSC(skb) != (uPD98401_TXPD_V | uPD98401_TXPD_DP | + skb_queue_head(&zatm_vcc->backlog,skb); + break; + } +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + wake_up(&zatm_vcc->tx_wait); + } + +diff --git a/drivers/base/bus.c b/drivers/base/bus.c +index 63c143e..fece183 100644 +--- a/drivers/base/bus.c ++++ b/drivers/base/bus.c +@@ -70,7 +70,7 @@ static ssize_t drv_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops driver_sysfs_ops = { ++static const struct sysfs_ops driver_sysfs_ops = { + .show = drv_attr_show, + .store = drv_attr_store, + }; +@@ -115,7 +115,7 @@ static ssize_t bus_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops bus_sysfs_ops = { ++static const struct sysfs_ops bus_sysfs_ops = { + .show = bus_attr_show, + .store = bus_attr_store, + }; +@@ -154,7 +154,7 @@ static int bus_uevent_filter(struct kset *kset, struct kobject *kobj) + return 0; + } + +-static struct kset_uevent_ops bus_uevent_ops = { ++static const struct kset_uevent_ops bus_uevent_ops = { + .filter = bus_uevent_filter, + }; + +diff --git a/drivers/base/class.c b/drivers/base/class.c +index 6e2c3b0..cb61871 100644 +--- a/drivers/base/class.c ++++ b/drivers/base/class.c +@@ -63,7 +63,7 @@ static void class_release(struct kobject *kobj) + kfree(cp); + } + +-static struct sysfs_ops class_sysfs_ops = { ++static const struct sysfs_ops class_sysfs_ops = { + .show = class_attr_show, + .store = class_attr_store, + }; +diff --git a/drivers/base/core.c b/drivers/base/core.c +index f33d768..a9358d0 100644 +--- a/drivers/base/core.c ++++ b/drivers/base/core.c +@@ -100,7 +100,7 @@ static ssize_t dev_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops dev_sysfs_ops = { ++static const struct sysfs_ops dev_sysfs_ops = { + .show = dev_attr_show, + .store = dev_attr_store, + }; +@@ -252,7 +252,7 @@ static int dev_uevent(struct kset *kset, struct kobject *kobj, + return retval; + } + +-static struct kset_uevent_ops device_uevent_ops = { ++static const struct kset_uevent_ops device_uevent_ops = { + .filter = dev_uevent_filter, + .name = dev_uevent_name, + .uevent = dev_uevent, +diff --git a/drivers/base/memory.c b/drivers/base/memory.c +index 989429c..2272b00 100644 +--- a/drivers/base/memory.c ++++ b/drivers/base/memory.c +@@ -44,7 +44,7 @@ static int memory_uevent(struct kset *kset, struct kobject *obj, struct kobj_uev + return retval; + } + +-static struct kset_uevent_ops memory_uevent_ops = { ++static const struct kset_uevent_ops memory_uevent_ops = { + .name = memory_uevent_name, + .uevent = memory_uevent, + }; +diff --git a/drivers/base/sys.c b/drivers/base/sys.c +index 3f202f7..61c4a6f 100644 +--- a/drivers/base/sys.c ++++ b/drivers/base/sys.c +@@ -54,7 +54,7 @@ sysdev_store(struct kobject *kobj, struct attribute *attr, + return -EIO; + } + +-static struct sysfs_ops sysfs_ops = { ++static const struct sysfs_ops sysfs_ops = { + .show = sysdev_show, + .store = sysdev_store, + }; +@@ -104,7 +104,7 @@ static ssize_t sysdev_class_store(struct kobject *kobj, struct attribute *attr, + return -EIO; + } + +-static struct sysfs_ops sysfs_class_ops = { ++static const struct sysfs_ops sysfs_class_ops = { + .show = sysdev_class_show, + .store = sysdev_class_store, + }; +diff --git a/drivers/block/DAC960.c b/drivers/block/DAC960.c +index eb4fa19..1954777 100644 +--- a/drivers/block/DAC960.c ++++ b/drivers/block/DAC960.c +@@ -1973,6 +1973,8 @@ static bool DAC960_V1_ReadDeviceConfiguration(DAC960_Controller_T + unsigned long flags; + int Channel, TargetID; + ++ pax_track_stack(); ++ + if (!init_dma_loaf(Controller->PCIDevice, &local_dma, + DAC960_V1_MaxChannels*(sizeof(DAC960_V1_DCDB_T) + + sizeof(DAC960_SCSI_Inquiry_T) + +diff --git a/drivers/block/cciss.c b/drivers/block/cciss.c +index 68b90d9..7e2e3f3 100644 +--- a/drivers/block/cciss.c ++++ b/drivers/block/cciss.c +@@ -1011,6 +1011,8 @@ static int cciss_ioctl32_passthru(struct block_device *bdev, fmode_t mode, + int err; + u32 cp; + ++ memset(&arg64, 0, sizeof(arg64)); ++ + err = 0; + err |= + copy_from_user(&arg64.LUN_info, &arg32->LUN_info, +@@ -2852,7 +2854,7 @@ static unsigned long pollcomplete(int ctlr) + /* Wait (up to 20 seconds) for a command to complete */ + + for (i = 20 * HZ; i > 0; i--) { +- done = hba[ctlr]->access.command_completed(hba[ctlr]); ++ done = hba[ctlr]->access->command_completed(hba[ctlr]); + if (done == FIFO_EMPTY) + schedule_timeout_uninterruptible(1); + else +@@ -2876,7 +2878,7 @@ static int sendcmd_core(ctlr_info_t *h, CommandList_struct *c) + resend_cmd1: + + /* Disable interrupt on the board. */ +- h->access.set_intr_mask(h, CCISS_INTR_OFF); ++ h->access->set_intr_mask(h, CCISS_INTR_OFF); + + /* Make sure there is room in the command FIFO */ + /* Actually it should be completely empty at this time */ +@@ -2884,13 +2886,13 @@ resend_cmd1: + /* tape side of the driver. */ + for (i = 200000; i > 0; i--) { + /* if fifo isn't full go */ +- if (!(h->access.fifo_full(h))) ++ if (!(h->access->fifo_full(h))) + break; + udelay(10); + printk(KERN_WARNING "cciss cciss%d: SendCmd FIFO full," + " waiting!\n", h->ctlr); + } +- h->access.submit_command(h, c); /* Send the cmd */ ++ h->access->submit_command(h, c); /* Send the cmd */ + do { + complete = pollcomplete(h->ctlr); + +@@ -3023,7 +3025,7 @@ static void start_io(ctlr_info_t *h) + while (!hlist_empty(&h->reqQ)) { + c = hlist_entry(h->reqQ.first, CommandList_struct, list); + /* can't do anything if fifo is full */ +- if ((h->access.fifo_full(h))) { ++ if ((h->access->fifo_full(h))) { + printk(KERN_WARNING "cciss: fifo full\n"); + break; + } +@@ -3033,7 +3035,7 @@ static void start_io(ctlr_info_t *h) + h->Qdepth--; + + /* Tell the controller execute command */ +- h->access.submit_command(h, c); ++ h->access->submit_command(h, c); + + /* Put job onto the completed Q */ + addQ(&h->cmpQ, c); +@@ -3393,17 +3395,17 @@ startio: + + static inline unsigned long get_next_completion(ctlr_info_t *h) + { +- return h->access.command_completed(h); ++ return h->access->command_completed(h); + } + + static inline int interrupt_pending(ctlr_info_t *h) + { +- return h->access.intr_pending(h); ++ return h->access->intr_pending(h); + } + + static inline long interrupt_not_for_us(ctlr_info_t *h) + { +- return (((h->access.intr_pending(h) == 0) || ++ return (((h->access->intr_pending(h) == 0) || + (h->interrupts_enabled == 0))); + } + +@@ -3892,7 +3894,7 @@ static int __devinit cciss_pci_init(ctlr_info_t *c, struct pci_dev *pdev) + */ + c->max_commands = readl(&(c->cfgtable->CmdsOutMax)); + c->product_name = products[prod_index].product_name; +- c->access = *(products[prod_index].access); ++ c->access = products[prod_index].access; + c->nr_cmds = c->max_commands - 4; + if ((readb(&c->cfgtable->Signature[0]) != 'C') || + (readb(&c->cfgtable->Signature[1]) != 'I') || +@@ -4291,7 +4293,7 @@ static int __devinit cciss_init_one(struct pci_dev *pdev, + } + + /* make sure the board interrupts are off */ +- hba[i]->access.set_intr_mask(hba[i], CCISS_INTR_OFF); ++ hba[i]->access->set_intr_mask(hba[i], CCISS_INTR_OFF); + if (request_irq(hba[i]->intr[SIMPLE_MODE_INT], do_cciss_intr, + IRQF_DISABLED | IRQF_SHARED, hba[i]->devname, hba[i])) { + printk(KERN_ERR "cciss: Unable to get irq %d for %s\n", +@@ -4341,7 +4343,7 @@ static int __devinit cciss_init_one(struct pci_dev *pdev, + cciss_scsi_setup(i); + + /* Turn the interrupts on so we can service requests */ +- hba[i]->access.set_intr_mask(hba[i], CCISS_INTR_ON); ++ hba[i]->access->set_intr_mask(hba[i], CCISS_INTR_ON); + + /* Get the firmware version */ + inq_buff = kzalloc(sizeof(InquiryData_struct), GFP_KERNEL); +diff --git a/drivers/block/cciss.h b/drivers/block/cciss.h +index 04d6bf8..36e712d 100644 +--- a/drivers/block/cciss.h ++++ b/drivers/block/cciss.h +@@ -90,7 +90,7 @@ struct ctlr_info + // information about each logical volume + drive_info_struct *drv[CISS_MAX_LUN]; + +- struct access_method access; ++ struct access_method *access; + + /* queue and queue Info */ + struct hlist_head reqQ; +diff --git a/drivers/block/cpqarray.c b/drivers/block/cpqarray.c +index 6422651..bb1bdef 100644 +--- a/drivers/block/cpqarray.c ++++ b/drivers/block/cpqarray.c +@@ -402,7 +402,7 @@ static int __init cpqarray_register_ctlr( int i, struct pci_dev *pdev) + if (register_blkdev(COMPAQ_SMART2_MAJOR+i, hba[i]->devname)) { + goto Enomem4; + } +- hba[i]->access.set_intr_mask(hba[i], 0); ++ hba[i]->access->set_intr_mask(hba[i], 0); + if (request_irq(hba[i]->intr, do_ida_intr, + IRQF_DISABLED|IRQF_SHARED, hba[i]->devname, hba[i])) + { +@@ -460,7 +460,7 @@ static int __init cpqarray_register_ctlr( int i, struct pci_dev *pdev) + add_timer(&hba[i]->timer); + + /* Enable IRQ now that spinlock and rate limit timer are set up */ +- hba[i]->access.set_intr_mask(hba[i], FIFO_NOT_EMPTY); ++ hba[i]->access->set_intr_mask(hba[i], FIFO_NOT_EMPTY); + + for(j=0; jproduct_name = products[i].product_name; +- c->access = *(products[i].access); ++ c->access = products[i].access; + break; + } + } +@@ -793,7 +793,7 @@ static int __init cpqarray_eisa_detect(void) + hba[ctlr]->intr = intr; + sprintf(hba[ctlr]->devname, "ida%d", nr_ctlr); + hba[ctlr]->product_name = products[j].product_name; +- hba[ctlr]->access = *(products[j].access); ++ hba[ctlr]->access = products[j].access; + hba[ctlr]->ctlr = ctlr; + hba[ctlr]->board_id = board_id; + hba[ctlr]->pci_dev = NULL; /* not PCI */ +@@ -896,6 +896,8 @@ static void do_ida_request(struct request_queue *q) + struct scatterlist tmp_sg[SG_MAX]; + int i, dir, seg; + ++ pax_track_stack(); ++ + if (blk_queue_plugged(q)) + goto startio; + +@@ -968,7 +970,7 @@ static void start_io(ctlr_info_t *h) + + while((c = h->reqQ) != NULL) { + /* Can't do anything if we're busy */ +- if (h->access.fifo_full(h) == 0) ++ if (h->access->fifo_full(h) == 0) + return; + + /* Get the first entry from the request Q */ +@@ -976,7 +978,7 @@ static void start_io(ctlr_info_t *h) + h->Qdepth--; + + /* Tell the controller to do our bidding */ +- h->access.submit_command(h, c); ++ h->access->submit_command(h, c); + + /* Get onto the completion Q */ + addQ(&h->cmpQ, c); +@@ -1038,7 +1040,7 @@ static irqreturn_t do_ida_intr(int irq, void *dev_id) + unsigned long flags; + __u32 a,a1; + +- istat = h->access.intr_pending(h); ++ istat = h->access->intr_pending(h); + /* Is this interrupt for us? */ + if (istat == 0) + return IRQ_NONE; +@@ -1049,7 +1051,7 @@ static irqreturn_t do_ida_intr(int irq, void *dev_id) + */ + spin_lock_irqsave(IDA_LOCK(h->ctlr), flags); + if (istat & FIFO_NOT_EMPTY) { +- while((a = h->access.command_completed(h))) { ++ while((a = h->access->command_completed(h))) { + a1 = a; a &= ~3; + if ((c = h->cmpQ) == NULL) + { +@@ -1434,11 +1436,11 @@ static int sendcmd( + /* + * Disable interrupt + */ +- info_p->access.set_intr_mask(info_p, 0); ++ info_p->access->set_intr_mask(info_p, 0); + /* Make sure there is room in the command FIFO */ + /* Actually it should be completely empty at this time. */ + for (i = 200000; i > 0; i--) { +- temp = info_p->access.fifo_full(info_p); ++ temp = info_p->access->fifo_full(info_p); + if (temp != 0) { + break; + } +@@ -1451,7 +1453,7 @@ DBG( + /* + * Send the cmd + */ +- info_p->access.submit_command(info_p, c); ++ info_p->access->submit_command(info_p, c); + complete = pollcomplete(ctlr); + + pci_unmap_single(info_p->pci_dev, (dma_addr_t) c->req.sg[0].addr, +@@ -1534,9 +1536,9 @@ static int revalidate_allvol(ctlr_info_t *host) + * we check the new geometry. Then turn interrupts back on when + * we're done. + */ +- host->access.set_intr_mask(host, 0); ++ host->access->set_intr_mask(host, 0); + getgeometry(ctlr); +- host->access.set_intr_mask(host, FIFO_NOT_EMPTY); ++ host->access->set_intr_mask(host, FIFO_NOT_EMPTY); + + for(i=0; i 0; i--) { +- done = hba[ctlr]->access.command_completed(hba[ctlr]); ++ done = hba[ctlr]->access->command_completed(hba[ctlr]); + if (done == 0) { + udelay(10); /* a short fixed delay */ + } else +diff --git a/drivers/block/cpqarray.h b/drivers/block/cpqarray.h +index be73e9d..7fbf140 100644 +--- a/drivers/block/cpqarray.h ++++ b/drivers/block/cpqarray.h +@@ -99,7 +99,7 @@ struct ctlr_info { + drv_info_t drv[NWD]; + struct proc_dir_entry *proc; + +- struct access_method access; ++ struct access_method *access; + + cmdlist_t *reqQ; + cmdlist_t *cmpQ; +diff --git a/drivers/block/loop.c b/drivers/block/loop.c +index 8ec2d70..2804b30 100644 +--- a/drivers/block/loop.c ++++ b/drivers/block/loop.c +@@ -282,7 +282,7 @@ static int __do_lo_send_write(struct file *file, + mm_segment_t old_fs = get_fs(); + + set_fs(get_ds()); +- bw = file->f_op->write(file, buf, len, &pos); ++ bw = file->f_op->write(file, (const char __force_user *)buf, len, &pos); + set_fs(old_fs); + if (likely(bw == len)) + return 0; +diff --git a/drivers/block/nbd.c b/drivers/block/nbd.c +index 26ada47..083c480 100644 +--- a/drivers/block/nbd.c ++++ b/drivers/block/nbd.c +@@ -155,6 +155,8 @@ static int sock_xmit(struct nbd_device *lo, int send, void *buf, int size, + struct kvec iov; + sigset_t blocked, oldset; + ++ pax_track_stack(); ++ + if (unlikely(!sock)) { + printk(KERN_ERR "%s: Attempted %s on closed socket in sock_xmit\n", + lo->disk->disk_name, (send ? "send" : "recv")); +@@ -569,6 +571,8 @@ static void do_nbd_request(struct request_queue *q) + static int __nbd_ioctl(struct block_device *bdev, struct nbd_device *lo, + unsigned int cmd, unsigned long arg) + { ++ pax_track_stack(); ++ + switch (cmd) { + case NBD_DISCONNECT: { + struct request sreq; +diff --git a/drivers/block/pktcdvd.c b/drivers/block/pktcdvd.c +index a5d585d..d087be3 100644 +--- a/drivers/block/pktcdvd.c ++++ b/drivers/block/pktcdvd.c +@@ -284,7 +284,7 @@ static ssize_t kobj_pkt_store(struct kobject *kobj, + return len; + } + +-static struct sysfs_ops kobj_pkt_ops = { ++static const struct sysfs_ops kobj_pkt_ops = { + .show = kobj_pkt_show, + .store = kobj_pkt_store + }; +diff --git a/drivers/char/Kconfig b/drivers/char/Kconfig +index 6aad99e..89cd142 100644 +--- a/drivers/char/Kconfig ++++ b/drivers/char/Kconfig +@@ -90,7 +90,8 @@ config VT_HW_CONSOLE_BINDING + + config DEVKMEM + bool "/dev/kmem virtual device support" +- default y ++ default n ++ depends on !GRKERNSEC_KMEM + help + Say Y here if you want to support the /dev/kmem device. The + /dev/kmem device is rarely used, but can be used for certain +@@ -1114,6 +1115,7 @@ config DEVPORT + bool + depends on !M68K + depends on ISA || PCI ++ depends on !GRKERNSEC_KMEM + default y + + source "drivers/s390/char/Kconfig" +diff --git a/drivers/char/agp/frontend.c b/drivers/char/agp/frontend.c +index a96f319..a778a5b 100644 +--- a/drivers/char/agp/frontend.c ++++ b/drivers/char/agp/frontend.c +@@ -824,7 +824,7 @@ static int agpioc_reserve_wrap(struct agp_file_private *priv, void __user *arg) + if (copy_from_user(&reserve, arg, sizeof(struct agp_region))) + return -EFAULT; + +- if ((unsigned) reserve.seg_count >= ~0U/sizeof(struct agp_segment)) ++ if ((unsigned) reserve.seg_count >= ~0U/sizeof(struct agp_segment_priv)) + return -EFAULT; + + client = agp_find_client_by_pid(reserve.pid); +diff --git a/drivers/char/briq_panel.c b/drivers/char/briq_panel.c +index d8cff90..9628e70 100644 +--- a/drivers/char/briq_panel.c ++++ b/drivers/char/briq_panel.c +@@ -10,6 +10,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -36,6 +37,7 @@ static int vfd_is_open; + static unsigned char vfd[40]; + static int vfd_cursor; + static unsigned char ledpb, led; ++static DEFINE_MUTEX(vfd_mutex); + + static void update_vfd(void) + { +@@ -142,12 +144,15 @@ static ssize_t briq_panel_write(struct file *file, const char __user *buf, size_ + if (!vfd_is_open) + return -EBUSY; + ++ mutex_lock(&vfd_mutex); + for (;;) { + char c; + if (!indx) + break; +- if (get_user(c, buf)) ++ if (get_user(c, buf)) { ++ mutex_unlock(&vfd_mutex); + return -EFAULT; ++ } + if (esc) { + set_led(c); + esc = 0; +@@ -177,6 +182,7 @@ static ssize_t briq_panel_write(struct file *file, const char __user *buf, size_ + buf++; + } + update_vfd(); ++ mutex_unlock(&vfd_mutex); + + return len; + } +diff --git a/drivers/char/genrtc.c b/drivers/char/genrtc.c +index 31e7c91..161afc0 100644 +--- a/drivers/char/genrtc.c ++++ b/drivers/char/genrtc.c +@@ -272,6 +272,7 @@ static int gen_rtc_ioctl(struct inode *inode, struct file *file, + switch (cmd) { + + case RTC_PLL_GET: ++ memset(&pll, 0, sizeof(pll)); + if (get_rtc_pll(&pll)) + return -EINVAL; + else +diff --git a/drivers/char/hpet.c b/drivers/char/hpet.c +index 006466d..a2bb21c 100644 +--- a/drivers/char/hpet.c ++++ b/drivers/char/hpet.c +@@ -430,7 +430,7 @@ static int hpet_release(struct inode *inode, struct file *file) + return 0; + } + +-static int hpet_ioctl_common(struct hpet_dev *, int, unsigned long, int); ++static int hpet_ioctl_common(struct hpet_dev *, unsigned int, unsigned long, int); + + static int + hpet_ioctl(struct inode *inode, struct file *file, unsigned int cmd, +@@ -565,7 +565,7 @@ static inline unsigned long hpet_time_div(struct hpets *hpets, + } + + static int +-hpet_ioctl_common(struct hpet_dev *devp, int cmd, unsigned long arg, int kernel) ++hpet_ioctl_common(struct hpet_dev *devp, unsigned int cmd, unsigned long arg, int kernel) + { + struct hpet_timer __iomem *timer; + struct hpet __iomem *hpet; +@@ -608,11 +608,11 @@ hpet_ioctl_common(struct hpet_dev *devp, int cmd, unsigned long arg, int kernel) + { + struct hpet_info info; + ++ memset(&info, 0, sizeof(info)); ++ + if (devp->hd_ireqfreq) + info.hi_ireqfreq = + hpet_time_div(hpetp, devp->hd_ireqfreq); +- else +- info.hi_ireqfreq = 0; + info.hi_flags = + readq(&timer->hpet_config) & Tn_PER_INT_CAP_MASK; + info.hi_hpet = hpetp->hp_which; +diff --git a/drivers/char/hvc_beat.c b/drivers/char/hvc_beat.c +index 0afc8b8..6913fc3 100644 +--- a/drivers/char/hvc_beat.c ++++ b/drivers/char/hvc_beat.c +@@ -84,7 +84,7 @@ static int hvc_beat_put_chars(uint32_t vtermno, const char *buf, int cnt) + return cnt; + } + +-static struct hv_ops hvc_beat_get_put_ops = { ++static const struct hv_ops hvc_beat_get_put_ops = { + .get_chars = hvc_beat_get_chars, + .put_chars = hvc_beat_put_chars, + }; +diff --git a/drivers/char/hvc_console.c b/drivers/char/hvc_console.c +index 98097f2..407dddc 100644 +--- a/drivers/char/hvc_console.c ++++ b/drivers/char/hvc_console.c +@@ -125,7 +125,7 @@ static struct hvc_struct *hvc_get_by_index(int index) + * console interfaces but can still be used as a tty device. This has to be + * static because kmalloc will not work during early console init. + */ +-static struct hv_ops *cons_ops[MAX_NR_HVC_CONSOLES]; ++static const struct hv_ops *cons_ops[MAX_NR_HVC_CONSOLES]; + static uint32_t vtermnos[MAX_NR_HVC_CONSOLES] = + {[0 ... MAX_NR_HVC_CONSOLES - 1] = -1}; + +@@ -249,7 +249,7 @@ static void destroy_hvc_struct(struct kref *kref) + * vty adapters do NOT get an hvc_instantiate() callback since they + * appear after early console init. + */ +-int hvc_instantiate(uint32_t vtermno, int index, struct hv_ops *ops) ++int hvc_instantiate(uint32_t vtermno, int index, const struct hv_ops *ops) + { + struct hvc_struct *hp; + +@@ -758,7 +758,7 @@ static const struct tty_operations hvc_ops = { + }; + + struct hvc_struct __devinit *hvc_alloc(uint32_t vtermno, int data, +- struct hv_ops *ops, int outbuf_size) ++ const struct hv_ops *ops, int outbuf_size) + { + struct hvc_struct *hp; + int i; +diff --git a/drivers/char/hvc_console.h b/drivers/char/hvc_console.h +index 10950ca..ed176c3 100644 +--- a/drivers/char/hvc_console.h ++++ b/drivers/char/hvc_console.h +@@ -55,7 +55,7 @@ struct hvc_struct { + int outbuf_size; + int n_outbuf; + uint32_t vtermno; +- struct hv_ops *ops; ++ const struct hv_ops *ops; + int irq_requested; + int data; + struct winsize ws; +@@ -76,11 +76,11 @@ struct hv_ops { + }; + + /* Register a vterm and a slot index for use as a console (console_init) */ +-extern int hvc_instantiate(uint32_t vtermno, int index, struct hv_ops *ops); ++extern int hvc_instantiate(uint32_t vtermno, int index, const struct hv_ops *ops); + + /* register a vterm for hvc tty operation (module_init or hotplug add) */ + extern struct hvc_struct * __devinit hvc_alloc(uint32_t vtermno, int data, +- struct hv_ops *ops, int outbuf_size); ++ const struct hv_ops *ops, int outbuf_size); + /* remove a vterm from hvc tty operation (module_exit or hotplug remove) */ + extern int hvc_remove(struct hvc_struct *hp); + +diff --git a/drivers/char/hvc_iseries.c b/drivers/char/hvc_iseries.c +index 936d05b..fd02426 100644 +--- a/drivers/char/hvc_iseries.c ++++ b/drivers/char/hvc_iseries.c +@@ -197,7 +197,7 @@ done: + return sent; + } + +-static struct hv_ops hvc_get_put_ops = { ++static const struct hv_ops hvc_get_put_ops = { + .get_chars = get_chars, + .put_chars = put_chars, + .notifier_add = notifier_add_irq, +diff --git a/drivers/char/hvc_iucv.c b/drivers/char/hvc_iucv.c +index b0e168f..69cda2a 100644 +--- a/drivers/char/hvc_iucv.c ++++ b/drivers/char/hvc_iucv.c +@@ -924,7 +924,7 @@ static int hvc_iucv_pm_restore_thaw(struct device *dev) + + + /* HVC operations */ +-static struct hv_ops hvc_iucv_ops = { ++static const struct hv_ops hvc_iucv_ops = { + .get_chars = hvc_iucv_get_chars, + .put_chars = hvc_iucv_put_chars, + .notifier_add = hvc_iucv_notifier_add, +diff --git a/drivers/char/hvc_rtas.c b/drivers/char/hvc_rtas.c +index 88590d0..61c4a61 100644 +--- a/drivers/char/hvc_rtas.c ++++ b/drivers/char/hvc_rtas.c +@@ -71,7 +71,7 @@ static int hvc_rtas_read_console(uint32_t vtermno, char *buf, int count) + return i; + } + +-static struct hv_ops hvc_rtas_get_put_ops = { ++static const struct hv_ops hvc_rtas_get_put_ops = { + .get_chars = hvc_rtas_read_console, + .put_chars = hvc_rtas_write_console, + }; +diff --git a/drivers/char/hvc_udbg.c b/drivers/char/hvc_udbg.c +index bd63ba8..b0957e6 100644 +--- a/drivers/char/hvc_udbg.c ++++ b/drivers/char/hvc_udbg.c +@@ -58,7 +58,7 @@ static int hvc_udbg_get(uint32_t vtermno, char *buf, int count) + return i; + } + +-static struct hv_ops hvc_udbg_ops = { ++static const struct hv_ops hvc_udbg_ops = { + .get_chars = hvc_udbg_get, + .put_chars = hvc_udbg_put, + }; +diff --git a/drivers/char/hvc_vio.c b/drivers/char/hvc_vio.c +index 10be343..27370e9 100644 +--- a/drivers/char/hvc_vio.c ++++ b/drivers/char/hvc_vio.c +@@ -77,7 +77,7 @@ static int filtered_get_chars(uint32_t vtermno, char *buf, int count) + return got; + } + +-static struct hv_ops hvc_get_put_ops = { ++static const struct hv_ops hvc_get_put_ops = { + .get_chars = filtered_get_chars, + .put_chars = hvc_put_chars, + .notifier_add = notifier_add_irq, +diff --git a/drivers/char/hvc_xen.c b/drivers/char/hvc_xen.c +index a6ee32b..94f8c26 100644 +--- a/drivers/char/hvc_xen.c ++++ b/drivers/char/hvc_xen.c +@@ -120,7 +120,7 @@ static int read_console(uint32_t vtermno, char *buf, int len) + return recv; + } + +-static struct hv_ops hvc_ops = { ++static const struct hv_ops hvc_ops = { + .get_chars = read_console, + .put_chars = write_console, + .notifier_add = notifier_add_irq, +diff --git a/drivers/char/hvcs.c b/drivers/char/hvcs.c +index 266b858..f3ee0bb 100644 +--- a/drivers/char/hvcs.c ++++ b/drivers/char/hvcs.c +@@ -82,6 +82,7 @@ + #include + #include + #include ++#include + + /* + * 1.3.0 -> 1.3.1 In hvcs_open memset(..,0x00,..) instead of memset(..,0x3F,00). +@@ -269,7 +270,7 @@ struct hvcs_struct { + unsigned int index; + + struct tty_struct *tty; +- int open_count; ++ local_t open_count; + + /* + * Used to tell the driver kernel_thread what operations need to take +@@ -419,7 +420,7 @@ static ssize_t hvcs_vterm_state_store(struct device *dev, struct device_attribut + + spin_lock_irqsave(&hvcsd->lock, flags); + +- if (hvcsd->open_count > 0) { ++ if (local_read(&hvcsd->open_count) > 0) { + spin_unlock_irqrestore(&hvcsd->lock, flags); + printk(KERN_INFO "HVCS: vterm state unchanged. " + "The hvcs device node is still in use.\n"); +@@ -1135,7 +1136,7 @@ static int hvcs_open(struct tty_struct *tty, struct file *filp) + if ((retval = hvcs_partner_connect(hvcsd))) + goto error_release; + +- hvcsd->open_count = 1; ++ local_set(&hvcsd->open_count, 1); + hvcsd->tty = tty; + tty->driver_data = hvcsd; + +@@ -1169,7 +1170,7 @@ fast_open: + + spin_lock_irqsave(&hvcsd->lock, flags); + kref_get(&hvcsd->kref); +- hvcsd->open_count++; ++ local_inc(&hvcsd->open_count); + hvcsd->todo_mask |= HVCS_SCHED_READ; + spin_unlock_irqrestore(&hvcsd->lock, flags); + +@@ -1213,7 +1214,7 @@ static void hvcs_close(struct tty_struct *tty, struct file *filp) + hvcsd = tty->driver_data; + + spin_lock_irqsave(&hvcsd->lock, flags); +- if (--hvcsd->open_count == 0) { ++ if (local_dec_and_test(&hvcsd->open_count)) { + + vio_disable_interrupts(hvcsd->vdev); + +@@ -1239,10 +1240,10 @@ static void hvcs_close(struct tty_struct *tty, struct file *filp) + free_irq(irq, hvcsd); + kref_put(&hvcsd->kref, destroy_hvcs_struct); + return; +- } else if (hvcsd->open_count < 0) { ++ } else if (local_read(&hvcsd->open_count) < 0) { + printk(KERN_ERR "HVCS: vty-server@%X open_count: %d" + " is missmanaged.\n", +- hvcsd->vdev->unit_address, hvcsd->open_count); ++ hvcsd->vdev->unit_address, local_read(&hvcsd->open_count)); + } + + spin_unlock_irqrestore(&hvcsd->lock, flags); +@@ -1258,7 +1259,7 @@ static void hvcs_hangup(struct tty_struct * tty) + + spin_lock_irqsave(&hvcsd->lock, flags); + /* Preserve this so that we know how many kref refs to put */ +- temp_open_count = hvcsd->open_count; ++ temp_open_count = local_read(&hvcsd->open_count); + + /* + * Don't kref put inside the spinlock because the destruction +@@ -1273,7 +1274,7 @@ static void hvcs_hangup(struct tty_struct * tty) + hvcsd->tty->driver_data = NULL; + hvcsd->tty = NULL; + +- hvcsd->open_count = 0; ++ local_set(&hvcsd->open_count, 0); + + /* This will drop any buffered data on the floor which is OK in a hangup + * scenario. */ +@@ -1344,7 +1345,7 @@ static int hvcs_write(struct tty_struct *tty, + * the middle of a write operation? This is a crummy place to do this + * but we want to keep it all in the spinlock. + */ +- if (hvcsd->open_count <= 0) { ++ if (local_read(&hvcsd->open_count) <= 0) { + spin_unlock_irqrestore(&hvcsd->lock, flags); + return -ENODEV; + } +@@ -1418,7 +1419,7 @@ static int hvcs_write_room(struct tty_struct *tty) + { + struct hvcs_struct *hvcsd = tty->driver_data; + +- if (!hvcsd || hvcsd->open_count <= 0) ++ if (!hvcsd || local_read(&hvcsd->open_count) <= 0) + return 0; + + return HVCS_BUFF_LEN - hvcsd->chars_in_buffer; +diff --git a/drivers/char/ipmi/ipmi_msghandler.c b/drivers/char/ipmi/ipmi_msghandler.c +index ec5e3f8..02455ba 100644 +--- a/drivers/char/ipmi/ipmi_msghandler.c ++++ b/drivers/char/ipmi/ipmi_msghandler.c +@@ -414,7 +414,7 @@ struct ipmi_smi { + struct proc_dir_entry *proc_dir; + char proc_dir_name[10]; + +- atomic_t stats[IPMI_NUM_STATS]; ++ atomic_unchecked_t stats[IPMI_NUM_STATS]; + + /* + * run_to_completion duplicate of smb_info, smi_info +@@ -447,9 +447,9 @@ static DEFINE_MUTEX(smi_watchers_mutex); + + + #define ipmi_inc_stat(intf, stat) \ +- atomic_inc(&(intf)->stats[IPMI_STAT_ ## stat]) ++ atomic_inc_unchecked(&(intf)->stats[IPMI_STAT_ ## stat]) + #define ipmi_get_stat(intf, stat) \ +- ((unsigned int) atomic_read(&(intf)->stats[IPMI_STAT_ ## stat])) ++ ((unsigned int) atomic_read_unchecked(&(intf)->stats[IPMI_STAT_ ## stat])) + + static int is_lan_addr(struct ipmi_addr *addr) + { +@@ -2808,7 +2808,7 @@ int ipmi_register_smi(struct ipmi_smi_handlers *handlers, + INIT_LIST_HEAD(&intf->cmd_rcvrs); + init_waitqueue_head(&intf->waitq); + for (i = 0; i < IPMI_NUM_STATS; i++) +- atomic_set(&intf->stats[i], 0); ++ atomic_set_unchecked(&intf->stats[i], 0); + + intf->proc_dir = NULL; + +@@ -4160,6 +4160,8 @@ static void send_panic_events(char *str) + struct ipmi_smi_msg smi_msg; + struct ipmi_recv_msg recv_msg; + ++ pax_track_stack(); ++ + si = (struct ipmi_system_interface_addr *) &addr; + si->addr_type = IPMI_SYSTEM_INTERFACE_ADDR_TYPE; + si->channel = IPMI_BMC_CHANNEL; +diff --git a/drivers/char/ipmi/ipmi_si_intf.c b/drivers/char/ipmi/ipmi_si_intf.c +index abae8c9..8021979 100644 +--- a/drivers/char/ipmi/ipmi_si_intf.c ++++ b/drivers/char/ipmi/ipmi_si_intf.c +@@ -277,7 +277,7 @@ struct smi_info { + unsigned char slave_addr; + + /* Counters and things for the proc filesystem. */ +- atomic_t stats[SI_NUM_STATS]; ++ atomic_unchecked_t stats[SI_NUM_STATS]; + + struct task_struct *thread; + +@@ -285,9 +285,9 @@ struct smi_info { + }; + + #define smi_inc_stat(smi, stat) \ +- atomic_inc(&(smi)->stats[SI_STAT_ ## stat]) ++ atomic_inc_unchecked(&(smi)->stats[SI_STAT_ ## stat]) + #define smi_get_stat(smi, stat) \ +- ((unsigned int) atomic_read(&(smi)->stats[SI_STAT_ ## stat])) ++ ((unsigned int) atomic_read_unchecked(&(smi)->stats[SI_STAT_ ## stat])) + + #define SI_MAX_PARMS 4 + +@@ -2931,7 +2931,7 @@ static int try_smi_init(struct smi_info *new_smi) + atomic_set(&new_smi->req_events, 0); + new_smi->run_to_completion = 0; + for (i = 0; i < SI_NUM_STATS; i++) +- atomic_set(&new_smi->stats[i], 0); ++ atomic_set_unchecked(&new_smi->stats[i], 0); + + new_smi->interrupt_disabled = 0; + atomic_set(&new_smi->stop_operation, 0); +diff --git a/drivers/char/istallion.c b/drivers/char/istallion.c +index 402838f..55e2200 100644 +--- a/drivers/char/istallion.c ++++ b/drivers/char/istallion.c +@@ -187,7 +187,6 @@ static struct ktermios stli_deftermios = { + * re-used for each stats call. + */ + static comstats_t stli_comstats; +-static combrd_t stli_brdstats; + static struct asystats stli_cdkstats; + + /*****************************************************************************/ +@@ -4058,6 +4057,7 @@ static int stli_getbrdstats(combrd_t __user *bp) + { + struct stlibrd *brdp; + unsigned int i; ++ combrd_t stli_brdstats; + + if (copy_from_user(&stli_brdstats, bp, sizeof(combrd_t))) + return -EFAULT; +@@ -4269,6 +4269,8 @@ static int stli_getportstruct(struct stliport __user *arg) + struct stliport stli_dummyport; + struct stliport *portp; + ++ pax_track_stack(); ++ + if (copy_from_user(&stli_dummyport, arg, sizeof(struct stliport))) + return -EFAULT; + portp = stli_getport(stli_dummyport.brdnr, stli_dummyport.panelnr, +@@ -4291,6 +4293,8 @@ static int stli_getbrdstruct(struct stlibrd __user *arg) + struct stlibrd stli_dummybrd; + struct stlibrd *brdp; + ++ pax_track_stack(); ++ + if (copy_from_user(&stli_dummybrd, arg, sizeof(struct stlibrd))) + return -EFAULT; + if (stli_dummybrd.brdnr >= STL_MAXBRDS) +diff --git a/drivers/char/keyboard.c b/drivers/char/keyboard.c +index 950837c..e55a288 100644 +--- a/drivers/char/keyboard.c ++++ b/drivers/char/keyboard.c +@@ -635,6 +635,16 @@ static void k_spec(struct vc_data *vc, unsigned char value, char up_flag) + kbd->kbdmode == VC_MEDIUMRAW) && + value != KVAL(K_SAK)) + return; /* SAK is allowed even in raw mode */ ++ ++#if defined(CONFIG_GRKERNSEC_PROC) || defined(CONFIG_GRKERNSEC_PROC_MEMMAP) ++ { ++ void *func = fn_handler[value]; ++ if (func == fn_show_state || func == fn_show_ptregs || ++ func == fn_show_mem) ++ return; ++ } ++#endif ++ + fn_handler[value](vc); + } + +@@ -1386,7 +1396,7 @@ static const struct input_device_id kbd_ids[] = { + .evbit = { BIT_MASK(EV_SND) }, + }, + +- { }, /* Terminating entry */ ++ { 0 }, /* Terminating entry */ + }; + + MODULE_DEVICE_TABLE(input, kbd_ids); +diff --git a/drivers/char/mbcs.c b/drivers/char/mbcs.c +index 87c67b4..230527a 100644 +--- a/drivers/char/mbcs.c ++++ b/drivers/char/mbcs.c +@@ -799,7 +799,7 @@ static int mbcs_remove(struct cx_dev *dev) + return 0; + } + +-static const struct cx_device_id __devinitdata mbcs_id_table[] = { ++static const struct cx_device_id __devinitconst mbcs_id_table[] = { + { + .part_num = MBCS_PART_NUM, + .mfg_num = MBCS_MFG_NUM, +diff --git a/drivers/char/mem.c b/drivers/char/mem.c +index 1270f64..8495f49 100644 +--- a/drivers/char/mem.c ++++ b/drivers/char/mem.c +@@ -18,6 +18,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -35,6 +36,10 @@ + # include + #endif + ++#if defined(CONFIG_GRKERNSEC) && !defined(CONFIG_GRKERNSEC_NO_RBAC) ++extern struct file_operations grsec_fops; ++#endif ++ + static inline unsigned long size_inside_page(unsigned long start, + unsigned long size) + { +@@ -102,9 +107,13 @@ static inline int range_is_allowed(unsigned long pfn, unsigned long size) + + while (cursor < to) { + if (!devmem_is_allowed(pfn)) { ++#ifdef CONFIG_GRKERNSEC_KMEM ++ gr_handle_mem_readwrite(from, to); ++#else + printk(KERN_INFO + "Program %s tried to access /dev/mem between %Lx->%Lx.\n", + current->comm, from, to); ++#endif + return 0; + } + cursor += PAGE_SIZE; +@@ -112,6 +121,11 @@ static inline int range_is_allowed(unsigned long pfn, unsigned long size) + } + return 1; + } ++#elif defined(CONFIG_GRKERNSEC_KMEM) ++static inline int range_is_allowed(unsigned long pfn, unsigned long size) ++{ ++ return 0; ++} + #else + static inline int range_is_allowed(unsigned long pfn, unsigned long size) + { +@@ -155,6 +169,8 @@ static ssize_t read_mem(struct file * file, char __user * buf, + #endif + + while (count > 0) { ++ char *temp; ++ + /* + * Handle first page in case it's not aligned + */ +@@ -177,11 +193,31 @@ static ssize_t read_mem(struct file * file, char __user * buf, + if (!ptr) + return -EFAULT; + +- if (copy_to_user(buf, ptr, sz)) { ++#ifdef CONFIG_PAX_USERCOPY ++ temp = kmalloc(sz, GFP_KERNEL); ++ if (!temp) { ++ unxlate_dev_mem_ptr(p, ptr); ++ return -ENOMEM; ++ } ++ memcpy(temp, ptr, sz); ++#else ++ temp = ptr; ++#endif ++ ++ if (copy_to_user(buf, temp, sz)) { ++ ++#ifdef CONFIG_PAX_USERCOPY ++ kfree(temp); ++#endif ++ + unxlate_dev_mem_ptr(p, ptr); + return -EFAULT; + } + ++#ifdef CONFIG_PAX_USERCOPY ++ kfree(temp); ++#endif ++ + unxlate_dev_mem_ptr(p, ptr); + + buf += sz; +@@ -419,9 +455,8 @@ static ssize_t read_kmem(struct file *file, char __user *buf, + size_t count, loff_t *ppos) + { + unsigned long p = *ppos; +- ssize_t low_count, read, sz; ++ ssize_t low_count, read, sz, err = 0; + char * kbuf; /* k-addr because vread() takes vmlist_lock rwlock */ +- int err = 0; + + read = 0; + if (p < (unsigned long) high_memory) { +@@ -444,6 +479,8 @@ static ssize_t read_kmem(struct file *file, char __user *buf, + } + #endif + while (low_count > 0) { ++ char *temp; ++ + sz = size_inside_page(p, low_count); + + /* +@@ -453,7 +490,22 @@ static ssize_t read_kmem(struct file *file, char __user *buf, + */ + kbuf = xlate_dev_kmem_ptr((char *)p); + +- if (copy_to_user(buf, kbuf, sz)) ++#ifdef CONFIG_PAX_USERCOPY ++ temp = kmalloc(sz, GFP_KERNEL); ++ if (!temp) ++ return -ENOMEM; ++ memcpy(temp, kbuf, sz); ++#else ++ temp = kbuf; ++#endif ++ ++ err = copy_to_user(buf, temp, sz); ++ ++#ifdef CONFIG_PAX_USERCOPY ++ kfree(temp); ++#endif ++ ++ if (err) + return -EFAULT; + buf += sz; + p += sz; +@@ -889,6 +941,9 @@ static const struct memdev { + #ifdef CONFIG_CRASH_DUMP + [12] = { "oldmem", 0, &oldmem_fops, NULL }, + #endif ++#if defined(CONFIG_GRKERNSEC) && !defined(CONFIG_GRKERNSEC_NO_RBAC) ++ [13] = { "grsec",S_IRUSR | S_IWUGO, &grsec_fops, NULL }, ++#endif + }; + + static int memory_open(struct inode *inode, struct file *filp) +diff --git a/drivers/char/mmtimer.c b/drivers/char/mmtimer.c +index 918711a..4ffaf5e 100644 +--- a/drivers/char/mmtimer.c ++++ b/drivers/char/mmtimer.c +@@ -756,7 +756,7 @@ static int sgi_timer_set(struct k_itimer *timr, int flags, + return err; + } + +-static struct k_clock sgi_clock = { ++static k_clock_no_const sgi_clock = { + .res = 0, + .clock_set = sgi_clock_set, + .clock_get = sgi_clock_get, +diff --git a/drivers/char/pcmcia/ipwireless/tty.c b/drivers/char/pcmcia/ipwireless/tty.c +index 674b3ab..a8d1970 100644 +--- a/drivers/char/pcmcia/ipwireless/tty.c ++++ b/drivers/char/pcmcia/ipwireless/tty.c +@@ -29,6 +29,7 @@ + #include + #include + #include ++#include + + #include "tty.h" + #include "network.h" +@@ -51,7 +52,7 @@ struct ipw_tty { + int tty_type; + struct ipw_network *network; + struct tty_struct *linux_tty; +- int open_count; ++ local_t open_count; + unsigned int control_lines; + struct mutex ipw_tty_mutex; + int tx_bytes_queued; +@@ -127,10 +128,10 @@ static int ipw_open(struct tty_struct *linux_tty, struct file *filp) + mutex_unlock(&tty->ipw_tty_mutex); + return -ENODEV; + } +- if (tty->open_count == 0) ++ if (local_read(&tty->open_count) == 0) + tty->tx_bytes_queued = 0; + +- tty->open_count++; ++ local_inc(&tty->open_count); + + tty->linux_tty = linux_tty; + linux_tty->driver_data = tty; +@@ -146,9 +147,7 @@ static int ipw_open(struct tty_struct *linux_tty, struct file *filp) + + static void do_ipw_close(struct ipw_tty *tty) + { +- tty->open_count--; +- +- if (tty->open_count == 0) { ++ if (local_dec_return(&tty->open_count) == 0) { + struct tty_struct *linux_tty = tty->linux_tty; + + if (linux_tty != NULL) { +@@ -169,7 +168,7 @@ static void ipw_hangup(struct tty_struct *linux_tty) + return; + + mutex_lock(&tty->ipw_tty_mutex); +- if (tty->open_count == 0) { ++ if (local_read(&tty->open_count) == 0) { + mutex_unlock(&tty->ipw_tty_mutex); + return; + } +@@ -198,7 +197,7 @@ void ipwireless_tty_received(struct ipw_tty *tty, unsigned char *data, + return; + } + +- if (!tty->open_count) { ++ if (!local_read(&tty->open_count)) { + mutex_unlock(&tty->ipw_tty_mutex); + return; + } +@@ -240,7 +239,7 @@ static int ipw_write(struct tty_struct *linux_tty, + return -ENODEV; + + mutex_lock(&tty->ipw_tty_mutex); +- if (!tty->open_count) { ++ if (!local_read(&tty->open_count)) { + mutex_unlock(&tty->ipw_tty_mutex); + return -EINVAL; + } +@@ -280,7 +279,7 @@ static int ipw_write_room(struct tty_struct *linux_tty) + if (!tty) + return -ENODEV; + +- if (!tty->open_count) ++ if (!local_read(&tty->open_count)) + return -EINVAL; + + room = IPWIRELESS_TX_QUEUE_SIZE - tty->tx_bytes_queued; +@@ -322,7 +321,7 @@ static int ipw_chars_in_buffer(struct tty_struct *linux_tty) + if (!tty) + return 0; + +- if (!tty->open_count) ++ if (!local_read(&tty->open_count)) + return 0; + + return tty->tx_bytes_queued; +@@ -403,7 +402,7 @@ static int ipw_tiocmget(struct tty_struct *linux_tty, struct file *file) + if (!tty) + return -ENODEV; + +- if (!tty->open_count) ++ if (!local_read(&tty->open_count)) + return -EINVAL; + + return get_control_lines(tty); +@@ -419,7 +418,7 @@ ipw_tiocmset(struct tty_struct *linux_tty, struct file *file, + if (!tty) + return -ENODEV; + +- if (!tty->open_count) ++ if (!local_read(&tty->open_count)) + return -EINVAL; + + return set_control_lines(tty, set, clear); +@@ -433,7 +432,7 @@ static int ipw_ioctl(struct tty_struct *linux_tty, struct file *file, + if (!tty) + return -ENODEV; + +- if (!tty->open_count) ++ if (!local_read(&tty->open_count)) + return -EINVAL; + + /* FIXME: Exactly how is the tty object locked here .. */ +@@ -591,7 +590,7 @@ void ipwireless_tty_free(struct ipw_tty *tty) + against a parallel ioctl etc */ + mutex_lock(&ttyj->ipw_tty_mutex); + } +- while (ttyj->open_count) ++ while (local_read(&ttyj->open_count)) + do_ipw_close(ttyj); + ipwireless_disassociate_network_ttys(network, + ttyj->channel_idx); +diff --git a/drivers/char/pty.c b/drivers/char/pty.c +index 62f282e..e45c45c 100644 +--- a/drivers/char/pty.c ++++ b/drivers/char/pty.c +@@ -736,8 +736,10 @@ static void __init unix98_pty_init(void) + register_sysctl_table(pty_root_table); + + /* Now create the /dev/ptmx special device */ ++ pax_open_kernel(); + tty_default_fops(&ptmx_fops); +- ptmx_fops.open = ptmx_open; ++ *(void **)&ptmx_fops.open = ptmx_open; ++ pax_close_kernel(); + + cdev_init(&ptmx_cdev, &ptmx_fops); + if (cdev_add(&ptmx_cdev, MKDEV(TTYAUX_MAJOR, 2), 1) || +diff --git a/drivers/char/random.c b/drivers/char/random.c +index 3a19e2d..6ed09d3 100644 +--- a/drivers/char/random.c ++++ b/drivers/char/random.c +@@ -254,8 +254,13 @@ + /* + * Configuration information + */ ++#ifdef CONFIG_GRKERNSEC_RANDNET ++#define INPUT_POOL_WORDS 512 ++#define OUTPUT_POOL_WORDS 128 ++#else + #define INPUT_POOL_WORDS 128 + #define OUTPUT_POOL_WORDS 32 ++#endif + #define SEC_XFER_SIZE 512 + + /* +@@ -292,10 +297,17 @@ static struct poolinfo { + int poolwords; + int tap1, tap2, tap3, tap4, tap5; + } poolinfo_table[] = { ++#ifdef CONFIG_GRKERNSEC_RANDNET ++ /* x^512 + x^411 + x^308 + x^208 +x^104 + x + 1 -- 225 */ ++ { 512, 411, 308, 208, 104, 1 }, ++ /* x^128 + x^103 + x^76 + x^51 + x^25 + x + 1 -- 105 */ ++ { 128, 103, 76, 51, 25, 1 }, ++#else + /* x^128 + x^103 + x^76 + x^51 +x^25 + x + 1 -- 105 */ + { 128, 103, 76, 51, 25, 1 }, + /* x^32 + x^26 + x^20 + x^14 + x^7 + x + 1 -- 15 */ + { 32, 26, 20, 14, 7, 1 }, ++#endif + #if 0 + /* x^2048 + x^1638 + x^1231 + x^819 + x^411 + x + 1 -- 115 */ + { 2048, 1638, 1231, 819, 411, 1 }, +@@ -1209,7 +1221,7 @@ EXPORT_SYMBOL(generate_random_uuid); + #include + + static int min_read_thresh = 8, min_write_thresh; +-static int max_read_thresh = INPUT_POOL_WORDS * 32; ++static int max_read_thresh = OUTPUT_POOL_WORDS * 32; + static int max_write_thresh = INPUT_POOL_WORDS * 32; + static char sysctl_bootid[16]; + +diff --git a/drivers/char/rocket.c b/drivers/char/rocket.c +index 0e29a23..0efc2c2 100644 +--- a/drivers/char/rocket.c ++++ b/drivers/char/rocket.c +@@ -1266,6 +1266,8 @@ static int get_ports(struct r_port *info, struct rocket_ports __user *retports) + struct rocket_ports tmp; + int board; + ++ pax_track_stack(); ++ + if (!retports) + return -EFAULT; + memset(&tmp, 0, sizeof (tmp)); +diff --git a/drivers/char/sonypi.c b/drivers/char/sonypi.c +index 8c262aa..4d3b058 100644 +--- a/drivers/char/sonypi.c ++++ b/drivers/char/sonypi.c +@@ -55,6 +55,7 @@ + #include + #include + #include ++#include + + #include + +@@ -491,7 +492,7 @@ static struct sonypi_device { + spinlock_t fifo_lock; + wait_queue_head_t fifo_proc_list; + struct fasync_struct *fifo_async; +- int open_count; ++ local_t open_count; + int model; + struct input_dev *input_jog_dev; + struct input_dev *input_key_dev; +@@ -895,7 +896,7 @@ static int sonypi_misc_fasync(int fd, struct file *filp, int on) + static int sonypi_misc_release(struct inode *inode, struct file *file) + { + mutex_lock(&sonypi_device.lock); +- sonypi_device.open_count--; ++ local_dec(&sonypi_device.open_count); + mutex_unlock(&sonypi_device.lock); + return 0; + } +@@ -905,9 +906,9 @@ static int sonypi_misc_open(struct inode *inode, struct file *file) + lock_kernel(); + mutex_lock(&sonypi_device.lock); + /* Flush input queue on first open */ +- if (!sonypi_device.open_count) ++ if (!local_read(&sonypi_device.open_count)) + kfifo_reset(sonypi_device.fifo); +- sonypi_device.open_count++; ++ local_inc(&sonypi_device.open_count); + mutex_unlock(&sonypi_device.lock); + unlock_kernel(); + return 0; +diff --git a/drivers/char/stallion.c b/drivers/char/stallion.c +index db6dcfa..13834cb 100644 +--- a/drivers/char/stallion.c ++++ b/drivers/char/stallion.c +@@ -2448,6 +2448,8 @@ static int stl_getportstruct(struct stlport __user *arg) + struct stlport stl_dummyport; + struct stlport *portp; + ++ pax_track_stack(); ++ + if (copy_from_user(&stl_dummyport, arg, sizeof(struct stlport))) + return -EFAULT; + portp = stl_getport(stl_dummyport.brdnr, stl_dummyport.panelnr, +diff --git a/drivers/char/tpm/tpm.c b/drivers/char/tpm/tpm.c +index a0789f6..cea3902 100644 +--- a/drivers/char/tpm/tpm.c ++++ b/drivers/char/tpm/tpm.c +@@ -405,7 +405,7 @@ static ssize_t tpm_transmit(struct tpm_chip *chip, const char *buf, + chip->vendor.req_complete_val) + goto out_recv; + +- if ((status == chip->vendor.req_canceled)) { ++ if (status == chip->vendor.req_canceled) { + dev_err(chip->dev, "Operation Canceled\n"); + rc = -ECANCELED; + goto out; +@@ -824,6 +824,8 @@ ssize_t tpm_show_pubek(struct device *dev, struct device_attribute *attr, + + struct tpm_chip *chip = dev_get_drvdata(dev); + ++ pax_track_stack(); ++ + tpm_cmd.header.in = tpm_readpubek_header; + err = transmit_cmd(chip, &tpm_cmd, READ_PUBEK_RESULT_SIZE, + "attempting to read the PUBEK"); +diff --git a/drivers/char/tpm/tpm_bios.c b/drivers/char/tpm/tpm_bios.c +index bf2170f..ce8cab9 100644 +--- a/drivers/char/tpm/tpm_bios.c ++++ b/drivers/char/tpm/tpm_bios.c +@@ -172,7 +172,7 @@ static void *tpm_bios_measurements_start(struct seq_file *m, loff_t *pos) + event = addr; + + if ((event->event_type == 0 && event->event_size == 0) || +- ((addr + sizeof(struct tcpa_event) + event->event_size) >= limit)) ++ (event->event_size >= limit - addr - sizeof(struct tcpa_event))) + return NULL; + + return addr; +@@ -197,7 +197,7 @@ static void *tpm_bios_measurements_next(struct seq_file *m, void *v, + return NULL; + + if ((event->event_type == 0 && event->event_size == 0) || +- ((v + sizeof(struct tcpa_event) + event->event_size) >= limit)) ++ (event->event_size >= limit - v - sizeof(struct tcpa_event))) + return NULL; + + (*pos)++; +@@ -290,7 +290,8 @@ static int tpm_binary_bios_measurements_show(struct seq_file *m, void *v) + int i; + + for (i = 0; i < sizeof(struct tcpa_event) + event->event_size; i++) +- seq_putc(m, data[i]); ++ if (!seq_putc(m, data[i])) ++ return -EFAULT; + + return 0; + } +@@ -409,8 +410,13 @@ static int read_log(struct tpm_bios_log *log) + log->bios_event_log_end = log->bios_event_log + len; + + virt = acpi_os_map_memory(start, len); ++ if (!virt) { ++ kfree(log->bios_event_log); ++ log->bios_event_log = NULL; ++ return -EFAULT; ++ } + +- memcpy(log->bios_event_log, virt, len); ++ memcpy(log->bios_event_log, (const char __force_kernel *)virt, len); + + acpi_os_unmap_memory(virt, len); + return 0; +diff --git a/drivers/char/tty_io.c b/drivers/char/tty_io.c +index 123cedf..6664cb4 100644 +--- a/drivers/char/tty_io.c ++++ b/drivers/char/tty_io.c +@@ -146,7 +146,7 @@ static int tty_open(struct inode *, struct file *); + static int tty_release(struct inode *, struct file *); + long tty_ioctl(struct file *file, unsigned int cmd, unsigned long arg); + #ifdef CONFIG_COMPAT +-static long tty_compat_ioctl(struct file *file, unsigned int cmd, ++long tty_compat_ioctl(struct file *file, unsigned int cmd, + unsigned long arg); + #else + #define tty_compat_ioctl NULL +@@ -1774,6 +1774,7 @@ got_driver: + + if (IS_ERR(tty)) { + mutex_unlock(&tty_mutex); ++ tty_driver_kref_put(driver); + return PTR_ERR(tty); + } + } +@@ -2603,8 +2604,10 @@ long tty_ioctl(struct file *file, unsigned int cmd, unsigned long arg) + return retval; + } + ++EXPORT_SYMBOL(tty_ioctl); ++ + #ifdef CONFIG_COMPAT +-static long tty_compat_ioctl(struct file *file, unsigned int cmd, ++long tty_compat_ioctl(struct file *file, unsigned int cmd, + unsigned long arg) + { + struct inode *inode = file->f_dentry->d_inode; +@@ -2628,6 +2631,8 @@ static long tty_compat_ioctl(struct file *file, unsigned int cmd, + + return retval; + } ++ ++EXPORT_SYMBOL(tty_compat_ioctl); + #endif + + /* +@@ -3073,7 +3078,7 @@ EXPORT_SYMBOL_GPL(get_current_tty); + + void tty_default_fops(struct file_operations *fops) + { +- *fops = tty_fops; ++ memcpy((void *)fops, &tty_fops, sizeof(tty_fops)); + } + + /* +diff --git a/drivers/char/tty_ldisc.c b/drivers/char/tty_ldisc.c +index d814a3d..b55b9c9 100644 +--- a/drivers/char/tty_ldisc.c ++++ b/drivers/char/tty_ldisc.c +@@ -74,7 +74,7 @@ static void put_ldisc(struct tty_ldisc *ld) + if (atomic_dec_and_lock(&ld->users, &tty_ldisc_lock)) { + struct tty_ldisc_ops *ldo = ld->ops; + +- ldo->refcount--; ++ atomic_dec(&ldo->refcount); + module_put(ldo->owner); + spin_unlock_irqrestore(&tty_ldisc_lock, flags); + +@@ -109,7 +109,7 @@ int tty_register_ldisc(int disc, struct tty_ldisc_ops *new_ldisc) + spin_lock_irqsave(&tty_ldisc_lock, flags); + tty_ldiscs[disc] = new_ldisc; + new_ldisc->num = disc; +- new_ldisc->refcount = 0; ++ atomic_set(&new_ldisc->refcount, 0); + spin_unlock_irqrestore(&tty_ldisc_lock, flags); + + return ret; +@@ -137,7 +137,7 @@ int tty_unregister_ldisc(int disc) + return -EINVAL; + + spin_lock_irqsave(&tty_ldisc_lock, flags); +- if (tty_ldiscs[disc]->refcount) ++ if (atomic_read(&tty_ldiscs[disc]->refcount)) + ret = -EBUSY; + else + tty_ldiscs[disc] = NULL; +@@ -158,7 +158,7 @@ static struct tty_ldisc_ops *get_ldops(int disc) + if (ldops) { + ret = ERR_PTR(-EAGAIN); + if (try_module_get(ldops->owner)) { +- ldops->refcount++; ++ atomic_inc(&ldops->refcount); + ret = ldops; + } + } +@@ -171,7 +171,7 @@ static void put_ldops(struct tty_ldisc_ops *ldops) + unsigned long flags; + + spin_lock_irqsave(&tty_ldisc_lock, flags); +- ldops->refcount--; ++ atomic_dec(&ldops->refcount); + module_put(ldops->owner); + spin_unlock_irqrestore(&tty_ldisc_lock, flags); + } +diff --git a/drivers/char/virtio_console.c b/drivers/char/virtio_console.c +index a035ae3..c27fe2c 100644 +--- a/drivers/char/virtio_console.c ++++ b/drivers/char/virtio_console.c +@@ -133,7 +133,9 @@ static int get_chars(u32 vtermno, char *buf, int count) + * virtqueue, so we let the drivers do some boutique early-output thing. */ + int __init virtio_cons_early_init(int (*put_chars)(u32, const char *, int)) + { +- virtio_cons.put_chars = put_chars; ++ pax_open_kernel(); ++ *(void **)&virtio_cons.put_chars = put_chars; ++ pax_close_kernel(); + return hvc_instantiate(0, 0, &virtio_cons); + } + +@@ -213,11 +215,13 @@ static int __devinit virtcons_probe(struct virtio_device *dev) + out_vq = vqs[1]; + + /* Start using the new console output. */ +- virtio_cons.get_chars = get_chars; +- virtio_cons.put_chars = put_chars; +- virtio_cons.notifier_add = notifier_add_vio; +- virtio_cons.notifier_del = notifier_del_vio; +- virtio_cons.notifier_hangup = notifier_del_vio; ++ pax_open_kernel(); ++ *(void **)&virtio_cons.get_chars = get_chars; ++ *(void **)&virtio_cons.put_chars = put_chars; ++ *(void **)&virtio_cons.notifier_add = notifier_add_vio; ++ *(void **)&virtio_cons.notifier_del = notifier_del_vio; ++ *(void **)&virtio_cons.notifier_hangup = notifier_del_vio; ++ pax_close_kernel(); + + /* The first argument of hvc_alloc() is the virtual console number, so + * we use zero. The second argument is the parameter for the +diff --git a/drivers/char/vt.c b/drivers/char/vt.c +index 0c80c68..53d59c1 100644 +--- a/drivers/char/vt.c ++++ b/drivers/char/vt.c +@@ -243,7 +243,7 @@ EXPORT_SYMBOL_GPL(unregister_vt_notifier); + + static void notify_write(struct vc_data *vc, unsigned int unicode) + { +- struct vt_notifier_param param = { .vc = vc, unicode = unicode }; ++ struct vt_notifier_param param = { .vc = vc, .c = unicode }; + atomic_notifier_call_chain(&vt_notifier_list, VT_WRITE, ¶m); + } + +diff --git a/drivers/char/vt_ioctl.c b/drivers/char/vt_ioctl.c +index 6351a26..999af95 100644 +--- a/drivers/char/vt_ioctl.c ++++ b/drivers/char/vt_ioctl.c +@@ -210,9 +210,6 @@ do_kdsk_ioctl(int cmd, struct kbentry __user *user_kbe, int perm, struct kbd_str + if (copy_from_user(&tmp, user_kbe, sizeof(struct kbentry))) + return -EFAULT; + +- if (!capable(CAP_SYS_TTY_CONFIG)) +- perm = 0; +- + switch (cmd) { + case KDGKBENT: + key_map = key_maps[s]; +@@ -224,8 +221,12 @@ do_kdsk_ioctl(int cmd, struct kbentry __user *user_kbe, int perm, struct kbd_str + val = (i ? K_HOLE : K_NOSUCHMAP); + return put_user(val, &user_kbe->kb_value); + case KDSKBENT: ++ if (!capable(CAP_SYS_TTY_CONFIG)) ++ perm = 0; ++ + if (!perm) + return -EPERM; ++ + if (!i && v == K_NOSUCHMAP) { + /* deallocate map */ + key_map = key_maps[s]; +@@ -325,9 +326,6 @@ do_kdgkb_ioctl(int cmd, struct kbsentry __user *user_kdgkb, int perm) + int i, j, k; + int ret; + +- if (!capable(CAP_SYS_TTY_CONFIG)) +- perm = 0; +- + kbs = kmalloc(sizeof(*kbs), GFP_KERNEL); + if (!kbs) { + ret = -ENOMEM; +@@ -361,6 +359,9 @@ do_kdgkb_ioctl(int cmd, struct kbsentry __user *user_kdgkb, int perm) + kfree(kbs); + return ((p && *p) ? -EOVERFLOW : 0); + case KDSKBSENT: ++ if (!capable(CAP_SYS_TTY_CONFIG)) ++ perm = 0; ++ + if (!perm) { + ret = -EPERM; + goto reterr; +diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c +index c7ae026..1769c1d 100644 +--- a/drivers/cpufreq/cpufreq.c ++++ b/drivers/cpufreq/cpufreq.c +@@ -750,7 +750,7 @@ static void cpufreq_sysfs_release(struct kobject *kobj) + complete(&policy->kobj_unregister); + } + +-static struct sysfs_ops sysfs_ops = { ++static const struct sysfs_ops sysfs_ops = { + .show = show, + .store = store, + }; +diff --git a/drivers/cpuidle/sysfs.c b/drivers/cpuidle/sysfs.c +index 97b0038..2056670 100644 +--- a/drivers/cpuidle/sysfs.c ++++ b/drivers/cpuidle/sysfs.c +@@ -191,7 +191,7 @@ static ssize_t cpuidle_store(struct kobject * kobj, struct attribute * attr, + return ret; + } + +-static struct sysfs_ops cpuidle_sysfs_ops = { ++static const struct sysfs_ops cpuidle_sysfs_ops = { + .show = cpuidle_show, + .store = cpuidle_store, + }; +@@ -277,7 +277,7 @@ static ssize_t cpuidle_state_show(struct kobject * kobj, + return ret; + } + +-static struct sysfs_ops cpuidle_state_sysfs_ops = { ++static const struct sysfs_ops cpuidle_state_sysfs_ops = { + .show = cpuidle_state_show, + }; + +@@ -294,7 +294,7 @@ static struct kobj_type ktype_state_cpuidle = { + .release = cpuidle_state_sysfs_release, + }; + +-static void inline cpuidle_free_state_kobj(struct cpuidle_device *device, int i) ++static inline void cpuidle_free_state_kobj(struct cpuidle_device *device, int i) + { + kobject_put(&device->kobjs[i]->kobj); + wait_for_completion(&device->kobjs[i]->kobj_unregister); +diff --git a/drivers/crypto/hifn_795x.c b/drivers/crypto/hifn_795x.c +index 5f753fc..0377ae9 100644 +--- a/drivers/crypto/hifn_795x.c ++++ b/drivers/crypto/hifn_795x.c +@@ -1655,6 +1655,8 @@ static int hifn_test(struct hifn_device *dev, int encdec, u8 snum) + 0xCA, 0x34, 0x2B, 0x2E}; + struct scatterlist sg; + ++ pax_track_stack(); ++ + memset(src, 0, sizeof(src)); + memset(ctx.key, 0, sizeof(ctx.key)); + +diff --git a/drivers/crypto/padlock-aes.c b/drivers/crypto/padlock-aes.c +index 71e6482..de8d96c 100644 +--- a/drivers/crypto/padlock-aes.c ++++ b/drivers/crypto/padlock-aes.c +@@ -108,6 +108,8 @@ static int aes_set_key(struct crypto_tfm *tfm, const u8 *in_key, + struct crypto_aes_ctx gen_aes; + int cpu; + ++ pax_track_stack(); ++ + if (key_len % 8) { + *flags |= CRYPTO_TFM_RES_BAD_KEY_LEN; + return -EINVAL; +diff --git a/drivers/dma/ioat/dma.c b/drivers/dma/ioat/dma.c +index dcc4ab7..cc834bb 100644 +--- a/drivers/dma/ioat/dma.c ++++ b/drivers/dma/ioat/dma.c +@@ -1146,7 +1146,7 @@ ioat_attr_show(struct kobject *kobj, struct attribute *attr, char *page) + return entry->show(&chan->common, page); + } + +-struct sysfs_ops ioat_sysfs_ops = { ++const struct sysfs_ops ioat_sysfs_ops = { + .show = ioat_attr_show, + }; + +diff --git a/drivers/dma/ioat/dma.h b/drivers/dma/ioat/dma.h +index bbc3e78..f2db62c 100644 +--- a/drivers/dma/ioat/dma.h ++++ b/drivers/dma/ioat/dma.h +@@ -347,7 +347,7 @@ bool ioat_cleanup_preamble(struct ioat_chan_common *chan, + unsigned long *phys_complete); + void ioat_kobject_add(struct ioatdma_device *device, struct kobj_type *type); + void ioat_kobject_del(struct ioatdma_device *device); +-extern struct sysfs_ops ioat_sysfs_ops; ++extern const struct sysfs_ops ioat_sysfs_ops; + extern struct ioat_sysfs_entry ioat_version_attr; + extern struct ioat_sysfs_entry ioat_cap_attr; + #endif /* IOATDMA_H */ +diff --git a/drivers/dma/ioat/dma_v3.c b/drivers/dma/ioat/dma_v3.c +index 9908c9e..3ceb0e5 100644 +--- a/drivers/dma/ioat/dma_v3.c ++++ b/drivers/dma/ioat/dma_v3.c +@@ -71,10 +71,10 @@ + /* provide a lookup table for setting the source address in the base or + * extended descriptor of an xor or pq descriptor + */ +-static const u8 xor_idx_to_desc __read_mostly = 0xd0; +-static const u8 xor_idx_to_field[] __read_mostly = { 1, 4, 5, 6, 7, 0, 1, 2 }; +-static const u8 pq_idx_to_desc __read_mostly = 0xf8; +-static const u8 pq_idx_to_field[] __read_mostly = { 1, 4, 5, 0, 1, 2, 4, 5 }; ++static const u8 xor_idx_to_desc = 0xd0; ++static const u8 xor_idx_to_field[] = { 1, 4, 5, 6, 7, 0, 1, 2 }; ++static const u8 pq_idx_to_desc = 0xf8; ++static const u8 pq_idx_to_field[] = { 1, 4, 5, 0, 1, 2, 4, 5 }; + + static dma_addr_t xor_get_src(struct ioat_raw_descriptor *descs[2], int idx) + { +diff --git a/drivers/edac/amd64_edac.c b/drivers/edac/amd64_edac.c +index 85c464a..afd1e73 100644 +--- a/drivers/edac/amd64_edac.c ++++ b/drivers/edac/amd64_edac.c +@@ -3099,7 +3099,7 @@ static void __devexit amd64_remove_one_instance(struct pci_dev *pdev) + * PCI core identifies what devices are on a system during boot, and then + * inquiry this table to see if this driver is for a given device found. + */ +-static const struct pci_device_id amd64_pci_table[] __devinitdata = { ++static const struct pci_device_id amd64_pci_table[] __devinitconst = { + { + .vendor = PCI_VENDOR_ID_AMD, + .device = PCI_DEVICE_ID_AMD_K8_NB_MEMCTL, +diff --git a/drivers/edac/amd76x_edac.c b/drivers/edac/amd76x_edac.c +index 2b95f1a..4f52793 100644 +--- a/drivers/edac/amd76x_edac.c ++++ b/drivers/edac/amd76x_edac.c +@@ -322,7 +322,7 @@ static void __devexit amd76x_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id amd76x_pci_tbl[] __devinitdata = { ++static const struct pci_device_id amd76x_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(AMD, FE_GATE_700C), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + AMD762}, +diff --git a/drivers/edac/e752x_edac.c b/drivers/edac/e752x_edac.c +index d205d49..74c9672 100644 +--- a/drivers/edac/e752x_edac.c ++++ b/drivers/edac/e752x_edac.c +@@ -1282,7 +1282,7 @@ static void __devexit e752x_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id e752x_pci_tbl[] __devinitdata = { ++static const struct pci_device_id e752x_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + E7520}, +diff --git a/drivers/edac/e7xxx_edac.c b/drivers/edac/e7xxx_edac.c +index c7d11cc..c59c1ca 100644 +--- a/drivers/edac/e7xxx_edac.c ++++ b/drivers/edac/e7xxx_edac.c +@@ -526,7 +526,7 @@ static void __devexit e7xxx_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id e7xxx_pci_tbl[] __devinitdata = { ++static const struct pci_device_id e7xxx_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 7205_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + E7205}, +diff --git a/drivers/edac/edac_device_sysfs.c b/drivers/edac/edac_device_sysfs.c +index 5376457..5fdedbc 100644 +--- a/drivers/edac/edac_device_sysfs.c ++++ b/drivers/edac/edac_device_sysfs.c +@@ -137,7 +137,7 @@ static ssize_t edac_dev_ctl_info_store(struct kobject *kobj, + } + + /* edac_dev file operations for an 'ctl_info' */ +-static struct sysfs_ops device_ctl_info_ops = { ++static const struct sysfs_ops device_ctl_info_ops = { + .show = edac_dev_ctl_info_show, + .store = edac_dev_ctl_info_store + }; +@@ -373,7 +373,7 @@ static ssize_t edac_dev_instance_store(struct kobject *kobj, + } + + /* edac_dev file operations for an 'instance' */ +-static struct sysfs_ops device_instance_ops = { ++static const struct sysfs_ops device_instance_ops = { + .show = edac_dev_instance_show, + .store = edac_dev_instance_store + }; +@@ -476,7 +476,7 @@ static ssize_t edac_dev_block_store(struct kobject *kobj, + } + + /* edac_dev file operations for a 'block' */ +-static struct sysfs_ops device_block_ops = { ++static const struct sysfs_ops device_block_ops = { + .show = edac_dev_block_show, + .store = edac_dev_block_store + }; +diff --git a/drivers/edac/edac_mc_sysfs.c b/drivers/edac/edac_mc_sysfs.c +index e1d4ce0..88840e9 100644 +--- a/drivers/edac/edac_mc_sysfs.c ++++ b/drivers/edac/edac_mc_sysfs.c +@@ -245,7 +245,7 @@ static ssize_t csrowdev_store(struct kobject *kobj, struct attribute *attr, + return -EIO; + } + +-static struct sysfs_ops csrowfs_ops = { ++static const struct sysfs_ops csrowfs_ops = { + .show = csrowdev_show, + .store = csrowdev_store + }; +@@ -575,7 +575,7 @@ static ssize_t mcidev_store(struct kobject *kobj, struct attribute *attr, + } + + /* Intermediate show/store table */ +-static struct sysfs_ops mci_ops = { ++static const struct sysfs_ops mci_ops = { + .show = mcidev_show, + .store = mcidev_store + }; +diff --git a/drivers/edac/edac_pci_sysfs.c b/drivers/edac/edac_pci_sysfs.c +index 422728c..d8d9c88 100644 +--- a/drivers/edac/edac_pci_sysfs.c ++++ b/drivers/edac/edac_pci_sysfs.c +@@ -25,8 +25,8 @@ static int edac_pci_log_pe = 1; /* log PCI parity errors */ + static int edac_pci_log_npe = 1; /* log PCI non-parity error errors */ + static int edac_pci_poll_msec = 1000; /* one second workq period */ + +-static atomic_t pci_parity_count = ATOMIC_INIT(0); +-static atomic_t pci_nonparity_count = ATOMIC_INIT(0); ++static atomic_unchecked_t pci_parity_count = ATOMIC_INIT(0); ++static atomic_unchecked_t pci_nonparity_count = ATOMIC_INIT(0); + + static struct kobject *edac_pci_top_main_kobj; + static atomic_t edac_pci_sysfs_refcount = ATOMIC_INIT(0); +@@ -121,7 +121,7 @@ static ssize_t edac_pci_instance_store(struct kobject *kobj, + } + + /* fs_ops table */ +-static struct sysfs_ops pci_instance_ops = { ++static const struct sysfs_ops pci_instance_ops = { + .show = edac_pci_instance_show, + .store = edac_pci_instance_store + }; +@@ -261,7 +261,7 @@ static ssize_t edac_pci_dev_store(struct kobject *kobj, + return -EIO; + } + +-static struct sysfs_ops edac_pci_sysfs_ops = { ++static const struct sysfs_ops edac_pci_sysfs_ops = { + .show = edac_pci_dev_show, + .store = edac_pci_dev_store + }; +@@ -579,7 +579,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + edac_printk(KERN_CRIT, EDAC_PCI, + "Signaled System Error on %s\n", + pci_name(dev)); +- atomic_inc(&pci_nonparity_count); ++ atomic_inc_unchecked(&pci_nonparity_count); + } + + if (status & (PCI_STATUS_PARITY)) { +@@ -587,7 +587,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + "Master Data Parity Error on %s\n", + pci_name(dev)); + +- atomic_inc(&pci_parity_count); ++ atomic_inc_unchecked(&pci_parity_count); + } + + if (status & (PCI_STATUS_DETECTED_PARITY)) { +@@ -595,7 +595,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + "Detected Parity Error on %s\n", + pci_name(dev)); + +- atomic_inc(&pci_parity_count); ++ atomic_inc_unchecked(&pci_parity_count); + } + } + +@@ -616,7 +616,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + edac_printk(KERN_CRIT, EDAC_PCI, "Bridge " + "Signaled System Error on %s\n", + pci_name(dev)); +- atomic_inc(&pci_nonparity_count); ++ atomic_inc_unchecked(&pci_nonparity_count); + } + + if (status & (PCI_STATUS_PARITY)) { +@@ -624,7 +624,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + "Master Data Parity Error on " + "%s\n", pci_name(dev)); + +- atomic_inc(&pci_parity_count); ++ atomic_inc_unchecked(&pci_parity_count); + } + + if (status & (PCI_STATUS_DETECTED_PARITY)) { +@@ -632,7 +632,7 @@ static void edac_pci_dev_parity_test(struct pci_dev *dev) + "Detected Parity Error on %s\n", + pci_name(dev)); + +- atomic_inc(&pci_parity_count); ++ atomic_inc_unchecked(&pci_parity_count); + } + } + } +@@ -674,7 +674,7 @@ void edac_pci_do_parity_check(void) + if (!check_pci_errors) + return; + +- before_count = atomic_read(&pci_parity_count); ++ before_count = atomic_read_unchecked(&pci_parity_count); + + /* scan all PCI devices looking for a Parity Error on devices and + * bridges. +@@ -686,7 +686,7 @@ void edac_pci_do_parity_check(void) + /* Only if operator has selected panic on PCI Error */ + if (edac_pci_get_panic_on_pe()) { + /* If the count is different 'after' from 'before' */ +- if (before_count != atomic_read(&pci_parity_count)) ++ if (before_count != atomic_read_unchecked(&pci_parity_count)) + panic("EDAC: PCI Parity Error"); + } + } +diff --git a/drivers/edac/i3000_edac.c b/drivers/edac/i3000_edac.c +index 6c9a0f2..9c1cf7e 100644 +--- a/drivers/edac/i3000_edac.c ++++ b/drivers/edac/i3000_edac.c +@@ -471,7 +471,7 @@ static void __devexit i3000_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i3000_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i3000_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 3000_HB), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + I3000}, +diff --git a/drivers/edac/i3200_edac.c b/drivers/edac/i3200_edac.c +index fde4db9..fe108f9 100644 +--- a/drivers/edac/i3200_edac.c ++++ b/drivers/edac/i3200_edac.c +@@ -444,7 +444,7 @@ static void __devexit i3200_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i3200_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i3200_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 3200_HB), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + I3200}, +diff --git a/drivers/edac/i5000_edac.c b/drivers/edac/i5000_edac.c +index adc10a2..57d4ccf 100644 +--- a/drivers/edac/i5000_edac.c ++++ b/drivers/edac/i5000_edac.c +@@ -1516,7 +1516,7 @@ static void __devexit i5000_remove_one(struct pci_dev *pdev) + * + * The "E500P" device is the first device supported. + */ +-static const struct pci_device_id i5000_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i5000_pci_tbl[] __devinitconst = { + {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_I5000_DEV16), + .driver_data = I5000P}, + +diff --git a/drivers/edac/i5100_edac.c b/drivers/edac/i5100_edac.c +index 22db05a..b2b5503 100644 +--- a/drivers/edac/i5100_edac.c ++++ b/drivers/edac/i5100_edac.c +@@ -944,7 +944,7 @@ static void __devexit i5100_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i5100_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i5100_pci_tbl[] __devinitconst = { + /* Device 16, Function 0, Channel 0 Memory Map, Error Flag/Mask, ... */ + { PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_5100_16) }, + { 0, } +diff --git a/drivers/edac/i5400_edac.c b/drivers/edac/i5400_edac.c +index f99d106..f050710 100644 +--- a/drivers/edac/i5400_edac.c ++++ b/drivers/edac/i5400_edac.c +@@ -1383,7 +1383,7 @@ static void __devexit i5400_remove_one(struct pci_dev *pdev) + * + * The "E500P" device is the first device supported. + */ +-static const struct pci_device_id i5400_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i5400_pci_tbl[] __devinitconst = { + {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_5400_ERR)}, + {0,} /* 0 terminated list. */ + }; +diff --git a/drivers/edac/i82443bxgx_edac.c b/drivers/edac/i82443bxgx_edac.c +index 577760a..9ce16ce 100644 +--- a/drivers/edac/i82443bxgx_edac.c ++++ b/drivers/edac/i82443bxgx_edac.c +@@ -381,7 +381,7 @@ static void __devexit i82443bxgx_edacmc_remove_one(struct pci_dev *pdev) + + EXPORT_SYMBOL_GPL(i82443bxgx_edacmc_remove_one); + +-static const struct pci_device_id i82443bxgx_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i82443bxgx_pci_tbl[] __devinitconst = { + {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82443BX_0)}, + {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82443BX_2)}, + {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82443GX_0)}, +diff --git a/drivers/edac/i82860_edac.c b/drivers/edac/i82860_edac.c +index c0088ba..64a7b98 100644 +--- a/drivers/edac/i82860_edac.c ++++ b/drivers/edac/i82860_edac.c +@@ -271,7 +271,7 @@ static void __devexit i82860_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i82860_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i82860_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 82860_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + I82860}, +diff --git a/drivers/edac/i82875p_edac.c b/drivers/edac/i82875p_edac.c +index b2d83b9..a34357b 100644 +--- a/drivers/edac/i82875p_edac.c ++++ b/drivers/edac/i82875p_edac.c +@@ -512,7 +512,7 @@ static void __devexit i82875p_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i82875p_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i82875p_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 82875_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + I82875P}, +diff --git a/drivers/edac/i82975x_edac.c b/drivers/edac/i82975x_edac.c +index 2eed3ea..87bbbd1 100644 +--- a/drivers/edac/i82975x_edac.c ++++ b/drivers/edac/i82975x_edac.c +@@ -586,7 +586,7 @@ static void __devexit i82975x_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id i82975x_pci_tbl[] __devinitdata = { ++static const struct pci_device_id i82975x_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, 82975_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + I82975X +diff --git a/drivers/edac/r82600_edac.c b/drivers/edac/r82600_edac.c +index 9900675..78ac2b6 100644 +--- a/drivers/edac/r82600_edac.c ++++ b/drivers/edac/r82600_edac.c +@@ -374,7 +374,7 @@ static void __devexit r82600_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id r82600_pci_tbl[] __devinitdata = { ++static const struct pci_device_id r82600_pci_tbl[] __devinitconst = { + { + PCI_DEVICE(PCI_VENDOR_ID_RADISYS, R82600_BRIDGE_ID) + }, +diff --git a/drivers/edac/x38_edac.c b/drivers/edac/x38_edac.c +index d4ec605..4cfec4e 100644 +--- a/drivers/edac/x38_edac.c ++++ b/drivers/edac/x38_edac.c +@@ -441,7 +441,7 @@ static void __devexit x38_remove_one(struct pci_dev *pdev) + edac_mc_free(mci); + } + +-static const struct pci_device_id x38_pci_tbl[] __devinitdata = { ++static const struct pci_device_id x38_pci_tbl[] __devinitconst = { + { + PCI_VEND_DEV(INTEL, X38_HB), PCI_ANY_ID, PCI_ANY_ID, 0, 0, + X38}, +diff --git a/drivers/firewire/core-card.c b/drivers/firewire/core-card.c +index 3fc2ceb..daf098f 100644 +--- a/drivers/firewire/core-card.c ++++ b/drivers/firewire/core-card.c +@@ -558,7 +558,7 @@ void fw_card_release(struct kref *kref) + + void fw_core_remove_card(struct fw_card *card) + { +- struct fw_card_driver dummy_driver = dummy_driver_template; ++ fw_card_driver_no_const dummy_driver = dummy_driver_template; + + card->driver->update_phy_reg(card, 4, + PHY_LINK_ACTIVE | PHY_CONTENDER, 0); +diff --git a/drivers/firewire/core-cdev.c b/drivers/firewire/core-cdev.c +index 4560d8f..36db24a 100644 +--- a/drivers/firewire/core-cdev.c ++++ b/drivers/firewire/core-cdev.c +@@ -1141,8 +1141,7 @@ static int init_iso_resource(struct client *client, + int ret; + + if ((request->channels == 0 && request->bandwidth == 0) || +- request->bandwidth > BANDWIDTH_AVAILABLE_INITIAL || +- request->bandwidth < 0) ++ request->bandwidth > BANDWIDTH_AVAILABLE_INITIAL) + return -EINVAL; + + r = kmalloc(sizeof(*r), GFP_KERNEL); +diff --git a/drivers/firewire/core-transaction.c b/drivers/firewire/core-transaction.c +index da628c7..cf54a2c 100644 +--- a/drivers/firewire/core-transaction.c ++++ b/drivers/firewire/core-transaction.c +@@ -36,6 +36,7 @@ + #include + #include + #include ++#include + + #include + +@@ -344,6 +345,8 @@ int fw_run_transaction(struct fw_card *card, int tcode, int destination_id, + struct transaction_callback_data d; + struct fw_transaction t; + ++ pax_track_stack(); ++ + init_completion(&d.done); + d.payload = payload; + fw_send_request(card, &t, tcode, destination_id, generation, speed, +diff --git a/drivers/firewire/core.h b/drivers/firewire/core.h +index 7ff6e75..a2965d9 100644 +--- a/drivers/firewire/core.h ++++ b/drivers/firewire/core.h +@@ -86,6 +86,7 @@ struct fw_card_driver { + + int (*stop_iso)(struct fw_iso_context *ctx); + }; ++typedef struct fw_card_driver __no_const fw_card_driver_no_const; + + void fw_card_initialize(struct fw_card *card, + const struct fw_card_driver *driver, struct device *device); +diff --git a/drivers/firmware/dmi_scan.c b/drivers/firmware/dmi_scan.c +index 3a2ccb0..82fd7c4 100644 +--- a/drivers/firmware/dmi_scan.c ++++ b/drivers/firmware/dmi_scan.c +@@ -391,11 +391,6 @@ void __init dmi_scan_machine(void) + } + } + else { +- /* +- * no iounmap() for that ioremap(); it would be a no-op, but +- * it's so early in setup that sucker gets confused into doing +- * what it shouldn't if we actually call it. +- */ + p = dmi_ioremap(0xF0000, 0x10000); + if (p == NULL) + goto error; +@@ -667,7 +662,7 @@ int dmi_walk(void (*decode)(const struct dmi_header *, void *), + if (buf == NULL) + return -1; + +- dmi_table(buf, dmi_len, dmi_num, decode, private_data); ++ dmi_table((char __force_kernel *)buf, dmi_len, dmi_num, decode, private_data); + + iounmap(buf); + return 0; +diff --git a/drivers/firmware/edd.c b/drivers/firmware/edd.c +index 9e4f59d..110e24e 100644 +--- a/drivers/firmware/edd.c ++++ b/drivers/firmware/edd.c +@@ -122,7 +122,7 @@ edd_attr_show(struct kobject * kobj, struct attribute *attr, char *buf) + return ret; + } + +-static struct sysfs_ops edd_attr_ops = { ++static const struct sysfs_ops edd_attr_ops = { + .show = edd_attr_show, + }; + +diff --git a/drivers/firmware/efivars.c b/drivers/firmware/efivars.c +index f4f709d..082f06e 100644 +--- a/drivers/firmware/efivars.c ++++ b/drivers/firmware/efivars.c +@@ -362,7 +362,7 @@ static ssize_t efivar_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops efivar_attr_ops = { ++static const struct sysfs_ops efivar_attr_ops = { + .show = efivar_attr_show, + .store = efivar_attr_store, + }; +diff --git a/drivers/firmware/iscsi_ibft.c b/drivers/firmware/iscsi_ibft.c +index 051d1eb..0a5d4e7 100644 +--- a/drivers/firmware/iscsi_ibft.c ++++ b/drivers/firmware/iscsi_ibft.c +@@ -525,7 +525,7 @@ static ssize_t ibft_show_attribute(struct kobject *kobj, + return ret; + } + +-static struct sysfs_ops ibft_attr_ops = { ++static const struct sysfs_ops ibft_attr_ops = { + .show = ibft_show_attribute, + }; + +diff --git a/drivers/firmware/memmap.c b/drivers/firmware/memmap.c +index 56f9234..8c58c7b 100644 +--- a/drivers/firmware/memmap.c ++++ b/drivers/firmware/memmap.c +@@ -74,7 +74,7 @@ static struct attribute *def_attrs[] = { + NULL + }; + +-static struct sysfs_ops memmap_attr_ops = { ++static const struct sysfs_ops memmap_attr_ops = { + .show = memmap_attr_show, + }; + +diff --git a/drivers/gpio/vr41xx_giu.c b/drivers/gpio/vr41xx_giu.c +index b16c9a8..2af7d3f 100644 +--- a/drivers/gpio/vr41xx_giu.c ++++ b/drivers/gpio/vr41xx_giu.c +@@ -204,7 +204,7 @@ static int giu_get_irq(unsigned int irq) + printk(KERN_ERR "spurious GIU interrupt: %04x(%04x),%04x(%04x)\n", + maskl, pendl, maskh, pendh); + +- atomic_inc(&irq_err_count); ++ atomic_inc_unchecked(&irq_err_count); + + return -EINVAL; + } +diff --git a/drivers/gpu/drm/drm_crtc.c b/drivers/gpu/drm/drm_crtc.c +index bea6efc..3dc0f42 100644 +--- a/drivers/gpu/drm/drm_crtc.c ++++ b/drivers/gpu/drm/drm_crtc.c +@@ -1323,7 +1323,7 @@ int drm_mode_getconnector(struct drm_device *dev, void *data, + */ + if ((out_resp->count_modes >= mode_count) && mode_count) { + copied = 0; +- mode_ptr = (struct drm_mode_modeinfo *)(unsigned long)out_resp->modes_ptr; ++ mode_ptr = (struct drm_mode_modeinfo __user *)(unsigned long)out_resp->modes_ptr; + list_for_each_entry(mode, &connector->modes, head) { + drm_crtc_convert_to_umode(&u_mode, mode); + if (copy_to_user(mode_ptr + copied, +@@ -1338,8 +1338,8 @@ int drm_mode_getconnector(struct drm_device *dev, void *data, + + if ((out_resp->count_props >= props_count) && props_count) { + copied = 0; +- prop_ptr = (uint32_t *)(unsigned long)(out_resp->props_ptr); +- prop_values = (uint64_t *)(unsigned long)(out_resp->prop_values_ptr); ++ prop_ptr = (uint32_t __user *)(unsigned long)(out_resp->props_ptr); ++ prop_values = (uint64_t __user *)(unsigned long)(out_resp->prop_values_ptr); + for (i = 0; i < DRM_CONNECTOR_MAX_PROPERTY; i++) { + if (connector->property_ids[i] != 0) { + if (put_user(connector->property_ids[i], +@@ -1361,7 +1361,7 @@ int drm_mode_getconnector(struct drm_device *dev, void *data, + + if ((out_resp->count_encoders >= encoders_count) && encoders_count) { + copied = 0; +- encoder_ptr = (uint32_t *)(unsigned long)(out_resp->encoders_ptr); ++ encoder_ptr = (uint32_t __user *)(unsigned long)(out_resp->encoders_ptr); + for (i = 0; i < DRM_CONNECTOR_MAX_ENCODER; i++) { + if (connector->encoder_ids[i] != 0) { + if (put_user(connector->encoder_ids[i], +@@ -1513,7 +1513,7 @@ int drm_mode_setcrtc(struct drm_device *dev, void *data, + } + + for (i = 0; i < crtc_req->count_connectors; i++) { +- set_connectors_ptr = (uint32_t *)(unsigned long)crtc_req->set_connectors_ptr; ++ set_connectors_ptr = (uint32_t __user *)(unsigned long)crtc_req->set_connectors_ptr; + if (get_user(out_id, &set_connectors_ptr[i])) { + ret = -EFAULT; + goto out; +@@ -2118,7 +2118,7 @@ int drm_mode_getproperty_ioctl(struct drm_device *dev, + out_resp->flags = property->flags; + + if ((out_resp->count_values >= value_count) && value_count) { +- values_ptr = (uint64_t *)(unsigned long)out_resp->values_ptr; ++ values_ptr = (uint64_t __user *)(unsigned long)out_resp->values_ptr; + for (i = 0; i < value_count; i++) { + if (copy_to_user(values_ptr + i, &property->values[i], sizeof(uint64_t))) { + ret = -EFAULT; +@@ -2131,7 +2131,7 @@ int drm_mode_getproperty_ioctl(struct drm_device *dev, + if (property->flags & DRM_MODE_PROP_ENUM) { + if ((out_resp->count_enum_blobs >= enum_count) && enum_count) { + copied = 0; +- enum_ptr = (struct drm_mode_property_enum *)(unsigned long)out_resp->enum_blob_ptr; ++ enum_ptr = (struct drm_mode_property_enum __user *)(unsigned long)out_resp->enum_blob_ptr; + list_for_each_entry(prop_enum, &property->enum_blob_list, head) { + + if (copy_to_user(&enum_ptr[copied].value, &prop_enum->value, sizeof(uint64_t))) { +@@ -2154,7 +2154,7 @@ int drm_mode_getproperty_ioctl(struct drm_device *dev, + if ((out_resp->count_enum_blobs >= blob_count) && blob_count) { + copied = 0; + blob_id_ptr = (uint32_t *)(unsigned long)out_resp->enum_blob_ptr; +- blob_length_ptr = (uint32_t *)(unsigned long)out_resp->values_ptr; ++ blob_length_ptr = (uint32_t __user *)(unsigned long)out_resp->values_ptr; + + list_for_each_entry(prop_blob, &property->enum_blob_list, head) { + if (put_user(prop_blob->base.id, blob_id_ptr + copied)) { +@@ -2226,7 +2226,7 @@ int drm_mode_getblob_ioctl(struct drm_device *dev, + blob = obj_to_blob(obj); + + if (out_resp->length == blob->length) { +- blob_ptr = (void *)(unsigned long)out_resp->data; ++ blob_ptr = (void __user *)(unsigned long)out_resp->data; + if (copy_to_user(blob_ptr, blob->data, blob->length)){ + ret = -EFAULT; + goto done; +diff --git a/drivers/gpu/drm/drm_crtc_helper.c b/drivers/gpu/drm/drm_crtc_helper.c +index 1b8745d..92fdbf6 100644 +--- a/drivers/gpu/drm/drm_crtc_helper.c ++++ b/drivers/gpu/drm/drm_crtc_helper.c +@@ -573,7 +573,7 @@ static bool drm_encoder_crtc_ok(struct drm_encoder *encoder, + struct drm_crtc *tmp; + int crtc_mask = 1; + +- WARN(!crtc, "checking null crtc?"); ++ BUG_ON(!crtc); + + dev = crtc->dev; + +@@ -642,6 +642,8 @@ bool drm_crtc_helper_set_mode(struct drm_crtc *crtc, + + adjusted_mode = drm_mode_duplicate(dev, mode); + ++ pax_track_stack(); ++ + crtc->enabled = drm_helper_crtc_in_use(crtc); + + if (!crtc->enabled) +diff --git a/drivers/gpu/drm/drm_drv.c b/drivers/gpu/drm/drm_drv.c +index 0e27d98..dec8768 100644 +--- a/drivers/gpu/drm/drm_drv.c ++++ b/drivers/gpu/drm/drm_drv.c +@@ -417,7 +417,7 @@ int drm_ioctl(struct inode *inode, struct file *filp, + char *kdata = NULL; + + atomic_inc(&dev->ioctl_count); +- atomic_inc(&dev->counts[_DRM_STAT_IOCTLS]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_IOCTLS]); + ++file_priv->ioctl_count; + + DRM_DEBUG("pid=%d, cmd=0x%02x, nr=0x%02x, dev 0x%lx, auth=%d\n", +diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c +index 519161e..98c840c 100644 +--- a/drivers/gpu/drm/drm_fops.c ++++ b/drivers/gpu/drm/drm_fops.c +@@ -66,7 +66,7 @@ static int drm_setup(struct drm_device * dev) + } + + for (i = 0; i < ARRAY_SIZE(dev->counts); i++) +- atomic_set(&dev->counts[i], 0); ++ atomic_set_unchecked(&dev->counts[i], 0); + + dev->sigdata.lock = NULL; + +@@ -130,9 +130,9 @@ int drm_open(struct inode *inode, struct file *filp) + + retcode = drm_open_helper(inode, filp, dev); + if (!retcode) { +- atomic_inc(&dev->counts[_DRM_STAT_OPENS]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_OPENS]); + spin_lock(&dev->count_lock); +- if (!dev->open_count++) { ++ if (local_inc_return(&dev->open_count) == 1) { + spin_unlock(&dev->count_lock); + retcode = drm_setup(dev); + goto out; +@@ -435,7 +435,7 @@ int drm_release(struct inode *inode, struct file *filp) + + lock_kernel(); + +- DRM_DEBUG("open_count = %d\n", dev->open_count); ++ DRM_DEBUG("open_count = %d\n", local_read(&dev->open_count)); + + if (dev->driver->preclose) + dev->driver->preclose(dev, file_priv); +@@ -447,7 +447,7 @@ int drm_release(struct inode *inode, struct file *filp) + DRM_DEBUG("pid = %d, device = 0x%lx, open_count = %d\n", + task_pid_nr(current), + (long)old_encode_dev(file_priv->minor->device), +- dev->open_count); ++ local_read(&dev->open_count)); + + /* Release any auth tokens that might point to this file_priv, + (do that under the drm_global_mutex) */ +@@ -529,9 +529,9 @@ int drm_release(struct inode *inode, struct file *filp) + * End inline drm_release + */ + +- atomic_inc(&dev->counts[_DRM_STAT_CLOSES]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_CLOSES]); + spin_lock(&dev->count_lock); +- if (!--dev->open_count) { ++ if (local_dec_and_test(&dev->open_count)) { + if (atomic_read(&dev->ioctl_count)) { + DRM_ERROR("Device busy: %d\n", + atomic_read(&dev->ioctl_count)); +diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c +index 8bf3770..79422805 100644 +--- a/drivers/gpu/drm/drm_gem.c ++++ b/drivers/gpu/drm/drm_gem.c +@@ -83,11 +83,11 @@ drm_gem_init(struct drm_device *dev) + spin_lock_init(&dev->object_name_lock); + idr_init(&dev->object_name_idr); + atomic_set(&dev->object_count, 0); +- atomic_set(&dev->object_memory, 0); ++ atomic_set_unchecked(&dev->object_memory, 0); + atomic_set(&dev->pin_count, 0); +- atomic_set(&dev->pin_memory, 0); ++ atomic_set_unchecked(&dev->pin_memory, 0); + atomic_set(&dev->gtt_count, 0); +- atomic_set(&dev->gtt_memory, 0); ++ atomic_set_unchecked(&dev->gtt_memory, 0); + + mm = kzalloc(sizeof(struct drm_gem_mm), GFP_KERNEL); + if (!mm) { +@@ -150,7 +150,7 @@ drm_gem_object_alloc(struct drm_device *dev, size_t size) + goto fput; + } + atomic_inc(&dev->object_count); +- atomic_add(obj->size, &dev->object_memory); ++ atomic_add_unchecked(obj->size, &dev->object_memory); + return obj; + fput: + fput(obj->filp); +@@ -429,7 +429,7 @@ drm_gem_object_free(struct kref *kref) + + fput(obj->filp); + atomic_dec(&dev->object_count); +- atomic_sub(obj->size, &dev->object_memory); ++ atomic_sub_unchecked(obj->size, &dev->object_memory); + kfree(obj); + } + EXPORT_SYMBOL(drm_gem_object_free); +diff --git a/drivers/gpu/drm/drm_info.c b/drivers/gpu/drm/drm_info.c +index f0f6c6b..34af322 100644 +--- a/drivers/gpu/drm/drm_info.c ++++ b/drivers/gpu/drm/drm_info.c +@@ -75,10 +75,14 @@ int drm_vm_info(struct seq_file *m, void *data) + struct drm_local_map *map; + struct drm_map_list *r_list; + +- /* Hardcoded from _DRM_FRAME_BUFFER, +- _DRM_REGISTERS, _DRM_SHM, _DRM_AGP, and +- _DRM_SCATTER_GATHER and _DRM_CONSISTENT */ +- const char *types[] = { "FB", "REG", "SHM", "AGP", "SG", "PCI" }; ++ static const char * const types[] = { ++ [_DRM_FRAME_BUFFER] = "FB", ++ [_DRM_REGISTERS] = "REG", ++ [_DRM_SHM] = "SHM", ++ [_DRM_AGP] = "AGP", ++ [_DRM_SCATTER_GATHER] = "SG", ++ [_DRM_CONSISTENT] = "PCI", ++ [_DRM_GEM] = "GEM" }; + const char *type; + int i; + +@@ -89,7 +93,7 @@ int drm_vm_info(struct seq_file *m, void *data) + map = r_list->map; + if (!map) + continue; +- if (map->type < 0 || map->type > 5) ++ if (map->type >= ARRAY_SIZE(types)) + type = "??"; + else + type = types[map->type]; +@@ -265,10 +269,10 @@ int drm_gem_object_info(struct seq_file *m, void* data) + struct drm_device *dev = node->minor->dev; + + seq_printf(m, "%d objects\n", atomic_read(&dev->object_count)); +- seq_printf(m, "%d object bytes\n", atomic_read(&dev->object_memory)); ++ seq_printf(m, "%d object bytes\n", atomic_read_unchecked(&dev->object_memory)); + seq_printf(m, "%d pinned\n", atomic_read(&dev->pin_count)); +- seq_printf(m, "%d pin bytes\n", atomic_read(&dev->pin_memory)); +- seq_printf(m, "%d gtt bytes\n", atomic_read(&dev->gtt_memory)); ++ seq_printf(m, "%d pin bytes\n", atomic_read_unchecked(&dev->pin_memory)); ++ seq_printf(m, "%d gtt bytes\n", atomic_read_unchecked(&dev->gtt_memory)); + seq_printf(m, "%d gtt total\n", dev->gtt_total); + return 0; + } +@@ -288,7 +292,11 @@ int drm_vma_info(struct seq_file *m, void *data) + mutex_lock(&dev->struct_mutex); + seq_printf(m, "vma use count: %d, high_memory = %p, 0x%08llx\n", + atomic_read(&dev->vma_count), ++#ifdef CONFIG_GRKERNSEC_HIDESYM ++ NULL, 0); ++#else + high_memory, (u64)virt_to_phys(high_memory)); ++#endif + + list_for_each_entry(pt, &dev->vmalist, head) { + vma = pt->vma; +@@ -296,14 +304,23 @@ int drm_vma_info(struct seq_file *m, void *data) + continue; + seq_printf(m, + "\n%5d 0x%08lx-0x%08lx %c%c%c%c%c%c 0x%08lx000", +- pt->pid, vma->vm_start, vma->vm_end, ++ pt->pid, ++#ifdef CONFIG_GRKERNSEC_HIDESYM ++ 0, 0, ++#else ++ vma->vm_start, vma->vm_end, ++#endif + vma->vm_flags & VM_READ ? 'r' : '-', + vma->vm_flags & VM_WRITE ? 'w' : '-', + vma->vm_flags & VM_EXEC ? 'x' : '-', + vma->vm_flags & VM_MAYSHARE ? 's' : 'p', + vma->vm_flags & VM_LOCKED ? 'l' : '-', + vma->vm_flags & VM_IO ? 'i' : '-', ++#ifdef CONFIG_GRKERNSEC_HIDESYM ++ 0); ++#else + vma->vm_pgoff); ++#endif + + #if defined(__i386__) + pgprot = pgprot_val(vma->vm_page_prot); +diff --git a/drivers/gpu/drm/drm_ioc32.c b/drivers/gpu/drm/drm_ioc32.c +index 282d9fd..71e5f11 100644 +--- a/drivers/gpu/drm/drm_ioc32.c ++++ b/drivers/gpu/drm/drm_ioc32.c +@@ -463,7 +463,7 @@ static int compat_drm_infobufs(struct file *file, unsigned int cmd, + request = compat_alloc_user_space(nbytes); + if (!access_ok(VERIFY_WRITE, request, nbytes)) + return -EFAULT; +- list = (struct drm_buf_desc *) (request + 1); ++ list = (struct drm_buf_desc __user *) (request + 1); + + if (__put_user(count, &request->count) + || __put_user(list, &request->list)) +@@ -525,7 +525,7 @@ static int compat_drm_mapbufs(struct file *file, unsigned int cmd, + request = compat_alloc_user_space(nbytes); + if (!access_ok(VERIFY_WRITE, request, nbytes)) + return -EFAULT; +- list = (struct drm_buf_pub *) (request + 1); ++ list = (struct drm_buf_pub __user *) (request + 1); + + if (__put_user(count, &request->count) + || __put_user(list, &request->list)) +diff --git a/drivers/gpu/drm/drm_ioctl.c b/drivers/gpu/drm/drm_ioctl.c +index 9b9ff46..4ea724c 100644 +--- a/drivers/gpu/drm/drm_ioctl.c ++++ b/drivers/gpu/drm/drm_ioctl.c +@@ -283,7 +283,7 @@ int drm_getstats(struct drm_device *dev, void *data, + stats->data[i].value = + (file_priv->master->lock.hw_lock ? file_priv->master->lock.hw_lock->lock : 0); + else +- stats->data[i].value = atomic_read(&dev->counts[i]); ++ stats->data[i].value = atomic_read_unchecked(&dev->counts[i]); + stats->data[i].type = dev->types[i]; + } + +diff --git a/drivers/gpu/drm/drm_lock.c b/drivers/gpu/drm/drm_lock.c +index e2f70a5..c703e86 100644 +--- a/drivers/gpu/drm/drm_lock.c ++++ b/drivers/gpu/drm/drm_lock.c +@@ -87,7 +87,7 @@ int drm_lock(struct drm_device *dev, void *data, struct drm_file *file_priv) + if (drm_lock_take(&master->lock, lock->context)) { + master->lock.file_priv = file_priv; + master->lock.lock_time = jiffies; +- atomic_inc(&dev->counts[_DRM_STAT_LOCKS]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_LOCKS]); + break; /* Got lock */ + } + +@@ -165,7 +165,7 @@ int drm_unlock(struct drm_device *dev, void *data, struct drm_file *file_priv) + return -EINVAL; + } + +- atomic_inc(&dev->counts[_DRM_STAT_UNLOCKS]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_UNLOCKS]); + + /* kernel_context_switch isn't used by any of the x86 drm + * modules but is required by the Sparc driver. +diff --git a/drivers/gpu/drm/i810/i810_dma.c b/drivers/gpu/drm/i810/i810_dma.c +index 7d1d88c..b9131b2 100644 +--- a/drivers/gpu/drm/i810/i810_dma.c ++++ b/drivers/gpu/drm/i810/i810_dma.c +@@ -952,8 +952,8 @@ static int i810_dma_vertex(struct drm_device *dev, void *data, + dma->buflist[vertex->idx], + vertex->discard, vertex->used); + +- atomic_add(vertex->used, &dev->counts[_DRM_STAT_SECONDARY]); +- atomic_inc(&dev->counts[_DRM_STAT_DMA]); ++ atomic_add_unchecked(vertex->used, &dev->counts[_DRM_STAT_SECONDARY]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_DMA]); + sarea_priv->last_enqueue = dev_priv->counter - 1; + sarea_priv->last_dispatch = (int)hw_status[5]; + +@@ -1115,8 +1115,8 @@ static int i810_dma_mc(struct drm_device *dev, void *data, + i810_dma_dispatch_mc(dev, dma->buflist[mc->idx], mc->used, + mc->last_render); + +- atomic_add(mc->used, &dev->counts[_DRM_STAT_SECONDARY]); +- atomic_inc(&dev->counts[_DRM_STAT_DMA]); ++ atomic_add_unchecked(mc->used, &dev->counts[_DRM_STAT_SECONDARY]); ++ atomic_inc_unchecked(&dev->counts[_DRM_STAT_DMA]); + sarea_priv->last_enqueue = dev_priv->counter - 1; + sarea_priv->last_dispatch = (int)hw_status[5]; + +diff --git a/drivers/gpu/drm/i810/i810_drv.h b/drivers/gpu/drm/i810/i810_drv.h +index 21e2691..7321edd 100644 +--- a/drivers/gpu/drm/i810/i810_drv.h ++++ b/drivers/gpu/drm/i810/i810_drv.h +@@ -108,8 +108,8 @@ typedef struct drm_i810_private { + int page_flipping; + + wait_queue_head_t irq_queue; +- atomic_t irq_received; +- atomic_t irq_emitted; ++ atomic_unchecked_t irq_received; ++ atomic_unchecked_t irq_emitted; + + int front_offset; + } drm_i810_private_t; +diff --git a/drivers/gpu/drm/i830/i830_drv.h b/drivers/gpu/drm/i830/i830_drv.h +index da82afe..48a45de 100644 +--- a/drivers/gpu/drm/i830/i830_drv.h ++++ b/drivers/gpu/drm/i830/i830_drv.h +@@ -115,8 +115,8 @@ typedef struct drm_i830_private { + int page_flipping; + + wait_queue_head_t irq_queue; +- atomic_t irq_received; +- atomic_t irq_emitted; ++ atomic_unchecked_t irq_received; ++ atomic_unchecked_t irq_emitted; + + int use_mi_batchbuffer_start; + +diff --git a/drivers/gpu/drm/i830/i830_irq.c b/drivers/gpu/drm/i830/i830_irq.c +index 91ec2bb..6f21fab 100644 +--- a/drivers/gpu/drm/i830/i830_irq.c ++++ b/drivers/gpu/drm/i830/i830_irq.c +@@ -47,7 +47,7 @@ irqreturn_t i830_driver_irq_handler(DRM_IRQ_ARGS) + + I830_WRITE16(I830REG_INT_IDENTITY_R, temp); + +- atomic_inc(&dev_priv->irq_received); ++ atomic_inc_unchecked(&dev_priv->irq_received); + wake_up_interruptible(&dev_priv->irq_queue); + + return IRQ_HANDLED; +@@ -60,14 +60,14 @@ static int i830_emit_irq(struct drm_device * dev) + + DRM_DEBUG("%s\n", __func__); + +- atomic_inc(&dev_priv->irq_emitted); ++ atomic_inc_unchecked(&dev_priv->irq_emitted); + + BEGIN_LP_RING(2); + OUT_RING(0); + OUT_RING(GFX_OP_USER_INTERRUPT); + ADVANCE_LP_RING(); + +- return atomic_read(&dev_priv->irq_emitted); ++ return atomic_read_unchecked(&dev_priv->irq_emitted); + } + + static int i830_wait_irq(struct drm_device * dev, int irq_nr) +@@ -79,7 +79,7 @@ static int i830_wait_irq(struct drm_device * dev, int irq_nr) + + DRM_DEBUG("%s\n", __func__); + +- if (atomic_read(&dev_priv->irq_received) >= irq_nr) ++ if (atomic_read_unchecked(&dev_priv->irq_received) >= irq_nr) + return 0; + + dev_priv->sarea_priv->perf_boxes |= I830_BOX_WAIT; +@@ -88,7 +88,7 @@ static int i830_wait_irq(struct drm_device * dev, int irq_nr) + + for (;;) { + __set_current_state(TASK_INTERRUPTIBLE); +- if (atomic_read(&dev_priv->irq_received) >= irq_nr) ++ if (atomic_read_unchecked(&dev_priv->irq_received) >= irq_nr) + break; + if ((signed)(end - jiffies) <= 0) { + DRM_ERROR("timeout iir %x imr %x ier %x hwstam %x\n", +@@ -163,8 +163,8 @@ void i830_driver_irq_preinstall(struct drm_device * dev) + I830_WRITE16(I830REG_HWSTAM, 0xffff); + I830_WRITE16(I830REG_INT_MASK_R, 0x0); + I830_WRITE16(I830REG_INT_ENABLE_R, 0x0); +- atomic_set(&dev_priv->irq_received, 0); +- atomic_set(&dev_priv->irq_emitted, 0); ++ atomic_set_unchecked(&dev_priv->irq_received, 0); ++ atomic_set_unchecked(&dev_priv->irq_emitted, 0); + init_waitqueue_head(&dev_priv->irq_queue); + } + +diff --git a/drivers/gpu/drm/i915/dvo.h b/drivers/gpu/drm/i915/dvo.h +index 288fc50..c6092055 100644 +--- a/drivers/gpu/drm/i915/dvo.h ++++ b/drivers/gpu/drm/i915/dvo.h +@@ -135,23 +135,23 @@ struct intel_dvo_dev_ops { + * + * \return singly-linked list of modes or NULL if no modes found. + */ +- struct drm_display_mode *(*get_modes)(struct intel_dvo_device *dvo); ++ struct drm_display_mode *(* const get_modes)(struct intel_dvo_device *dvo); + + /** + * Clean up driver-specific bits of the output + */ +- void (*destroy) (struct intel_dvo_device *dvo); ++ void (* const destroy) (struct intel_dvo_device *dvo); + + /** + * Debugging hook to dump device registers to log file + */ +- void (*dump_regs)(struct intel_dvo_device *dvo); ++ void (* const dump_regs)(struct intel_dvo_device *dvo); + }; + +-extern struct intel_dvo_dev_ops sil164_ops; +-extern struct intel_dvo_dev_ops ch7xxx_ops; +-extern struct intel_dvo_dev_ops ivch_ops; +-extern struct intel_dvo_dev_ops tfp410_ops; +-extern struct intel_dvo_dev_ops ch7017_ops; ++extern const struct intel_dvo_dev_ops sil164_ops; ++extern const struct intel_dvo_dev_ops ch7xxx_ops; ++extern const struct intel_dvo_dev_ops ivch_ops; ++extern const struct intel_dvo_dev_ops tfp410_ops; ++extern const struct intel_dvo_dev_ops ch7017_ops; + + #endif /* _INTEL_DVO_H */ +diff --git a/drivers/gpu/drm/i915/dvo_ch7017.c b/drivers/gpu/drm/i915/dvo_ch7017.c +index 621815b..499d82e 100644 +--- a/drivers/gpu/drm/i915/dvo_ch7017.c ++++ b/drivers/gpu/drm/i915/dvo_ch7017.c +@@ -443,7 +443,7 @@ static void ch7017_destroy(struct intel_dvo_device *dvo) + } + } + +-struct intel_dvo_dev_ops ch7017_ops = { ++const struct intel_dvo_dev_ops ch7017_ops = { + .init = ch7017_init, + .detect = ch7017_detect, + .mode_valid = ch7017_mode_valid, +diff --git a/drivers/gpu/drm/i915/dvo_ch7xxx.c b/drivers/gpu/drm/i915/dvo_ch7xxx.c +index a9b8962..ac769ba 100644 +--- a/drivers/gpu/drm/i915/dvo_ch7xxx.c ++++ b/drivers/gpu/drm/i915/dvo_ch7xxx.c +@@ -356,7 +356,7 @@ static void ch7xxx_destroy(struct intel_dvo_device *dvo) + } + } + +-struct intel_dvo_dev_ops ch7xxx_ops = { ++const struct intel_dvo_dev_ops ch7xxx_ops = { + .init = ch7xxx_init, + .detect = ch7xxx_detect, + .mode_valid = ch7xxx_mode_valid, +diff --git a/drivers/gpu/drm/i915/dvo_ivch.c b/drivers/gpu/drm/i915/dvo_ivch.c +index aa176f9..ed2930c 100644 +--- a/drivers/gpu/drm/i915/dvo_ivch.c ++++ b/drivers/gpu/drm/i915/dvo_ivch.c +@@ -430,7 +430,7 @@ static void ivch_destroy(struct intel_dvo_device *dvo) + } + } + +-struct intel_dvo_dev_ops ivch_ops= { ++const struct intel_dvo_dev_ops ivch_ops= { + .init = ivch_init, + .dpms = ivch_dpms, + .save = ivch_save, +diff --git a/drivers/gpu/drm/i915/dvo_sil164.c b/drivers/gpu/drm/i915/dvo_sil164.c +index e1c1f73..7dbebcf 100644 +--- a/drivers/gpu/drm/i915/dvo_sil164.c ++++ b/drivers/gpu/drm/i915/dvo_sil164.c +@@ -290,7 +290,7 @@ static void sil164_destroy(struct intel_dvo_device *dvo) + } + } + +-struct intel_dvo_dev_ops sil164_ops = { ++const struct intel_dvo_dev_ops sil164_ops = { + .init = sil164_init, + .detect = sil164_detect, + .mode_valid = sil164_mode_valid, +diff --git a/drivers/gpu/drm/i915/dvo_tfp410.c b/drivers/gpu/drm/i915/dvo_tfp410.c +index 16dce84..7e1b6f8 100644 +--- a/drivers/gpu/drm/i915/dvo_tfp410.c ++++ b/drivers/gpu/drm/i915/dvo_tfp410.c +@@ -323,7 +323,7 @@ static void tfp410_destroy(struct intel_dvo_device *dvo) + } + } + +-struct intel_dvo_dev_ops tfp410_ops = { ++const struct intel_dvo_dev_ops tfp410_ops = { + .init = tfp410_init, + .detect = tfp410_detect, + .mode_valid = tfp410_mode_valid, +diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c +index 7e859d6..7d1cf2b 100644 +--- a/drivers/gpu/drm/i915/i915_debugfs.c ++++ b/drivers/gpu/drm/i915/i915_debugfs.c +@@ -192,7 +192,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) + I915_READ(GTIMR)); + } + seq_printf(m, "Interrupts received: %d\n", +- atomic_read(&dev_priv->irq_received)); ++ atomic_read_unchecked(&dev_priv->irq_received)); + if (dev_priv->hw_status_page != NULL) { + seq_printf(m, "Current sequence: %d\n", + i915_get_gem_seqno(dev)); +diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c +index 5449239..7e4f68d 100644 +--- a/drivers/gpu/drm/i915/i915_drv.c ++++ b/drivers/gpu/drm/i915/i915_drv.c +@@ -285,7 +285,7 @@ i915_pci_resume(struct pci_dev *pdev) + return i915_resume(dev); + } + +-static struct vm_operations_struct i915_gem_vm_ops = { ++static const struct vm_operations_struct i915_gem_vm_ops = { + .fault = i915_gem_fault, + .open = drm_gem_vm_open, + .close = drm_gem_vm_close, +diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h +index 97163f7..c24c7c7 100644 +--- a/drivers/gpu/drm/i915/i915_drv.h ++++ b/drivers/gpu/drm/i915/i915_drv.h +@@ -168,7 +168,7 @@ struct drm_i915_display_funcs { + /* display clock increase/decrease */ + /* pll clock increase/decrease */ + /* clock gating init */ +-}; ++} __no_const; + + typedef struct drm_i915_private { + struct drm_device *dev; +@@ -197,7 +197,7 @@ typedef struct drm_i915_private { + int page_flipping; + + wait_queue_head_t irq_queue; +- atomic_t irq_received; ++ atomic_unchecked_t irq_received; + /** Protects user_irq_refcount and irq_mask_reg */ + spinlock_t user_irq_lock; + /** Refcount for i915_user_irq_get() versus i915_user_irq_put(). */ +diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c +index 27a3074..eb3f959 100644 +--- a/drivers/gpu/drm/i915/i915_gem.c ++++ b/drivers/gpu/drm/i915/i915_gem.c +@@ -102,7 +102,7 @@ i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, + + args->aper_size = dev->gtt_total; + args->aper_available_size = (args->aper_size - +- atomic_read(&dev->pin_memory)); ++ atomic_read_unchecked(&dev->pin_memory)); + + return 0; + } +@@ -2058,7 +2058,7 @@ i915_gem_object_unbind(struct drm_gem_object *obj) + + if (obj_priv->gtt_space) { + atomic_dec(&dev->gtt_count); +- atomic_sub(obj->size, &dev->gtt_memory); ++ atomic_sub_unchecked(obj->size, &dev->gtt_memory); + + drm_mm_put_block(obj_priv->gtt_space); + obj_priv->gtt_space = NULL; +@@ -2701,7 +2701,7 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment) + goto search_free; + } + atomic_inc(&dev->gtt_count); +- atomic_add(obj->size, &dev->gtt_memory); ++ atomic_add_unchecked(obj->size, &dev->gtt_memory); + + /* Assert that the object is not currently in any GPU domain. As it + * wasn't in the GTT, there shouldn't be any way it could have been in +@@ -3755,9 +3755,9 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, + "%d/%d gtt bytes\n", + atomic_read(&dev->object_count), + atomic_read(&dev->pin_count), +- atomic_read(&dev->object_memory), +- atomic_read(&dev->pin_memory), +- atomic_read(&dev->gtt_memory), ++ atomic_read_unchecked(&dev->object_memory), ++ atomic_read_unchecked(&dev->pin_memory), ++ atomic_read_unchecked(&dev->gtt_memory), + dev->gtt_total); + } + goto err; +@@ -3989,7 +3989,7 @@ i915_gem_object_pin(struct drm_gem_object *obj, uint32_t alignment) + */ + if (obj_priv->pin_count == 1) { + atomic_inc(&dev->pin_count); +- atomic_add(obj->size, &dev->pin_memory); ++ atomic_add_unchecked(obj->size, &dev->pin_memory); + if (!obj_priv->active && + (obj->write_domain & I915_GEM_GPU_DOMAINS) == 0 && + !list_empty(&obj_priv->list)) +@@ -4022,7 +4022,7 @@ i915_gem_object_unpin(struct drm_gem_object *obj) + list_move_tail(&obj_priv->list, + &dev_priv->mm.inactive_list); + atomic_dec(&dev->pin_count); +- atomic_sub(obj->size, &dev->pin_memory); ++ atomic_sub_unchecked(obj->size, &dev->pin_memory); + } + i915_verify_inactive(dev, __FILE__, __LINE__); + } +diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c +index 63f28ad..f5469da 100644 +--- a/drivers/gpu/drm/i915/i915_irq.c ++++ b/drivers/gpu/drm/i915/i915_irq.c +@@ -528,7 +528,7 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) + int irq_received; + int ret = IRQ_NONE; + +- atomic_inc(&dev_priv->irq_received); ++ atomic_inc_unchecked(&dev_priv->irq_received); + + if (IS_IGDNG(dev)) + return igdng_irq_handler(dev); +@@ -1021,7 +1021,7 @@ void i915_driver_irq_preinstall(struct drm_device * dev) + { + drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; + +- atomic_set(&dev_priv->irq_received, 0); ++ atomic_set_unchecked(&dev_priv->irq_received, 0); + + INIT_WORK(&dev_priv->hotplug_work, i915_hotplug_work_func); + INIT_WORK(&dev_priv->error_work, i915_error_work_func); +diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c +index 5d9c6a7..d1b0e29 100644 +--- a/drivers/gpu/drm/i915/intel_sdvo.c ++++ b/drivers/gpu/drm/i915/intel_sdvo.c +@@ -2795,7 +2795,9 @@ bool intel_sdvo_init(struct drm_device *dev, int output_device) + sdvo_priv->slave_addr = intel_sdvo_get_slave_addr(dev, output_device); + + /* Save the bit-banging i2c functionality for use by the DDC wrapper */ +- intel_sdvo_i2c_bit_algo.functionality = intel_output->i2c_bus->algo->functionality; ++ pax_open_kernel(); ++ *(void **)&intel_sdvo_i2c_bit_algo.functionality = intel_output->i2c_bus->algo->functionality; ++ pax_close_kernel(); + + /* Read the regs to test if we can talk to the device */ + for (i = 0; i < 0x40; i++) { +diff --git a/drivers/gpu/drm/mga/mga_drv.h b/drivers/gpu/drm/mga/mga_drv.h +index be6c6b9..8615d9c 100644 +--- a/drivers/gpu/drm/mga/mga_drv.h ++++ b/drivers/gpu/drm/mga/mga_drv.h +@@ -120,9 +120,9 @@ typedef struct drm_mga_private { + u32 clear_cmd; + u32 maccess; + +- atomic_t vbl_received; /**< Number of vblanks received. */ ++ atomic_unchecked_t vbl_received; /**< Number of vblanks received. */ + wait_queue_head_t fence_queue; +- atomic_t last_fence_retired; ++ atomic_unchecked_t last_fence_retired; + u32 next_fence_to_post; + + unsigned int fb_cpp; +diff --git a/drivers/gpu/drm/mga/mga_irq.c b/drivers/gpu/drm/mga/mga_irq.c +index daa6041..a28a5da 100644 +--- a/drivers/gpu/drm/mga/mga_irq.c ++++ b/drivers/gpu/drm/mga/mga_irq.c +@@ -44,7 +44,7 @@ u32 mga_get_vblank_counter(struct drm_device *dev, int crtc) + if (crtc != 0) + return 0; + +- return atomic_read(&dev_priv->vbl_received); ++ return atomic_read_unchecked(&dev_priv->vbl_received); + } + + +@@ -60,7 +60,7 @@ irqreturn_t mga_driver_irq_handler(DRM_IRQ_ARGS) + /* VBLANK interrupt */ + if (status & MGA_VLINEPEN) { + MGA_WRITE(MGA_ICLEAR, MGA_VLINEICLR); +- atomic_inc(&dev_priv->vbl_received); ++ atomic_inc_unchecked(&dev_priv->vbl_received); + drm_handle_vblank(dev, 0); + handled = 1; + } +@@ -80,7 +80,7 @@ irqreturn_t mga_driver_irq_handler(DRM_IRQ_ARGS) + MGA_WRITE(MGA_PRIMEND, prim_end); + } + +- atomic_inc(&dev_priv->last_fence_retired); ++ atomic_inc_unchecked(&dev_priv->last_fence_retired); + DRM_WAKEUP(&dev_priv->fence_queue); + handled = 1; + } +@@ -131,7 +131,7 @@ int mga_driver_fence_wait(struct drm_device * dev, unsigned int *sequence) + * using fences. + */ + DRM_WAIT_ON(ret, dev_priv->fence_queue, 3 * DRM_HZ, +- (((cur_fence = atomic_read(&dev_priv->last_fence_retired)) ++ (((cur_fence = atomic_read_unchecked(&dev_priv->last_fence_retired)) + - *sequence) <= (1 << 23))); + + *sequence = cur_fence; +diff --git a/drivers/gpu/drm/r128/r128_cce.c b/drivers/gpu/drm/r128/r128_cce.c +index 4c39a40..b22a9ea 100644 +--- a/drivers/gpu/drm/r128/r128_cce.c ++++ b/drivers/gpu/drm/r128/r128_cce.c +@@ -377,7 +377,7 @@ static int r128_do_init_cce(struct drm_device * dev, drm_r128_init_t * init) + + /* GH: Simple idle check. + */ +- atomic_set(&dev_priv->idle_count, 0); ++ atomic_set_unchecked(&dev_priv->idle_count, 0); + + /* We don't support anything other than bus-mastering ring mode, + * but the ring can be in either AGP or PCI space for the ring +diff --git a/drivers/gpu/drm/r128/r128_drv.h b/drivers/gpu/drm/r128/r128_drv.h +index 3c60829..4faf484 100644 +--- a/drivers/gpu/drm/r128/r128_drv.h ++++ b/drivers/gpu/drm/r128/r128_drv.h +@@ -90,14 +90,14 @@ typedef struct drm_r128_private { + int is_pci; + unsigned long cce_buffers_offset; + +- atomic_t idle_count; ++ atomic_unchecked_t idle_count; + + int page_flipping; + int current_page; + u32 crtc_offset; + u32 crtc_offset_cntl; + +- atomic_t vbl_received; ++ atomic_unchecked_t vbl_received; + + u32 color_fmt; + unsigned int front_offset; +diff --git a/drivers/gpu/drm/r128/r128_irq.c b/drivers/gpu/drm/r128/r128_irq.c +index 69810fb..97bf17a 100644 +--- a/drivers/gpu/drm/r128/r128_irq.c ++++ b/drivers/gpu/drm/r128/r128_irq.c +@@ -42,7 +42,7 @@ u32 r128_get_vblank_counter(struct drm_device *dev, int crtc) + if (crtc != 0) + return 0; + +- return atomic_read(&dev_priv->vbl_received); ++ return atomic_read_unchecked(&dev_priv->vbl_received); + } + + irqreturn_t r128_driver_irq_handler(DRM_IRQ_ARGS) +@@ -56,7 +56,7 @@ irqreturn_t r128_driver_irq_handler(DRM_IRQ_ARGS) + /* VBLANK interrupt */ + if (status & R128_CRTC_VBLANK_INT) { + R128_WRITE(R128_GEN_INT_STATUS, R128_CRTC_VBLANK_INT_AK); +- atomic_inc(&dev_priv->vbl_received); ++ atomic_inc_unchecked(&dev_priv->vbl_received); + drm_handle_vblank(dev, 0); + return IRQ_HANDLED; + } +diff --git a/drivers/gpu/drm/r128/r128_state.c b/drivers/gpu/drm/r128/r128_state.c +index af2665c..51922d2 100644 +--- a/drivers/gpu/drm/r128/r128_state.c ++++ b/drivers/gpu/drm/r128/r128_state.c +@@ -323,10 +323,10 @@ static void r128_clear_box(drm_r128_private_t * dev_priv, + + static void r128_cce_performance_boxes(drm_r128_private_t * dev_priv) + { +- if (atomic_read(&dev_priv->idle_count) == 0) { ++ if (atomic_read_unchecked(&dev_priv->idle_count) == 0) { + r128_clear_box(dev_priv, 64, 4, 8, 8, 0, 255, 0); + } else { +- atomic_set(&dev_priv->idle_count, 0); ++ atomic_set_unchecked(&dev_priv->idle_count, 0); + } + } + +diff --git a/drivers/gpu/drm/radeon/atom.c b/drivers/gpu/drm/radeon/atom.c +index dd72b91..8644b3c 100644 +--- a/drivers/gpu/drm/radeon/atom.c ++++ b/drivers/gpu/drm/radeon/atom.c +@@ -1115,6 +1115,8 @@ struct atom_context *atom_parse(struct card_info *card, void *bios) + char name[512]; + int i; + ++ pax_track_stack(); ++ + ctx->card = card; + ctx->bios = bios; + +diff --git a/drivers/gpu/drm/radeon/mkregtable.c b/drivers/gpu/drm/radeon/mkregtable.c +index 0d79577..efaa7a5 100644 +--- a/drivers/gpu/drm/radeon/mkregtable.c ++++ b/drivers/gpu/drm/radeon/mkregtable.c +@@ -637,14 +637,14 @@ static int parser_auth(struct table *t, const char *filename) + regex_t mask_rex; + regmatch_t match[4]; + char buf[1024]; +- size_t end; ++ long end; + int len; + int done = 0; + int r; + unsigned o; + struct offset *offset; + char last_reg_s[10]; +- int last_reg; ++ unsigned long last_reg; + + if (regcomp + (&mask_rex, "(0x[0-9a-fA-F]*) *([_a-zA-Z0-9]*)", REG_EXTENDED)) { +diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h +index 6735213..38c2c67 100644 +--- a/drivers/gpu/drm/radeon/radeon.h ++++ b/drivers/gpu/drm/radeon/radeon.h +@@ -149,7 +149,7 @@ int radeon_pm_init(struct radeon_device *rdev); + */ + struct radeon_fence_driver { + uint32_t scratch_reg; +- atomic_t seq; ++ atomic_unchecked_t seq; + uint32_t last_seq; + unsigned long count_timeout; + wait_queue_head_t queue; +@@ -640,7 +640,7 @@ struct radeon_asic { + uint32_t offset, uint32_t obj_size); + int (*clear_surface_reg)(struct radeon_device *rdev, int reg); + void (*bandwidth_update)(struct radeon_device *rdev); +-}; ++} __no_const; + + /* + * Asic structures +diff --git a/drivers/gpu/drm/radeon/radeon_atombios.c b/drivers/gpu/drm/radeon/radeon_atombios.c +index 4e928b9..d8b6008 100644 +--- a/drivers/gpu/drm/radeon/radeon_atombios.c ++++ b/drivers/gpu/drm/radeon/radeon_atombios.c +@@ -275,6 +275,8 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) + bool linkb; + struct radeon_i2c_bus_rec ddc_bus; + ++ pax_track_stack(); ++ + atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); + + if (data_offset == 0) +@@ -520,13 +522,13 @@ static uint16_t atombios_get_connector_object_id(struct drm_device *dev, + } + } + +-struct bios_connector { ++static struct bios_connector { + bool valid; + uint16_t line_mux; + uint16_t devices; + int connector_type; + struct radeon_i2c_bus_rec ddc_bus; +-}; ++} bios_connectors[ATOM_MAX_SUPPORTED_DEVICE]; + + bool radeon_get_atom_connector_info_from_supported_devices_table(struct + drm_device +@@ -542,7 +544,6 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct + uint8_t dac; + union atom_supported_devices *supported_devices; + int i, j; +- struct bios_connector bios_connectors[ATOM_MAX_SUPPORTED_DEVICE]; + + atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); + +diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c +index 083a181..ccccae0 100644 +--- a/drivers/gpu/drm/radeon/radeon_display.c ++++ b/drivers/gpu/drm/radeon/radeon_display.c +@@ -482,7 +482,7 @@ void radeon_compute_pll(struct radeon_pll *pll, + + if (flags & RADEON_PLL_PREFER_CLOSEST_LOWER) { + error = freq - current_freq; +- error = error < 0 ? 0xffffffff : error; ++ error = (int32_t)error < 0 ? 0xffffffff : error; + } else + error = abs(current_freq - freq); + vco_diff = abs(vco - best_vco); +diff --git a/drivers/gpu/drm/radeon/radeon_drv.h b/drivers/gpu/drm/radeon/radeon_drv.h +index 76e4070..193fa7f 100644 +--- a/drivers/gpu/drm/radeon/radeon_drv.h ++++ b/drivers/gpu/drm/radeon/radeon_drv.h +@@ -253,7 +253,7 @@ typedef struct drm_radeon_private { + + /* SW interrupt */ + wait_queue_head_t swi_queue; +- atomic_t swi_emitted; ++ atomic_unchecked_t swi_emitted; + int vblank_crtc; + uint32_t irq_enable_reg; + uint32_t r500_disp_irq_reg; +diff --git a/drivers/gpu/drm/radeon/radeon_fence.c b/drivers/gpu/drm/radeon/radeon_fence.c +index 3beb26d..6ce9c4a 100644 +--- a/drivers/gpu/drm/radeon/radeon_fence.c ++++ b/drivers/gpu/drm/radeon/radeon_fence.c +@@ -47,7 +47,7 @@ int radeon_fence_emit(struct radeon_device *rdev, struct radeon_fence *fence) + write_unlock_irqrestore(&rdev->fence_drv.lock, irq_flags); + return 0; + } +- fence->seq = atomic_add_return(1, &rdev->fence_drv.seq); ++ fence->seq = atomic_add_return_unchecked(1, &rdev->fence_drv.seq); + if (!rdev->cp.ready) { + /* FIXME: cp is not running assume everythings is done right + * away +@@ -364,7 +364,7 @@ int radeon_fence_driver_init(struct radeon_device *rdev) + return r; + } + WREG32(rdev->fence_drv.scratch_reg, 0); +- atomic_set(&rdev->fence_drv.seq, 0); ++ atomic_set_unchecked(&rdev->fence_drv.seq, 0); + INIT_LIST_HEAD(&rdev->fence_drv.created); + INIT_LIST_HEAD(&rdev->fence_drv.emited); + INIT_LIST_HEAD(&rdev->fence_drv.signaled); +diff --git a/drivers/gpu/drm/radeon/radeon_ioc32.c b/drivers/gpu/drm/radeon/radeon_ioc32.c +index a1bf11d..4a123c0 100644 +--- a/drivers/gpu/drm/radeon/radeon_ioc32.c ++++ b/drivers/gpu/drm/radeon/radeon_ioc32.c +@@ -368,7 +368,7 @@ static int compat_radeon_cp_setparam(struct file *file, unsigned int cmd, + request = compat_alloc_user_space(sizeof(*request)); + if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) + || __put_user(req32.param, &request->param) +- || __put_user((void __user *)(unsigned long)req32.value, ++ || __put_user((unsigned long)req32.value, + &request->value)) + return -EFAULT; + +diff --git a/drivers/gpu/drm/radeon/radeon_irq.c b/drivers/gpu/drm/radeon/radeon_irq.c +index b79ecc4..8dab92d 100644 +--- a/drivers/gpu/drm/radeon/radeon_irq.c ++++ b/drivers/gpu/drm/radeon/radeon_irq.c +@@ -225,8 +225,8 @@ static int radeon_emit_irq(struct drm_device * dev) + unsigned int ret; + RING_LOCALS; + +- atomic_inc(&dev_priv->swi_emitted); +- ret = atomic_read(&dev_priv->swi_emitted); ++ atomic_inc_unchecked(&dev_priv->swi_emitted); ++ ret = atomic_read_unchecked(&dev_priv->swi_emitted); + + BEGIN_RING(4); + OUT_RING_REG(RADEON_LAST_SWI_REG, ret); +@@ -352,7 +352,7 @@ int radeon_driver_irq_postinstall(struct drm_device *dev) + drm_radeon_private_t *dev_priv = + (drm_radeon_private_t *) dev->dev_private; + +- atomic_set(&dev_priv->swi_emitted, 0); ++ atomic_set_unchecked(&dev_priv->swi_emitted, 0); + DRM_INIT_WAITQUEUE(&dev_priv->swi_queue); + + dev->max_vblank_count = 0x001fffff; +diff --git a/drivers/gpu/drm/radeon/radeon_state.c b/drivers/gpu/drm/radeon/radeon_state.c +index 4747910..48ca4b3 100644 +--- a/drivers/gpu/drm/radeon/radeon_state.c ++++ b/drivers/gpu/drm/radeon/radeon_state.c +@@ -3021,7 +3021,7 @@ static int radeon_cp_getparam(struct drm_device *dev, void *data, struct drm_fil + { + drm_radeon_private_t *dev_priv = dev->dev_private; + drm_radeon_getparam_t *param = data; +- int value; ++ int value = 0; + + DRM_DEBUG("pid=%d\n", DRM_CURRENTPID); + +diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c +index 1381e06..0e53b17 100644 +--- a/drivers/gpu/drm/radeon/radeon_ttm.c ++++ b/drivers/gpu/drm/radeon/radeon_ttm.c +@@ -535,27 +535,10 @@ void radeon_ttm_fini(struct radeon_device *rdev) + DRM_INFO("radeon: ttm finalized\n"); + } + +-static struct vm_operations_struct radeon_ttm_vm_ops; +-static const struct vm_operations_struct *ttm_vm_ops = NULL; +- +-static int radeon_ttm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) +-{ +- struct ttm_buffer_object *bo; +- int r; +- +- bo = (struct ttm_buffer_object *)vma->vm_private_data; +- if (bo == NULL) { +- return VM_FAULT_NOPAGE; +- } +- r = ttm_vm_ops->fault(vma, vmf); +- return r; +-} +- + int radeon_mmap(struct file *filp, struct vm_area_struct *vma) + { + struct drm_file *file_priv; + struct radeon_device *rdev; +- int r; + + if (unlikely(vma->vm_pgoff < DRM_FILE_PAGE_OFFSET)) { + return drm_mmap(filp, vma); +@@ -563,20 +546,9 @@ int radeon_mmap(struct file *filp, struct vm_area_struct *vma) + + file_priv = (struct drm_file *)filp->private_data; + rdev = file_priv->minor->dev->dev_private; +- if (rdev == NULL) { ++ if (!rdev) + return -EINVAL; +- } +- r = ttm_bo_mmap(filp, vma, &rdev->mman.bdev); +- if (unlikely(r != 0)) { +- return r; +- } +- if (unlikely(ttm_vm_ops == NULL)) { +- ttm_vm_ops = vma->vm_ops; +- radeon_ttm_vm_ops = *ttm_vm_ops; +- radeon_ttm_vm_ops.fault = &radeon_ttm_fault; +- } +- vma->vm_ops = &radeon_ttm_vm_ops; +- return 0; ++ return ttm_bo_mmap(filp, vma, &rdev->mman.bdev); + } + + +diff --git a/drivers/gpu/drm/radeon/rs690.c b/drivers/gpu/drm/radeon/rs690.c +index b12ff76..0bd0c6e 100644 +--- a/drivers/gpu/drm/radeon/rs690.c ++++ b/drivers/gpu/drm/radeon/rs690.c +@@ -302,9 +302,11 @@ void rs690_crtc_bandwidth_compute(struct radeon_device *rdev, + if (rdev->pm.max_bandwidth.full > rdev->pm.sideport_bandwidth.full && + rdev->pm.sideport_bandwidth.full) + rdev->pm.max_bandwidth = rdev->pm.sideport_bandwidth; +- read_delay_latency.full = rfixed_const(370 * 800 * 1000); ++ read_delay_latency.full = rfixed_const(800 * 1000); + read_delay_latency.full = rfixed_div(read_delay_latency, + rdev->pm.igp_sideport_mclk); ++ a.full = rfixed_const(370); ++ read_delay_latency.full = rfixed_mul(read_delay_latency, a); + } else { + if (rdev->pm.max_bandwidth.full > rdev->pm.k8_bandwidth.full && + rdev->pm.k8_bandwidth.full) +diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c +index 0ed436e..e6e7ce3 100644 +--- a/drivers/gpu/drm/ttm/ttm_bo.c ++++ b/drivers/gpu/drm/ttm/ttm_bo.c +@@ -67,7 +67,7 @@ static struct attribute *ttm_bo_global_attrs[] = { + NULL + }; + +-static struct sysfs_ops ttm_bo_global_ops = { ++static const struct sysfs_ops ttm_bo_global_ops = { + .show = &ttm_bo_global_show + }; + +diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c +index 1c040d0..f9e4af8 100644 +--- a/drivers/gpu/drm/ttm/ttm_bo_vm.c ++++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c +@@ -73,7 +73,7 @@ static int ttm_bo_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) + { + struct ttm_buffer_object *bo = (struct ttm_buffer_object *) + vma->vm_private_data; +- struct ttm_bo_device *bdev = bo->bdev; ++ struct ttm_bo_device *bdev; + unsigned long bus_base; + unsigned long bus_offset; + unsigned long bus_size; +@@ -88,6 +88,10 @@ static int ttm_bo_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) + unsigned long address = (unsigned long)vmf->virtual_address; + int retval = VM_FAULT_NOPAGE; + ++ if (!bo) ++ return VM_FAULT_NOPAGE; ++ bdev = bo->bdev; ++ + /* + * Work around locking order reversal in fault / nopfn + * between mmap_sem and bo_reserve: Perform a trylock operation +diff --git a/drivers/gpu/drm/ttm/ttm_global.c b/drivers/gpu/drm/ttm/ttm_global.c +index b170071..28ae90e 100644 +--- a/drivers/gpu/drm/ttm/ttm_global.c ++++ b/drivers/gpu/drm/ttm/ttm_global.c +@@ -36,7 +36,7 @@ + struct ttm_global_item { + struct mutex mutex; + void *object; +- int refcount; ++ atomic_t refcount; + }; + + static struct ttm_global_item glob[TTM_GLOBAL_NUM]; +@@ -49,7 +49,7 @@ void ttm_global_init(void) + struct ttm_global_item *item = &glob[i]; + mutex_init(&item->mutex); + item->object = NULL; +- item->refcount = 0; ++ atomic_set(&item->refcount, 0); + } + } + +@@ -59,7 +59,7 @@ void ttm_global_release(void) + for (i = 0; i < TTM_GLOBAL_NUM; ++i) { + struct ttm_global_item *item = &glob[i]; + BUG_ON(item->object != NULL); +- BUG_ON(item->refcount != 0); ++ BUG_ON(atomic_read(&item->refcount) != 0); + } + } + +@@ -70,7 +70,7 @@ int ttm_global_item_ref(struct ttm_global_reference *ref) + void *object; + + mutex_lock(&item->mutex); +- if (item->refcount == 0) { ++ if (atomic_read(&item->refcount) == 0) { + item->object = kzalloc(ref->size, GFP_KERNEL); + if (unlikely(item->object == NULL)) { + ret = -ENOMEM; +@@ -83,7 +83,7 @@ int ttm_global_item_ref(struct ttm_global_reference *ref) + goto out_err; + + } +- ++item->refcount; ++ atomic_inc(&item->refcount); + ref->object = item->object; + object = item->object; + mutex_unlock(&item->mutex); +@@ -100,9 +100,9 @@ void ttm_global_item_unref(struct ttm_global_reference *ref) + struct ttm_global_item *item = &glob[ref->global_type]; + + mutex_lock(&item->mutex); +- BUG_ON(item->refcount == 0); ++ BUG_ON(atomic_read(&item->refcount) == 0); + BUG_ON(ref->object != item->object); +- if (--item->refcount == 0) { ++ if (atomic_dec_and_test(&item->refcount)) { + ref->release(ref); + item->object = NULL; + } +diff --git a/drivers/gpu/drm/ttm/ttm_memory.c b/drivers/gpu/drm/ttm/ttm_memory.c +index 072c281..d8ef483 100644 +--- a/drivers/gpu/drm/ttm/ttm_memory.c ++++ b/drivers/gpu/drm/ttm/ttm_memory.c +@@ -152,7 +152,7 @@ static struct attribute *ttm_mem_zone_attrs[] = { + NULL + }; + +-static struct sysfs_ops ttm_mem_zone_ops = { ++static const struct sysfs_ops ttm_mem_zone_ops = { + .show = &ttm_mem_zone_show, + .store = &ttm_mem_zone_store + }; +diff --git a/drivers/gpu/drm/via/via_drv.h b/drivers/gpu/drm/via/via_drv.h +index cafcb84..b8e66cc 100644 +--- a/drivers/gpu/drm/via/via_drv.h ++++ b/drivers/gpu/drm/via/via_drv.h +@@ -51,7 +51,7 @@ typedef struct drm_via_ring_buffer { + typedef uint32_t maskarray_t[5]; + + typedef struct drm_via_irq { +- atomic_t irq_received; ++ atomic_unchecked_t irq_received; + uint32_t pending_mask; + uint32_t enable_mask; + wait_queue_head_t irq_queue; +@@ -75,7 +75,7 @@ typedef struct drm_via_private { + struct timeval last_vblank; + int last_vblank_valid; + unsigned usec_per_vblank; +- atomic_t vbl_received; ++ atomic_unchecked_t vbl_received; + drm_via_state_t hc_state; + char pci_buf[VIA_PCI_BUF_SIZE]; + const uint32_t *fire_offsets[VIA_FIRE_BUF_SIZE]; +diff --git a/drivers/gpu/drm/via/via_irq.c b/drivers/gpu/drm/via/via_irq.c +index 5935b88..127a8a6 100644 +--- a/drivers/gpu/drm/via/via_irq.c ++++ b/drivers/gpu/drm/via/via_irq.c +@@ -102,7 +102,7 @@ u32 via_get_vblank_counter(struct drm_device *dev, int crtc) + if (crtc != 0) + return 0; + +- return atomic_read(&dev_priv->vbl_received); ++ return atomic_read_unchecked(&dev_priv->vbl_received); + } + + irqreturn_t via_driver_irq_handler(DRM_IRQ_ARGS) +@@ -117,8 +117,8 @@ irqreturn_t via_driver_irq_handler(DRM_IRQ_ARGS) + + status = VIA_READ(VIA_REG_INTERRUPT); + if (status & VIA_IRQ_VBLANK_PENDING) { +- atomic_inc(&dev_priv->vbl_received); +- if (!(atomic_read(&dev_priv->vbl_received) & 0x0F)) { ++ atomic_inc_unchecked(&dev_priv->vbl_received); ++ if (!(atomic_read_unchecked(&dev_priv->vbl_received) & 0x0F)) { + do_gettimeofday(&cur_vblank); + if (dev_priv->last_vblank_valid) { + dev_priv->usec_per_vblank = +@@ -128,7 +128,7 @@ irqreturn_t via_driver_irq_handler(DRM_IRQ_ARGS) + dev_priv->last_vblank = cur_vblank; + dev_priv->last_vblank_valid = 1; + } +- if (!(atomic_read(&dev_priv->vbl_received) & 0xFF)) { ++ if (!(atomic_read_unchecked(&dev_priv->vbl_received) & 0xFF)) { + DRM_DEBUG("US per vblank is: %u\n", + dev_priv->usec_per_vblank); + } +@@ -138,7 +138,7 @@ irqreturn_t via_driver_irq_handler(DRM_IRQ_ARGS) + + for (i = 0; i < dev_priv->num_irqs; ++i) { + if (status & cur_irq->pending_mask) { +- atomic_inc(&cur_irq->irq_received); ++ atomic_inc_unchecked(&cur_irq->irq_received); + DRM_WAKEUP(&cur_irq->irq_queue); + handled = 1; + if (dev_priv->irq_map[drm_via_irq_dma0_td] == i) { +@@ -244,11 +244,11 @@ via_driver_irq_wait(struct drm_device * dev, unsigned int irq, int force_sequenc + DRM_WAIT_ON(ret, cur_irq->irq_queue, 3 * DRM_HZ, + ((VIA_READ(masks[irq][2]) & masks[irq][3]) == + masks[irq][4])); +- cur_irq_sequence = atomic_read(&cur_irq->irq_received); ++ cur_irq_sequence = atomic_read_unchecked(&cur_irq->irq_received); + } else { + DRM_WAIT_ON(ret, cur_irq->irq_queue, 3 * DRM_HZ, + (((cur_irq_sequence = +- atomic_read(&cur_irq->irq_received)) - ++ atomic_read_unchecked(&cur_irq->irq_received)) - + *sequence) <= (1 << 23))); + } + *sequence = cur_irq_sequence; +@@ -286,7 +286,7 @@ void via_driver_irq_preinstall(struct drm_device * dev) + } + + for (i = 0; i < dev_priv->num_irqs; ++i) { +- atomic_set(&cur_irq->irq_received, 0); ++ atomic_set_unchecked(&cur_irq->irq_received, 0); + cur_irq->enable_mask = dev_priv->irq_masks[i][0]; + cur_irq->pending_mask = dev_priv->irq_masks[i][1]; + DRM_INIT_WAITQUEUE(&cur_irq->irq_queue); +@@ -368,7 +368,7 @@ int via_wait_irq(struct drm_device *dev, void *data, struct drm_file *file_priv) + switch (irqwait->request.type & ~VIA_IRQ_FLAGS_MASK) { + case VIA_IRQ_RELATIVE: + irqwait->request.sequence += +- atomic_read(&cur_irq->irq_received); ++ atomic_read_unchecked(&cur_irq->irq_received); + irqwait->request.type &= ~_DRM_VBLANK_RELATIVE; + case VIA_IRQ_ABSOLUTE: + break; +diff --git a/drivers/gpu/vga/vgaarb.c b/drivers/gpu/vga/vgaarb.c +index aa8688d..6a0140c 100644 +--- a/drivers/gpu/vga/vgaarb.c ++++ b/drivers/gpu/vga/vgaarb.c +@@ -894,14 +894,20 @@ static ssize_t vga_arb_write(struct file *file, const char __user * buf, + uc = &priv->cards[i]; + } + +- if (!uc) +- return -EINVAL; ++ if (!uc) { ++ ret_val = -EINVAL; ++ goto done; ++ } + +- if (io_state & VGA_RSRC_LEGACY_IO && uc->io_cnt == 0) +- return -EINVAL; ++ if (io_state & VGA_RSRC_LEGACY_IO && uc->io_cnt == 0) { ++ ret_val = -EINVAL; ++ goto done; ++ } + +- if (io_state & VGA_RSRC_LEGACY_MEM && uc->mem_cnt == 0) +- return -EINVAL; ++ if (io_state & VGA_RSRC_LEGACY_MEM && uc->mem_cnt == 0) { ++ ret_val = -EINVAL; ++ goto done; ++ } + + vga_put(pdev, io_state); + +diff --git a/drivers/hid/hid-core.c b/drivers/hid/hid-core.c +index 11f8069..4783396 100644 +--- a/drivers/hid/hid-core.c ++++ b/drivers/hid/hid-core.c +@@ -1752,7 +1752,7 @@ static bool hid_ignore(struct hid_device *hdev) + + int hid_add_device(struct hid_device *hdev) + { +- static atomic_t id = ATOMIC_INIT(0); ++ static atomic_unchecked_t id = ATOMIC_INIT(0); + int ret; + + if (WARN_ON(hdev->status & HID_STAT_ADDED)) +@@ -1766,7 +1766,7 @@ int hid_add_device(struct hid_device *hdev) + /* XXX hack, any other cleaner solution after the driver core + * is converted to allow more than 20 bytes as the device name? */ + dev_set_name(&hdev->dev, "%04X:%04X:%04X.%04X", hdev->bus, +- hdev->vendor, hdev->product, atomic_inc_return(&id)); ++ hdev->vendor, hdev->product, atomic_inc_return_unchecked(&id)); + + ret = device_add(&hdev->dev); + if (!ret) +diff --git a/drivers/hid/usbhid/hiddev.c b/drivers/hid/usbhid/hiddev.c +index 8b6ee24..70f657d 100644 +--- a/drivers/hid/usbhid/hiddev.c ++++ b/drivers/hid/usbhid/hiddev.c +@@ -617,7 +617,7 @@ static long hiddev_ioctl(struct file *file, unsigned int cmd, unsigned long arg) + return put_user(HID_VERSION, (int __user *)arg); + + case HIDIOCAPPLICATION: +- if (arg < 0 || arg >= hid->maxapplication) ++ if (arg >= hid->maxapplication) + return -EINVAL; + + for (i = 0; i < hid->maxcollection; i++) +diff --git a/drivers/hwmon/lis3lv02d.c b/drivers/hwmon/lis3lv02d.c +index 5d5ed69..f40533e 100644 +--- a/drivers/hwmon/lis3lv02d.c ++++ b/drivers/hwmon/lis3lv02d.c +@@ -146,7 +146,7 @@ static irqreturn_t lis302dl_interrupt(int irq, void *dummy) + * the lid is closed. This leads to interrupts as soon as a little move + * is done. + */ +- atomic_inc(&lis3_dev.count); ++ atomic_inc_unchecked(&lis3_dev.count); + + wake_up_interruptible(&lis3_dev.misc_wait); + kill_fasync(&lis3_dev.async_queue, SIGIO, POLL_IN); +@@ -160,7 +160,7 @@ static int lis3lv02d_misc_open(struct inode *inode, struct file *file) + if (test_and_set_bit(0, &lis3_dev.misc_opened)) + return -EBUSY; /* already open */ + +- atomic_set(&lis3_dev.count, 0); ++ atomic_set_unchecked(&lis3_dev.count, 0); + + /* + * The sensor can generate interrupts for free-fall and direction +@@ -206,7 +206,7 @@ static ssize_t lis3lv02d_misc_read(struct file *file, char __user *buf, + add_wait_queue(&lis3_dev.misc_wait, &wait); + while (true) { + set_current_state(TASK_INTERRUPTIBLE); +- data = atomic_xchg(&lis3_dev.count, 0); ++ data = atomic_xchg_unchecked(&lis3_dev.count, 0); + if (data) + break; + +@@ -244,7 +244,7 @@ out: + static unsigned int lis3lv02d_misc_poll(struct file *file, poll_table *wait) + { + poll_wait(file, &lis3_dev.misc_wait, wait); +- if (atomic_read(&lis3_dev.count)) ++ if (atomic_read_unchecked(&lis3_dev.count)) + return POLLIN | POLLRDNORM; + return 0; + } +diff --git a/drivers/hwmon/lis3lv02d.h b/drivers/hwmon/lis3lv02d.h +index 7cdd76f..fe0efdf 100644 +--- a/drivers/hwmon/lis3lv02d.h ++++ b/drivers/hwmon/lis3lv02d.h +@@ -201,7 +201,7 @@ struct lis3lv02d { + + struct input_polled_dev *idev; /* input device */ + struct platform_device *pdev; /* platform device */ +- atomic_t count; /* interrupt count after last read */ ++ atomic_unchecked_t count; /* interrupt count after last read */ + int xcalib; /* calibrated null value for x */ + int ycalib; /* calibrated null value for y */ + int zcalib; /* calibrated null value for z */ +diff --git a/drivers/hwmon/sht15.c b/drivers/hwmon/sht15.c +index 740785e..5a5c6c6 100644 +--- a/drivers/hwmon/sht15.c ++++ b/drivers/hwmon/sht15.c +@@ -112,7 +112,7 @@ struct sht15_data { + int supply_uV; + int supply_uV_valid; + struct work_struct update_supply_work; +- atomic_t interrupt_handled; ++ atomic_unchecked_t interrupt_handled; + }; + + /** +@@ -245,13 +245,13 @@ static inline int sht15_update_single_val(struct sht15_data *data, + return ret; + + gpio_direction_input(data->pdata->gpio_data); +- atomic_set(&data->interrupt_handled, 0); ++ atomic_set_unchecked(&data->interrupt_handled, 0); + + enable_irq(gpio_to_irq(data->pdata->gpio_data)); + if (gpio_get_value(data->pdata->gpio_data) == 0) { + disable_irq_nosync(gpio_to_irq(data->pdata->gpio_data)); + /* Only relevant if the interrupt hasn't occured. */ +- if (!atomic_read(&data->interrupt_handled)) ++ if (!atomic_read_unchecked(&data->interrupt_handled)) + schedule_work(&data->read_work); + } + ret = wait_event_timeout(data->wait_queue, +@@ -398,7 +398,7 @@ static irqreturn_t sht15_interrupt_fired(int irq, void *d) + struct sht15_data *data = d; + /* First disable the interrupt */ + disable_irq_nosync(irq); +- atomic_inc(&data->interrupt_handled); ++ atomic_inc_unchecked(&data->interrupt_handled); + /* Then schedule a reading work struct */ + if (data->flag != SHT15_READING_NOTHING) + schedule_work(&data->read_work); +@@ -449,11 +449,11 @@ static void sht15_bh_read_data(struct work_struct *work_s) + here as could have gone low in meantime so verify + it hasn't! + */ +- atomic_set(&data->interrupt_handled, 0); ++ atomic_set_unchecked(&data->interrupt_handled, 0); + enable_irq(gpio_to_irq(data->pdata->gpio_data)); + /* If still not occured or another handler has been scheduled */ + if (gpio_get_value(data->pdata->gpio_data) +- || atomic_read(&data->interrupt_handled)) ++ || atomic_read_unchecked(&data->interrupt_handled)) + return; + } + /* Read the data back from the device */ +diff --git a/drivers/hwmon/w83791d.c b/drivers/hwmon/w83791d.c +index 97851c5..cb40626 100644 +--- a/drivers/hwmon/w83791d.c ++++ b/drivers/hwmon/w83791d.c +@@ -330,8 +330,8 @@ static int w83791d_detect(struct i2c_client *client, int kind, + struct i2c_board_info *info); + static int w83791d_remove(struct i2c_client *client); + +-static int w83791d_read(struct i2c_client *client, u8 register); +-static int w83791d_write(struct i2c_client *client, u8 register, u8 value); ++static int w83791d_read(struct i2c_client *client, u8 reg); ++static int w83791d_write(struct i2c_client *client, u8 reg, u8 value); + static struct w83791d_data *w83791d_update_device(struct device *dev); + + #ifdef DEBUG +diff --git a/drivers/i2c/busses/i2c-amd756-s4882.c b/drivers/i2c/busses/i2c-amd756-s4882.c +index 378fcb5..5e91fa8 100644 +--- a/drivers/i2c/busses/i2c-amd756-s4882.c ++++ b/drivers/i2c/busses/i2c-amd756-s4882.c +@@ -43,7 +43,7 @@ + extern struct i2c_adapter amd756_smbus; + + static struct i2c_adapter *s4882_adapter; +-static struct i2c_algorithm *s4882_algo; ++static i2c_algorithm_no_const *s4882_algo; + + /* Wrapper access functions for multiplexed SMBus */ + static DEFINE_MUTEX(amd756_lock); +diff --git a/drivers/i2c/busses/i2c-nforce2-s4985.c b/drivers/i2c/busses/i2c-nforce2-s4985.c +index 29015eb..af2d8e9 100644 +--- a/drivers/i2c/busses/i2c-nforce2-s4985.c ++++ b/drivers/i2c/busses/i2c-nforce2-s4985.c +@@ -41,7 +41,7 @@ + extern struct i2c_adapter *nforce2_smbus; + + static struct i2c_adapter *s4985_adapter; +-static struct i2c_algorithm *s4985_algo; ++static i2c_algorithm_no_const *s4985_algo; + + /* Wrapper access functions for multiplexed SMBus */ + static DEFINE_MUTEX(nforce2_lock); +diff --git a/drivers/ide/aec62xx.c b/drivers/ide/aec62xx.c +index 878f8ec..12376fc 100644 +--- a/drivers/ide/aec62xx.c ++++ b/drivers/ide/aec62xx.c +@@ -180,7 +180,7 @@ static const struct ide_port_ops atp86x_port_ops = { + .cable_detect = atp86x_cable_detect, + }; + +-static const struct ide_port_info aec62xx_chipsets[] __devinitdata = { ++static const struct ide_port_info aec62xx_chipsets[] __devinitconst = { + { /* 0: AEC6210 */ + .name = DRV_NAME, + .init_chipset = init_chipset_aec62xx, +diff --git a/drivers/ide/alim15x3.c b/drivers/ide/alim15x3.c +index e59b6de..4b4fc65 100644 +--- a/drivers/ide/alim15x3.c ++++ b/drivers/ide/alim15x3.c +@@ -509,7 +509,7 @@ static const struct ide_dma_ops ali_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info ali15x3_chipset __devinitdata = { ++static const struct ide_port_info ali15x3_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_ali15x3, + .init_hwif = init_hwif_ali15x3, +diff --git a/drivers/ide/amd74xx.c b/drivers/ide/amd74xx.c +index 628cd2e..087a414 100644 +--- a/drivers/ide/amd74xx.c ++++ b/drivers/ide/amd74xx.c +@@ -221,7 +221,7 @@ static const struct ide_port_ops amd_port_ops = { + .udma_mask = udma, \ + } + +-static const struct ide_port_info amd74xx_chipsets[] __devinitdata = { ++static const struct ide_port_info amd74xx_chipsets[] __devinitconst = { + /* 0: AMD7401 */ DECLARE_AMD_DEV(0x00, ATA_UDMA2), + /* 1: AMD7409 */ DECLARE_AMD_DEV(ATA_SWDMA2, ATA_UDMA4), + /* 2: AMD7411/7441 */ DECLARE_AMD_DEV(ATA_SWDMA2, ATA_UDMA5), +diff --git a/drivers/ide/atiixp.c b/drivers/ide/atiixp.c +index 837322b..837fd71 100644 +--- a/drivers/ide/atiixp.c ++++ b/drivers/ide/atiixp.c +@@ -137,7 +137,7 @@ static const struct ide_port_ops atiixp_port_ops = { + .cable_detect = atiixp_cable_detect, + }; + +-static const struct ide_port_info atiixp_pci_info[] __devinitdata = { ++static const struct ide_port_info atiixp_pci_info[] __devinitconst = { + { /* 0: IXP200/300/400/700 */ + .name = DRV_NAME, + .enablebits = {{0x48,0x01,0x00}, {0x48,0x08,0x00}}, +diff --git a/drivers/ide/cmd64x.c b/drivers/ide/cmd64x.c +index ca0c46f..d55318a 100644 +--- a/drivers/ide/cmd64x.c ++++ b/drivers/ide/cmd64x.c +@@ -372,7 +372,7 @@ static const struct ide_dma_ops cmd646_rev1_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info cmd64x_chipsets[] __devinitdata = { ++static const struct ide_port_info cmd64x_chipsets[] __devinitconst = { + { /* 0: CMD643 */ + .name = DRV_NAME, + .init_chipset = init_chipset_cmd64x, +diff --git a/drivers/ide/cs5520.c b/drivers/ide/cs5520.c +index 09f98ed..cebc5bc 100644 +--- a/drivers/ide/cs5520.c ++++ b/drivers/ide/cs5520.c +@@ -93,7 +93,7 @@ static const struct ide_port_ops cs5520_port_ops = { + .set_dma_mode = cs5520_set_dma_mode, + }; + +-static const struct ide_port_info cyrix_chipset __devinitdata = { ++static const struct ide_port_info cyrix_chipset __devinitconst = { + .name = DRV_NAME, + .enablebits = { { 0x60, 0x01, 0x01 }, { 0x60, 0x02, 0x02 } }, + .port_ops = &cs5520_port_ops, +diff --git a/drivers/ide/cs5530.c b/drivers/ide/cs5530.c +index 40bf05e..7d58ca0 100644 +--- a/drivers/ide/cs5530.c ++++ b/drivers/ide/cs5530.c +@@ -244,7 +244,7 @@ static const struct ide_port_ops cs5530_port_ops = { + .udma_filter = cs5530_udma_filter, + }; + +-static const struct ide_port_info cs5530_chipset __devinitdata = { ++static const struct ide_port_info cs5530_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_cs5530, + .init_hwif = init_hwif_cs5530, +diff --git a/drivers/ide/cs5535.c b/drivers/ide/cs5535.c +index 983d957..53e6172 100644 +--- a/drivers/ide/cs5535.c ++++ b/drivers/ide/cs5535.c +@@ -170,7 +170,7 @@ static const struct ide_port_ops cs5535_port_ops = { + .cable_detect = cs5535_cable_detect, + }; + +-static const struct ide_port_info cs5535_chipset __devinitdata = { ++static const struct ide_port_info cs5535_chipset __devinitconst = { + .name = DRV_NAME, + .port_ops = &cs5535_port_ops, + .host_flags = IDE_HFLAG_SINGLE | IDE_HFLAG_POST_SET_MODE, +diff --git a/drivers/ide/cy82c693.c b/drivers/ide/cy82c693.c +index 74fc540..8e933d8 100644 +--- a/drivers/ide/cy82c693.c ++++ b/drivers/ide/cy82c693.c +@@ -288,7 +288,7 @@ static const struct ide_port_ops cy82c693_port_ops = { + .set_dma_mode = cy82c693_set_dma_mode, + }; + +-static const struct ide_port_info cy82c693_chipset __devinitdata = { ++static const struct ide_port_info cy82c693_chipset __devinitconst = { + .name = DRV_NAME, + .init_iops = init_iops_cy82c693, + .port_ops = &cy82c693_port_ops, +diff --git a/drivers/ide/hpt366.c b/drivers/ide/hpt366.c +index 7ce68ef..e78197d 100644 +--- a/drivers/ide/hpt366.c ++++ b/drivers/ide/hpt366.c +@@ -507,7 +507,7 @@ static struct hpt_timings hpt37x_timings = { + } + }; + +-static const struct hpt_info hpt36x __devinitdata = { ++static const struct hpt_info hpt36x __devinitconst = { + .chip_name = "HPT36x", + .chip_type = HPT36x, + .udma_mask = HPT366_ALLOW_ATA66_3 ? (HPT366_ALLOW_ATA66_4 ? ATA_UDMA4 : ATA_UDMA3) : ATA_UDMA2, +@@ -515,7 +515,7 @@ static const struct hpt_info hpt36x __devinitdata = { + .timings = &hpt36x_timings + }; + +-static const struct hpt_info hpt370 __devinitdata = { ++static const struct hpt_info hpt370 __devinitconst = { + .chip_name = "HPT370", + .chip_type = HPT370, + .udma_mask = HPT370_ALLOW_ATA100_5 ? ATA_UDMA5 : ATA_UDMA4, +@@ -523,7 +523,7 @@ static const struct hpt_info hpt370 __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt370a __devinitdata = { ++static const struct hpt_info hpt370a __devinitconst = { + .chip_name = "HPT370A", + .chip_type = HPT370A, + .udma_mask = HPT370_ALLOW_ATA100_5 ? ATA_UDMA5 : ATA_UDMA4, +@@ -531,7 +531,7 @@ static const struct hpt_info hpt370a __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt374 __devinitdata = { ++static const struct hpt_info hpt374 __devinitconst = { + .chip_name = "HPT374", + .chip_type = HPT374, + .udma_mask = ATA_UDMA5, +@@ -539,7 +539,7 @@ static const struct hpt_info hpt374 __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt372 __devinitdata = { ++static const struct hpt_info hpt372 __devinitconst = { + .chip_name = "HPT372", + .chip_type = HPT372, + .udma_mask = HPT372_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -547,7 +547,7 @@ static const struct hpt_info hpt372 __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt372a __devinitdata = { ++static const struct hpt_info hpt372a __devinitconst = { + .chip_name = "HPT372A", + .chip_type = HPT372A, + .udma_mask = HPT372_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -555,7 +555,7 @@ static const struct hpt_info hpt372a __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt302 __devinitdata = { ++static const struct hpt_info hpt302 __devinitconst = { + .chip_name = "HPT302", + .chip_type = HPT302, + .udma_mask = HPT302_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -563,7 +563,7 @@ static const struct hpt_info hpt302 __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt371 __devinitdata = { ++static const struct hpt_info hpt371 __devinitconst = { + .chip_name = "HPT371", + .chip_type = HPT371, + .udma_mask = HPT371_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -571,7 +571,7 @@ static const struct hpt_info hpt371 __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt372n __devinitdata = { ++static const struct hpt_info hpt372n __devinitconst = { + .chip_name = "HPT372N", + .chip_type = HPT372N, + .udma_mask = HPT372_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -579,7 +579,7 @@ static const struct hpt_info hpt372n __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt302n __devinitdata = { ++static const struct hpt_info hpt302n __devinitconst = { + .chip_name = "HPT302N", + .chip_type = HPT302N, + .udma_mask = HPT302_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -587,7 +587,7 @@ static const struct hpt_info hpt302n __devinitdata = { + .timings = &hpt37x_timings + }; + +-static const struct hpt_info hpt371n __devinitdata = { ++static const struct hpt_info hpt371n __devinitconst = { + .chip_name = "HPT371N", + .chip_type = HPT371N, + .udma_mask = HPT371_ALLOW_ATA133_6 ? ATA_UDMA6 : ATA_UDMA5, +@@ -1422,7 +1422,7 @@ static const struct ide_dma_ops hpt36x_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info hpt366_chipsets[] __devinitdata = { ++static const struct ide_port_info hpt366_chipsets[] __devinitconst = { + { /* 0: HPT36x */ + .name = DRV_NAME, + .init_chipset = init_chipset_hpt366, +diff --git a/drivers/ide/ide-cd.c b/drivers/ide/ide-cd.c +index 2de76cc..74186a1 100644 +--- a/drivers/ide/ide-cd.c ++++ b/drivers/ide/ide-cd.c +@@ -774,7 +774,7 @@ static void cdrom_do_block_pc(ide_drive_t *drive, struct request *rq) + alignment = queue_dma_alignment(q) | q->dma_pad_mask; + if ((unsigned long)buf & alignment + || blk_rq_bytes(rq) & q->dma_pad_mask +- || object_is_on_stack(buf)) ++ || object_starts_on_stack(buf)) + drive->dma = 0; + } + } +diff --git a/drivers/ide/ide-floppy.c b/drivers/ide/ide-floppy.c +index fefbdfc..62ff465 100644 +--- a/drivers/ide/ide-floppy.c ++++ b/drivers/ide/ide-floppy.c +@@ -373,6 +373,8 @@ static int ide_floppy_get_capacity(ide_drive_t *drive) + u8 pc_buf[256], header_len, desc_cnt; + int i, rc = 1, blocks, length; + ++ pax_track_stack(); ++ + ide_debug_log(IDE_DBG_FUNC, "enter"); + + drive->bios_cyl = 0; +diff --git a/drivers/ide/ide-pci-generic.c b/drivers/ide/ide-pci-generic.c +index 39d4e01..11538ce 100644 +--- a/drivers/ide/ide-pci-generic.c ++++ b/drivers/ide/ide-pci-generic.c +@@ -53,7 +53,7 @@ static const struct ide_port_ops netcell_port_ops = { + .udma_mask = ATA_UDMA6, \ + } + +-static const struct ide_port_info generic_chipsets[] __devinitdata = { ++static const struct ide_port_info generic_chipsets[] __devinitconst = { + /* 0: Unknown */ + DECLARE_GENERIC_PCI_DEV(0), + +diff --git a/drivers/ide/it8172.c b/drivers/ide/it8172.c +index 0d266a5..aaca790 100644 +--- a/drivers/ide/it8172.c ++++ b/drivers/ide/it8172.c +@@ -115,7 +115,7 @@ static const struct ide_port_ops it8172_port_ops = { + .set_dma_mode = it8172_set_dma_mode, + }; + +-static const struct ide_port_info it8172_port_info __devinitdata = { ++static const struct ide_port_info it8172_port_info __devinitconst = { + .name = DRV_NAME, + .port_ops = &it8172_port_ops, + .enablebits = { {0x41, 0x80, 0x80}, {0x00, 0x00, 0x00} }, +diff --git a/drivers/ide/it8213.c b/drivers/ide/it8213.c +index 4797616..4be488a 100644 +--- a/drivers/ide/it8213.c ++++ b/drivers/ide/it8213.c +@@ -156,7 +156,7 @@ static const struct ide_port_ops it8213_port_ops = { + .cable_detect = it8213_cable_detect, + }; + +-static const struct ide_port_info it8213_chipset __devinitdata = { ++static const struct ide_port_info it8213_chipset __devinitconst = { + .name = DRV_NAME, + .enablebits = { {0x41, 0x80, 0x80} }, + .port_ops = &it8213_port_ops, +diff --git a/drivers/ide/it821x.c b/drivers/ide/it821x.c +index 51aa745..146ee60 100644 +--- a/drivers/ide/it821x.c ++++ b/drivers/ide/it821x.c +@@ -627,7 +627,7 @@ static const struct ide_port_ops it821x_port_ops = { + .cable_detect = it821x_cable_detect, + }; + +-static const struct ide_port_info it821x_chipset __devinitdata = { ++static const struct ide_port_info it821x_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_it821x, + .init_hwif = init_hwif_it821x, +diff --git a/drivers/ide/jmicron.c b/drivers/ide/jmicron.c +index bf2be64..9270098 100644 +--- a/drivers/ide/jmicron.c ++++ b/drivers/ide/jmicron.c +@@ -102,7 +102,7 @@ static const struct ide_port_ops jmicron_port_ops = { + .cable_detect = jmicron_cable_detect, + }; + +-static const struct ide_port_info jmicron_chipset __devinitdata = { ++static const struct ide_port_info jmicron_chipset __devinitconst = { + .name = DRV_NAME, + .enablebits = { { 0x40, 0x01, 0x01 }, { 0x40, 0x10, 0x10 } }, + .port_ops = &jmicron_port_ops, +diff --git a/drivers/ide/ns87415.c b/drivers/ide/ns87415.c +index 95327a2..73f78d8 100644 +--- a/drivers/ide/ns87415.c ++++ b/drivers/ide/ns87415.c +@@ -293,7 +293,7 @@ static const struct ide_dma_ops ns87415_dma_ops = { + .dma_sff_read_status = superio_dma_sff_read_status, + }; + +-static const struct ide_port_info ns87415_chipset __devinitdata = { ++static const struct ide_port_info ns87415_chipset __devinitconst = { + .name = DRV_NAME, + .init_hwif = init_hwif_ns87415, + .tp_ops = &ns87415_tp_ops, +diff --git a/drivers/ide/opti621.c b/drivers/ide/opti621.c +index f1d70d6..e1de05b 100644 +--- a/drivers/ide/opti621.c ++++ b/drivers/ide/opti621.c +@@ -202,7 +202,7 @@ static const struct ide_port_ops opti621_port_ops = { + .set_pio_mode = opti621_set_pio_mode, + }; + +-static const struct ide_port_info opti621_chipset __devinitdata = { ++static const struct ide_port_info opti621_chipset __devinitconst = { + .name = DRV_NAME, + .enablebits = { {0x45, 0x80, 0x00}, {0x40, 0x08, 0x00} }, + .port_ops = &opti621_port_ops, +diff --git a/drivers/ide/pdc202xx_new.c b/drivers/ide/pdc202xx_new.c +index 65ba823..7311f4d 100644 +--- a/drivers/ide/pdc202xx_new.c ++++ b/drivers/ide/pdc202xx_new.c +@@ -465,7 +465,7 @@ static const struct ide_port_ops pdcnew_port_ops = { + .udma_mask = udma, \ + } + +-static const struct ide_port_info pdcnew_chipsets[] __devinitdata = { ++static const struct ide_port_info pdcnew_chipsets[] __devinitconst = { + /* 0: PDC202{68,70} */ DECLARE_PDCNEW_DEV(ATA_UDMA5), + /* 1: PDC202{69,71,75,76,77} */ DECLARE_PDCNEW_DEV(ATA_UDMA6), + }; +diff --git a/drivers/ide/pdc202xx_old.c b/drivers/ide/pdc202xx_old.c +index cb812f3..af816ef 100644 +--- a/drivers/ide/pdc202xx_old.c ++++ b/drivers/ide/pdc202xx_old.c +@@ -285,7 +285,7 @@ static const struct ide_dma_ops pdc2026x_dma_ops = { + .max_sectors = sectors, \ + } + +-static const struct ide_port_info pdc202xx_chipsets[] __devinitdata = { ++static const struct ide_port_info pdc202xx_chipsets[] __devinitconst = { + { /* 0: PDC20246 */ + .name = DRV_NAME, + .init_chipset = init_chipset_pdc202xx, +diff --git a/drivers/ide/piix.c b/drivers/ide/piix.c +index bf14f39..15c4b98 100644 +--- a/drivers/ide/piix.c ++++ b/drivers/ide/piix.c +@@ -344,7 +344,7 @@ static const struct ide_port_ops ich_port_ops = { + .udma_mask = udma, \ + } + +-static const struct ide_port_info piix_pci_info[] __devinitdata = { ++static const struct ide_port_info piix_pci_info[] __devinitconst = { + /* 0: MPIIX */ + { /* + * MPIIX actually has only a single IDE channel mapped to +diff --git a/drivers/ide/rz1000.c b/drivers/ide/rz1000.c +index a6414a8..c04173e 100644 +--- a/drivers/ide/rz1000.c ++++ b/drivers/ide/rz1000.c +@@ -38,7 +38,7 @@ static int __devinit rz1000_disable_readahead(struct pci_dev *dev) + } + } + +-static const struct ide_port_info rz1000_chipset __devinitdata = { ++static const struct ide_port_info rz1000_chipset __devinitconst = { + .name = DRV_NAME, + .host_flags = IDE_HFLAG_NO_DMA, + }; +diff --git a/drivers/ide/sc1200.c b/drivers/ide/sc1200.c +index d467478..9203942 100644 +--- a/drivers/ide/sc1200.c ++++ b/drivers/ide/sc1200.c +@@ -290,7 +290,7 @@ static const struct ide_dma_ops sc1200_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info sc1200_chipset __devinitdata = { ++static const struct ide_port_info sc1200_chipset __devinitconst = { + .name = DRV_NAME, + .port_ops = &sc1200_port_ops, + .dma_ops = &sc1200_dma_ops, +diff --git a/drivers/ide/scc_pata.c b/drivers/ide/scc_pata.c +index 1104bb3..59c5194 100644 +--- a/drivers/ide/scc_pata.c ++++ b/drivers/ide/scc_pata.c +@@ -811,7 +811,7 @@ static const struct ide_dma_ops scc_dma_ops = { + .dma_sff_read_status = scc_dma_sff_read_status, + }; + +-static const struct ide_port_info scc_chipset __devinitdata = { ++static const struct ide_port_info scc_chipset __devinitconst = { + .name = "sccIDE", + .init_iops = init_iops_scc, + .init_dma = scc_init_dma, +diff --git a/drivers/ide/serverworks.c b/drivers/ide/serverworks.c +index b6554ef..6cc2cc3 100644 +--- a/drivers/ide/serverworks.c ++++ b/drivers/ide/serverworks.c +@@ -353,7 +353,7 @@ static const struct ide_port_ops svwks_port_ops = { + .cable_detect = svwks_cable_detect, + }; + +-static const struct ide_port_info serverworks_chipsets[] __devinitdata = { ++static const struct ide_port_info serverworks_chipsets[] __devinitconst = { + { /* 0: OSB4 */ + .name = DRV_NAME, + .init_chipset = init_chipset_svwks, +diff --git a/drivers/ide/setup-pci.c b/drivers/ide/setup-pci.c +index ab3db61..afed580 100644 +--- a/drivers/ide/setup-pci.c ++++ b/drivers/ide/setup-pci.c +@@ -542,6 +542,8 @@ int ide_pci_init_two(struct pci_dev *dev1, struct pci_dev *dev2, + int ret, i, n_ports = dev2 ? 4 : 2; + struct ide_hw hw[4], *hws[] = { NULL, NULL, NULL, NULL }; + ++ pax_track_stack(); ++ + for (i = 0; i < n_ports / 2; i++) { + ret = ide_setup_pci_controller(pdev[i], d, !i); + if (ret < 0) +diff --git a/drivers/ide/siimage.c b/drivers/ide/siimage.c +index d95df52..0b03a39 100644 +--- a/drivers/ide/siimage.c ++++ b/drivers/ide/siimage.c +@@ -719,7 +719,7 @@ static const struct ide_dma_ops sil_dma_ops = { + .udma_mask = ATA_UDMA6, \ + } + +-static const struct ide_port_info siimage_chipsets[] __devinitdata = { ++static const struct ide_port_info siimage_chipsets[] __devinitconst = { + /* 0: SiI680 */ DECLARE_SII_DEV(&sil_pata_port_ops), + /* 1: SiI3112 */ DECLARE_SII_DEV(&sil_sata_port_ops) + }; +diff --git a/drivers/ide/sis5513.c b/drivers/ide/sis5513.c +index 3b88eba..ca8699d 100644 +--- a/drivers/ide/sis5513.c ++++ b/drivers/ide/sis5513.c +@@ -561,7 +561,7 @@ static const struct ide_port_ops sis_ata133_port_ops = { + .cable_detect = sis_cable_detect, + }; + +-static const struct ide_port_info sis5513_chipset __devinitdata = { ++static const struct ide_port_info sis5513_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_sis5513, + .enablebits = { {0x4a, 0x02, 0x02}, {0x4a, 0x04, 0x04} }, +diff --git a/drivers/ide/sl82c105.c b/drivers/ide/sl82c105.c +index d698da4..fca42a4 100644 +--- a/drivers/ide/sl82c105.c ++++ b/drivers/ide/sl82c105.c +@@ -319,7 +319,7 @@ static const struct ide_dma_ops sl82c105_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info sl82c105_chipset __devinitdata = { ++static const struct ide_port_info sl82c105_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_sl82c105, + .enablebits = {{0x40,0x01,0x01}, {0x40,0x10,0x10}}, +diff --git a/drivers/ide/slc90e66.c b/drivers/ide/slc90e66.c +index 1ccfb40..83d5779 100644 +--- a/drivers/ide/slc90e66.c ++++ b/drivers/ide/slc90e66.c +@@ -131,7 +131,7 @@ static const struct ide_port_ops slc90e66_port_ops = { + .cable_detect = slc90e66_cable_detect, + }; + +-static const struct ide_port_info slc90e66_chipset __devinitdata = { ++static const struct ide_port_info slc90e66_chipset __devinitconst = { + .name = DRV_NAME, + .enablebits = { {0x41, 0x80, 0x80}, {0x43, 0x80, 0x80} }, + .port_ops = &slc90e66_port_ops, +diff --git a/drivers/ide/tc86c001.c b/drivers/ide/tc86c001.c +index 05a93d6..5f9e325 100644 +--- a/drivers/ide/tc86c001.c ++++ b/drivers/ide/tc86c001.c +@@ -190,7 +190,7 @@ static const struct ide_dma_ops tc86c001_dma_ops = { + .dma_sff_read_status = ide_dma_sff_read_status, + }; + +-static const struct ide_port_info tc86c001_chipset __devinitdata = { ++static const struct ide_port_info tc86c001_chipset __devinitconst = { + .name = DRV_NAME, + .init_hwif = init_hwif_tc86c001, + .port_ops = &tc86c001_port_ops, +diff --git a/drivers/ide/triflex.c b/drivers/ide/triflex.c +index 8773c3b..7907d6c 100644 +--- a/drivers/ide/triflex.c ++++ b/drivers/ide/triflex.c +@@ -92,7 +92,7 @@ static const struct ide_port_ops triflex_port_ops = { + .set_dma_mode = triflex_set_mode, + }; + +-static const struct ide_port_info triflex_device __devinitdata = { ++static const struct ide_port_info triflex_device __devinitconst = { + .name = DRV_NAME, + .enablebits = {{0x80, 0x01, 0x01}, {0x80, 0x02, 0x02}}, + .port_ops = &triflex_port_ops, +diff --git a/drivers/ide/trm290.c b/drivers/ide/trm290.c +index 4b42ca0..e494a98 100644 +--- a/drivers/ide/trm290.c ++++ b/drivers/ide/trm290.c +@@ -324,7 +324,7 @@ static struct ide_dma_ops trm290_dma_ops = { + .dma_check = trm290_dma_check, + }; + +-static const struct ide_port_info trm290_chipset __devinitdata = { ++static const struct ide_port_info trm290_chipset __devinitconst = { + .name = DRV_NAME, + .init_hwif = init_hwif_trm290, + .tp_ops = &trm290_tp_ops, +diff --git a/drivers/ide/via82cxxx.c b/drivers/ide/via82cxxx.c +index 028de26..520d5d5 100644 +--- a/drivers/ide/via82cxxx.c ++++ b/drivers/ide/via82cxxx.c +@@ -374,7 +374,7 @@ static const struct ide_port_ops via_port_ops = { + .cable_detect = via82cxxx_cable_detect, + }; + +-static const struct ide_port_info via82cxxx_chipset __devinitdata = { ++static const struct ide_port_info via82cxxx_chipset __devinitconst = { + .name = DRV_NAME, + .init_chipset = init_chipset_via82cxxx, + .enablebits = { { 0x40, 0x02, 0x02 }, { 0x40, 0x01, 0x01 } }, +diff --git a/drivers/ieee1394/dv1394.c b/drivers/ieee1394/dv1394.c +index 2cd00b5..14de699 100644 +--- a/drivers/ieee1394/dv1394.c ++++ b/drivers/ieee1394/dv1394.c +@@ -739,7 +739,7 @@ static void frame_prepare(struct video_card *video, unsigned int this_frame) + based upon DIF section and sequence + */ + +-static void inline ++static inline void + frame_put_packet (struct frame *f, struct packet *p) + { + int section_type = p->data[0] >> 5; /* section type is in bits 5 - 7 */ +diff --git a/drivers/ieee1394/hosts.c b/drivers/ieee1394/hosts.c +index e947d8f..6a966b9 100644 +--- a/drivers/ieee1394/hosts.c ++++ b/drivers/ieee1394/hosts.c +@@ -78,6 +78,7 @@ static int dummy_isoctl(struct hpsb_iso *iso, enum isoctl_cmd command, + } + + static struct hpsb_host_driver dummy_driver = { ++ .name = "dummy", + .transmit_packet = dummy_transmit_packet, + .devctl = dummy_devctl, + .isoctl = dummy_isoctl +diff --git a/drivers/ieee1394/init_ohci1394_dma.c b/drivers/ieee1394/init_ohci1394_dma.c +index ddaab6e..8d37435 100644 +--- a/drivers/ieee1394/init_ohci1394_dma.c ++++ b/drivers/ieee1394/init_ohci1394_dma.c +@@ -257,7 +257,7 @@ void __init init_ohci1394_dma_on_all_controllers(void) + for (func = 0; func < 8; func++) { + u32 class = read_pci_config(num,slot,func, + PCI_CLASS_REVISION); +- if ((class == 0xffffffff)) ++ if (class == 0xffffffff) + continue; /* No device at this func */ + + if (class>>8 != PCI_CLASS_SERIAL_FIREWIRE_OHCI) +diff --git a/drivers/ieee1394/ohci1394.c b/drivers/ieee1394/ohci1394.c +index 65c1429..5d8c11f 100644 +--- a/drivers/ieee1394/ohci1394.c ++++ b/drivers/ieee1394/ohci1394.c +@@ -147,9 +147,9 @@ printk(level "%s: " fmt "\n" , OHCI1394_DRIVER_NAME , ## args) + printk(level "%s: fw-host%d: " fmt "\n" , OHCI1394_DRIVER_NAME, ohci->host->id , ## args) + + /* Module Parameters */ +-static int phys_dma = 1; ++static int phys_dma; + module_param(phys_dma, int, 0444); +-MODULE_PARM_DESC(phys_dma, "Enable physical DMA (default = 1)."); ++MODULE_PARM_DESC(phys_dma, "Enable physical DMA (default = 0)."); + + static void dma_trm_tasklet(unsigned long data); + static void dma_trm_reset(struct dma_trm_ctx *d); +diff --git a/drivers/ieee1394/sbp2.c b/drivers/ieee1394/sbp2.c +index f199896..78c9fc8 100644 +--- a/drivers/ieee1394/sbp2.c ++++ b/drivers/ieee1394/sbp2.c +@@ -2111,7 +2111,7 @@ MODULE_DESCRIPTION("IEEE-1394 SBP-2 protocol driver"); + MODULE_SUPPORTED_DEVICE(SBP2_DEVICE_NAME); + MODULE_LICENSE("GPL"); + +-static int sbp2_module_init(void) ++static int __init sbp2_module_init(void) + { + int ret; + +diff --git a/drivers/infiniband/core/cm.c b/drivers/infiniband/core/cm.c +index a5dea6b..0cefe8f 100644 +--- a/drivers/infiniband/core/cm.c ++++ b/drivers/infiniband/core/cm.c +@@ -112,7 +112,7 @@ static char const counter_group_names[CM_COUNTER_GROUPS] + + struct cm_counter_group { + struct kobject obj; +- atomic_long_t counter[CM_ATTR_COUNT]; ++ atomic_long_unchecked_t counter[CM_ATTR_COUNT]; + }; + + struct cm_counter_attribute { +@@ -1386,7 +1386,7 @@ static void cm_dup_req_handler(struct cm_work *work, + struct ib_mad_send_buf *msg = NULL; + int ret; + +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_REQ_COUNTER]); + + /* Quick state check to discard duplicate REQs. */ +@@ -1764,7 +1764,7 @@ static void cm_dup_rep_handler(struct cm_work *work) + if (!cm_id_priv) + return; + +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_REP_COUNTER]); + ret = cm_alloc_response_msg(work->port, work->mad_recv_wc, &msg); + if (ret) +@@ -1931,7 +1931,7 @@ static int cm_rtu_handler(struct cm_work *work) + if (cm_id_priv->id.state != IB_CM_REP_SENT && + cm_id_priv->id.state != IB_CM_MRA_REP_RCVD) { + spin_unlock_irq(&cm_id_priv->lock); +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_RTU_COUNTER]); + goto out; + } +@@ -2110,7 +2110,7 @@ static int cm_dreq_handler(struct cm_work *work) + cm_id_priv = cm_acquire_id(dreq_msg->remote_comm_id, + dreq_msg->local_comm_id); + if (!cm_id_priv) { +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_DREQ_COUNTER]); + cm_issue_drep(work->port, work->mad_recv_wc); + return -EINVAL; +@@ -2131,7 +2131,7 @@ static int cm_dreq_handler(struct cm_work *work) + case IB_CM_MRA_REP_RCVD: + break; + case IB_CM_TIMEWAIT: +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_DREQ_COUNTER]); + if (cm_alloc_response_msg(work->port, work->mad_recv_wc, &msg)) + goto unlock; +@@ -2145,7 +2145,7 @@ static int cm_dreq_handler(struct cm_work *work) + cm_free_msg(msg); + goto deref; + case IB_CM_DREQ_RCVD: +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_DREQ_COUNTER]); + goto unlock; + default: +@@ -2501,7 +2501,7 @@ static int cm_mra_handler(struct cm_work *work) + ib_modify_mad(cm_id_priv->av.port->mad_agent, + cm_id_priv->msg, timeout)) { + if (cm_id_priv->id.lap_state == IB_CM_MRA_LAP_RCVD) +- atomic_long_inc(&work->port-> ++ atomic_long_inc_unchecked(&work->port-> + counter_group[CM_RECV_DUPLICATES]. + counter[CM_MRA_COUNTER]); + goto out; +@@ -2510,7 +2510,7 @@ static int cm_mra_handler(struct cm_work *work) + break; + case IB_CM_MRA_REQ_RCVD: + case IB_CM_MRA_REP_RCVD: +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_MRA_COUNTER]); + /* fall through */ + default: +@@ -2672,7 +2672,7 @@ static int cm_lap_handler(struct cm_work *work) + case IB_CM_LAP_IDLE: + break; + case IB_CM_MRA_LAP_SENT: +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_LAP_COUNTER]); + if (cm_alloc_response_msg(work->port, work->mad_recv_wc, &msg)) + goto unlock; +@@ -2688,7 +2688,7 @@ static int cm_lap_handler(struct cm_work *work) + cm_free_msg(msg); + goto deref; + case IB_CM_LAP_RCVD: +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_LAP_COUNTER]); + goto unlock; + default: +@@ -2972,7 +2972,7 @@ static int cm_sidr_req_handler(struct cm_work *work) + cur_cm_id_priv = cm_insert_remote_sidr(cm_id_priv); + if (cur_cm_id_priv) { + spin_unlock_irq(&cm.lock); +- atomic_long_inc(&work->port->counter_group[CM_RECV_DUPLICATES]. ++ atomic_long_inc_unchecked(&work->port->counter_group[CM_RECV_DUPLICATES]. + counter[CM_SIDR_REQ_COUNTER]); + goto out; /* Duplicate message. */ + } +@@ -3184,10 +3184,10 @@ static void cm_send_handler(struct ib_mad_agent *mad_agent, + if (!msg->context[0] && (attr_index != CM_REJ_COUNTER)) + msg->retries = 1; + +- atomic_long_add(1 + msg->retries, ++ atomic_long_add_unchecked(1 + msg->retries, + &port->counter_group[CM_XMIT].counter[attr_index]); + if (msg->retries) +- atomic_long_add(msg->retries, ++ atomic_long_add_unchecked(msg->retries, + &port->counter_group[CM_XMIT_RETRIES]. + counter[attr_index]); + +@@ -3397,7 +3397,7 @@ static void cm_recv_handler(struct ib_mad_agent *mad_agent, + } + + attr_id = be16_to_cpu(mad_recv_wc->recv_buf.mad->mad_hdr.attr_id); +- atomic_long_inc(&port->counter_group[CM_RECV]. ++ atomic_long_inc_unchecked(&port->counter_group[CM_RECV]. + counter[attr_id - CM_ATTR_ID_OFFSET]); + + work = kmalloc(sizeof *work + sizeof(struct ib_sa_path_rec) * paths, +@@ -3595,10 +3595,10 @@ static ssize_t cm_show_counter(struct kobject *obj, struct attribute *attr, + cm_attr = container_of(attr, struct cm_counter_attribute, attr); + + return sprintf(buf, "%ld\n", +- atomic_long_read(&group->counter[cm_attr->index])); ++ atomic_long_read_unchecked(&group->counter[cm_attr->index])); + } + +-static struct sysfs_ops cm_counter_ops = { ++static const struct sysfs_ops cm_counter_ops = { + .show = cm_show_counter + }; + +diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c +index 8fd3a6f..61d8075 100644 +--- a/drivers/infiniband/core/cma.c ++++ b/drivers/infiniband/core/cma.c +@@ -2267,6 +2267,9 @@ static int cma_resolve_ib_udp(struct rdma_id_private *id_priv, + + req.private_data_len = sizeof(struct cma_hdr) + + conn_param->private_data_len; ++ if (req.private_data_len < conn_param->private_data_len) ++ return -EINVAL; ++ + req.private_data = kzalloc(req.private_data_len, GFP_ATOMIC); + if (!req.private_data) + return -ENOMEM; +@@ -2314,6 +2317,9 @@ static int cma_connect_ib(struct rdma_id_private *id_priv, + memset(&req, 0, sizeof req); + offset = cma_user_data_offset(id_priv->id.ps); + req.private_data_len = offset + conn_param->private_data_len; ++ if (req.private_data_len < conn_param->private_data_len) ++ return -EINVAL; ++ + private_data = kzalloc(req.private_data_len, GFP_ATOMIC); + if (!private_data) + return -ENOMEM; +diff --git a/drivers/infiniband/core/fmr_pool.c b/drivers/infiniband/core/fmr_pool.c +index 4507043..14ad522 100644 +--- a/drivers/infiniband/core/fmr_pool.c ++++ b/drivers/infiniband/core/fmr_pool.c +@@ -97,8 +97,8 @@ struct ib_fmr_pool { + + struct task_struct *thread; + +- atomic_t req_ser; +- atomic_t flush_ser; ++ atomic_unchecked_t req_ser; ++ atomic_unchecked_t flush_ser; + + wait_queue_head_t force_wait; + }; +@@ -179,10 +179,10 @@ static int ib_fmr_cleanup_thread(void *pool_ptr) + struct ib_fmr_pool *pool = pool_ptr; + + do { +- if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) < 0) { ++ if (atomic_read_unchecked(&pool->flush_ser) - atomic_read_unchecked(&pool->req_ser) < 0) { + ib_fmr_batch_release(pool); + +- atomic_inc(&pool->flush_ser); ++ atomic_inc_unchecked(&pool->flush_ser); + wake_up_interruptible(&pool->force_wait); + + if (pool->flush_function) +@@ -190,7 +190,7 @@ static int ib_fmr_cleanup_thread(void *pool_ptr) + } + + set_current_state(TASK_INTERRUPTIBLE); +- if (atomic_read(&pool->flush_ser) - atomic_read(&pool->req_ser) >= 0 && ++ if (atomic_read_unchecked(&pool->flush_ser) - atomic_read_unchecked(&pool->req_ser) >= 0 && + !kthread_should_stop()) + schedule(); + __set_current_state(TASK_RUNNING); +@@ -282,8 +282,8 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd, + pool->dirty_watermark = params->dirty_watermark; + pool->dirty_len = 0; + spin_lock_init(&pool->pool_lock); +- atomic_set(&pool->req_ser, 0); +- atomic_set(&pool->flush_ser, 0); ++ atomic_set_unchecked(&pool->req_ser, 0); ++ atomic_set_unchecked(&pool->flush_ser, 0); + init_waitqueue_head(&pool->force_wait); + + pool->thread = kthread_run(ib_fmr_cleanup_thread, +@@ -411,11 +411,11 @@ int ib_flush_fmr_pool(struct ib_fmr_pool *pool) + } + spin_unlock_irq(&pool->pool_lock); + +- serial = atomic_inc_return(&pool->req_ser); ++ serial = atomic_inc_return_unchecked(&pool->req_ser); + wake_up_process(pool->thread); + + if (wait_event_interruptible(pool->force_wait, +- atomic_read(&pool->flush_ser) - serial >= 0)) ++ atomic_read_unchecked(&pool->flush_ser) - serial >= 0)) + return -EINTR; + + return 0; +@@ -525,7 +525,7 @@ int ib_fmr_pool_unmap(struct ib_pool_fmr *fmr) + } else { + list_add_tail(&fmr->list, &pool->dirty_list); + if (++pool->dirty_len >= pool->dirty_watermark) { +- atomic_inc(&pool->req_ser); ++ atomic_inc_unchecked(&pool->req_ser); + wake_up_process(pool->thread); + } + } +diff --git a/drivers/infiniband/core/sysfs.c b/drivers/infiniband/core/sysfs.c +index 158a214..1558bb7 100644 +--- a/drivers/infiniband/core/sysfs.c ++++ b/drivers/infiniband/core/sysfs.c +@@ -79,7 +79,7 @@ static ssize_t port_attr_show(struct kobject *kobj, + return port_attr->show(p, port_attr, buf); + } + +-static struct sysfs_ops port_sysfs_ops = { ++static const struct sysfs_ops port_sysfs_ops = { + .show = port_attr_show + }; + +diff --git a/drivers/infiniband/core/uverbs_marshall.c b/drivers/infiniband/core/uverbs_marshall.c +index 5440da0..1194ecb 100644 +--- a/drivers/infiniband/core/uverbs_marshall.c ++++ b/drivers/infiniband/core/uverbs_marshall.c +@@ -40,18 +40,21 @@ void ib_copy_ah_attr_to_user(struct ib_uverbs_ah_attr *dst, + dst->grh.sgid_index = src->grh.sgid_index; + dst->grh.hop_limit = src->grh.hop_limit; + dst->grh.traffic_class = src->grh.traffic_class; ++ memset(&dst->grh.reserved, 0, sizeof(dst->grh.reserved)); + dst->dlid = src->dlid; + dst->sl = src->sl; + dst->src_path_bits = src->src_path_bits; + dst->static_rate = src->static_rate; + dst->is_global = src->ah_flags & IB_AH_GRH ? 1 : 0; + dst->port_num = src->port_num; ++ dst->reserved = 0; + } + EXPORT_SYMBOL(ib_copy_ah_attr_to_user); + + void ib_copy_qp_attr_to_user(struct ib_uverbs_qp_attr *dst, + struct ib_qp_attr *src) + { ++ dst->qp_state = src->qp_state; + dst->cur_qp_state = src->cur_qp_state; + dst->path_mtu = src->path_mtu; + dst->path_mig_state = src->path_mig_state; +@@ -83,6 +86,7 @@ void ib_copy_qp_attr_to_user(struct ib_uverbs_qp_attr *dst, + dst->rnr_retry = src->rnr_retry; + dst->alt_port_num = src->alt_port_num; + dst->alt_timeout = src->alt_timeout; ++ memset(dst->reserved, 0, sizeof(dst->reserved)); + } + EXPORT_SYMBOL(ib_copy_qp_attr_to_user); + +diff --git a/drivers/infiniband/hw/ipath/ipath_fs.c b/drivers/infiniband/hw/ipath/ipath_fs.c +index 100da85..62e6b88 100644 +--- a/drivers/infiniband/hw/ipath/ipath_fs.c ++++ b/drivers/infiniband/hw/ipath/ipath_fs.c +@@ -110,6 +110,8 @@ static ssize_t atomic_counters_read(struct file *file, char __user *buf, + struct infinipath_counters counters; + struct ipath_devdata *dd; + ++ pax_track_stack(); ++ + dd = file->f_path.dentry->d_inode->i_private; + dd->ipath_f_read_counters(dd, &counters); + +diff --git a/drivers/infiniband/hw/nes/nes.c b/drivers/infiniband/hw/nes/nes.c +index cbde0cf..afaf55c 100644 +--- a/drivers/infiniband/hw/nes/nes.c ++++ b/drivers/infiniband/hw/nes/nes.c +@@ -102,7 +102,7 @@ MODULE_PARM_DESC(limit_maxrdreqsz, "Limit max read request size to 256 Bytes"); + LIST_HEAD(nes_adapter_list); + static LIST_HEAD(nes_dev_list); + +-atomic_t qps_destroyed; ++atomic_unchecked_t qps_destroyed; + + static unsigned int ee_flsh_adapter; + static unsigned int sysfs_nonidx_addr; +@@ -259,7 +259,7 @@ static void nes_cqp_rem_ref_callback(struct nes_device *nesdev, struct nes_cqp_r + struct nes_adapter *nesadapter = nesdev->nesadapter; + u32 qp_id; + +- atomic_inc(&qps_destroyed); ++ atomic_inc_unchecked(&qps_destroyed); + + /* Free the control structures */ + +diff --git a/drivers/infiniband/hw/nes/nes.h b/drivers/infiniband/hw/nes/nes.h +index bcc6abc..9c76b2f 100644 +--- a/drivers/infiniband/hw/nes/nes.h ++++ b/drivers/infiniband/hw/nes/nes.h +@@ -174,17 +174,17 @@ extern unsigned int nes_debug_level; + extern unsigned int wqm_quanta; + extern struct list_head nes_adapter_list; + +-extern atomic_t cm_connects; +-extern atomic_t cm_accepts; +-extern atomic_t cm_disconnects; +-extern atomic_t cm_closes; +-extern atomic_t cm_connecteds; +-extern atomic_t cm_connect_reqs; +-extern atomic_t cm_rejects; +-extern atomic_t mod_qp_timouts; +-extern atomic_t qps_created; +-extern atomic_t qps_destroyed; +-extern atomic_t sw_qps_destroyed; ++extern atomic_unchecked_t cm_connects; ++extern atomic_unchecked_t cm_accepts; ++extern atomic_unchecked_t cm_disconnects; ++extern atomic_unchecked_t cm_closes; ++extern atomic_unchecked_t cm_connecteds; ++extern atomic_unchecked_t cm_connect_reqs; ++extern atomic_unchecked_t cm_rejects; ++extern atomic_unchecked_t mod_qp_timouts; ++extern atomic_unchecked_t qps_created; ++extern atomic_unchecked_t qps_destroyed; ++extern atomic_unchecked_t sw_qps_destroyed; + extern u32 mh_detected; + extern u32 mh_pauses_sent; + extern u32 cm_packets_sent; +@@ -196,11 +196,11 @@ extern u32 cm_packets_retrans; + extern u32 cm_listens_created; + extern u32 cm_listens_destroyed; + extern u32 cm_backlog_drops; +-extern atomic_t cm_loopbacks; +-extern atomic_t cm_nodes_created; +-extern atomic_t cm_nodes_destroyed; +-extern atomic_t cm_accel_dropped_pkts; +-extern atomic_t cm_resets_recvd; ++extern atomic_unchecked_t cm_loopbacks; ++extern atomic_unchecked_t cm_nodes_created; ++extern atomic_unchecked_t cm_nodes_destroyed; ++extern atomic_unchecked_t cm_accel_dropped_pkts; ++extern atomic_unchecked_t cm_resets_recvd; + + extern u32 int_mod_timer_init; + extern u32 int_mod_cq_depth_256; +diff --git a/drivers/infiniband/hw/nes/nes_cm.c b/drivers/infiniband/hw/nes/nes_cm.c +index 73473db..5ed06e8 100644 +--- a/drivers/infiniband/hw/nes/nes_cm.c ++++ b/drivers/infiniband/hw/nes/nes_cm.c +@@ -69,11 +69,11 @@ u32 cm_packets_received; + u32 cm_listens_created; + u32 cm_listens_destroyed; + u32 cm_backlog_drops; +-atomic_t cm_loopbacks; +-atomic_t cm_nodes_created; +-atomic_t cm_nodes_destroyed; +-atomic_t cm_accel_dropped_pkts; +-atomic_t cm_resets_recvd; ++atomic_unchecked_t cm_loopbacks; ++atomic_unchecked_t cm_nodes_created; ++atomic_unchecked_t cm_nodes_destroyed; ++atomic_unchecked_t cm_accel_dropped_pkts; ++atomic_unchecked_t cm_resets_recvd; + + static inline int mini_cm_accelerated(struct nes_cm_core *, + struct nes_cm_node *); +@@ -149,13 +149,13 @@ static struct nes_cm_ops nes_cm_api = { + + static struct nes_cm_core *g_cm_core; + +-atomic_t cm_connects; +-atomic_t cm_accepts; +-atomic_t cm_disconnects; +-atomic_t cm_closes; +-atomic_t cm_connecteds; +-atomic_t cm_connect_reqs; +-atomic_t cm_rejects; ++atomic_unchecked_t cm_connects; ++atomic_unchecked_t cm_accepts; ++atomic_unchecked_t cm_disconnects; ++atomic_unchecked_t cm_closes; ++atomic_unchecked_t cm_connecteds; ++atomic_unchecked_t cm_connect_reqs; ++atomic_unchecked_t cm_rejects; + + + /** +@@ -1195,7 +1195,7 @@ static struct nes_cm_node *make_cm_node(struct nes_cm_core *cm_core, + cm_node->rem_mac); + + add_hte_node(cm_core, cm_node); +- atomic_inc(&cm_nodes_created); ++ atomic_inc_unchecked(&cm_nodes_created); + + return cm_node; + } +@@ -1253,7 +1253,7 @@ static int rem_ref_cm_node(struct nes_cm_core *cm_core, + } + + atomic_dec(&cm_core->node_cnt); +- atomic_inc(&cm_nodes_destroyed); ++ atomic_inc_unchecked(&cm_nodes_destroyed); + nesqp = cm_node->nesqp; + if (nesqp) { + nesqp->cm_node = NULL; +@@ -1320,7 +1320,7 @@ static int process_options(struct nes_cm_node *cm_node, u8 *optionsloc, + + static void drop_packet(struct sk_buff *skb) + { +- atomic_inc(&cm_accel_dropped_pkts); ++ atomic_inc_unchecked(&cm_accel_dropped_pkts); + dev_kfree_skb_any(skb); + } + +@@ -1377,7 +1377,7 @@ static void handle_rst_pkt(struct nes_cm_node *cm_node, struct sk_buff *skb, + + int reset = 0; /* whether to send reset in case of err.. */ + int passive_state; +- atomic_inc(&cm_resets_recvd); ++ atomic_inc_unchecked(&cm_resets_recvd); + nes_debug(NES_DBG_CM, "Received Reset, cm_node = %p, state = %u." + " refcnt=%d\n", cm_node, cm_node->state, + atomic_read(&cm_node->ref_count)); +@@ -2000,7 +2000,7 @@ static struct nes_cm_node *mini_cm_connect(struct nes_cm_core *cm_core, + rem_ref_cm_node(cm_node->cm_core, cm_node); + return NULL; + } +- atomic_inc(&cm_loopbacks); ++ atomic_inc_unchecked(&cm_loopbacks); + loopbackremotenode->loopbackpartner = cm_node; + loopbackremotenode->tcp_cntxt.rcv_wscale = + NES_CM_DEFAULT_RCV_WND_SCALE; +@@ -2262,7 +2262,7 @@ static int mini_cm_recv_pkt(struct nes_cm_core *cm_core, + add_ref_cm_node(cm_node); + } else if (cm_node->state == NES_CM_STATE_TSA) { + rem_ref_cm_node(cm_core, cm_node); +- atomic_inc(&cm_accel_dropped_pkts); ++ atomic_inc_unchecked(&cm_accel_dropped_pkts); + dev_kfree_skb_any(skb); + break; + } +@@ -2568,7 +2568,7 @@ static int nes_cm_disconn_true(struct nes_qp *nesqp) + + if ((cm_id) && (cm_id->event_handler)) { + if (issue_disconn) { +- atomic_inc(&cm_disconnects); ++ atomic_inc_unchecked(&cm_disconnects); + cm_event.event = IW_CM_EVENT_DISCONNECT; + cm_event.status = disconn_status; + cm_event.local_addr = cm_id->local_addr; +@@ -2590,7 +2590,7 @@ static int nes_cm_disconn_true(struct nes_qp *nesqp) + } + + if (issue_close) { +- atomic_inc(&cm_closes); ++ atomic_inc_unchecked(&cm_closes); + nes_disconnect(nesqp, 1); + + cm_id->provider_data = nesqp; +@@ -2710,7 +2710,7 @@ int nes_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) + + nes_debug(NES_DBG_CM, "QP%u, cm_node=%p, jiffies = %lu listener = %p\n", + nesqp->hwqp.qp_id, cm_node, jiffies, cm_node->listener); +- atomic_inc(&cm_accepts); ++ atomic_inc_unchecked(&cm_accepts); + + nes_debug(NES_DBG_CM, "netdev refcnt = %u.\n", + atomic_read(&nesvnic->netdev->refcnt)); +@@ -2919,7 +2919,7 @@ int nes_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) + + struct nes_cm_core *cm_core; + +- atomic_inc(&cm_rejects); ++ atomic_inc_unchecked(&cm_rejects); + cm_node = (struct nes_cm_node *) cm_id->provider_data; + loopback = cm_node->loopbackpartner; + cm_core = cm_node->cm_core; +@@ -2982,7 +2982,7 @@ int nes_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) + ntohl(cm_id->local_addr.sin_addr.s_addr), + ntohs(cm_id->local_addr.sin_port)); + +- atomic_inc(&cm_connects); ++ atomic_inc_unchecked(&cm_connects); + nesqp->active_conn = 1; + + /* cache the cm_id in the qp */ +@@ -3195,7 +3195,7 @@ static void cm_event_connected(struct nes_cm_event *event) + if (nesqp->destroyed) { + return; + } +- atomic_inc(&cm_connecteds); ++ atomic_inc_unchecked(&cm_connecteds); + nes_debug(NES_DBG_CM, "QP%u attempting to connect to 0x%08X:0x%04X on" + " local port 0x%04X. jiffies = %lu.\n", + nesqp->hwqp.qp_id, +@@ -3403,7 +3403,7 @@ static void cm_event_reset(struct nes_cm_event *event) + + ret = cm_id->event_handler(cm_id, &cm_event); + cm_id->add_ref(cm_id); +- atomic_inc(&cm_closes); ++ atomic_inc_unchecked(&cm_closes); + cm_event.event = IW_CM_EVENT_CLOSE; + cm_event.status = IW_CM_EVENT_STATUS_OK; + cm_event.provider_data = cm_id->provider_data; +@@ -3439,7 +3439,7 @@ static void cm_event_mpa_req(struct nes_cm_event *event) + return; + cm_id = cm_node->cm_id; + +- atomic_inc(&cm_connect_reqs); ++ atomic_inc_unchecked(&cm_connect_reqs); + nes_debug(NES_DBG_CM, "cm_node = %p - cm_id = %p, jiffies = %lu\n", + cm_node, cm_id, jiffies); + +@@ -3477,7 +3477,7 @@ static void cm_event_mpa_reject(struct nes_cm_event *event) + return; + cm_id = cm_node->cm_id; + +- atomic_inc(&cm_connect_reqs); ++ atomic_inc_unchecked(&cm_connect_reqs); + nes_debug(NES_DBG_CM, "cm_node = %p - cm_id = %p, jiffies = %lu\n", + cm_node, cm_id, jiffies); + +diff --git a/drivers/infiniband/hw/nes/nes_nic.c b/drivers/infiniband/hw/nes/nes_nic.c +index e593af3..870694a 100644 +--- a/drivers/infiniband/hw/nes/nes_nic.c ++++ b/drivers/infiniband/hw/nes/nes_nic.c +@@ -1210,17 +1210,17 @@ static void nes_netdev_get_ethtool_stats(struct net_device *netdev, + target_stat_values[++index] = mh_detected; + target_stat_values[++index] = mh_pauses_sent; + target_stat_values[++index] = nesvnic->endnode_ipv4_tcp_retransmits; +- target_stat_values[++index] = atomic_read(&cm_connects); +- target_stat_values[++index] = atomic_read(&cm_accepts); +- target_stat_values[++index] = atomic_read(&cm_disconnects); +- target_stat_values[++index] = atomic_read(&cm_connecteds); +- target_stat_values[++index] = atomic_read(&cm_connect_reqs); +- target_stat_values[++index] = atomic_read(&cm_rejects); +- target_stat_values[++index] = atomic_read(&mod_qp_timouts); +- target_stat_values[++index] = atomic_read(&qps_created); +- target_stat_values[++index] = atomic_read(&sw_qps_destroyed); +- target_stat_values[++index] = atomic_read(&qps_destroyed); +- target_stat_values[++index] = atomic_read(&cm_closes); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_connects); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_accepts); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_disconnects); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_connecteds); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_connect_reqs); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_rejects); ++ target_stat_values[++index] = atomic_read_unchecked(&mod_qp_timouts); ++ target_stat_values[++index] = atomic_read_unchecked(&qps_created); ++ target_stat_values[++index] = atomic_read_unchecked(&sw_qps_destroyed); ++ target_stat_values[++index] = atomic_read_unchecked(&qps_destroyed); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_closes); + target_stat_values[++index] = cm_packets_sent; + target_stat_values[++index] = cm_packets_bounced; + target_stat_values[++index] = cm_packets_created; +@@ -1230,11 +1230,11 @@ static void nes_netdev_get_ethtool_stats(struct net_device *netdev, + target_stat_values[++index] = cm_listens_created; + target_stat_values[++index] = cm_listens_destroyed; + target_stat_values[++index] = cm_backlog_drops; +- target_stat_values[++index] = atomic_read(&cm_loopbacks); +- target_stat_values[++index] = atomic_read(&cm_nodes_created); +- target_stat_values[++index] = atomic_read(&cm_nodes_destroyed); +- target_stat_values[++index] = atomic_read(&cm_accel_dropped_pkts); +- target_stat_values[++index] = atomic_read(&cm_resets_recvd); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_loopbacks); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_nodes_created); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_nodes_destroyed); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_accel_dropped_pkts); ++ target_stat_values[++index] = atomic_read_unchecked(&cm_resets_recvd); + target_stat_values[++index] = int_mod_timer_init; + target_stat_values[++index] = int_mod_cq_depth_1; + target_stat_values[++index] = int_mod_cq_depth_4; +diff --git a/drivers/infiniband/hw/nes/nes_verbs.c b/drivers/infiniband/hw/nes/nes_verbs.c +index a680c42..f914deb 100644 +--- a/drivers/infiniband/hw/nes/nes_verbs.c ++++ b/drivers/infiniband/hw/nes/nes_verbs.c +@@ -45,9 +45,9 @@ + + #include + +-atomic_t mod_qp_timouts; +-atomic_t qps_created; +-atomic_t sw_qps_destroyed; ++atomic_unchecked_t mod_qp_timouts; ++atomic_unchecked_t qps_created; ++atomic_unchecked_t sw_qps_destroyed; + + static void nes_unregister_ofa_device(struct nes_ib_device *nesibdev); + +@@ -1240,7 +1240,7 @@ static struct ib_qp *nes_create_qp(struct ib_pd *ibpd, + if (init_attr->create_flags) + return ERR_PTR(-EINVAL); + +- atomic_inc(&qps_created); ++ atomic_inc_unchecked(&qps_created); + switch (init_attr->qp_type) { + case IB_QPT_RC: + if (nes_drv_opt & NES_DRV_OPT_NO_INLINE_DATA) { +@@ -1568,7 +1568,7 @@ static int nes_destroy_qp(struct ib_qp *ibqp) + struct iw_cm_event cm_event; + int ret; + +- atomic_inc(&sw_qps_destroyed); ++ atomic_inc_unchecked(&sw_qps_destroyed); + nesqp->destroyed = 1; + + /* Blow away the connection if it exists. */ +diff --git a/drivers/input/gameport/gameport.c b/drivers/input/gameport/gameport.c +index ac11be0..3883c04 100644 +--- a/drivers/input/gameport/gameport.c ++++ b/drivers/input/gameport/gameport.c +@@ -515,13 +515,13 @@ EXPORT_SYMBOL(gameport_set_phys); + */ + static void gameport_init_port(struct gameport *gameport) + { +- static atomic_t gameport_no = ATOMIC_INIT(0); ++ static atomic_unchecked_t gameport_no = ATOMIC_INIT(0); + + __module_get(THIS_MODULE); + + mutex_init(&gameport->drv_mutex); + device_initialize(&gameport->dev); +- dev_set_name(&gameport->dev, "gameport%lu", (unsigned long)atomic_inc_return(&gameport_no) - 1); ++ dev_set_name(&gameport->dev, "gameport%lu", (unsigned long)atomic_inc_return_unchecked(&gameport_no) - 1); + gameport->dev.bus = &gameport_bus; + gameport->dev.release = gameport_release_port; + if (gameport->parent) +diff --git a/drivers/input/input.c b/drivers/input/input.c +index c82ae82..8cfb9cb 100644 +--- a/drivers/input/input.c ++++ b/drivers/input/input.c +@@ -1558,7 +1558,7 @@ EXPORT_SYMBOL(input_set_capability); + */ + int input_register_device(struct input_dev *dev) + { +- static atomic_t input_no = ATOMIC_INIT(0); ++ static atomic_unchecked_t input_no = ATOMIC_INIT(0); + struct input_handler *handler; + const char *path; + int error; +@@ -1585,7 +1585,7 @@ int input_register_device(struct input_dev *dev) + dev->setkeycode = input_default_setkeycode; + + dev_set_name(&dev->dev, "input%ld", +- (unsigned long) atomic_inc_return(&input_no) - 1); ++ (unsigned long) atomic_inc_return_unchecked(&input_no) - 1); + + error = device_add(&dev->dev); + if (error) +diff --git a/drivers/input/joystick/sidewinder.c b/drivers/input/joystick/sidewinder.c +index ca13a6b..b032b0c 100644 +--- a/drivers/input/joystick/sidewinder.c ++++ b/drivers/input/joystick/sidewinder.c +@@ -30,6 +30,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -428,6 +429,8 @@ static int sw_read(struct sw *sw) + unsigned char buf[SW_LENGTH]; + int i; + ++ pax_track_stack(); ++ + i = sw_read_packet(sw->gameport, buf, sw->length, 0); + + if (sw->type == SW_ID_3DP && sw->length == 66 && i != 66) { /* Broken packet, try to fix */ +diff --git a/drivers/input/joystick/xpad.c b/drivers/input/joystick/xpad.c +index 79e3edc..01412b9 100644 +--- a/drivers/input/joystick/xpad.c ++++ b/drivers/input/joystick/xpad.c +@@ -621,7 +621,7 @@ static void xpad_led_set(struct led_classdev *led_cdev, + + static int xpad_led_probe(struct usb_xpad *xpad) + { +- static atomic_t led_seq = ATOMIC_INIT(0); ++ static atomic_unchecked_t led_seq = ATOMIC_INIT(0); + long led_no; + struct xpad_led *led; + struct led_classdev *led_cdev; +@@ -634,7 +634,7 @@ static int xpad_led_probe(struct usb_xpad *xpad) + if (!led) + return -ENOMEM; + +- led_no = (long)atomic_inc_return(&led_seq) - 1; ++ led_no = (long)atomic_inc_return_unchecked(&led_seq) - 1; + + snprintf(led->name, sizeof(led->name), "xpad%ld", led_no); + led->xpad = xpad; +diff --git a/drivers/input/serio/serio.c b/drivers/input/serio/serio.c +index 0236f0d..c7327f1 100644 +--- a/drivers/input/serio/serio.c ++++ b/drivers/input/serio/serio.c +@@ -527,7 +527,7 @@ static void serio_release_port(struct device *dev) + */ + static void serio_init_port(struct serio *serio) + { +- static atomic_t serio_no = ATOMIC_INIT(0); ++ static atomic_unchecked_t serio_no = ATOMIC_INIT(0); + + __module_get(THIS_MODULE); + +@@ -536,7 +536,7 @@ static void serio_init_port(struct serio *serio) + mutex_init(&serio->drv_mutex); + device_initialize(&serio->dev); + dev_set_name(&serio->dev, "serio%ld", +- (long)atomic_inc_return(&serio_no) - 1); ++ (long)atomic_inc_return_unchecked(&serio_no) - 1); + serio->dev.bus = &serio_bus; + serio->dev.release = serio_release_port; + if (serio->parent) { +diff --git a/drivers/isdn/gigaset/common.c b/drivers/isdn/gigaset/common.c +index 33dcd8d..2783d25 100644 +--- a/drivers/isdn/gigaset/common.c ++++ b/drivers/isdn/gigaset/common.c +@@ -712,7 +712,7 @@ struct cardstate *gigaset_initcs(struct gigaset_driver *drv, int channels, + cs->commands_pending = 0; + cs->cur_at_seq = 0; + cs->gotfwver = -1; +- cs->open_count = 0; ++ local_set(&cs->open_count, 0); + cs->dev = NULL; + cs->tty = NULL; + cs->tty_dev = NULL; +diff --git a/drivers/isdn/gigaset/gigaset.h b/drivers/isdn/gigaset/gigaset.h +index a2f6125..6a70677 100644 +--- a/drivers/isdn/gigaset/gigaset.h ++++ b/drivers/isdn/gigaset/gigaset.h +@@ -34,6 +34,7 @@ + #include + #include + #include ++#include + + #define GIG_VERSION {0,5,0,0} + #define GIG_COMPAT {0,4,0,0} +@@ -446,7 +447,7 @@ struct cardstate { + spinlock_t cmdlock; + unsigned curlen, cmdbytes; + +- unsigned open_count; ++ local_t open_count; + struct tty_struct *tty; + struct tasklet_struct if_wake_tasklet; + unsigned control_state; +diff --git a/drivers/isdn/gigaset/interface.c b/drivers/isdn/gigaset/interface.c +index b3065b8..c7e8cc9 100644 +--- a/drivers/isdn/gigaset/interface.c ++++ b/drivers/isdn/gigaset/interface.c +@@ -165,9 +165,7 @@ static int if_open(struct tty_struct *tty, struct file *filp) + return -ERESTARTSYS; // FIXME -EINTR? + tty->driver_data = cs; + +- ++cs->open_count; +- +- if (cs->open_count == 1) { ++ if (local_inc_return(&cs->open_count) == 1) { + spin_lock_irqsave(&cs->lock, flags); + cs->tty = tty; + spin_unlock_irqrestore(&cs->lock, flags); +@@ -195,10 +193,10 @@ static void if_close(struct tty_struct *tty, struct file *filp) + + if (!cs->connected) + gig_dbg(DEBUG_IF, "not connected"); /* nothing to do */ +- else if (!cs->open_count) ++ else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else { +- if (!--cs->open_count) { ++ if (!local_dec_return(&cs->open_count)) { + spin_lock_irqsave(&cs->lock, flags); + cs->tty = NULL; + spin_unlock_irqrestore(&cs->lock, flags); +@@ -233,7 +231,7 @@ static int if_ioctl(struct tty_struct *tty, struct file *file, + if (!cs->connected) { + gig_dbg(DEBUG_IF, "not connected"); + retval = -ENODEV; +- } else if (!cs->open_count) ++ } else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else { + retval = 0; +@@ -361,7 +359,7 @@ static int if_write(struct tty_struct *tty, const unsigned char *buf, int count) + if (!cs->connected) { + gig_dbg(DEBUG_IF, "not connected"); + retval = -ENODEV; +- } else if (!cs->open_count) ++ } else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else if (cs->mstate != MS_LOCKED) { + dev_warn(cs->dev, "can't write to unlocked device\n"); +@@ -395,7 +393,7 @@ static int if_write_room(struct tty_struct *tty) + if (!cs->connected) { + gig_dbg(DEBUG_IF, "not connected"); + retval = -ENODEV; +- } else if (!cs->open_count) ++ } else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else if (cs->mstate != MS_LOCKED) { + dev_warn(cs->dev, "can't write to unlocked device\n"); +@@ -425,7 +423,7 @@ static int if_chars_in_buffer(struct tty_struct *tty) + + if (!cs->connected) + gig_dbg(DEBUG_IF, "not connected"); +- else if (!cs->open_count) ++ else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else if (cs->mstate != MS_LOCKED) + dev_warn(cs->dev, "can't write to unlocked device\n"); +@@ -453,7 +451,7 @@ static void if_throttle(struct tty_struct *tty) + + if (!cs->connected) + gig_dbg(DEBUG_IF, "not connected"); /* nothing to do */ +- else if (!cs->open_count) ++ else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else { + //FIXME +@@ -478,7 +476,7 @@ static void if_unthrottle(struct tty_struct *tty) + + if (!cs->connected) + gig_dbg(DEBUG_IF, "not connected"); /* nothing to do */ +- else if (!cs->open_count) ++ else if (!local_read(&cs->open_count)) + dev_warn(cs->dev, "%s: device not opened\n", __func__); + else { + //FIXME +@@ -510,7 +508,7 @@ static void if_set_termios(struct tty_struct *tty, struct ktermios *old) + goto out; + } + +- if (!cs->open_count) { ++ if (!local_read(&cs->open_count)) { + dev_warn(cs->dev, "%s: device not opened\n", __func__); + goto out; + } +diff --git a/drivers/isdn/hardware/avm/b1.c b/drivers/isdn/hardware/avm/b1.c +index a7c0083..62a7cb6 100644 +--- a/drivers/isdn/hardware/avm/b1.c ++++ b/drivers/isdn/hardware/avm/b1.c +@@ -173,7 +173,7 @@ int b1_load_t4file(avmcard *card, capiloaddatapart * t4file) + } + if (left) { + if (t4file->user) { +- if (copy_from_user(buf, dp, left)) ++ if (left > sizeof buf || copy_from_user(buf, dp, left)) + return -EFAULT; + } else { + memcpy(buf, dp, left); +@@ -221,7 +221,7 @@ int b1_load_config(avmcard *card, capiloaddatapart * config) + } + if (left) { + if (config->user) { +- if (copy_from_user(buf, dp, left)) ++ if (left > sizeof buf || copy_from_user(buf, dp, left)) + return -EFAULT; + } else { + memcpy(buf, dp, left); +diff --git a/drivers/isdn/hardware/eicon/capidtmf.c b/drivers/isdn/hardware/eicon/capidtmf.c +index f130724..c373c68 100644 +--- a/drivers/isdn/hardware/eicon/capidtmf.c ++++ b/drivers/isdn/hardware/eicon/capidtmf.c +@@ -498,6 +498,7 @@ void capidtmf_recv_block (t_capidtmf_state *p_state, byte *buffer, word leng + byte goertzel_result_buffer[CAPIDTMF_RECV_TOTAL_FREQUENCY_COUNT]; + short windowed_sample_buffer[CAPIDTMF_RECV_WINDOWED_SAMPLES]; + ++ pax_track_stack(); + + if (p_state->recv.state & CAPIDTMF_RECV_STATE_DTMF_ACTIVE) + { +diff --git a/drivers/isdn/hardware/eicon/capifunc.c b/drivers/isdn/hardware/eicon/capifunc.c +index 4d425c6..a9be6c4 100644 +--- a/drivers/isdn/hardware/eicon/capifunc.c ++++ b/drivers/isdn/hardware/eicon/capifunc.c +@@ -1055,6 +1055,8 @@ static int divacapi_connect_didd(void) + IDI_SYNC_REQ req; + DESCRIPTOR DIDD_Table[MAX_DESCRIPTORS]; + ++ pax_track_stack(); ++ + DIVA_DIDD_Read(DIDD_Table, sizeof(DIDD_Table)); + + for (x = 0; x < MAX_DESCRIPTORS; x++) { +diff --git a/drivers/isdn/hardware/eicon/diddfunc.c b/drivers/isdn/hardware/eicon/diddfunc.c +index 3029234..ef0d9e2 100644 +--- a/drivers/isdn/hardware/eicon/diddfunc.c ++++ b/drivers/isdn/hardware/eicon/diddfunc.c +@@ -54,6 +54,8 @@ static int DIVA_INIT_FUNCTION connect_didd(void) + IDI_SYNC_REQ req; + DESCRIPTOR DIDD_Table[MAX_DESCRIPTORS]; + ++ pax_track_stack(); ++ + DIVA_DIDD_Read(DIDD_Table, sizeof(DIDD_Table)); + + for (x = 0; x < MAX_DESCRIPTORS; x++) { +diff --git a/drivers/isdn/hardware/eicon/divasfunc.c b/drivers/isdn/hardware/eicon/divasfunc.c +index d36a4c0..11e7d1a 100644 +--- a/drivers/isdn/hardware/eicon/divasfunc.c ++++ b/drivers/isdn/hardware/eicon/divasfunc.c +@@ -161,6 +161,8 @@ static int DIVA_INIT_FUNCTION connect_didd(void) + IDI_SYNC_REQ req; + DESCRIPTOR DIDD_Table[MAX_DESCRIPTORS]; + ++ pax_track_stack(); ++ + DIVA_DIDD_Read(DIDD_Table, sizeof(DIDD_Table)); + + for (x = 0; x < MAX_DESCRIPTORS; x++) { +diff --git a/drivers/isdn/hardware/eicon/divasync.h b/drivers/isdn/hardware/eicon/divasync.h +index 85784a7..a19ca98 100644 +--- a/drivers/isdn/hardware/eicon/divasync.h ++++ b/drivers/isdn/hardware/eicon/divasync.h +@@ -146,7 +146,7 @@ typedef struct _diva_didd_add_adapter { + } diva_didd_add_adapter_t; + typedef struct _diva_didd_remove_adapter { + IDI_CALL p_request; +-} diva_didd_remove_adapter_t; ++} __no_const diva_didd_remove_adapter_t; + typedef struct _diva_didd_read_adapter_array { + void * buffer; + dword length; +diff --git a/drivers/isdn/hardware/eicon/idifunc.c b/drivers/isdn/hardware/eicon/idifunc.c +index db87d51..7d09acf 100644 +--- a/drivers/isdn/hardware/eicon/idifunc.c ++++ b/drivers/isdn/hardware/eicon/idifunc.c +@@ -188,6 +188,8 @@ static int DIVA_INIT_FUNCTION connect_didd(void) + IDI_SYNC_REQ req; + DESCRIPTOR DIDD_Table[MAX_DESCRIPTORS]; + ++ pax_track_stack(); ++ + DIVA_DIDD_Read(DIDD_Table, sizeof(DIDD_Table)); + + for (x = 0; x < MAX_DESCRIPTORS; x++) { +diff --git a/drivers/isdn/hardware/eicon/message.c b/drivers/isdn/hardware/eicon/message.c +index ae89fb8..0fab299 100644 +--- a/drivers/isdn/hardware/eicon/message.c ++++ b/drivers/isdn/hardware/eicon/message.c +@@ -4889,6 +4889,8 @@ static void sig_ind(PLCI *plci) + dword d; + word w; + ++ pax_track_stack(); ++ + a = plci->adapter; + Id = ((word)plci->Id<<8)|a->Id; + PUT_WORD(&SS_Ind[4],0x0000); +@@ -7484,6 +7486,8 @@ static word add_b1(PLCI *plci, API_PARSE *bp, word b_channel_info, + word j, n, w; + dword d; + ++ pax_track_stack(); ++ + + for(i=0;i<8;i++) bp_parms[i].length = 0; + for(i=0;i<2;i++) global_config[i].length = 0; +@@ -7958,6 +7962,8 @@ static word add_b23(PLCI *plci, API_PARSE *bp) + const byte llc3[] = {4,3,2,2,6,6,0}; + const byte header[] = {0,2,3,3,0,0,0}; + ++ pax_track_stack(); ++ + for(i=0;i<8;i++) bp_parms[i].length = 0; + for(i=0;i<6;i++) b2_config_parms[i].length = 0; + for(i=0;i<5;i++) b3_config_parms[i].length = 0; +@@ -14761,6 +14767,8 @@ static void group_optimization(DIVA_CAPI_ADAPTER * a, PLCI * plci) + word appl_number_group_type[MAX_APPL]; + PLCI *auxplci; + ++ pax_track_stack(); ++ + set_group_ind_mask (plci); /* all APPLs within this inc. call are allowed to dial in */ + + if(!a->group_optimization_enabled) +diff --git a/drivers/isdn/hardware/eicon/mntfunc.c b/drivers/isdn/hardware/eicon/mntfunc.c +index a564b75..f3cf8b5 100644 +--- a/drivers/isdn/hardware/eicon/mntfunc.c ++++ b/drivers/isdn/hardware/eicon/mntfunc.c +@@ -79,6 +79,8 @@ static int DIVA_INIT_FUNCTION connect_didd(void) + IDI_SYNC_REQ req; + DESCRIPTOR DIDD_Table[MAX_DESCRIPTORS]; + ++ pax_track_stack(); ++ + DIVA_DIDD_Read(DIDD_Table, sizeof(DIDD_Table)); + + for (x = 0; x < MAX_DESCRIPTORS; x++) { +diff --git a/drivers/isdn/hardware/eicon/xdi_adapter.h b/drivers/isdn/hardware/eicon/xdi_adapter.h +index a3bd163..8956575 100644 +--- a/drivers/isdn/hardware/eicon/xdi_adapter.h ++++ b/drivers/isdn/hardware/eicon/xdi_adapter.h +@@ -44,7 +44,7 @@ typedef struct _xdi_mbox_t { + typedef struct _diva_os_idi_adapter_interface { + diva_init_card_proc_t cleanup_adapter_proc; + diva_cmd_card_proc_t cmd_proc; +-} diva_os_idi_adapter_interface_t; ++} __no_const diva_os_idi_adapter_interface_t; + + typedef struct _diva_os_xdi_adapter { + struct list_head link; +diff --git a/drivers/isdn/i4l/isdn_common.c b/drivers/isdn/i4l/isdn_common.c +index adb1e8c..21b590b 100644 +--- a/drivers/isdn/i4l/isdn_common.c ++++ b/drivers/isdn/i4l/isdn_common.c +@@ -1290,6 +1290,8 @@ isdn_ioctl(struct inode *inode, struct file *file, uint cmd, ulong arg) + } iocpar; + void __user *argp = (void __user *)arg; + ++ pax_track_stack(); ++ + #define name iocpar.name + #define bname iocpar.bname + #define iocts iocpar.iocts +diff --git a/drivers/isdn/i4l/isdn_net.c b/drivers/isdn/i4l/isdn_net.c +index 90b56ed..5ed3305 100644 +--- a/drivers/isdn/i4l/isdn_net.c ++++ b/drivers/isdn/i4l/isdn_net.c +@@ -1902,7 +1902,7 @@ static int isdn_net_header(struct sk_buff *skb, struct net_device *dev, + { + isdn_net_local *lp = netdev_priv(dev); + unsigned char *p; +- ushort len = 0; ++ int len = 0; + + switch (lp->p_encap) { + case ISDN_NET_ENCAP_ETHER: +diff --git a/drivers/isdn/icn/icn.c b/drivers/isdn/icn/icn.c +index bf7997a..cf091db 100644 +--- a/drivers/isdn/icn/icn.c ++++ b/drivers/isdn/icn/icn.c +@@ -1044,7 +1044,7 @@ icn_writecmd(const u_char * buf, int len, int user, icn_card * card) + if (count > len) + count = len; + if (user) { +- if (copy_from_user(msg, buf, count)) ++ if (count > sizeof msg || copy_from_user(msg, buf, count)) + return -EFAULT; + } else + memcpy(msg, buf, count); +diff --git a/drivers/isdn/mISDN/socket.c b/drivers/isdn/mISDN/socket.c +index feb0fa4..f76f830 100644 +--- a/drivers/isdn/mISDN/socket.c ++++ b/drivers/isdn/mISDN/socket.c +@@ -391,6 +391,7 @@ data_sock_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg) + if (dev) { + struct mISDN_devinfo di; + ++ memset(&di, 0, sizeof(di)); + di.id = dev->id; + di.Dprotocols = dev->Dprotocols; + di.Bprotocols = dev->Bprotocols | get_all_Bprotocols(); +@@ -671,6 +672,7 @@ base_sock_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg) + if (dev) { + struct mISDN_devinfo di; + ++ memset(&di, 0, sizeof(di)); + di.id = dev->id; + di.Dprotocols = dev->Dprotocols; + di.Bprotocols = dev->Bprotocols | get_all_Bprotocols(); +diff --git a/drivers/isdn/sc/interrupt.c b/drivers/isdn/sc/interrupt.c +index 485be8b..f0225bc 100644 +--- a/drivers/isdn/sc/interrupt.c ++++ b/drivers/isdn/sc/interrupt.c +@@ -112,11 +112,19 @@ irqreturn_t interrupt_handler(int dummy, void *card_inst) + } + else if(callid>=0x0000 && callid<=0x7FFF) + { ++ int len; ++ + pr_debug("%s: Got Incoming Call\n", + sc_adapter[card]->devicename); +- strcpy(setup.phone,&(rcvmsg.msg_data.byte_array[4])); +- strcpy(setup.eazmsn, +- sc_adapter[card]->channel[rcvmsg.phy_link_no-1].dn); ++ len = strlcpy(setup.phone, &(rcvmsg.msg_data.byte_array[4]), ++ sizeof(setup.phone)); ++ if (len >= sizeof(setup.phone)) ++ continue; ++ len = strlcpy(setup.eazmsn, ++ sc_adapter[card]->channel[rcvmsg.phy_link_no - 1].dn, ++ sizeof(setup.eazmsn)); ++ if (len >= sizeof(setup.eazmsn)) ++ continue; + setup.si1 = 7; + setup.si2 = 0; + setup.plan = 0; +@@ -176,7 +184,9 @@ irqreturn_t interrupt_handler(int dummy, void *card_inst) + * Handle a GetMyNumber Rsp + */ + if (IS_CE_MESSAGE(rcvmsg,Call,0,GetMyNumber)){ +- strcpy(sc_adapter[card]->channel[rcvmsg.phy_link_no-1].dn,rcvmsg.msg_data.byte_array); ++ strlcpy(sc_adapter[card]->channel[rcvmsg.phy_link_no - 1].dn, ++ rcvmsg.msg_data.byte_array, ++ sizeof(rcvmsg.msg_data.byte_array)); + continue; + } + +diff --git a/drivers/lguest/core.c b/drivers/lguest/core.c +index 8744d24..d1f9a9a 100644 +--- a/drivers/lguest/core.c ++++ b/drivers/lguest/core.c +@@ -91,9 +91,17 @@ static __init int map_switcher(void) + * it's worked so far. The end address needs +1 because __get_vm_area + * allocates an extra guard page, so we need space for that. + */ ++ ++#if defined(CONFIG_MODULES) && defined(CONFIG_X86_32) && defined(CONFIG_PAX_KERNEXEC) ++ switcher_vma = __get_vm_area(TOTAL_SWITCHER_PAGES * PAGE_SIZE, ++ VM_ALLOC | VM_KERNEXEC, SWITCHER_ADDR, SWITCHER_ADDR ++ + (TOTAL_SWITCHER_PAGES+1) * PAGE_SIZE); ++#else + switcher_vma = __get_vm_area(TOTAL_SWITCHER_PAGES * PAGE_SIZE, + VM_ALLOC, SWITCHER_ADDR, SWITCHER_ADDR + + (TOTAL_SWITCHER_PAGES+1) * PAGE_SIZE); ++#endif ++ + if (!switcher_vma) { + err = -ENOMEM; + printk("lguest: could not map switcher pages high\n"); +@@ -118,7 +126,7 @@ static __init int map_switcher(void) + * Now the Switcher is mapped at the right address, we can't fail! + * Copy in the compiled-in Switcher code (from _switcher.S). + */ +- memcpy(switcher_vma->addr, start_switcher_text, ++ memcpy(switcher_vma->addr, ktla_ktva(start_switcher_text), + end_switcher_text - start_switcher_text); + + printk(KERN_INFO "lguest: mapped switcher at %p\n", +diff --git a/drivers/lguest/x86/core.c b/drivers/lguest/x86/core.c +index 6ae3888..8b38145 100644 +--- a/drivers/lguest/x86/core.c ++++ b/drivers/lguest/x86/core.c +@@ -59,7 +59,7 @@ static struct { + /* Offset from where switcher.S was compiled to where we've copied it */ + static unsigned long switcher_offset(void) + { +- return SWITCHER_ADDR - (unsigned long)start_switcher_text; ++ return SWITCHER_ADDR - (unsigned long)ktla_ktva(start_switcher_text); + } + + /* This cpu's struct lguest_pages. */ +@@ -100,7 +100,13 @@ static void copy_in_guest_info(struct lg_cpu *cpu, struct lguest_pages *pages) + * These copies are pretty cheap, so we do them unconditionally: */ + /* Save the current Host top-level page directory. + */ ++ ++#ifdef CONFIG_PAX_PER_CPU_PGD ++ pages->state.host_cr3 = read_cr3(); ++#else + pages->state.host_cr3 = __pa(current->mm->pgd); ++#endif ++ + /* + * Set up the Guest's page tables to see this CPU's pages (and no + * other CPU's pages). +@@ -535,7 +541,7 @@ void __init lguest_arch_host_init(void) + * compiled-in switcher code and the high-mapped copy we just made. + */ + for (i = 0; i < IDT_ENTRIES; i++) +- default_idt_entries[i] += switcher_offset(); ++ default_idt_entries[i] = ktla_ktva(default_idt_entries[i]) + switcher_offset(); + + /* + * Set up the Switcher's per-cpu areas. +@@ -618,7 +624,7 @@ void __init lguest_arch_host_init(void) + * it will be undisturbed when we switch. To change %cs and jump we + * need this structure to feed to Intel's "lcall" instruction. + */ +- lguest_entry.offset = (long)switch_to_guest + switcher_offset(); ++ lguest_entry.offset = (long)ktla_ktva(switch_to_guest) + switcher_offset(); + lguest_entry.segment = LGUEST_CS; + + /* +diff --git a/drivers/lguest/x86/switcher_32.S b/drivers/lguest/x86/switcher_32.S +index 40634b0..4f5855e 100644 +--- a/drivers/lguest/x86/switcher_32.S ++++ b/drivers/lguest/x86/switcher_32.S +@@ -87,6 +87,7 @@ + #include + #include + #include ++#include + + // We mark the start of the code to copy + // It's placed in .text tho it's never run here +@@ -149,6 +150,13 @@ ENTRY(switch_to_guest) + // Changes type when we load it: damn Intel! + // For after we switch over our page tables + // That entry will be read-only: we'd crash. ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ mov %cr0, %edx ++ xor $X86_CR0_WP, %edx ++ mov %edx, %cr0 ++#endif ++ + movl $(GDT_ENTRY_TSS*8), %edx + ltr %dx + +@@ -157,9 +165,15 @@ ENTRY(switch_to_guest) + // Let's clear it again for our return. + // The GDT descriptor of the Host + // Points to the table after two "size" bytes +- movl (LGUEST_PAGES_host_gdt_desc+2)(%eax), %edx ++ movl (LGUEST_PAGES_host_gdt_desc+2)(%eax), %eax + // Clear "used" from type field (byte 5, bit 2) +- andb $0xFD, (GDT_ENTRY_TSS*8 + 5)(%edx) ++ andb $0xFD, (GDT_ENTRY_TSS*8 + 5)(%eax) ++ ++#ifdef CONFIG_PAX_KERNEXEC ++ mov %cr0, %eax ++ xor $X86_CR0_WP, %eax ++ mov %eax, %cr0 ++#endif + + // Once our page table's switched, the Guest is live! + // The Host fades as we run this final step. +@@ -295,13 +309,12 @@ deliver_to_host: + // I consulted gcc, and it gave + // These instructions, which I gladly credit: + leal (%edx,%ebx,8), %eax +- movzwl (%eax),%edx +- movl 4(%eax), %eax +- xorw %ax, %ax +- orl %eax, %edx ++ movl 4(%eax), %edx ++ movw (%eax), %dx + // Now the address of the handler's in %edx + // We call it now: its "iret" drops us home. +- jmp *%edx ++ ljmp $__KERNEL_CS, $1f ++1: jmp *%edx + + // Every interrupt can come to us here + // But we must truly tell each apart. +diff --git a/drivers/macintosh/macio_asic.c b/drivers/macintosh/macio_asic.c +index 588a5b0..b71db89 100644 +--- a/drivers/macintosh/macio_asic.c ++++ b/drivers/macintosh/macio_asic.c +@@ -701,7 +701,7 @@ static void __devexit macio_pci_remove(struct pci_dev* pdev) + * MacIO is matched against any Apple ID, it's probe() function + * will then decide wether it applies or not + */ +-static const struct pci_device_id __devinitdata pci_ids [] = { { ++static const struct pci_device_id __devinitconst pci_ids [] = { { + .vendor = PCI_VENDOR_ID_APPLE, + .device = PCI_ANY_ID, + .subvendor = PCI_ANY_ID, +diff --git a/drivers/macintosh/via-pmu-backlight.c b/drivers/macintosh/via-pmu-backlight.c +index a348bb0..ecd9b3f 100644 +--- a/drivers/macintosh/via-pmu-backlight.c ++++ b/drivers/macintosh/via-pmu-backlight.c +@@ -15,7 +15,7 @@ + + #define MAX_PMU_LEVEL 0xFF + +-static struct backlight_ops pmu_backlight_data; ++static const struct backlight_ops pmu_backlight_data; + static DEFINE_SPINLOCK(pmu_backlight_lock); + static int sleeping, uses_pmu_bl; + static u8 bl_curve[FB_BACKLIGHT_LEVELS]; +@@ -115,7 +115,7 @@ static int pmu_backlight_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops pmu_backlight_data = { ++static const struct backlight_ops pmu_backlight_data = { + .get_brightness = pmu_backlight_get_brightness, + .update_status = pmu_backlight_update_status, + +diff --git a/drivers/macintosh/via-pmu.c b/drivers/macintosh/via-pmu.c +index 6f308a4..b5f7ff7 100644 +--- a/drivers/macintosh/via-pmu.c ++++ b/drivers/macintosh/via-pmu.c +@@ -2232,7 +2232,7 @@ static int pmu_sleep_valid(suspend_state_t state) + && (pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, -1) >= 0); + } + +-static struct platform_suspend_ops pmu_pm_ops = { ++static const struct platform_suspend_ops pmu_pm_ops = { + .enter = powerbook_sleep, + .valid = pmu_sleep_valid, + }; +diff --git a/drivers/md/dm-ioctl.c b/drivers/md/dm-ioctl.c +index 818b617..4656e38 100644 +--- a/drivers/md/dm-ioctl.c ++++ b/drivers/md/dm-ioctl.c +@@ -1437,7 +1437,7 @@ static int validate_params(uint cmd, struct dm_ioctl *param) + cmd == DM_LIST_VERSIONS_CMD) + return 0; + +- if ((cmd == DM_DEV_CREATE_CMD)) { ++ if (cmd == DM_DEV_CREATE_CMD) { + if (!*param->name) { + DMWARN("name not supplied when creating device"); + return -EINVAL; +diff --git a/drivers/md/dm-raid1.c b/drivers/md/dm-raid1.c +index 6021d0a..a878643 100644 +--- a/drivers/md/dm-raid1.c ++++ b/drivers/md/dm-raid1.c +@@ -41,7 +41,7 @@ enum dm_raid1_error { + + struct mirror { + struct mirror_set *ms; +- atomic_t error_count; ++ atomic_unchecked_t error_count; + unsigned long error_type; + struct dm_dev *dev; + sector_t offset; +@@ -203,7 +203,7 @@ static void fail_mirror(struct mirror *m, enum dm_raid1_error error_type) + * simple way to tell if a device has encountered + * errors. + */ +- atomic_inc(&m->error_count); ++ atomic_inc_unchecked(&m->error_count); + + if (test_and_set_bit(error_type, &m->error_type)) + return; +@@ -225,7 +225,7 @@ static void fail_mirror(struct mirror *m, enum dm_raid1_error error_type) + } + + for (new = ms->mirror; new < ms->mirror + ms->nr_mirrors; new++) +- if (!atomic_read(&new->error_count)) { ++ if (!atomic_read_unchecked(&new->error_count)) { + set_default_mirror(new); + break; + } +@@ -363,7 +363,7 @@ static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) + struct mirror *m = get_default_mirror(ms); + + do { +- if (likely(!atomic_read(&m->error_count))) ++ if (likely(!atomic_read_unchecked(&m->error_count))) + return m; + + if (m-- == ms->mirror) +@@ -377,7 +377,7 @@ static int default_ok(struct mirror *m) + { + struct mirror *default_mirror = get_default_mirror(m->ms); + +- return !atomic_read(&default_mirror->error_count); ++ return !atomic_read_unchecked(&default_mirror->error_count); + } + + static int mirror_available(struct mirror_set *ms, struct bio *bio) +@@ -484,7 +484,7 @@ static void do_reads(struct mirror_set *ms, struct bio_list *reads) + */ + if (likely(region_in_sync(ms, region, 1))) + m = choose_mirror(ms, bio->bi_sector); +- else if (m && atomic_read(&m->error_count)) ++ else if (m && atomic_read_unchecked(&m->error_count)) + m = NULL; + + if (likely(m)) +@@ -855,7 +855,7 @@ static int get_mirror(struct mirror_set *ms, struct dm_target *ti, + } + + ms->mirror[mirror].ms = ms; +- atomic_set(&(ms->mirror[mirror].error_count), 0); ++ atomic_set_unchecked(&(ms->mirror[mirror].error_count), 0); + ms->mirror[mirror].error_type = 0; + ms->mirror[mirror].offset = offset; + +@@ -1241,7 +1241,7 @@ static void mirror_resume(struct dm_target *ti) + */ + static char device_status_char(struct mirror *m) + { +- if (!atomic_read(&(m->error_count))) ++ if (!atomic_read_unchecked(&(m->error_count))) + return 'A'; + + return (test_bit(DM_RAID1_WRITE_ERROR, &(m->error_type))) ? 'D' : +diff --git a/drivers/md/dm-stripe.c b/drivers/md/dm-stripe.c +index bd58703..9f26571 100644 +--- a/drivers/md/dm-stripe.c ++++ b/drivers/md/dm-stripe.c +@@ -20,7 +20,7 @@ struct stripe { + struct dm_dev *dev; + sector_t physical_start; + +- atomic_t error_count; ++ atomic_unchecked_t error_count; + }; + + struct stripe_c { +@@ -188,7 +188,7 @@ static int stripe_ctr(struct dm_target *ti, unsigned int argc, char **argv) + kfree(sc); + return r; + } +- atomic_set(&(sc->stripe[i].error_count), 0); ++ atomic_set_unchecked(&(sc->stripe[i].error_count), 0); + } + + ti->private = sc; +@@ -257,7 +257,7 @@ static int stripe_status(struct dm_target *ti, + DMEMIT("%d ", sc->stripes); + for (i = 0; i < sc->stripes; i++) { + DMEMIT("%s ", sc->stripe[i].dev->name); +- buffer[i] = atomic_read(&(sc->stripe[i].error_count)) ? ++ buffer[i] = atomic_read_unchecked(&(sc->stripe[i].error_count)) ? + 'D' : 'A'; + } + buffer[i] = '\0'; +@@ -304,8 +304,8 @@ static int stripe_end_io(struct dm_target *ti, struct bio *bio, + */ + for (i = 0; i < sc->stripes; i++) + if (!strcmp(sc->stripe[i].dev->name, major_minor)) { +- atomic_inc(&(sc->stripe[i].error_count)); +- if (atomic_read(&(sc->stripe[i].error_count)) < ++ atomic_inc_unchecked(&(sc->stripe[i].error_count)); ++ if (atomic_read_unchecked(&(sc->stripe[i].error_count)) < + DM_IO_ERROR_THRESHOLD) + queue_work(kstriped, &sc->kstriped_ws); + } +diff --git a/drivers/md/dm-sysfs.c b/drivers/md/dm-sysfs.c +index 4b04590..13a77b2 100644 +--- a/drivers/md/dm-sysfs.c ++++ b/drivers/md/dm-sysfs.c +@@ -75,7 +75,7 @@ static struct attribute *dm_attrs[] = { + NULL, + }; + +-static struct sysfs_ops dm_sysfs_ops = { ++static const struct sysfs_ops dm_sysfs_ops = { + .show = dm_attr_show, + }; + +diff --git a/drivers/md/dm-table.c b/drivers/md/dm-table.c +index 03345bb..332250d 100644 +--- a/drivers/md/dm-table.c ++++ b/drivers/md/dm-table.c +@@ -376,7 +376,7 @@ static int device_area_is_invalid(struct dm_target *ti, struct dm_dev *dev, + if (!dev_size) + return 0; + +- if ((start >= dev_size) || (start + len > dev_size)) { ++ if ((start >= dev_size) || (len > dev_size - start)) { + DMWARN("%s: %s too small for target: " + "start=%llu, len=%llu, dev_size=%llu", + dm_device_name(ti->table->md), bdevname(bdev, b), +diff --git a/drivers/md/dm.c b/drivers/md/dm.c +index c988ac2..c418141 100644 +--- a/drivers/md/dm.c ++++ b/drivers/md/dm.c +@@ -165,9 +165,9 @@ struct mapped_device { + /* + * Event handling. + */ +- atomic_t event_nr; ++ atomic_unchecked_t event_nr; + wait_queue_head_t eventq; +- atomic_t uevent_seq; ++ atomic_unchecked_t uevent_seq; + struct list_head uevent_list; + spinlock_t uevent_lock; /* Protect access to uevent_list */ + +@@ -1776,8 +1776,8 @@ static struct mapped_device *alloc_dev(int minor) + rwlock_init(&md->map_lock); + atomic_set(&md->holders, 1); + atomic_set(&md->open_count, 0); +- atomic_set(&md->event_nr, 0); +- atomic_set(&md->uevent_seq, 0); ++ atomic_set_unchecked(&md->event_nr, 0); ++ atomic_set_unchecked(&md->uevent_seq, 0); + INIT_LIST_HEAD(&md->uevent_list); + spin_lock_init(&md->uevent_lock); + +@@ -1927,7 +1927,7 @@ static void event_callback(void *context) + + dm_send_uevents(&uevents, &disk_to_dev(md->disk)->kobj); + +- atomic_inc(&md->event_nr); ++ atomic_inc_unchecked(&md->event_nr); + wake_up(&md->eventq); + } + +@@ -2562,18 +2562,18 @@ void dm_kobject_uevent(struct mapped_device *md, enum kobject_action action, + + uint32_t dm_next_uevent_seq(struct mapped_device *md) + { +- return atomic_add_return(1, &md->uevent_seq); ++ return atomic_add_return_unchecked(1, &md->uevent_seq); + } + + uint32_t dm_get_event_nr(struct mapped_device *md) + { +- return atomic_read(&md->event_nr); ++ return atomic_read_unchecked(&md->event_nr); + } + + int dm_wait_event(struct mapped_device *md, int event_nr) + { + return wait_event_interruptible(md->eventq, +- (event_nr != atomic_read(&md->event_nr))); ++ (event_nr != atomic_read_unchecked(&md->event_nr))); + } + + void dm_uevent_add(struct mapped_device *md, struct list_head *elist) +diff --git a/drivers/md/md.c b/drivers/md/md.c +index 4ce6e2f..7a9530a 100644 +--- a/drivers/md/md.c ++++ b/drivers/md/md.c +@@ -153,10 +153,10 @@ static int start_readonly; + * start build, activate spare + */ + static DECLARE_WAIT_QUEUE_HEAD(md_event_waiters); +-static atomic_t md_event_count; ++static atomic_unchecked_t md_event_count; + void md_new_event(mddev_t *mddev) + { +- atomic_inc(&md_event_count); ++ atomic_inc_unchecked(&md_event_count); + wake_up(&md_event_waiters); + } + EXPORT_SYMBOL_GPL(md_new_event); +@@ -166,7 +166,7 @@ EXPORT_SYMBOL_GPL(md_new_event); + */ + static void md_new_event_inintr(mddev_t *mddev) + { +- atomic_inc(&md_event_count); ++ atomic_inc_unchecked(&md_event_count); + wake_up(&md_event_waiters); + } + +@@ -1226,7 +1226,7 @@ static int super_1_load(mdk_rdev_t *rdev, mdk_rdev_t *refdev, int minor_version) + + rdev->preferred_minor = 0xffff; + rdev->data_offset = le64_to_cpu(sb->data_offset); +- atomic_set(&rdev->corrected_errors, le32_to_cpu(sb->cnt_corrected_read)); ++ atomic_set_unchecked(&rdev->corrected_errors, le32_to_cpu(sb->cnt_corrected_read)); + + rdev->sb_size = le32_to_cpu(sb->max_dev) * 2 + 256; + bmask = queue_logical_block_size(rdev->bdev->bd_disk->queue)-1; +@@ -1400,7 +1400,7 @@ static void super_1_sync(mddev_t *mddev, mdk_rdev_t *rdev) + else + sb->resync_offset = cpu_to_le64(0); + +- sb->cnt_corrected_read = cpu_to_le32(atomic_read(&rdev->corrected_errors)); ++ sb->cnt_corrected_read = cpu_to_le32(atomic_read_unchecked(&rdev->corrected_errors)); + + sb->raid_disks = cpu_to_le32(mddev->raid_disks); + sb->size = cpu_to_le64(mddev->dev_sectors); +@@ -2222,7 +2222,7 @@ __ATTR(state, S_IRUGO|S_IWUSR, state_show, state_store); + static ssize_t + errors_show(mdk_rdev_t *rdev, char *page) + { +- return sprintf(page, "%d\n", atomic_read(&rdev->corrected_errors)); ++ return sprintf(page, "%d\n", atomic_read_unchecked(&rdev->corrected_errors)); + } + + static ssize_t +@@ -2231,7 +2231,7 @@ errors_store(mdk_rdev_t *rdev, const char *buf, size_t len) + char *e; + unsigned long n = simple_strtoul(buf, &e, 10); + if (*buf && (*e == 0 || *e == '\n')) { +- atomic_set(&rdev->corrected_errors, n); ++ atomic_set_unchecked(&rdev->corrected_errors, n); + return len; + } + return -EINVAL; +@@ -2525,7 +2525,7 @@ static void rdev_free(struct kobject *ko) + mdk_rdev_t *rdev = container_of(ko, mdk_rdev_t, kobj); + kfree(rdev); + } +-static struct sysfs_ops rdev_sysfs_ops = { ++static const struct sysfs_ops rdev_sysfs_ops = { + .show = rdev_attr_show, + .store = rdev_attr_store, + }; +@@ -2574,8 +2574,8 @@ static mdk_rdev_t *md_import_device(dev_t newdev, int super_format, int super_mi + rdev->data_offset = 0; + rdev->sb_events = 0; + atomic_set(&rdev->nr_pending, 0); +- atomic_set(&rdev->read_errors, 0); +- atomic_set(&rdev->corrected_errors, 0); ++ atomic_set_unchecked(&rdev->read_errors, 0); ++ atomic_set_unchecked(&rdev->corrected_errors, 0); + + size = rdev->bdev->bd_inode->i_size >> BLOCK_SIZE_BITS; + if (!size) { +@@ -3895,7 +3895,7 @@ static void md_free(struct kobject *ko) + kfree(mddev); + } + +-static struct sysfs_ops md_sysfs_ops = { ++static const struct sysfs_ops md_sysfs_ops = { + .show = md_attr_show, + .store = md_attr_store, + }; +@@ -4482,7 +4482,8 @@ out: + err = 0; + blk_integrity_unregister(disk); + md_new_event(mddev); +- sysfs_notify_dirent(mddev->sysfs_state); ++ if (mddev->sysfs_state) ++ sysfs_notify_dirent(mddev->sysfs_state); + return err; + } + +@@ -5962,7 +5963,7 @@ static int md_seq_show(struct seq_file *seq, void *v) + + spin_unlock(&pers_lock); + seq_printf(seq, "\n"); +- mi->event = atomic_read(&md_event_count); ++ mi->event = atomic_read_unchecked(&md_event_count); + return 0; + } + if (v == (void*)2) { +@@ -6051,7 +6052,7 @@ static int md_seq_show(struct seq_file *seq, void *v) + chunk_kb ? "KB" : "B"); + if (bitmap->file) { + seq_printf(seq, ", file: "); +- seq_path(seq, &bitmap->file->f_path, " \t\n"); ++ seq_path(seq, &bitmap->file->f_path, " \t\n\\"); + } + + seq_printf(seq, "\n"); +@@ -6085,7 +6086,7 @@ static int md_seq_open(struct inode *inode, struct file *file) + else { + struct seq_file *p = file->private_data; + p->private = mi; +- mi->event = atomic_read(&md_event_count); ++ mi->event = atomic_read_unchecked(&md_event_count); + } + return error; + } +@@ -6101,7 +6102,7 @@ static unsigned int mdstat_poll(struct file *filp, poll_table *wait) + /* always allow read */ + mask = POLLIN | POLLRDNORM; + +- if (mi->event != atomic_read(&md_event_count)) ++ if (mi->event != atomic_read_unchecked(&md_event_count)) + mask |= POLLERR | POLLPRI; + return mask; + } +@@ -6145,7 +6146,7 @@ static int is_mddev_idle(mddev_t *mddev, int init) + struct gendisk *disk = rdev->bdev->bd_contains->bd_disk; + curr_events = (int)part_stat_read(&disk->part0, sectors[0]) + + (int)part_stat_read(&disk->part0, sectors[1]) - +- atomic_read(&disk->sync_io); ++ atomic_read_unchecked(&disk->sync_io); + /* sync IO will cause sync_io to increase before the disk_stats + * as sync_io is counted when a request starts, and + * disk_stats is counted when it completes. +diff --git a/drivers/md/md.h b/drivers/md/md.h +index 87430fe..0024a4c 100644 +--- a/drivers/md/md.h ++++ b/drivers/md/md.h +@@ -94,10 +94,10 @@ struct mdk_rdev_s + * only maintained for arrays that + * support hot removal + */ +- atomic_t read_errors; /* number of consecutive read errors that ++ atomic_unchecked_t read_errors; /* number of consecutive read errors that + * we have tried to ignore. + */ +- atomic_t corrected_errors; /* number of corrected read errors, ++ atomic_unchecked_t corrected_errors; /* number of corrected read errors, + * for reporting to userspace and storing + * in superblock. + */ +@@ -304,7 +304,7 @@ static inline void rdev_dec_pending(mdk_rdev_t *rdev, mddev_t *mddev) + + static inline void md_sync_acct(struct block_device *bdev, unsigned long nr_sectors) + { +- atomic_add(nr_sectors, &bdev->bd_contains->bd_disk->sync_io); ++ atomic_add_unchecked(nr_sectors, &bdev->bd_contains->bd_disk->sync_io); + } + + struct mdk_personality +diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c +index 968cb14..f0ad2e4 100644 +--- a/drivers/md/raid1.c ++++ b/drivers/md/raid1.c +@@ -1415,7 +1415,7 @@ static void sync_request_write(mddev_t *mddev, r1bio_t *r1_bio) + if (r1_bio->bios[d]->bi_end_io != end_sync_read) + continue; + rdev = conf->mirrors[d].rdev; +- atomic_add(s, &rdev->corrected_errors); ++ atomic_add_unchecked(s, &rdev->corrected_errors); + if (sync_page_io(rdev->bdev, + sect + rdev->data_offset, + s<<9, +@@ -1564,7 +1564,7 @@ static void fix_read_error(conf_t *conf, int read_disk, + /* Well, this device is dead */ + md_error(mddev, rdev); + else { +- atomic_add(s, &rdev->corrected_errors); ++ atomic_add_unchecked(s, &rdev->corrected_errors); + printk(KERN_INFO + "raid1:%s: read error corrected " + "(%d sectors at %llu on %s)\n", +diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c +index 1b4e232..cf0f534b 100644 +--- a/drivers/md/raid10.c ++++ b/drivers/md/raid10.c +@@ -1255,7 +1255,7 @@ static void end_sync_read(struct bio *bio, int error) + if (test_bit(BIO_UPTODATE, &bio->bi_flags)) + set_bit(R10BIO_Uptodate, &r10_bio->state); + else { +- atomic_add(r10_bio->sectors, ++ atomic_add_unchecked(r10_bio->sectors, + &conf->mirrors[d].rdev->corrected_errors); + if (!test_bit(MD_RECOVERY_SYNC, &conf->mddev->recovery)) + md_error(r10_bio->mddev, +@@ -1520,7 +1520,7 @@ static void fix_read_error(conf_t *conf, mddev_t *mddev, r10bio_t *r10_bio) + test_bit(In_sync, &rdev->flags)) { + atomic_inc(&rdev->nr_pending); + rcu_read_unlock(); +- atomic_add(s, &rdev->corrected_errors); ++ atomic_add_unchecked(s, &rdev->corrected_errors); + if (sync_page_io(rdev->bdev, + r10_bio->devs[sl].addr + + sect + rdev->data_offset, +diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c +index 883215d..675bf47 100644 +--- a/drivers/md/raid5.c ++++ b/drivers/md/raid5.c +@@ -482,7 +482,7 @@ static void ops_run_io(struct stripe_head *sh, struct stripe_head_state *s) + bi->bi_next = NULL; + if ((rw & WRITE) && + test_bit(R5_ReWrite, &sh->dev[i].flags)) +- atomic_add(STRIPE_SECTORS, ++ atomic_add_unchecked(STRIPE_SECTORS, + &rdev->corrected_errors); + generic_make_request(bi); + } else { +@@ -1517,15 +1517,15 @@ static void raid5_end_read_request(struct bio * bi, int error) + clear_bit(R5_ReadError, &sh->dev[i].flags); + clear_bit(R5_ReWrite, &sh->dev[i].flags); + } +- if (atomic_read(&conf->disks[i].rdev->read_errors)) +- atomic_set(&conf->disks[i].rdev->read_errors, 0); ++ if (atomic_read_unchecked(&conf->disks[i].rdev->read_errors)) ++ atomic_set_unchecked(&conf->disks[i].rdev->read_errors, 0); + } else { + const char *bdn = bdevname(conf->disks[i].rdev->bdev, b); + int retry = 0; + rdev = conf->disks[i].rdev; + + clear_bit(R5_UPTODATE, &sh->dev[i].flags); +- atomic_inc(&rdev->read_errors); ++ atomic_inc_unchecked(&rdev->read_errors); + if (conf->mddev->degraded >= conf->max_degraded) + printk_rl(KERN_WARNING + "raid5:%s: read error not correctable " +@@ -1543,7 +1543,7 @@ static void raid5_end_read_request(struct bio * bi, int error) + (unsigned long long)(sh->sector + + rdev->data_offset), + bdn); +- else if (atomic_read(&rdev->read_errors) ++ else if (atomic_read_unchecked(&rdev->read_errors) + > conf->max_nr_stripes) + printk(KERN_WARNING + "raid5:%s: Too many read errors, failing device %s.\n", +@@ -1870,6 +1870,7 @@ static sector_t compute_blocknr(struct stripe_head *sh, int i, int previous) + sector_t r_sector; + struct stripe_head sh2; + ++ pax_track_stack(); + + chunk_offset = sector_div(new_sector, sectors_per_chunk); + stripe = new_sector; +diff --git a/drivers/media/common/saa7146_hlp.c b/drivers/media/common/saa7146_hlp.c +index 05bde9c..2f31d40 100644 +--- a/drivers/media/common/saa7146_hlp.c ++++ b/drivers/media/common/saa7146_hlp.c +@@ -353,6 +353,8 @@ static void calculate_clipping_registers_rect(struct saa7146_dev *dev, struct sa + + int x[32], y[32], w[32], h[32]; + ++ pax_track_stack(); ++ + /* clear out memory */ + memset(&line_list[0], 0x00, sizeof(u32)*32); + memset(&pixel_list[0], 0x00, sizeof(u32)*32); +diff --git a/drivers/media/dvb/dvb-core/dvb_ca_en50221.c b/drivers/media/dvb/dvb-core/dvb_ca_en50221.c +index cb22da5..82b686e 100644 +--- a/drivers/media/dvb/dvb-core/dvb_ca_en50221.c ++++ b/drivers/media/dvb/dvb-core/dvb_ca_en50221.c +@@ -590,6 +590,8 @@ static int dvb_ca_en50221_read_data(struct dvb_ca_private *ca, int slot, u8 * eb + u8 buf[HOST_LINK_BUF_SIZE]; + int i; + ++ pax_track_stack(); ++ + dprintk("%s\n", __func__); + + /* check if we have space for a link buf in the rx_buffer */ +@@ -1285,6 +1287,8 @@ static ssize_t dvb_ca_en50221_io_write(struct file *file, + unsigned long timeout; + int written; + ++ pax_track_stack(); ++ + dprintk("%s\n", __func__); + + /* Incoming packet has a 2 byte header. hdr[0] = slot_id, hdr[1] = connection_id */ +diff --git a/drivers/media/dvb/dvb-core/dvb_demux.h b/drivers/media/dvb/dvb-core/dvb_demux.h +index 2fe05d0..a3289c4 100644 +--- a/drivers/media/dvb/dvb-core/dvb_demux.h ++++ b/drivers/media/dvb/dvb-core/dvb_demux.h +@@ -71,7 +71,7 @@ struct dvb_demux_feed { + union { + dmx_ts_cb ts; + dmx_section_cb sec; +- } cb; ++ } __no_const cb; + + struct dvb_demux *demux; + void *priv; +diff --git a/drivers/media/dvb/dvb-core/dvbdev.c b/drivers/media/dvb/dvb-core/dvbdev.c +index 94159b9..376bd8e 100644 +--- a/drivers/media/dvb/dvb-core/dvbdev.c ++++ b/drivers/media/dvb/dvb-core/dvbdev.c +@@ -191,7 +191,7 @@ int dvb_register_device(struct dvb_adapter *adap, struct dvb_device **pdvbdev, + const struct dvb_device *template, void *priv, int type) + { + struct dvb_device *dvbdev; +- struct file_operations *dvbdevfops; ++ file_operations_no_const *dvbdevfops; + struct device *clsdev; + int minor; + int id; +diff --git a/drivers/media/dvb/dvb-usb/cxusb.c b/drivers/media/dvb/dvb-usb/cxusb.c +index 2a53dd0..db8c07a 100644 +--- a/drivers/media/dvb/dvb-usb/cxusb.c ++++ b/drivers/media/dvb/dvb-usb/cxusb.c +@@ -1040,7 +1040,7 @@ static struct dib0070_config dib7070p_dib0070_config = { + struct dib0700_adapter_state { + int (*set_param_save) (struct dvb_frontend *, + struct dvb_frontend_parameters *); +-}; ++} __no_const; + + static int dib7070_set_param_override(struct dvb_frontend *fe, + struct dvb_frontend_parameters *fep) +diff --git a/drivers/media/dvb/dvb-usb/dib0700_core.c b/drivers/media/dvb/dvb-usb/dib0700_core.c +index db7f7f7..f55e96f 100644 +--- a/drivers/media/dvb/dvb-usb/dib0700_core.c ++++ b/drivers/media/dvb/dvb-usb/dib0700_core.c +@@ -332,6 +332,8 @@ int dib0700_download_firmware(struct usb_device *udev, const struct firmware *fw + + u8 buf[260]; + ++ pax_track_stack(); ++ + while ((ret = dvb_usb_get_hexline(fw, &hx, &pos)) > 0) { + deb_fwdata("writing to address 0x%08x (buffer: 0x%02x %02x)\n",hx.addr, hx.len, hx.chk); + +diff --git a/drivers/media/dvb/dvb-usb/dib0700_devices.c b/drivers/media/dvb/dvb-usb/dib0700_devices.c +index 524acf5..5ffc403 100644 +--- a/drivers/media/dvb/dvb-usb/dib0700_devices.c ++++ b/drivers/media/dvb/dvb-usb/dib0700_devices.c +@@ -28,7 +28,7 @@ MODULE_PARM_DESC(force_lna_activation, "force the activation of Low-Noise-Amplif + + struct dib0700_adapter_state { + int (*set_param_save) (struct dvb_frontend *, struct dvb_frontend_parameters *); +-}; ++} __no_const; + + /* Hauppauge Nova-T 500 (aka Bristol) + * has a LNA on GPIO0 which is enabled by setting 1 */ +diff --git a/drivers/media/dvb/frontends/dib3000.h b/drivers/media/dvb/frontends/dib3000.h +index ba91735..4261d84 100644 +--- a/drivers/media/dvb/frontends/dib3000.h ++++ b/drivers/media/dvb/frontends/dib3000.h +@@ -39,7 +39,7 @@ struct dib_fe_xfer_ops + int (*fifo_ctrl)(struct dvb_frontend *fe, int onoff); + int (*pid_ctrl)(struct dvb_frontend *fe, int index, int pid, int onoff); + int (*tuner_pass_ctrl)(struct dvb_frontend *fe, int onoff, u8 pll_ctrl); +-}; ++} __no_const; + + #if defined(CONFIG_DVB_DIB3000MB) || (defined(CONFIG_DVB_DIB3000MB_MODULE) && defined(MODULE)) + extern struct dvb_frontend* dib3000mb_attach(const struct dib3000_config* config, +diff --git a/drivers/media/dvb/frontends/or51211.c b/drivers/media/dvb/frontends/or51211.c +index c709ce6..b3fe620 100644 +--- a/drivers/media/dvb/frontends/or51211.c ++++ b/drivers/media/dvb/frontends/or51211.c +@@ -113,6 +113,8 @@ static int or51211_load_firmware (struct dvb_frontend* fe, + u8 tudata[585]; + int i; + ++ pax_track_stack(); ++ + dprintk("Firmware is %zd bytes\n",fw->size); + + /* Get eprom data */ +diff --git a/drivers/media/radio/radio-cadet.c b/drivers/media/radio/radio-cadet.c +index 482d0f3..ee1e202 100644 +--- a/drivers/media/radio/radio-cadet.c ++++ b/drivers/media/radio/radio-cadet.c +@@ -347,7 +347,7 @@ static ssize_t cadet_read(struct file *file, char __user *data, size_t count, lo + while (i < count && dev->rdsin != dev->rdsout) + readbuf[i++] = dev->rdsbuf[dev->rdsout++]; + +- if (copy_to_user(data, readbuf, i)) ++ if (i > sizeof readbuf || copy_to_user(data, readbuf, i)) + return -EFAULT; + return i; + } +diff --git a/drivers/media/video/cx18/cx18-driver.c b/drivers/media/video/cx18/cx18-driver.c +index 6dd51e2..0359b92 100644 +--- a/drivers/media/video/cx18/cx18-driver.c ++++ b/drivers/media/video/cx18/cx18-driver.c +@@ -56,7 +56,7 @@ static struct pci_device_id cx18_pci_tbl[] __devinitdata = { + + MODULE_DEVICE_TABLE(pci, cx18_pci_tbl); + +-static atomic_t cx18_instance = ATOMIC_INIT(0); ++static atomic_unchecked_t cx18_instance = ATOMIC_INIT(0); + + /* Parameter declarations */ + static int cardtype[CX18_MAX_CARDS]; +@@ -288,6 +288,8 @@ void cx18_read_eeprom(struct cx18 *cx, struct tveeprom *tv) + struct i2c_client c; + u8 eedata[256]; + ++ pax_track_stack(); ++ + memset(&c, 0, sizeof(c)); + strlcpy(c.name, "cx18 tveeprom tmp", sizeof(c.name)); + c.adapter = &cx->i2c_adap[0]; +@@ -800,7 +802,7 @@ static int __devinit cx18_probe(struct pci_dev *pci_dev, + struct cx18 *cx; + + /* FIXME - module parameter arrays constrain max instances */ +- i = atomic_inc_return(&cx18_instance) - 1; ++ i = atomic_inc_return_unchecked(&cx18_instance) - 1; + if (i >= CX18_MAX_CARDS) { + printk(KERN_ERR "cx18: cannot manage card %d, driver has a " + "limit of 0 - %d\n", i, CX18_MAX_CARDS - 1); +diff --git a/drivers/media/video/ivtv/ivtv-driver.c b/drivers/media/video/ivtv/ivtv-driver.c +index 463ec34..2f4625a 100644 +--- a/drivers/media/video/ivtv/ivtv-driver.c ++++ b/drivers/media/video/ivtv/ivtv-driver.c +@@ -79,7 +79,7 @@ static struct pci_device_id ivtv_pci_tbl[] __devinitdata = { + MODULE_DEVICE_TABLE(pci,ivtv_pci_tbl); + + /* ivtv instance counter */ +-static atomic_t ivtv_instance = ATOMIC_INIT(0); ++static atomic_unchecked_t ivtv_instance = ATOMIC_INIT(0); + + /* Parameter declarations */ + static int cardtype[IVTV_MAX_CARDS]; +diff --git a/drivers/media/video/omap24xxcam.c b/drivers/media/video/omap24xxcam.c +index 5fc4ac0..652a54a 100644 +--- a/drivers/media/video/omap24xxcam.c ++++ b/drivers/media/video/omap24xxcam.c +@@ -401,7 +401,7 @@ static void omap24xxcam_vbq_complete(struct omap24xxcam_sgdma *sgdma, + spin_unlock_irqrestore(&cam->core_enable_disable_lock, flags); + + do_gettimeofday(&vb->ts); +- vb->field_count = atomic_add_return(2, &fh->field_count); ++ vb->field_count = atomic_add_return_unchecked(2, &fh->field_count); + if (csr & csr_error) { + vb->state = VIDEOBUF_ERROR; + if (!atomic_read(&fh->cam->in_reset)) { +diff --git a/drivers/media/video/omap24xxcam.h b/drivers/media/video/omap24xxcam.h +index 2ce67f5..cf26a5b 100644 +--- a/drivers/media/video/omap24xxcam.h ++++ b/drivers/media/video/omap24xxcam.h +@@ -533,7 +533,7 @@ struct omap24xxcam_fh { + spinlock_t vbq_lock; /* spinlock for the videobuf queue */ + struct videobuf_queue vbq; + struct v4l2_pix_format pix; /* serialise pix by vbq->lock */ +- atomic_t field_count; /* field counter for videobuf_buffer */ ++ atomic_unchecked_t field_count; /* field counter for videobuf_buffer */ + /* accessing cam here doesn't need serialisation: it's constant */ + struct omap24xxcam_device *cam; + }; +diff --git a/drivers/media/video/pvrusb2/pvrusb2-eeprom.c b/drivers/media/video/pvrusb2/pvrusb2-eeprom.c +index 299afa4..eb47459 100644 +--- a/drivers/media/video/pvrusb2/pvrusb2-eeprom.c ++++ b/drivers/media/video/pvrusb2/pvrusb2-eeprom.c +@@ -119,6 +119,8 @@ int pvr2_eeprom_analyze(struct pvr2_hdw *hdw) + u8 *eeprom; + struct tveeprom tvdata; + ++ pax_track_stack(); ++ + memset(&tvdata,0,sizeof(tvdata)); + + eeprom = pvr2_eeprom_fetch(hdw); +diff --git a/drivers/media/video/pvrusb2/pvrusb2-hdw-internal.h b/drivers/media/video/pvrusb2/pvrusb2-hdw-internal.h +index 5b152ff..3320638 100644 +--- a/drivers/media/video/pvrusb2/pvrusb2-hdw-internal.h ++++ b/drivers/media/video/pvrusb2/pvrusb2-hdw-internal.h +@@ -195,7 +195,7 @@ struct pvr2_hdw { + + /* I2C stuff */ + struct i2c_adapter i2c_adap; +- struct i2c_algorithm i2c_algo; ++ i2c_algorithm_no_const i2c_algo; + pvr2_i2c_func i2c_func[PVR2_I2C_FUNC_CNT]; + int i2c_cx25840_hack_state; + int i2c_linked; +diff --git a/drivers/media/video/saa7134/saa6752hs.c b/drivers/media/video/saa7134/saa6752hs.c +index 1eabff6..8e2313a 100644 +--- a/drivers/media/video/saa7134/saa6752hs.c ++++ b/drivers/media/video/saa7134/saa6752hs.c +@@ -683,6 +683,8 @@ static int saa6752hs_init(struct v4l2_subdev *sd, u32 leading_null_bytes) + unsigned char localPAT[256]; + unsigned char localPMT[256]; + ++ pax_track_stack(); ++ + /* Set video format - must be done first as it resets other settings */ + set_reg8(client, 0x41, h->video_format); + +diff --git a/drivers/media/video/saa7164/saa7164-cmd.c b/drivers/media/video/saa7164/saa7164-cmd.c +index 9c1d3ac..b1b49e9 100644 +--- a/drivers/media/video/saa7164/saa7164-cmd.c ++++ b/drivers/media/video/saa7164/saa7164-cmd.c +@@ -87,6 +87,8 @@ int saa7164_irq_dequeue(struct saa7164_dev *dev) + wait_queue_head_t *q = 0; + dprintk(DBGLVL_CMD, "%s()\n", __func__); + ++ pax_track_stack(); ++ + /* While any outstand message on the bus exists... */ + do { + +@@ -126,6 +128,8 @@ int saa7164_cmd_dequeue(struct saa7164_dev *dev) + u8 tmp[512]; + dprintk(DBGLVL_CMD, "%s()\n", __func__); + ++ pax_track_stack(); ++ + while (loop) { + + tmComResInfo_t tRsp = { 0, 0, 0, 0, 0, 0 }; +diff --git a/drivers/media/video/usbvideo/ibmcam.c b/drivers/media/video/usbvideo/ibmcam.c +index b085496..cde0270 100644 +--- a/drivers/media/video/usbvideo/ibmcam.c ++++ b/drivers/media/video/usbvideo/ibmcam.c +@@ -3947,15 +3947,15 @@ static struct usb_device_id id_table[] = { + static int __init ibmcam_init(void) + { + struct usbvideo_cb cbTbl; +- memset(&cbTbl, 0, sizeof(cbTbl)); +- cbTbl.probe = ibmcam_probe; +- cbTbl.setupOnOpen = ibmcam_setup_on_open; +- cbTbl.videoStart = ibmcam_video_start; +- cbTbl.videoStop = ibmcam_video_stop; +- cbTbl.processData = ibmcam_ProcessIsocData; +- cbTbl.postProcess = usbvideo_DeinterlaceFrame; +- cbTbl.adjustPicture = ibmcam_adjust_picture; +- cbTbl.getFPS = ibmcam_calculate_fps; ++ memset((void *)&cbTbl, 0, sizeof(cbTbl)); ++ *(void **)&cbTbl.probe = ibmcam_probe; ++ *(void **)&cbTbl.setupOnOpen = ibmcam_setup_on_open; ++ *(void **)&cbTbl.videoStart = ibmcam_video_start; ++ *(void **)&cbTbl.videoStop = ibmcam_video_stop; ++ *(void **)&cbTbl.processData = ibmcam_ProcessIsocData; ++ *(void **)&cbTbl.postProcess = usbvideo_DeinterlaceFrame; ++ *(void **)&cbTbl.adjustPicture = ibmcam_adjust_picture; ++ *(void **)&cbTbl.getFPS = ibmcam_calculate_fps; + return usbvideo_register( + &cams, + MAX_IBMCAM, +diff --git a/drivers/media/video/usbvideo/konicawc.c b/drivers/media/video/usbvideo/konicawc.c +index 31d57f2..600b735 100644 +--- a/drivers/media/video/usbvideo/konicawc.c ++++ b/drivers/media/video/usbvideo/konicawc.c +@@ -225,7 +225,7 @@ static void konicawc_register_input(struct konicawc *cam, struct usb_device *dev + int error; + + usb_make_path(dev, cam->input_physname, sizeof(cam->input_physname)); +- strncat(cam->input_physname, "/input0", sizeof(cam->input_physname)); ++ strlcat(cam->input_physname, "/input0", sizeof(cam->input_physname)); + + cam->input = input_dev = input_allocate_device(); + if (!input_dev) { +@@ -935,16 +935,16 @@ static int __init konicawc_init(void) + struct usbvideo_cb cbTbl; + printk(KERN_INFO KBUILD_MODNAME ": " DRIVER_VERSION ":" + DRIVER_DESC "\n"); +- memset(&cbTbl, 0, sizeof(cbTbl)); +- cbTbl.probe = konicawc_probe; +- cbTbl.setupOnOpen = konicawc_setup_on_open; +- cbTbl.processData = konicawc_process_isoc; +- cbTbl.getFPS = konicawc_calculate_fps; +- cbTbl.setVideoMode = konicawc_set_video_mode; +- cbTbl.startDataPump = konicawc_start_data; +- cbTbl.stopDataPump = konicawc_stop_data; +- cbTbl.adjustPicture = konicawc_adjust_picture; +- cbTbl.userFree = konicawc_free_uvd; ++ memset((void * )&cbTbl, 0, sizeof(cbTbl)); ++ *(void **)&cbTbl.probe = konicawc_probe; ++ *(void **)&cbTbl.setupOnOpen = konicawc_setup_on_open; ++ *(void **)&cbTbl.processData = konicawc_process_isoc; ++ *(void **)&cbTbl.getFPS = konicawc_calculate_fps; ++ *(void **)&cbTbl.setVideoMode = konicawc_set_video_mode; ++ *(void **)&cbTbl.startDataPump = konicawc_start_data; ++ *(void **)&cbTbl.stopDataPump = konicawc_stop_data; ++ *(void **)&cbTbl.adjustPicture = konicawc_adjust_picture; ++ *(void **)&cbTbl.userFree = konicawc_free_uvd; + return usbvideo_register( + &cams, + MAX_CAMERAS, +diff --git a/drivers/media/video/usbvideo/quickcam_messenger.c b/drivers/media/video/usbvideo/quickcam_messenger.c +index 803d3e4..c4d1b96 100644 +--- a/drivers/media/video/usbvideo/quickcam_messenger.c ++++ b/drivers/media/video/usbvideo/quickcam_messenger.c +@@ -89,7 +89,7 @@ static void qcm_register_input(struct qcm *cam, struct usb_device *dev) + int error; + + usb_make_path(dev, cam->input_physname, sizeof(cam->input_physname)); +- strncat(cam->input_physname, "/input0", sizeof(cam->input_physname)); ++ strlcat(cam->input_physname, "/input0", sizeof(cam->input_physname)); + + cam->input = input_dev = input_allocate_device(); + if (!input_dev) { +diff --git a/drivers/media/video/usbvideo/ultracam.c b/drivers/media/video/usbvideo/ultracam.c +index fbd1b63..292f9f0 100644 +--- a/drivers/media/video/usbvideo/ultracam.c ++++ b/drivers/media/video/usbvideo/ultracam.c +@@ -655,14 +655,14 @@ static int __init ultracam_init(void) + { + struct usbvideo_cb cbTbl; + memset(&cbTbl, 0, sizeof(cbTbl)); +- cbTbl.probe = ultracam_probe; +- cbTbl.setupOnOpen = ultracam_setup_on_open; +- cbTbl.videoStart = ultracam_video_start; +- cbTbl.videoStop = ultracam_video_stop; +- cbTbl.processData = ultracam_ProcessIsocData; +- cbTbl.postProcess = usbvideo_DeinterlaceFrame; +- cbTbl.adjustPicture = ultracam_adjust_picture; +- cbTbl.getFPS = ultracam_calculate_fps; ++ *(void **)&cbTbl.probe = ultracam_probe; ++ *(void **)&cbTbl.setupOnOpen = ultracam_setup_on_open; ++ *(void **)&cbTbl.videoStart = ultracam_video_start; ++ *(void **)&cbTbl.videoStop = ultracam_video_stop; ++ *(void **)&cbTbl.processData = ultracam_ProcessIsocData; ++ *(void **)&cbTbl.postProcess = usbvideo_DeinterlaceFrame; ++ *(void **)&cbTbl.adjustPicture = ultracam_adjust_picture; ++ *(void **)&cbTbl.getFPS = ultracam_calculate_fps; + return usbvideo_register( + &cams, + MAX_CAMERAS, +diff --git a/drivers/media/video/usbvideo/usbvideo.c b/drivers/media/video/usbvideo/usbvideo.c +index dea8b32..34f6878 100644 +--- a/drivers/media/video/usbvideo/usbvideo.c ++++ b/drivers/media/video/usbvideo/usbvideo.c +@@ -697,15 +697,15 @@ int usbvideo_register( + __func__, cams, base_size, num_cams); + + /* Copy callbacks, apply defaults for those that are not set */ +- memmove(&cams->cb, cbTbl, sizeof(cams->cb)); ++ memmove((void *)&cams->cb, cbTbl, sizeof(cams->cb)); + if (cams->cb.getFrame == NULL) +- cams->cb.getFrame = usbvideo_GetFrame; ++ *(void **)&cams->cb.getFrame = usbvideo_GetFrame; + if (cams->cb.disconnect == NULL) +- cams->cb.disconnect = usbvideo_Disconnect; ++ *(void **)&cams->cb.disconnect = usbvideo_Disconnect; + if (cams->cb.startDataPump == NULL) +- cams->cb.startDataPump = usbvideo_StartDataPump; ++ *(void **)&cams->cb.startDataPump = usbvideo_StartDataPump; + if (cams->cb.stopDataPump == NULL) +- cams->cb.stopDataPump = usbvideo_StopDataPump; ++ *(void **)&cams->cb.stopDataPump = usbvideo_StopDataPump; + + cams->num_cameras = num_cams; + cams->cam = (struct uvd *) &cams[1]; +diff --git a/drivers/media/video/usbvideo/usbvideo.h b/drivers/media/video/usbvideo/usbvideo.h +index c66985b..7fa143a 100644 +--- a/drivers/media/video/usbvideo/usbvideo.h ++++ b/drivers/media/video/usbvideo/usbvideo.h +@@ -268,7 +268,7 @@ struct usbvideo_cb { + int (*startDataPump)(struct uvd *uvd); + void (*stopDataPump)(struct uvd *uvd); + int (*setVideoMode)(struct uvd *uvd, struct video_window *vw); +-}; ++} __no_const; + + struct usbvideo { + int num_cameras; /* As allocated */ +diff --git a/drivers/media/video/usbvision/usbvision-core.c b/drivers/media/video/usbvision/usbvision-core.c +index e0f91e4..37554ea 100644 +--- a/drivers/media/video/usbvision/usbvision-core.c ++++ b/drivers/media/video/usbvision/usbvision-core.c +@@ -820,6 +820,8 @@ static enum ParseState usbvision_parse_compress(struct usb_usbvision *usbvision, + unsigned char rv, gv, bv; + static unsigned char *Y, *U, *V; + ++ pax_track_stack(); ++ + frame = usbvision->curFrame; + imageSize = frame->frmwidth * frame->frmheight; + if ( (frame->v4l2_format.format == V4L2_PIX_FMT_YUV422P) || +diff --git a/drivers/media/video/v4l2-device.c b/drivers/media/video/v4l2-device.c +index 0d06e7c..3d17d24 100644 +--- a/drivers/media/video/v4l2-device.c ++++ b/drivers/media/video/v4l2-device.c +@@ -50,9 +50,9 @@ int v4l2_device_register(struct device *dev, struct v4l2_device *v4l2_dev) + EXPORT_SYMBOL_GPL(v4l2_device_register); + + int v4l2_device_set_name(struct v4l2_device *v4l2_dev, const char *basename, +- atomic_t *instance) ++ atomic_unchecked_t *instance) + { +- int num = atomic_inc_return(instance) - 1; ++ int num = atomic_inc_return_unchecked(instance) - 1; + int len = strlen(basename); + + if (basename[len - 1] >= '0' && basename[len - 1] <= '9') +diff --git a/drivers/media/video/videobuf-dma-sg.c b/drivers/media/video/videobuf-dma-sg.c +index 032ebae..6a3532c 100644 +--- a/drivers/media/video/videobuf-dma-sg.c ++++ b/drivers/media/video/videobuf-dma-sg.c +@@ -693,6 +693,8 @@ void *videobuf_sg_alloc(size_t size) + { + struct videobuf_queue q; + ++ pax_track_stack(); ++ + /* Required to make generic handler to call __videobuf_alloc */ + q.int_ops = &sg_ops; + +diff --git a/drivers/message/fusion/mptbase.c b/drivers/message/fusion/mptbase.c +index b6992b7..9fa7547 100644 +--- a/drivers/message/fusion/mptbase.c ++++ b/drivers/message/fusion/mptbase.c +@@ -6709,8 +6709,14 @@ procmpt_iocinfo_read(char *buf, char **start, off_t offset, int request, int *eo + len += sprintf(buf+len, " MaxChainDepth = 0x%02x frames\n", ioc->facts.MaxChainDepth); + len += sprintf(buf+len, " MinBlockSize = 0x%02x bytes\n", 4*ioc->facts.BlockSize); + ++#ifdef CONFIG_GRKERNSEC_HIDESYM ++ len += sprintf(buf+len, " RequestFrames @ 0x%p (Dma @ 0x%p)\n", ++ NULL, NULL); ++#else + len += sprintf(buf+len, " RequestFrames @ 0x%p (Dma @ 0x%p)\n", + (void *)ioc->req_frames, (void *)(ulong)ioc->req_frames_dma); ++#endif ++ + /* + * Rounding UP to nearest 4-kB boundary here... + */ +diff --git a/drivers/message/fusion/mptsas.c b/drivers/message/fusion/mptsas.c +index 83873e3..e360e9a 100644 +--- a/drivers/message/fusion/mptsas.c ++++ b/drivers/message/fusion/mptsas.c +@@ -436,6 +436,23 @@ mptsas_is_end_device(struct mptsas_devinfo * attached) + return 0; + } + ++static inline void ++mptsas_set_rphy(MPT_ADAPTER *ioc, struct mptsas_phyinfo *phy_info, struct sas_rphy *rphy) ++{ ++ if (phy_info->port_details) { ++ phy_info->port_details->rphy = rphy; ++ dsaswideprintk(ioc, printk(MYIOC_s_DEBUG_FMT "sas_rphy_add: rphy=%p\n", ++ ioc->name, rphy)); ++ } ++ ++ if (rphy) { ++ dsaswideprintk(ioc, dev_printk(KERN_DEBUG, ++ &rphy->dev, MYIOC_s_FMT "add:", ioc->name)); ++ dsaswideprintk(ioc, printk(MYIOC_s_DEBUG_FMT "rphy=%p release=%p\n", ++ ioc->name, rphy, rphy->dev.release)); ++ } ++} ++ + /* no mutex */ + static void + mptsas_port_delete(MPT_ADAPTER *ioc, struct mptsas_portinfo_details * port_details) +@@ -474,23 +491,6 @@ mptsas_get_rphy(struct mptsas_phyinfo *phy_info) + return NULL; + } + +-static inline void +-mptsas_set_rphy(MPT_ADAPTER *ioc, struct mptsas_phyinfo *phy_info, struct sas_rphy *rphy) +-{ +- if (phy_info->port_details) { +- phy_info->port_details->rphy = rphy; +- dsaswideprintk(ioc, printk(MYIOC_s_DEBUG_FMT "sas_rphy_add: rphy=%p\n", +- ioc->name, rphy)); +- } +- +- if (rphy) { +- dsaswideprintk(ioc, dev_printk(KERN_DEBUG, +- &rphy->dev, MYIOC_s_FMT "add:", ioc->name)); +- dsaswideprintk(ioc, printk(MYIOC_s_DEBUG_FMT "rphy=%p release=%p\n", +- ioc->name, rphy, rphy->dev.release)); +- } +-} +- + static inline struct sas_port * + mptsas_get_port(struct mptsas_phyinfo *phy_info) + { +diff --git a/drivers/message/fusion/mptscsih.c b/drivers/message/fusion/mptscsih.c +index bd096ca..332cf76 100644 +--- a/drivers/message/fusion/mptscsih.c ++++ b/drivers/message/fusion/mptscsih.c +@@ -1248,15 +1248,16 @@ mptscsih_info(struct Scsi_Host *SChost) + + h = shost_priv(SChost); + +- if (h) { +- if (h->info_kbuf == NULL) +- if ((h->info_kbuf = kmalloc(0x1000 /* 4Kb */, GFP_KERNEL)) == NULL) +- return h->info_kbuf; +- h->info_kbuf[0] = '\0'; ++ if (!h) ++ return NULL; + +- mpt_print_ioc_summary(h->ioc, h->info_kbuf, &size, 0, 0); +- h->info_kbuf[size-1] = '\0'; +- } ++ if (h->info_kbuf == NULL) ++ if ((h->info_kbuf = kmalloc(0x1000 /* 4Kb */, GFP_KERNEL)) == NULL) ++ return h->info_kbuf; ++ h->info_kbuf[0] = '\0'; ++ ++ mpt_print_ioc_summary(h->ioc, h->info_kbuf, &size, 0, 0); ++ h->info_kbuf[size-1] = '\0'; + + return h->info_kbuf; + } +diff --git a/drivers/message/i2o/i2o_config.c b/drivers/message/i2o/i2o_config.c +index efba702..59b2c0f 100644 +--- a/drivers/message/i2o/i2o_config.c ++++ b/drivers/message/i2o/i2o_config.c +@@ -787,6 +787,8 @@ static int i2o_cfg_passthru(unsigned long arg) + struct i2o_message *msg; + unsigned int iop; + ++ pax_track_stack(); ++ + if (get_user(iop, &cmd->iop) || get_user(user_msg, &cmd->msg)) + return -EFAULT; + +diff --git a/drivers/message/i2o/i2o_proc.c b/drivers/message/i2o/i2o_proc.c +index 7045c45..c07b170 100644 +--- a/drivers/message/i2o/i2o_proc.c ++++ b/drivers/message/i2o/i2o_proc.c +@@ -259,13 +259,6 @@ static char *scsi_devices[] = { + "Array Controller Device" + }; + +-static char *chtostr(u8 * chars, int n) +-{ +- char tmp[256]; +- tmp[0] = 0; +- return strncat(tmp, (char *)chars, n); +-} +- + static int i2o_report_query_status(struct seq_file *seq, int block_status, + char *group) + { +@@ -842,8 +835,7 @@ static int i2o_seq_show_ddm_table(struct seq_file *seq, void *v) + + seq_printf(seq, "%-#7x", ddm_table.i2o_vendor_id); + seq_printf(seq, "%-#8x", ddm_table.module_id); +- seq_printf(seq, "%-29s", +- chtostr(ddm_table.module_name_version, 28)); ++ seq_printf(seq, "%-.28s", ddm_table.module_name_version); + seq_printf(seq, "%9d ", ddm_table.data_size); + seq_printf(seq, "%8d", ddm_table.code_size); + +@@ -944,8 +936,8 @@ static int i2o_seq_show_drivers_stored(struct seq_file *seq, void *v) + + seq_printf(seq, "%-#7x", dst->i2o_vendor_id); + seq_printf(seq, "%-#8x", dst->module_id); +- seq_printf(seq, "%-29s", chtostr(dst->module_name_version, 28)); +- seq_printf(seq, "%-9s", chtostr(dst->date, 8)); ++ seq_printf(seq, "%-.28s", dst->module_name_version); ++ seq_printf(seq, "%-.8s", dst->date); + seq_printf(seq, "%8d ", dst->module_size); + seq_printf(seq, "%8d ", dst->mpb_size); + seq_printf(seq, "0x%04x", dst->module_flags); +@@ -1276,14 +1268,10 @@ static int i2o_seq_show_dev_identity(struct seq_file *seq, void *v) + seq_printf(seq, "Device Class : %s\n", i2o_get_class_name(work16[0])); + seq_printf(seq, "Owner TID : %0#5x\n", work16[2]); + seq_printf(seq, "Parent TID : %0#5x\n", work16[3]); +- seq_printf(seq, "Vendor info : %s\n", +- chtostr((u8 *) (work32 + 2), 16)); +- seq_printf(seq, "Product info : %s\n", +- chtostr((u8 *) (work32 + 6), 16)); +- seq_printf(seq, "Description : %s\n", +- chtostr((u8 *) (work32 + 10), 16)); +- seq_printf(seq, "Product rev. : %s\n", +- chtostr((u8 *) (work32 + 14), 8)); ++ seq_printf(seq, "Vendor info : %.16s\n", (u8 *) (work32 + 2)); ++ seq_printf(seq, "Product info : %.16s\n", (u8 *) (work32 + 6)); ++ seq_printf(seq, "Description : %.16s\n", (u8 *) (work32 + 10)); ++ seq_printf(seq, "Product rev. : %.8s\n", (u8 *) (work32 + 14)); + + seq_printf(seq, "Serial number : "); + print_serial_number(seq, (u8 *) (work32 + 16), +@@ -1328,10 +1316,8 @@ static int i2o_seq_show_ddm_identity(struct seq_file *seq, void *v) + } + + seq_printf(seq, "Registering DDM TID : 0x%03x\n", result.ddm_tid); +- seq_printf(seq, "Module name : %s\n", +- chtostr(result.module_name, 24)); +- seq_printf(seq, "Module revision : %s\n", +- chtostr(result.module_rev, 8)); ++ seq_printf(seq, "Module name : %.24s\n", result.module_name); ++ seq_printf(seq, "Module revision : %.8s\n", result.module_rev); + + seq_printf(seq, "Serial number : "); + print_serial_number(seq, result.serial_number, sizeof(result) - 36); +@@ -1362,14 +1348,10 @@ static int i2o_seq_show_uinfo(struct seq_file *seq, void *v) + return 0; + } + +- seq_printf(seq, "Device name : %s\n", +- chtostr(result.device_name, 64)); +- seq_printf(seq, "Service name : %s\n", +- chtostr(result.service_name, 64)); +- seq_printf(seq, "Physical name : %s\n", +- chtostr(result.physical_location, 64)); +- seq_printf(seq, "Instance number : %s\n", +- chtostr(result.instance_number, 4)); ++ seq_printf(seq, "Device name : %.64s\n", result.device_name); ++ seq_printf(seq, "Service name : %.64s\n", result.service_name); ++ seq_printf(seq, "Physical name : %.64s\n", result.physical_location); ++ seq_printf(seq, "Instance number : %.4s\n", result.instance_number); + + return 0; + } +diff --git a/drivers/message/i2o/iop.c b/drivers/message/i2o/iop.c +index 27cf4af..b1205b8 100644 +--- a/drivers/message/i2o/iop.c ++++ b/drivers/message/i2o/iop.c +@@ -110,10 +110,10 @@ u32 i2o_cntxt_list_add(struct i2o_controller * c, void *ptr) + + spin_lock_irqsave(&c->context_list_lock, flags); + +- if (unlikely(atomic_inc_and_test(&c->context_list_counter))) +- atomic_inc(&c->context_list_counter); ++ if (unlikely(atomic_inc_and_test_unchecked(&c->context_list_counter))) ++ atomic_inc_unchecked(&c->context_list_counter); + +- entry->context = atomic_read(&c->context_list_counter); ++ entry->context = atomic_read_unchecked(&c->context_list_counter); + + list_add(&entry->list, &c->context_list); + +@@ -1076,7 +1076,7 @@ struct i2o_controller *i2o_iop_alloc(void) + + #if BITS_PER_LONG == 64 + spin_lock_init(&c->context_list_lock); +- atomic_set(&c->context_list_counter, 0); ++ atomic_set_unchecked(&c->context_list_counter, 0); + INIT_LIST_HEAD(&c->context_list); + #endif + +diff --git a/drivers/mfd/ab3100-core.c b/drivers/mfd/ab3100-core.c +index 78e3e85..66c9a0d 100644 +--- a/drivers/mfd/ab3100-core.c ++++ b/drivers/mfd/ab3100-core.c +@@ -777,7 +777,7 @@ struct ab_family_id { + char *name; + }; + +-static const struct ab_family_id ids[] __initdata = { ++static const struct ab_family_id ids[] __initconst = { + /* AB3100 */ + { + .id = 0xc0, +diff --git a/drivers/mfd/wm8350-i2c.c b/drivers/mfd/wm8350-i2c.c +index 8d8c932..8104515 100644 +--- a/drivers/mfd/wm8350-i2c.c ++++ b/drivers/mfd/wm8350-i2c.c +@@ -43,6 +43,8 @@ static int wm8350_i2c_write_device(struct wm8350 *wm8350, char reg, + u8 msg[(WM8350_MAX_REGISTER << 1) + 1]; + int ret; + ++ pax_track_stack(); ++ + if (bytes > ((WM8350_MAX_REGISTER << 1) + 1)) + return -EINVAL; + +diff --git a/drivers/misc/kgdbts.c b/drivers/misc/kgdbts.c +index e4ff50b..4cc3f04 100644 +--- a/drivers/misc/kgdbts.c ++++ b/drivers/misc/kgdbts.c +@@ -118,7 +118,7 @@ + } while (0) + #define MAX_CONFIG_LEN 40 + +-static struct kgdb_io kgdbts_io_ops; ++static const struct kgdb_io kgdbts_io_ops; + static char get_buf[BUFMAX]; + static int get_buf_cnt; + static char put_buf[BUFMAX]; +@@ -1102,7 +1102,7 @@ static void kgdbts_post_exp_handler(void) + module_put(THIS_MODULE); + } + +-static struct kgdb_io kgdbts_io_ops = { ++static const struct kgdb_io kgdbts_io_ops = { + .name = "kgdbts", + .read_char = kgdbts_get_char, + .write_char = kgdbts_put_char, +diff --git a/drivers/misc/sgi-gru/gruhandles.c b/drivers/misc/sgi-gru/gruhandles.c +index 37e7cfc..67cfb76 100644 +--- a/drivers/misc/sgi-gru/gruhandles.c ++++ b/drivers/misc/sgi-gru/gruhandles.c +@@ -39,8 +39,8 @@ struct mcs_op_statistic mcs_op_statistics[mcsop_last]; + + static void update_mcs_stats(enum mcs_op op, unsigned long clks) + { +- atomic_long_inc(&mcs_op_statistics[op].count); +- atomic_long_add(clks, &mcs_op_statistics[op].total); ++ atomic_long_inc_unchecked(&mcs_op_statistics[op].count); ++ atomic_long_add_unchecked(clks, &mcs_op_statistics[op].total); + if (mcs_op_statistics[op].max < clks) + mcs_op_statistics[op].max = clks; + } +diff --git a/drivers/misc/sgi-gru/gruprocfs.c b/drivers/misc/sgi-gru/gruprocfs.c +index 3f2375c..467c6e6 100644 +--- a/drivers/misc/sgi-gru/gruprocfs.c ++++ b/drivers/misc/sgi-gru/gruprocfs.c +@@ -32,9 +32,9 @@ + + #define printstat(s, f) printstat_val(s, &gru_stats.f, #f) + +-static void printstat_val(struct seq_file *s, atomic_long_t *v, char *id) ++static void printstat_val(struct seq_file *s, atomic_long_unchecked_t *v, char *id) + { +- unsigned long val = atomic_long_read(v); ++ unsigned long val = atomic_long_read_unchecked(v); + + if (val) + seq_printf(s, "%16lu %s\n", val, id); +@@ -136,8 +136,8 @@ static int mcs_statistics_show(struct seq_file *s, void *p) + "cch_interrupt_sync", "cch_deallocate", "tgh_invalidate"}; + + for (op = 0; op < mcsop_last; op++) { +- count = atomic_long_read(&mcs_op_statistics[op].count); +- total = atomic_long_read(&mcs_op_statistics[op].total); ++ count = atomic_long_read_unchecked(&mcs_op_statistics[op].count); ++ total = atomic_long_read_unchecked(&mcs_op_statistics[op].total); + max = mcs_op_statistics[op].max; + seq_printf(s, "%-20s%12ld%12ld%12ld\n", id[op], count, + count ? total / count : 0, max); +diff --git a/drivers/misc/sgi-gru/grutables.h b/drivers/misc/sgi-gru/grutables.h +index 46990bc..4a251b5 100644 +--- a/drivers/misc/sgi-gru/grutables.h ++++ b/drivers/misc/sgi-gru/grutables.h +@@ -167,84 +167,84 @@ extern unsigned int gru_max_gids; + * GRU statistics. + */ + struct gru_stats_s { +- atomic_long_t vdata_alloc; +- atomic_long_t vdata_free; +- atomic_long_t gts_alloc; +- atomic_long_t gts_free; +- atomic_long_t vdata_double_alloc; +- atomic_long_t gts_double_allocate; +- atomic_long_t assign_context; +- atomic_long_t assign_context_failed; +- atomic_long_t free_context; +- atomic_long_t load_user_context; +- atomic_long_t load_kernel_context; +- atomic_long_t lock_kernel_context; +- atomic_long_t unlock_kernel_context; +- atomic_long_t steal_user_context; +- atomic_long_t steal_kernel_context; +- atomic_long_t steal_context_failed; +- atomic_long_t nopfn; +- atomic_long_t break_cow; +- atomic_long_t asid_new; +- atomic_long_t asid_next; +- atomic_long_t asid_wrap; +- atomic_long_t asid_reuse; +- atomic_long_t intr; +- atomic_long_t intr_mm_lock_failed; +- atomic_long_t call_os; +- atomic_long_t call_os_offnode_reference; +- atomic_long_t call_os_check_for_bug; +- atomic_long_t call_os_wait_queue; +- atomic_long_t user_flush_tlb; +- atomic_long_t user_unload_context; +- atomic_long_t user_exception; +- atomic_long_t set_context_option; +- atomic_long_t migrate_check; +- atomic_long_t migrated_retarget; +- atomic_long_t migrated_unload; +- atomic_long_t migrated_unload_delay; +- atomic_long_t migrated_nopfn_retarget; +- atomic_long_t migrated_nopfn_unload; +- atomic_long_t tlb_dropin; +- atomic_long_t tlb_dropin_fail_no_asid; +- atomic_long_t tlb_dropin_fail_upm; +- atomic_long_t tlb_dropin_fail_invalid; +- atomic_long_t tlb_dropin_fail_range_active; +- atomic_long_t tlb_dropin_fail_idle; +- atomic_long_t tlb_dropin_fail_fmm; +- atomic_long_t tlb_dropin_fail_no_exception; +- atomic_long_t tlb_dropin_fail_no_exception_war; +- atomic_long_t tfh_stale_on_fault; +- atomic_long_t mmu_invalidate_range; +- atomic_long_t mmu_invalidate_page; +- atomic_long_t mmu_clear_flush_young; +- atomic_long_t flush_tlb; +- atomic_long_t flush_tlb_gru; +- atomic_long_t flush_tlb_gru_tgh; +- atomic_long_t flush_tlb_gru_zero_asid; ++ atomic_long_unchecked_t vdata_alloc; ++ atomic_long_unchecked_t vdata_free; ++ atomic_long_unchecked_t gts_alloc; ++ atomic_long_unchecked_t gts_free; ++ atomic_long_unchecked_t vdata_double_alloc; ++ atomic_long_unchecked_t gts_double_allocate; ++ atomic_long_unchecked_t assign_context; ++ atomic_long_unchecked_t assign_context_failed; ++ atomic_long_unchecked_t free_context; ++ atomic_long_unchecked_t load_user_context; ++ atomic_long_unchecked_t load_kernel_context; ++ atomic_long_unchecked_t lock_kernel_context; ++ atomic_long_unchecked_t unlock_kernel_context; ++ atomic_long_unchecked_t steal_user_context; ++ atomic_long_unchecked_t steal_kernel_context; ++ atomic_long_unchecked_t steal_context_failed; ++ atomic_long_unchecked_t nopfn; ++ atomic_long_unchecked_t break_cow; ++ atomic_long_unchecked_t asid_new; ++ atomic_long_unchecked_t asid_next; ++ atomic_long_unchecked_t asid_wrap; ++ atomic_long_unchecked_t asid_reuse; ++ atomic_long_unchecked_t intr; ++ atomic_long_unchecked_t intr_mm_lock_failed; ++ atomic_long_unchecked_t call_os; ++ atomic_long_unchecked_t call_os_offnode_reference; ++ atomic_long_unchecked_t call_os_check_for_bug; ++ atomic_long_unchecked_t call_os_wait_queue; ++ atomic_long_unchecked_t user_flush_tlb; ++ atomic_long_unchecked_t user_unload_context; ++ atomic_long_unchecked_t user_exception; ++ atomic_long_unchecked_t set_context_option; ++ atomic_long_unchecked_t migrate_check; ++ atomic_long_unchecked_t migrated_retarget; ++ atomic_long_unchecked_t migrated_unload; ++ atomic_long_unchecked_t migrated_unload_delay; ++ atomic_long_unchecked_t migrated_nopfn_retarget; ++ atomic_long_unchecked_t migrated_nopfn_unload; ++ atomic_long_unchecked_t tlb_dropin; ++ atomic_long_unchecked_t tlb_dropin_fail_no_asid; ++ atomic_long_unchecked_t tlb_dropin_fail_upm; ++ atomic_long_unchecked_t tlb_dropin_fail_invalid; ++ atomic_long_unchecked_t tlb_dropin_fail_range_active; ++ atomic_long_unchecked_t tlb_dropin_fail_idle; ++ atomic_long_unchecked_t tlb_dropin_fail_fmm; ++ atomic_long_unchecked_t tlb_dropin_fail_no_exception; ++ atomic_long_unchecked_t tlb_dropin_fail_no_exception_war; ++ atomic_long_unchecked_t tfh_stale_on_fault; ++ atomic_long_unchecked_t mmu_invalidate_range; ++ atomic_long_unchecked_t mmu_invalidate_page; ++ atomic_long_unchecked_t mmu_clear_flush_young; ++ atomic_long_unchecked_t flush_tlb; ++ atomic_long_unchecked_t flush_tlb_gru; ++ atomic_long_unchecked_t flush_tlb_gru_tgh; ++ atomic_long_unchecked_t flush_tlb_gru_zero_asid; + +- atomic_long_t copy_gpa; ++ atomic_long_unchecked_t copy_gpa; + +- atomic_long_t mesq_receive; +- atomic_long_t mesq_receive_none; +- atomic_long_t mesq_send; +- atomic_long_t mesq_send_failed; +- atomic_long_t mesq_noop; +- atomic_long_t mesq_send_unexpected_error; +- atomic_long_t mesq_send_lb_overflow; +- atomic_long_t mesq_send_qlimit_reached; +- atomic_long_t mesq_send_amo_nacked; +- atomic_long_t mesq_send_put_nacked; +- atomic_long_t mesq_qf_not_full; +- atomic_long_t mesq_qf_locked; +- atomic_long_t mesq_qf_noop_not_full; +- atomic_long_t mesq_qf_switch_head_failed; +- atomic_long_t mesq_qf_unexpected_error; +- atomic_long_t mesq_noop_unexpected_error; +- atomic_long_t mesq_noop_lb_overflow; +- atomic_long_t mesq_noop_qlimit_reached; +- atomic_long_t mesq_noop_amo_nacked; +- atomic_long_t mesq_noop_put_nacked; ++ atomic_long_unchecked_t mesq_receive; ++ atomic_long_unchecked_t mesq_receive_none; ++ atomic_long_unchecked_t mesq_send; ++ atomic_long_unchecked_t mesq_send_failed; ++ atomic_long_unchecked_t mesq_noop; ++ atomic_long_unchecked_t mesq_send_unexpected_error; ++ atomic_long_unchecked_t mesq_send_lb_overflow; ++ atomic_long_unchecked_t mesq_send_qlimit_reached; ++ atomic_long_unchecked_t mesq_send_amo_nacked; ++ atomic_long_unchecked_t mesq_send_put_nacked; ++ atomic_long_unchecked_t mesq_qf_not_full; ++ atomic_long_unchecked_t mesq_qf_locked; ++ atomic_long_unchecked_t mesq_qf_noop_not_full; ++ atomic_long_unchecked_t mesq_qf_switch_head_failed; ++ atomic_long_unchecked_t mesq_qf_unexpected_error; ++ atomic_long_unchecked_t mesq_noop_unexpected_error; ++ atomic_long_unchecked_t mesq_noop_lb_overflow; ++ atomic_long_unchecked_t mesq_noop_qlimit_reached; ++ atomic_long_unchecked_t mesq_noop_amo_nacked; ++ atomic_long_unchecked_t mesq_noop_put_nacked; + + }; + +@@ -252,8 +252,8 @@ enum mcs_op {cchop_allocate, cchop_start, cchop_interrupt, cchop_interrupt_sync, + cchop_deallocate, tghop_invalidate, mcsop_last}; + + struct mcs_op_statistic { +- atomic_long_t count; +- atomic_long_t total; ++ atomic_long_unchecked_t count; ++ atomic_long_unchecked_t total; + unsigned long max; + }; + +@@ -276,7 +276,7 @@ extern struct mcs_op_statistic mcs_op_statistics[mcsop_last]; + + #define STAT(id) do { \ + if (gru_options & OPT_STATS) \ +- atomic_long_inc(&gru_stats.id); \ ++ atomic_long_inc_unchecked(&gru_stats.id); \ + } while (0) + + #ifdef CONFIG_SGI_GRU_DEBUG +diff --git a/drivers/misc/sgi-xp/xp.h b/drivers/misc/sgi-xp/xp.h +index 2275126..12a9dbfb 100644 +--- a/drivers/misc/sgi-xp/xp.h ++++ b/drivers/misc/sgi-xp/xp.h +@@ -289,7 +289,7 @@ struct xpc_interface { + xpc_notify_func, void *); + void (*received) (short, int, void *); + enum xp_retval (*partid_to_nasids) (short, void *); +-}; ++} __no_const; + + extern struct xpc_interface xpc_interface; + +diff --git a/drivers/misc/sgi-xp/xpc.h b/drivers/misc/sgi-xp/xpc.h +index b94d5f7..7f494c5 100644 +--- a/drivers/misc/sgi-xp/xpc.h ++++ b/drivers/misc/sgi-xp/xpc.h +@@ -835,6 +835,7 @@ struct xpc_arch_operations { + void (*received_payload) (struct xpc_channel *, void *); + void (*notify_senders_of_disconnect) (struct xpc_channel *); + }; ++typedef struct xpc_arch_operations __no_const xpc_arch_operations_no_const; + + /* struct xpc_partition act_state values (for XPC HB) */ + +@@ -876,7 +877,7 @@ extern struct xpc_registration xpc_registrations[]; + /* found in xpc_main.c */ + extern struct device *xpc_part; + extern struct device *xpc_chan; +-extern struct xpc_arch_operations xpc_arch_ops; ++extern xpc_arch_operations_no_const xpc_arch_ops; + extern int xpc_disengage_timelimit; + extern int xpc_disengage_timedout; + extern int xpc_activate_IRQ_rcvd; +diff --git a/drivers/misc/sgi-xp/xpc_main.c b/drivers/misc/sgi-xp/xpc_main.c +index fd3688a..7e211a4 100644 +--- a/drivers/misc/sgi-xp/xpc_main.c ++++ b/drivers/misc/sgi-xp/xpc_main.c +@@ -169,7 +169,7 @@ static struct notifier_block xpc_die_notifier = { + .notifier_call = xpc_system_die, + }; + +-struct xpc_arch_operations xpc_arch_ops; ++xpc_arch_operations_no_const xpc_arch_ops; + + /* + * Timer function to enforce the timelimit on the partition disengage. +diff --git a/drivers/misc/sgi-xp/xpc_sn2.c b/drivers/misc/sgi-xp/xpc_sn2.c +index 8b70e03..700bda6 100644 +--- a/drivers/misc/sgi-xp/xpc_sn2.c ++++ b/drivers/misc/sgi-xp/xpc_sn2.c +@@ -2350,7 +2350,7 @@ xpc_received_payload_sn2(struct xpc_channel *ch, void *payload) + xpc_acknowledge_msgs_sn2(ch, get, msg->flags); + } + +-static struct xpc_arch_operations xpc_arch_ops_sn2 = { ++static const struct xpc_arch_operations xpc_arch_ops_sn2 = { + .setup_partitions = xpc_setup_partitions_sn2, + .teardown_partitions = xpc_teardown_partitions_sn2, + .process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_sn2, +@@ -2413,7 +2413,9 @@ xpc_init_sn2(void) + int ret; + size_t buf_size; + +- xpc_arch_ops = xpc_arch_ops_sn2; ++ pax_open_kernel(); ++ memcpy((void *)&xpc_arch_ops, &xpc_arch_ops_sn2, sizeof(xpc_arch_ops_sn2)); ++ pax_close_kernel(); + + if (offsetof(struct xpc_msg_sn2, payload) > XPC_MSG_HDR_MAX_SIZE) { + dev_err(xpc_part, "header portion of struct xpc_msg_sn2 is " +diff --git a/drivers/misc/sgi-xp/xpc_uv.c b/drivers/misc/sgi-xp/xpc_uv.c +index 8e08d71..7cb8c9b 100644 +--- a/drivers/misc/sgi-xp/xpc_uv.c ++++ b/drivers/misc/sgi-xp/xpc_uv.c +@@ -1669,7 +1669,7 @@ xpc_received_payload_uv(struct xpc_channel *ch, void *payload) + XPC_DEACTIVATE_PARTITION(&xpc_partitions[ch->partid], ret); + } + +-static struct xpc_arch_operations xpc_arch_ops_uv = { ++static const struct xpc_arch_operations xpc_arch_ops_uv = { + .setup_partitions = xpc_setup_partitions_uv, + .teardown_partitions = xpc_teardown_partitions_uv, + .process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_uv, +@@ -1729,7 +1729,9 @@ static struct xpc_arch_operations xpc_arch_ops_uv = { + int + xpc_init_uv(void) + { +- xpc_arch_ops = xpc_arch_ops_uv; ++ pax_open_kernel(); ++ memcpy((void *)&xpc_arch_ops, &xpc_arch_ops_uv, sizeof(xpc_arch_ops_uv)); ++ pax_close_kernel(); + + if (sizeof(struct xpc_notify_mq_msghdr_uv) > XPC_MSG_HDR_MAX_SIZE) { + dev_err(xpc_part, "xpc_notify_mq_msghdr_uv is larger than %d\n", +diff --git a/drivers/mmc/host/sdhci-pci.c b/drivers/mmc/host/sdhci-pci.c +index 6fd20b42..650efe3 100644 +--- a/drivers/mmc/host/sdhci-pci.c ++++ b/drivers/mmc/host/sdhci-pci.c +@@ -297,7 +297,7 @@ static const struct sdhci_pci_fixes sdhci_via = { + .probe = via_probe, + }; + +-static const struct pci_device_id pci_ids[] __devinitdata = { ++static const struct pci_device_id pci_ids[] __devinitconst = { + { + .vendor = PCI_VENDOR_ID_RICOH, + .device = PCI_DEVICE_ID_RICOH_R5C822, +diff --git a/drivers/mtd/chips/cfi_cmdset_0001.c b/drivers/mtd/chips/cfi_cmdset_0001.c +index e7563a9..5f90ce5 100644 +--- a/drivers/mtd/chips/cfi_cmdset_0001.c ++++ b/drivers/mtd/chips/cfi_cmdset_0001.c +@@ -743,6 +743,8 @@ static int chip_ready (struct map_info *map, struct flchip *chip, unsigned long + struct cfi_pri_intelext *cfip = cfi->cmdset_priv; + unsigned long timeo = jiffies + HZ; + ++ pax_track_stack(); ++ + /* Prevent setting state FL_SYNCING for chip in suspended state. */ + if (mode == FL_SYNCING && chip->oldstate != FL_READY) + goto sleep; +@@ -1642,6 +1644,8 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip, + unsigned long initial_adr; + int initial_len = len; + ++ pax_track_stack(); ++ + wbufsize = cfi_interleave(cfi) << cfi->cfiq->MaxBufWriteSize; + adr += chip->start; + initial_adr = adr; +@@ -1860,6 +1864,8 @@ static int __xipram do_erase_oneblock(struct map_info *map, struct flchip *chip, + int retries = 3; + int ret; + ++ pax_track_stack(); ++ + adr += chip->start; + + retry: +diff --git a/drivers/mtd/chips/cfi_cmdset_0020.c b/drivers/mtd/chips/cfi_cmdset_0020.c +index 0667a67..3ab97ed 100644 +--- a/drivers/mtd/chips/cfi_cmdset_0020.c ++++ b/drivers/mtd/chips/cfi_cmdset_0020.c +@@ -255,6 +255,8 @@ static inline int do_read_onechip(struct map_info *map, struct flchip *chip, lof + unsigned long cmd_addr; + struct cfi_private *cfi = map->fldrv_priv; + ++ pax_track_stack(); ++ + adr += chip->start; + + /* Ensure cmd read/writes are aligned. */ +@@ -428,6 +430,8 @@ static inline int do_write_buffer(struct map_info *map, struct flchip *chip, + DECLARE_WAITQUEUE(wait, current); + int wbufsize, z; + ++ pax_track_stack(); ++ + /* M58LW064A requires bus alignment for buffer wriets -- saw */ + if (adr & (map_bankwidth(map)-1)) + return -EINVAL; +@@ -742,6 +746,8 @@ static inline int do_erase_oneblock(struct map_info *map, struct flchip *chip, u + DECLARE_WAITQUEUE(wait, current); + int ret = 0; + ++ pax_track_stack(); ++ + adr += chip->start; + + /* Let's determine this according to the interleave only once */ +@@ -1047,6 +1053,8 @@ static inline int do_lock_oneblock(struct map_info *map, struct flchip *chip, un + unsigned long timeo = jiffies + HZ; + DECLARE_WAITQUEUE(wait, current); + ++ pax_track_stack(); ++ + adr += chip->start; + + /* Let's determine this according to the interleave only once */ +@@ -1196,6 +1204,8 @@ static inline int do_unlock_oneblock(struct map_info *map, struct flchip *chip, + unsigned long timeo = jiffies + HZ; + DECLARE_WAITQUEUE(wait, current); + ++ pax_track_stack(); ++ + adr += chip->start; + + /* Let's determine this according to the interleave only once */ +diff --git a/drivers/mtd/devices/doc2000.c b/drivers/mtd/devices/doc2000.c +index 5bf5f46..c5de373 100644 +--- a/drivers/mtd/devices/doc2000.c ++++ b/drivers/mtd/devices/doc2000.c +@@ -776,7 +776,7 @@ static int doc_write(struct mtd_info *mtd, loff_t to, size_t len, + + /* The ECC will not be calculated correctly if less than 512 is written */ + /* DBB- +- if (len != 0x200 && eccbuf) ++ if (len != 0x200) + printk(KERN_WARNING + "ECC needs a full sector write (adr: %lx size %lx)\n", + (long) to, (long) len); +diff --git a/drivers/mtd/devices/doc2001.c b/drivers/mtd/devices/doc2001.c +index 0990f78..bb4e8a4 100644 +--- a/drivers/mtd/devices/doc2001.c ++++ b/drivers/mtd/devices/doc2001.c +@@ -393,7 +393,7 @@ static int doc_read (struct mtd_info *mtd, loff_t from, size_t len, + struct Nand *mychip = &this->chips[from >> (this->chipshift)]; + + /* Don't allow read past end of device */ +- if (from >= this->totlen) ++ if (from >= this->totlen || !len) + return -EINVAL; + + /* Don't allow a single read to cross a 512-byte block boundary */ +diff --git a/drivers/mtd/ftl.c b/drivers/mtd/ftl.c +index e56d6b4..f07e6cf 100644 +--- a/drivers/mtd/ftl.c ++++ b/drivers/mtd/ftl.c +@@ -474,6 +474,8 @@ static int copy_erase_unit(partition_t *part, uint16_t srcunit, + loff_t offset; + uint16_t srcunitswap = cpu_to_le16(srcunit); + ++ pax_track_stack(); ++ + eun = &part->EUNInfo[srcunit]; + xfer = &part->XferInfo[xferunit]; + DEBUG(2, "ftl_cs: copying block 0x%x to 0x%x\n", +diff --git a/drivers/mtd/inftlcore.c b/drivers/mtd/inftlcore.c +index 8aca552..146446e 100755 +--- a/drivers/mtd/inftlcore.c ++++ b/drivers/mtd/inftlcore.c +@@ -260,6 +260,8 @@ static u16 INFTL_foldchain(struct INFTLrecord *inftl, unsigned thisVUC, unsigned + struct inftl_oob oob; + size_t retlen; + ++ pax_track_stack(); ++ + DEBUG(MTD_DEBUG_LEVEL3, "INFTL: INFTL_foldchain(inftl=%p,thisVUC=%d," + "pending=%d)\n", inftl, thisVUC, pendingblock); + +diff --git a/drivers/mtd/inftlmount.c b/drivers/mtd/inftlmount.c +index 32e82ae..ed50953 100644 +--- a/drivers/mtd/inftlmount.c ++++ b/drivers/mtd/inftlmount.c +@@ -54,6 +54,8 @@ static int find_boot_record(struct INFTLrecord *inftl) + struct INFTLPartition *ip; + size_t retlen; + ++ pax_track_stack(); ++ + DEBUG(MTD_DEBUG_LEVEL3, "INFTL: find_boot_record(inftl=%p)\n", inftl); + + /* +diff --git a/drivers/mtd/lpddr/qinfo_probe.c b/drivers/mtd/lpddr/qinfo_probe.c +index 79bf40f..fe5f8fd 100644 +--- a/drivers/mtd/lpddr/qinfo_probe.c ++++ b/drivers/mtd/lpddr/qinfo_probe.c +@@ -106,6 +106,8 @@ static int lpddr_pfow_present(struct map_info *map, struct lpddr_private *lpddr) + { + map_word pfow_val[4]; + ++ pax_track_stack(); ++ + /* Check identification string */ + pfow_val[0] = map_read(map, map->pfow_base + PFOW_QUERY_STRING_P); + pfow_val[1] = map_read(map, map->pfow_base + PFOW_QUERY_STRING_F); +diff --git a/drivers/mtd/mtdchar.c b/drivers/mtd/mtdchar.c +index 726a1b8..f46b460 100644 +--- a/drivers/mtd/mtdchar.c ++++ b/drivers/mtd/mtdchar.c +@@ -461,6 +461,8 @@ static int mtd_ioctl(struct inode *inode, struct file *file, + u_long size; + struct mtd_info_user info; + ++ pax_track_stack(); ++ + DEBUG(MTD_DEBUG_LEVEL0, "MTD_ioctl\n"); + + size = (cmd & IOCSIZE_MASK) >> IOCSIZE_SHIFT; +diff --git a/drivers/mtd/nftlcore.c b/drivers/mtd/nftlcore.c +index 1002e18..26d82d5 100644 +--- a/drivers/mtd/nftlcore.c ++++ b/drivers/mtd/nftlcore.c +@@ -254,6 +254,8 @@ static u16 NFTL_foldchain (struct NFTLrecord *nftl, unsigned thisVUC, unsigned p + int inplace = 1; + size_t retlen; + ++ pax_track_stack(); ++ + memset(BlockMap, 0xff, sizeof(BlockMap)); + memset(BlockFreeFound, 0, sizeof(BlockFreeFound)); + +diff --git a/drivers/mtd/nftlmount.c b/drivers/mtd/nftlmount.c +index 8b22b18..6fada85 100644 +--- a/drivers/mtd/nftlmount.c ++++ b/drivers/mtd/nftlmount.c +@@ -23,6 +23,7 @@ + #include + #include + #include ++#include + #include + #include + #include +@@ -44,6 +45,8 @@ static int find_boot_record(struct NFTLrecord *nftl) + struct mtd_info *mtd = nftl->mbd.mtd; + unsigned int i; + ++ pax_track_stack(); ++ + /* Assume logical EraseSize == physical erasesize for starting the scan. + We'll sort it out later if we find a MediaHeader which says otherwise */ + /* Actually, we won't. The new DiskOnChip driver has already scanned +diff --git a/drivers/mtd/ubi/build.c b/drivers/mtd/ubi/build.c +index 14cec04..d775b87 100644 +--- a/drivers/mtd/ubi/build.c ++++ b/drivers/mtd/ubi/build.c +@@ -1255,7 +1255,7 @@ module_exit(ubi_exit); + static int __init bytes_str_to_int(const char *str) + { + char *endp; +- unsigned long result; ++ unsigned long result, scale = 1; + + result = simple_strtoul(str, &endp, 0); + if (str == endp || result >= INT_MAX) { +@@ -1266,11 +1266,11 @@ static int __init bytes_str_to_int(const char *str) + + switch (*endp) { + case 'G': +- result *= 1024; ++ scale *= 1024; + case 'M': +- result *= 1024; ++ scale *= 1024; + case 'K': +- result *= 1024; ++ scale *= 1024; + if (endp[1] == 'i' && endp[2] == 'B') + endp += 2; + case '\0': +@@ -1281,7 +1281,13 @@ static int __init bytes_str_to_int(const char *str) + return -EINVAL; + } + +- return result; ++ if ((intoverflow_t)result*scale >= INT_MAX) { ++ printk(KERN_ERR "UBI error: incorrect bytes count: \"%s\"\n", ++ str); ++ return -EINVAL; ++ } ++ ++ return result*scale; + } + + /** +diff --git a/drivers/net/atlx/atl2.c b/drivers/net/atlx/atl2.c +index ab68886..ca405e8 100644 +--- a/drivers/net/atlx/atl2.c ++++ b/drivers/net/atlx/atl2.c +@@ -2845,7 +2845,7 @@ static void atl2_force_ps(struct atl2_hw *hw) + */ + + #define ATL2_PARAM(X, desc) \ +- static const int __devinitdata X[ATL2_MAX_NIC + 1] = ATL2_PARAM_INIT; \ ++ static const int __devinitconst X[ATL2_MAX_NIC + 1] = ATL2_PARAM_INIT; \ + MODULE_PARM(X, "1-" __MODULE_STRING(ATL2_MAX_NIC) "i"); \ + MODULE_PARM_DESC(X, desc); + #else +diff --git a/drivers/net/benet/Makefile b/drivers/net/benet/Makefile +index a60cd80..0ed11ef 100644 +--- a/drivers/net/benet/Makefile ++++ b/drivers/net/benet/Makefile +@@ -1,7 +1,9 @@ + # +-# Makefile to build the network driver for ServerEngine's BladeEngine. ++# Makefile to build the be2net network driver + # + ++EXTRA_CFLAGS += -DCONFIG_PALAU ++ + obj-$(CONFIG_BE2NET) += be2net.o + +-be2net-y := be_main.o be_cmds.o be_ethtool.o ++be2net-y := be_main.o be_cmds.o be_ethtool.o be_compat.o be_misc.o +diff --git a/drivers/net/benet/be.h b/drivers/net/benet/be.h +index 5c74ff0..7382603 100644 +--- a/drivers/net/benet/be.h ++++ b/drivers/net/benet/be.h +@@ -1,18 +1,18 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + + #ifndef BE_H +@@ -29,32 +29,53 @@ + #include + #include + #include ++#include ++#ifndef CONFIG_PALAU ++#include ++#endif + ++#ifdef CONFIG_PALAU ++#include "be_compat.h" ++#endif + #include "be_hw.h" + +-#define DRV_VER "2.101.205" ++#ifdef CONFIG_PALAU ++#include "version.h" ++#define DRV_VER STR_BE_MAJOR "." STR_BE_MINOR "."\ ++ STR_BE_BUILD "." STR_BE_BRANCH ++#else ++#define DRV_VER "2.0.348" ++#endif + #define DRV_NAME "be2net" +-#define BE_NAME "ServerEngines BladeEngine2 10Gbps NIC" +-#define BE3_NAME "ServerEngines BladeEngine3 10Gbps NIC" +-#define OC_NAME "Emulex OneConnect 10Gbps NIC" +-#define OC_NAME1 "Emulex OneConnect 10Gbps NIC (be3)" +-#define DRV_DESC BE_NAME "Driver" ++#define BE_NAME "Emulex BladeEngine2" ++#define BE3_NAME "Emulex BladeEngine3" ++#define OC_NAME "Emulex OneConnect" ++#define OC_NAME_BE OC_NAME "(be3)" ++#define OC_NAME_LANCER OC_NAME "(Lancer)" ++#define DRV_DESC "Emulex OneConnect 10Gbps NIC Driver" + +-#define BE_VENDOR_ID 0x19a2 ++#define BE_VENDOR_ID 0x19a2 ++#define EMULEX_VENDOR_ID 0x10df + #define BE_DEVICE_ID1 0x211 + #define BE_DEVICE_ID2 0x221 +-#define OC_DEVICE_ID1 0x700 +-#define OC_DEVICE_ID2 0x701 +-#define OC_DEVICE_ID3 0x710 ++#define OC_DEVICE_ID1 0x700 /* Device Id for BE2 cards */ ++#define OC_DEVICE_ID2 0x710 /* Device Id for BE3 cards */ ++#define OC_DEVICE_ID3 0xe220 /* Device id for Lancer cards */ ++ ++#define OC_SUBSYS_DEVICE_ID1 0xE602 ++#define OC_SUBSYS_DEVICE_ID2 0xE642 ++#define OC_SUBSYS_DEVICE_ID3 0xE612 ++#define OC_SUBSYS_DEVICE_ID4 0xE652 + + static inline char *nic_name(struct pci_dev *pdev) + { + switch (pdev->device) { + case OC_DEVICE_ID1: +- case OC_DEVICE_ID2: + return OC_NAME; ++ case OC_DEVICE_ID2: ++ return OC_NAME_BE; + case OC_DEVICE_ID3: +- return OC_NAME1; ++ return OC_NAME_LANCER; + case BE_DEVICE_ID2: + return BE3_NAME; + default: +@@ -63,7 +84,7 @@ static inline char *nic_name(struct pci_dev *pdev) + } + + /* Number of bytes of an RX frame that are copied to skb->data */ +-#define BE_HDR_LEN 64 ++#define BE_HDR_LEN ((u16) 64) + #define BE_MAX_JUMBO_FRAME_SIZE 9018 + #define BE_MIN_MTU 256 + +@@ -79,10 +100,24 @@ static inline char *nic_name(struct pci_dev *pdev) + #define MCC_Q_LEN 128 /* total size not to exceed 8 pages */ + #define MCC_CQ_LEN 256 + ++#define MAX_RSS_QS 4 /* BE limit is 4 queues/port */ ++ ++#define MAX_RX_QS (MAX_RSS_QS + 1) ++ ++#ifdef MQ_TX ++#define MAX_TX_QS 8 ++#else ++#define MAX_TX_QS 1 ++#endif ++ ++#define BE_MAX_MSIX_VECTORS (MAX_RX_QS + 1)/* RSS qs + 1 def Rx + Tx */ + #define BE_NAPI_WEIGHT 64 +-#define MAX_RX_POST BE_NAPI_WEIGHT /* Frags posted at a time */ ++#define MAX_RX_POST BE_NAPI_WEIGHT /* Frags posted at a time */ + #define RX_FRAGS_REFILL_WM (RX_Q_LEN - MAX_RX_POST) + ++#define BE_MAX_LRO_DESCRIPTORS 16 ++#define BE_MAX_FRAGS_PER_FRAME (min((u32) 16, (u32) MAX_SKB_FRAGS)) ++ + #define FW_VER_LEN 32 + + struct be_dma_mem { +@@ -127,6 +162,11 @@ static inline void *queue_tail_node(struct be_queue_info *q) + return q->dma_mem.va + q->tail * q->entry_size; + } + ++static inline void *queue_index_node(struct be_queue_info *q, u16 index) ++{ ++ return q->dma_mem.va + index * q->entry_size; ++} ++ + static inline void queue_head_inc(struct be_queue_info *q) + { + index_inc(&q->head, q->len); +@@ -137,6 +177,7 @@ static inline void queue_tail_inc(struct be_queue_info *q) + index_inc(&q->tail, q->len); + } + ++ + struct be_eq_obj { + struct be_queue_info q; + char desc[32]; +@@ -146,6 +187,7 @@ struct be_eq_obj { + u16 min_eqd; /* in usecs */ + u16 max_eqd; /* in usecs */ + u16 cur_eqd; /* in usecs */ ++ u8 eq_idx; + + struct napi_struct napi; + }; +@@ -153,49 +195,20 @@ struct be_eq_obj { + struct be_mcc_obj { + struct be_queue_info q; + struct be_queue_info cq; ++ bool rearm_cq; + }; + +-struct be_drvr_stats { ++struct be_tx_stats { + u32 be_tx_reqs; /* number of TX requests initiated */ + u32 be_tx_stops; /* number of times TX Q was stopped */ +- u32 be_fwd_reqs; /* number of send reqs through forwarding i/f */ + u32 be_tx_wrbs; /* number of tx WRBs used */ +- u32 be_tx_events; /* number of tx completion events */ + u32 be_tx_compl; /* number of tx completion entries processed */ + ulong be_tx_jiffies; + u64 be_tx_bytes; + u64 be_tx_bytes_prev; + u64 be_tx_pkts; + u32 be_tx_rate; +- +- u32 cache_barrier[16]; +- +- u32 be_ethrx_post_fail;/* number of ethrx buffer alloc failures */ +- u32 be_polls; /* number of times NAPI called poll function */ +- u32 be_rx_events; /* number of ucast rx completion events */ +- u32 be_rx_compl; /* number of rx completion entries processed */ +- ulong be_rx_jiffies; +- u64 be_rx_bytes; +- u64 be_rx_bytes_prev; +- u64 be_rx_pkts; +- u32 be_rx_rate; +- /* number of non ether type II frames dropped where +- * frame len > length field of Mac Hdr */ +- u32 be_802_3_dropped_frames; +- /* number of non ether type II frames malformed where +- * in frame len < length field of Mac Hdr */ +- u32 be_802_3_malformed_frames; +- u32 be_rxcp_err; /* Num rx completion entries w/ err set. */ +- ulong rx_fps_jiffies; /* jiffies at last FPS calc */ +- u32 be_rx_frags; +- u32 be_prev_rx_frags; +- u32 be_rx_fps; /* Rx frags per second */ +-}; +- +-struct be_stats_obj { +- struct be_drvr_stats drvr_stats; +- struct net_device_stats net_stats; +- struct be_dma_mem cmd; ++ u32 be_ipv6_ext_hdr_tx_drop; + }; + + struct be_tx_obj { +@@ -203,23 +216,124 @@ struct be_tx_obj { + struct be_queue_info cq; + /* Remember the skbs that were transmitted */ + struct sk_buff *sent_skb_list[TX_Q_LEN]; ++ struct be_tx_stats stats; + }; + + /* Struct to remember the pages posted for rx frags */ + struct be_rx_page_info { + struct page *page; +- dma_addr_t bus; ++ DEFINE_DMA_UNMAP_ADDR(bus); + u16 page_offset; + bool last_page_user; + }; + ++struct be_rx_stats { ++ u32 rx_post_fail;/* number of ethrx buffer alloc failures */ ++ u32 rx_polls; /* number of times NAPI called poll function */ ++ u32 rx_events; /* number of ucast rx completion events */ ++ u32 rx_compl; /* number of rx completion entries processed */ ++ ulong rx_jiffies; ++ u64 rx_bytes; ++ u64 rx_bytes_prev; ++ u64 rx_pkts; ++ u32 rx_rate; ++ u32 rx_mcast_pkts; ++ u32 rxcp_err; /* Num rx completion entries w/ err set. */ ++ ulong rx_fps_jiffies; /* jiffies at last FPS calc */ ++ u32 rx_frags; ++ u32 prev_rx_frags; ++ u32 rx_fps; /* Rx frags per second */ ++ u32 rx_drops_no_frags; ++}; ++ ++struct be_rx_compl_info { ++ u32 rss_hash; ++ u16 vlan_tag; ++ u16 pkt_size; ++ u16 rxq_idx; ++ u16 port; ++ u8 vlanf; ++ u8 num_rcvd; ++ u8 err; ++ u8 ipf; ++ u8 tcpf; ++ u8 udpf; ++ u8 ip_csum; ++ u8 l4_csum; ++ u8 ipv6; ++ u8 vtm; ++ u8 pkt_type; ++}; ++ + struct be_rx_obj { ++ struct be_adapter *adapter; + struct be_queue_info q; + struct be_queue_info cq; +- struct be_rx_page_info page_info_tbl[RX_Q_LEN]; ++ struct be_rx_compl_info rxcp; ++ struct be_rx_page_info *page_info_tbl; ++ struct net_lro_mgr lro_mgr; ++ struct net_lro_desc lro_desc[BE_MAX_LRO_DESCRIPTORS]; ++ struct be_eq_obj rx_eq; ++ struct be_rx_stats stats; ++ u8 rss_id; ++ bool rx_post_starved; /* Zero rx frags have been posted to BE */ ++ u16 prev_frag_idx; ++ u32 cache_line_barrier[16]; + }; + +-#define BE_NUM_MSIX_VECTORS 2 /* 1 each for Tx and Rx */ ++struct be_drv_stats { ++ u32 be_on_die_temperature; ++ u32 be_tx_events; ++ u32 eth_red_drops; ++ u32 rx_drops_no_pbuf; ++ u32 rx_drops_no_txpb; ++ u32 rx_drops_no_erx_descr; ++ u32 rx_drops_no_tpre_descr; ++ u32 rx_drops_too_many_frags; ++ u32 rx_drops_invalid_ring; ++ u32 forwarded_packets; ++ u32 rx_drops_mtu; ++ u32 rx_crc_errors; ++ u32 rx_alignment_symbol_errors; ++ u32 rx_pause_frames; ++ u32 rx_priority_pause_frames; ++ u32 rx_control_frames; ++ u32 rx_in_range_errors; ++ u32 rx_out_range_errors; ++ u32 rx_frame_too_long; ++ u32 rx_address_match_errors; ++ u32 rx_dropped_too_small; ++ u32 rx_dropped_too_short; ++ u32 rx_dropped_header_too_small; ++ u32 rx_dropped_tcp_length; ++ u32 rx_dropped_runt; ++ u32 rx_ip_checksum_errs; ++ u32 rx_tcp_checksum_errs; ++ u32 rx_udp_checksum_errs; ++ u32 rx_switched_unicast_packets; ++ u32 rx_switched_multicast_packets; ++ u32 rx_switched_broadcast_packets; ++ u32 tx_pauseframes; ++ u32 tx_priority_pauseframes; ++ u32 tx_controlframes; ++ u32 rxpp_fifo_overflow_drop; ++ u32 rx_input_fifo_overflow_drop; ++ u32 pmem_fifo_overflow_drop; ++ u32 jabber_events; ++}; ++ ++struct be_vf_cfg { ++ unsigned char vf_mac_addr[ETH_ALEN]; ++ u32 vf_if_handle; ++ u32 vf_pmac_id; ++ u16 vf_def_vid; ++ u16 vf_vlan_tag; ++ u32 vf_tx_rate; ++}; ++ ++#define BE_INVALID_PMAC_ID 0xffffffff ++#define BE_FLAGS_DCBX (1 << 16) ++ + struct be_adapter { + struct pci_dev *pdev; + struct net_device *netdev; +@@ -228,7 +342,7 @@ struct be_adapter { + u8 __iomem *db; /* Door Bell */ + u8 __iomem *pcicfg; /* PCI config space */ + +- spinlock_t mbox_lock; /* For serializing mbox cmds to BE card */ ++ struct mutex mbox_lock; /* For serializing mbox cmds to BE card */ + struct be_dma_mem mbox_mem; + /* Mbox mem is adjusted to align to 16 bytes. The allocated addr + * is stored for freeing purpose */ +@@ -238,66 +352,121 @@ struct be_adapter { + spinlock_t mcc_lock; /* For serializing mcc cmds to BE card */ + spinlock_t mcc_cq_lock; + +- struct msix_entry msix_entries[BE_NUM_MSIX_VECTORS]; +- bool msix_enabled; ++ struct msix_entry msix_entries[BE_MAX_MSIX_VECTORS]; ++ u32 num_msix_vec; + bool isr_registered; + + /* TX Rings */ + struct be_eq_obj tx_eq; +- struct be_tx_obj tx_obj; ++ struct be_tx_obj tx_obj[MAX_TX_QS]; ++ u8 num_tx_qs; ++ u8 prio_tc_map[MAX_TX_QS]; /* prio_tc_map[prio] => tc-id */ ++ u8 tc_txq_map[MAX_TX_QS]; /* tc_txq_map[tc-id] => txq index */ + + u32 cache_line_break[8]; + + /* Rx rings */ +- struct be_eq_obj rx_eq; +- struct be_rx_obj rx_obj; ++ struct be_rx_obj rx_obj[MAX_RX_QS]; /* one default non-rss Q */ ++ u32 num_rx_qs; ++ ++ struct be_dma_mem stats_cmd; ++ struct net_device_stats net_stats; ++ struct be_drv_stats drv_stats; + u32 big_page_size; /* Compounded page size shared by rx wrbs */ +- bool rx_post_starved; /* Zero rx frags have been posted to BE */ + + struct vlan_group *vlan_grp; +- u16 num_vlans; ++ u16 vlans_added; ++ u16 max_vlans; /* Number of vlans supported */ + u8 vlan_tag[VLAN_GROUP_ARRAY_LEN]; ++ u8 vlan_prio_bmap; /* Available priority BitMap */ ++ u16 recommended_prio; /* Recommended Priority */ ++ struct be_dma_mem rx_filter; + +- struct be_stats_obj stats; + /* Work queue used to perform periodic tasks like getting statistics */ + struct delayed_work work; ++ u16 work_counter; + +- /* Ethtool knobs and info */ +- bool rx_csum; /* BE card must perform rx-checksumming */ ++ u32 flags; ++ bool rx_csum; /* BE card must perform rx-checksumming */ ++ u32 max_rx_coal; + char fw_ver[FW_VER_LEN]; + u32 if_handle; /* Used to configure filtering */ + u32 pmac_id; /* MAC addr handle used by BE card */ ++ u32 beacon_state; /* for set_phys_id */ + +- bool link_up; ++ bool eeh_err; ++ int link_status; + u32 port_num; ++ u32 hba_port_num; + bool promiscuous; +- u32 cap; ++ bool wol; ++ u32 function_mode; ++ u32 function_caps; + u32 rx_fc; /* Rx flow control */ + u32 tx_fc; /* Tx flow control */ ++ bool ue_detected; ++ bool stats_cmd_sent; ++ bool gro_supported; ++ int link_speed; ++ u8 port_type; ++ u8 transceiver; ++ u8 autoneg; + u8 generation; /* BladeEngine ASIC generation */ ++ u32 flash_status; ++ struct completion flash_compl; ++ ++ u8 eq_next_idx; ++ bool be3_native; ++ u16 num_vfs; ++ struct be_vf_cfg *vf_cfg; ++ u8 is_virtfn; ++ u16 pvid; ++ u32 sli_family; ++ u8 port_name[4]; ++ char model_number[32]; + }; + + /* BladeEngine Generation numbers */ + #define BE_GEN2 2 + #define BE_GEN3 3 + +-extern const struct ethtool_ops be_ethtool_ops; ++#define ON 1 ++#define OFF 0 ++#define lancer_chip(adapter) (adapter->pdev->device == OC_DEVICE_ID3) ++#define lancer_A0_chip(adapter) \ ++ (adapter->sli_family == LANCER_A0_SLI_FAMILY) + +-#define drvr_stats(adapter) (&adapter->stats.drvr_stats) ++extern struct ethtool_ops be_ethtool_ops; + +-static inline unsigned int be_pci_func(struct be_adapter *adapter) +-{ +- return PCI_FUNC(adapter->pdev->devfn); +-} ++#define msix_enabled(adapter) (adapter->num_msix_vec > 0) ++#define tx_stats(txo) (&txo->stats) ++#define rx_stats(rxo) (&rxo->stats) + ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 29) ++#define BE_SET_NETDEV_OPS(netdev, ops) be_netdev_ops_init(netdev, ops) ++#else + #define BE_SET_NETDEV_OPS(netdev, ops) (netdev->netdev_ops = ops) ++#endif ++ ++#define for_all_rx_queues(adapter, rxo, i) \ ++ for (i = 0, rxo = &adapter->rx_obj[i]; i < adapter->num_rx_qs; \ ++ i++, rxo++) ++ ++/* Just skip the first default non-rss queue */ ++#define for_all_rss_queues(adapter, rxo, i) \ ++ for (i = 0, rxo = &adapter->rx_obj[i+1]; i < (adapter->num_rx_qs - 1);\ ++ i++, rxo++) ++ ++#define for_all_tx_queues(adapter, txo, i) \ ++ for (i = 0, txo = &adapter->tx_obj[i]; i < adapter->num_tx_qs; \ ++ i++, txo++) + + #define PAGE_SHIFT_4K 12 + #define PAGE_SIZE_4K (1 << PAGE_SHIFT_4K) + + /* Returns number of pages spanned by the data starting at the given addr */ +-#define PAGES_4K_SPANNED(_address, size) \ +- ((u32)((((size_t)(_address) & (PAGE_SIZE_4K - 1)) + \ ++#define PAGES_4K_SPANNED(_address, size) \ ++ ((u32)((((size_t)(_address) & (PAGE_SIZE_4K - 1)) + \ + (size) + (PAGE_SIZE_4K - 1)) >> PAGE_SHIFT_4K)) + + /* Byte offset into the page corresponding to given address */ +@@ -305,7 +474,7 @@ static inline unsigned int be_pci_func(struct be_adapter *adapter) + ((size_t)(addr) & (PAGE_SIZE_4K-1)) + + /* Returns bit offset within a DWORD of a bitfield */ +-#define AMAP_BIT_OFFSET(_struct, field) \ ++#define AMAP_BIT_OFFSET(_struct, field) \ + (((size_t)&(((_struct *)0)->field))%32) + + /* Returns the bit mask of the field that is NOT shifted into location. */ +@@ -356,6 +525,11 @@ static inline void swap_dws(void *wrb, int len) + #endif /* __BIG_ENDIAN */ + } + ++static inline bool vlan_configured(struct be_adapter *adapter) ++{ ++ return adapter->vlan_grp && adapter->vlans_added; ++} ++ + static inline u8 is_tcp_pkt(struct sk_buff *skb) + { + u8 val = 0; +@@ -380,9 +554,65 @@ static inline u8 is_udp_pkt(struct sk_buff *skb) + return val; + } + ++static inline u8 is_ipv6_ext_hdr(struct sk_buff *skb) ++{ ++#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE) ++ if (ip_hdr(skb)->version == 6) ++ return ipv6_ext_hdr(ipv6_hdr(skb)->nexthdr); ++ else ++#endif ++ return 0; ++} ++ ++static inline void be_check_sriov_fn_type(struct be_adapter *adapter) ++{ ++ u32 sli_intf; ++ ++ pci_read_config_dword(adapter->pdev, SLI_INTF_REG_OFFSET, &sli_intf); ++ adapter->is_virtfn = (sli_intf & SLI_INTF_FT_MASK) ? 1 : 0; ++} ++ ++static inline void be_vf_eth_addr_generate(struct be_adapter *adapter, u8 *mac) ++{ ++ u32 addr; ++ ++ addr = jhash(adapter->netdev->dev_addr, ETH_ALEN, 0); ++ ++ mac[5] = (u8)(addr & 0xFF); ++ mac[4] = (u8)((addr >> 8) & 0xFF); ++ mac[3] = (u8)((addr >> 16) & 0xFF); ++ /* Use the OUI programmed in hardware */ ++ memcpy(mac, adapter->netdev->dev_addr, 3); ++} ++ ++static inline u16 be_get_tx_vlan_tag(struct be_adapter *adapter, ++ struct sk_buff *skb) ++{ ++ u8 vlan_prio = 0; ++ u16 vlan_tag = 0; ++ ++ vlan_tag = vlan_tx_tag_get(skb); ++ vlan_prio = (vlan_tag & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT; ++ /* If vlan priority provided by OS is NOT in available bmap */ ++ if (!(adapter->vlan_prio_bmap & (1 << vlan_prio))) ++ vlan_tag = (vlan_tag & ~VLAN_PRIO_MASK) | ++ adapter->recommended_prio; ++ ++ return vlan_tag; ++} ++ ++#define be_physfn(adapter) (!adapter->is_virtfn) ++ + extern void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm, + u16 num_popped); +-extern void be_link_status_update(struct be_adapter *adapter, bool link_up); ++extern void be_link_status_update(struct be_adapter *adapter, int link_status); + extern void netdev_stats_update(struct be_adapter *adapter); ++extern void be_parse_stats(struct be_adapter *adapter); + extern int be_load_fw(struct be_adapter *adapter, u8 *func); ++ ++#ifdef CONFIG_PALAU ++extern void be_sysfs_create_group(struct be_adapter *adapter); ++extern void be_sysfs_remove_group(struct be_adapter *adapter); ++#endif ++ + #endif /* BE_H */ +diff --git a/drivers/net/benet/be_cmds.c b/drivers/net/benet/be_cmds.c +index 28a0eda..b4ca89c 100644 +--- a/drivers/net/benet/be_cmds.c ++++ b/drivers/net/benet/be_cmds.c +@@ -1,30 +1,45 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + + #include "be.h" + #include "be_cmds.h" + ++/* Must be a power of 2 or else MODULO will BUG_ON */ ++static int be_get_temp_freq = 64; ++ ++static inline void *embedded_payload(struct be_mcc_wrb *wrb) ++{ ++ return wrb->payload.embedded_payload; ++} ++ + static void be_mcc_notify(struct be_adapter *adapter) + { + struct be_queue_info *mccq = &adapter->mcc_obj.q; + u32 val = 0; + ++ if (adapter->eeh_err) { ++ dev_info(&adapter->pdev->dev, "Error in Card Detected! Cannot issue commands\n"); ++ return; ++ } ++ + val |= mccq->id & DB_MCCQ_RING_ID_MASK; + val |= 1 << DB_MCCQ_NUM_POSTED_SHIFT; ++ ++ wmb(); + iowrite32(val, adapter->db + DB_MCCQ_OFFSET); + } + +@@ -59,21 +74,67 @@ static int be_mcc_compl_process(struct be_adapter *adapter, + + compl_status = (compl->status >> CQE_STATUS_COMPL_SHIFT) & + CQE_STATUS_COMPL_MASK; ++ ++ if ((compl->tag0 == OPCODE_COMMON_WRITE_FLASHROM) && ++ (compl->tag1 == CMD_SUBSYSTEM_COMMON)) { ++ adapter->flash_status = compl_status; ++ complete(&adapter->flash_compl); ++ } ++ + if (compl_status == MCC_STATUS_SUCCESS) { +- if (compl->tag0 == OPCODE_ETH_GET_STATISTICS) { +- struct be_cmd_resp_get_stats *resp = +- adapter->stats.cmd.va; +- be_dws_le_to_cpu(&resp->hw_stats, +- sizeof(resp->hw_stats)); ++ if ((compl->tag0 == OPCODE_ETH_GET_STATISTICS) && ++ (compl->tag1 == CMD_SUBSYSTEM_ETH)) { ++ if (adapter->generation == BE_GEN3) { ++ struct be_cmd_resp_get_stats_v1 *resp = ++ adapter->stats_cmd.va; ++ ++ be_dws_le_to_cpu(&resp->hw_stats, ++ sizeof(resp->hw_stats)); ++ } else { ++ struct be_cmd_resp_get_stats_v0 *resp = ++ adapter->stats_cmd.va; ++ ++ be_dws_le_to_cpu(&resp->hw_stats, ++ sizeof(resp->hw_stats)); ++ } ++ be_parse_stats(adapter); + netdev_stats_update(adapter); ++ adapter->stats_cmd_sent = false; ++ } ++ if (compl->tag0 == ++ OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES) { ++ struct be_mcc_wrb *mcc_wrb = ++ queue_index_node(&adapter->mcc_obj.q, ++ compl->tag1); ++ struct be_cmd_resp_get_cntl_addnl_attribs *resp = ++ embedded_payload(mcc_wrb); ++ adapter->drv_stats.be_on_die_temperature = ++ resp->on_die_temperature; ++ } ++ } else { ++ if (compl->tag0 == OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES) ++ be_get_temp_freq = 0; ++ ++ if (compl->tag1 == MCC_WRB_PASS_THRU) ++ goto done; ++ ++ if (compl_status == MCC_STATUS_NOT_SUPPORTED || ++ compl_status == MCC_STATUS_ILLEGAL_REQUEST) ++ goto done; ++ ++ if (compl_status == MCC_STATUS_UNAUTHORIZED_REQUEST) { ++ dev_warn(&adapter->pdev->dev, "This domain(VM) is not " ++ "permitted to execute this cmd (opcode %d)\n", ++ compl->tag0); ++ } else { ++ extd_status = (compl->status >> CQE_STATUS_EXTD_SHIFT) & ++ CQE_STATUS_EXTD_MASK; ++ dev_err(&adapter->pdev->dev, "Cmd (opcode %d) failed:" ++ "status %d, extd-status %d\n", ++ compl->tag0, compl_status, extd_status); + } +- } else if (compl_status != MCC_STATUS_NOT_SUPPORTED) { +- extd_status = (compl->status >> CQE_STATUS_EXTD_SHIFT) & +- CQE_STATUS_EXTD_MASK; +- dev_warn(&adapter->pdev->dev, +- "Error in cmd completion: status(compl/extd)=%d/%d\n", +- compl_status, extd_status); + } ++done: + return compl_status; + } + +@@ -82,7 +143,70 @@ static void be_async_link_state_process(struct be_adapter *adapter, + struct be_async_event_link_state *evt) + { + be_link_status_update(adapter, +- evt->port_link_status == ASYNC_EVENT_LINK_UP); ++ ((evt->port_link_status & ~ASYNC_EVENT_LOGICAL) == ++ ASYNC_EVENT_LINK_UP ? LINK_UP : LINK_DOWN)); ++} ++ ++/* Grp5 CoS Priority evt */ ++static void be_async_grp5_cos_priority_process(struct be_adapter *adapter, ++ struct be_async_event_grp5_cos_priority *evt) ++{ ++ if (evt->valid) { ++ adapter->vlan_prio_bmap = evt->available_priority_bmap; ++ adapter->recommended_prio &= ~VLAN_PRIO_MASK; ++ adapter->recommended_prio = ++ evt->reco_default_priority << VLAN_PRIO_SHIFT; ++ } ++} ++ ++/* Grp5 QOS Speed evt */ ++static void be_async_grp5_qos_speed_process(struct be_adapter *adapter, ++ struct be_async_event_grp5_qos_link_speed *evt) ++{ ++ if (evt->physical_port == adapter->hba_port_num) { ++ /* qos_link_speed is in units of 10 Mbps */ ++ adapter->link_speed = evt->qos_link_speed * 10; ++ } ++} ++ ++/*Grp5 PVID evt*/ ++static void be_async_grp5_pvid_state_process(struct be_adapter *adapter, ++ struct be_async_event_grp5_pvid_state *evt) ++{ ++ if (evt->enabled) ++ adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK ; ++ else ++ adapter->pvid = 0; ++} ++ ++static void be_async_grp5_evt_process(struct be_adapter *adapter, ++ u32 trailer, struct be_mcc_compl *evt) ++{ ++ u8 event_type = 0; ++ ++ event_type = (trailer >> ASYNC_TRAILER_EVENT_TYPE_SHIFT) & ++ ASYNC_TRAILER_EVENT_TYPE_MASK; ++ ++ switch (event_type) { ++ case ASYNC_EVENT_COS_PRIORITY: ++ be_async_grp5_cos_priority_process(adapter, ++ (struct be_async_event_grp5_cos_priority *)evt); ++ break; ++ case ASYNC_EVENT_QOS_SPEED: ++ be_async_grp5_qos_speed_process(adapter, ++ (struct be_async_event_grp5_qos_link_speed *)evt); ++ break; ++ case ASYNC_EVENT_PVID_STATE: ++ be_async_grp5_pvid_state_process(adapter, ++ (struct be_async_event_grp5_pvid_state *)evt); ++ break; ++ case GRP5_TYPE_PRIO_TC_MAP: ++ memcpy(adapter->prio_tc_map, evt, MAX_TX_QS); ++ break; ++ default: ++ printk(KERN_WARNING "Unknown grp5 event!\n"); ++ break; ++ } + } + + static inline bool is_link_state_evt(u32 trailer) +@@ -92,6 +216,13 @@ static inline bool is_link_state_evt(u32 trailer) + ASYNC_EVENT_CODE_LINK_STATE); + } + ++static inline bool is_grp5_evt(u32 trailer) ++{ ++ return (((trailer >> ASYNC_TRAILER_EVENT_CODE_SHIFT) & ++ ASYNC_TRAILER_EVENT_CODE_MASK) == ++ ASYNC_EVENT_CODE_GRP_5); ++} ++ + static struct be_mcc_compl *be_mcc_compl_get(struct be_adapter *adapter) + { + struct be_queue_info *mcc_cq = &adapter->mcc_obj.cq; +@@ -104,46 +235,67 @@ static struct be_mcc_compl *be_mcc_compl_get(struct be_adapter *adapter) + return NULL; + } + +-int be_process_mcc(struct be_adapter *adapter) ++void be_async_mcc_enable(struct be_adapter *adapter) ++{ ++ spin_lock_bh(&adapter->mcc_cq_lock); ++ ++ be_cq_notify(adapter, adapter->mcc_obj.cq.id, true, 0); ++ adapter->mcc_obj.rearm_cq = true; ++ ++ spin_unlock_bh(&adapter->mcc_cq_lock); ++} ++ ++void be_async_mcc_disable(struct be_adapter *adapter) ++{ ++ adapter->mcc_obj.rearm_cq = false; ++} ++ ++int be_process_mcc(struct be_adapter *adapter, int *status) + { + struct be_mcc_compl *compl; +- int num = 0, status = 0; ++ int num = 0; ++ struct be_mcc_obj *mcc_obj = &adapter->mcc_obj; + + spin_lock_bh(&adapter->mcc_cq_lock); + while ((compl = be_mcc_compl_get(adapter))) { + if (compl->flags & CQE_FLAGS_ASYNC_MASK) { + /* Interpret flags as an async trailer */ +- BUG_ON(!is_link_state_evt(compl->flags)); +- +- /* Interpret compl as a async link evt */ +- be_async_link_state_process(adapter, ++ if (is_link_state_evt(compl->flags)) ++ be_async_link_state_process(adapter, + (struct be_async_event_link_state *) compl); ++ else if (is_grp5_evt(compl->flags)) ++ be_async_grp5_evt_process(adapter, ++ compl->flags, compl); ++ + } else if (compl->flags & CQE_FLAGS_COMPLETED_MASK) { +- status = be_mcc_compl_process(adapter, compl); +- atomic_dec(&adapter->mcc_obj.q.used); ++ *status = be_mcc_compl_process(adapter, compl); ++ atomic_dec(&mcc_obj->q.used); + } + be_mcc_compl_use(compl); + num++; + } + +- if (num) +- be_cq_notify(adapter, adapter->mcc_obj.cq.id, true, num); +- + spin_unlock_bh(&adapter->mcc_cq_lock); +- return status; ++ return num; + } + + /* Wait till no more pending mcc requests are present */ + static int be_mcc_wait_compl(struct be_adapter *adapter) + { + #define mcc_timeout 120000 /* 12s timeout */ +- int i, status; ++ int i, num, status = 0; ++ struct be_mcc_obj *mcc_obj = &adapter->mcc_obj; ++ ++ if (adapter->eeh_err) ++ return -EIO; ++ + for (i = 0; i < mcc_timeout; i++) { +- status = be_process_mcc(adapter); +- if (status) +- return status; ++ num = be_process_mcc(adapter, &status); ++ if (num) ++ be_cq_notify(adapter, mcc_obj->cq.id, ++ mcc_obj->rearm_cq, num); + +- if (atomic_read(&adapter->mcc_obj.q.used) == 0) ++ if (atomic_read(&mcc_obj->q.used) == 0) + break; + udelay(100); + } +@@ -151,7 +303,7 @@ static int be_mcc_wait_compl(struct be_adapter *adapter) + dev_err(&adapter->pdev->dev, "mccq poll timed out\n"); + return -1; + } +- return 0; ++ return status; + } + + /* Notify MCC requests and wait for completion */ +@@ -163,23 +315,34 @@ static int be_mcc_notify_wait(struct be_adapter *adapter) + + static int be_mbox_db_ready_wait(struct be_adapter *adapter, void __iomem *db) + { +- int cnt = 0, wait = 5; ++ int msecs = 0; + u32 ready; + ++ if (adapter->eeh_err) { ++ dev_err(&adapter->pdev->dev, "Error detected in card.Cannot issue commands\n"); ++ return -EIO; ++ } + do { +- ready = ioread32(db) & MPU_MAILBOX_DB_RDY_MASK; ++ ready = ioread32(db); ++ if (ready == 0xffffffff) { ++ dev_err(&adapter->pdev->dev, ++ "pci slot disconnected\n"); ++ return -1; ++ } ++ ++ ready &= MPU_MAILBOX_DB_RDY_MASK; + if (ready) + break; + +- if (cnt > 4000000) { ++ if (msecs > 4000) { + dev_err(&adapter->pdev->dev, "mbox poll timed out\n"); ++ be_detect_dump_ue(adapter); + return -1; + } + +- if (cnt > 50) +- wait = 200; +- cnt += wait; +- udelay(wait); ++ set_current_state(TASK_UNINTERRUPTIBLE); ++ schedule_timeout(msecs_to_jiffies(1)); ++ msecs++; + } while (true); + + return 0; +@@ -198,6 +361,11 @@ static int be_mbox_notify_wait(struct be_adapter *adapter) + struct be_mcc_mailbox *mbox = mbox_mem->va; + struct be_mcc_compl *compl = &mbox->compl; + ++ /* wait for ready to be set */ ++ status = be_mbox_db_ready_wait(adapter, db); ++ if (status != 0) ++ return status; ++ + val |= MPU_MAILBOX_DB_HI_MASK; + /* at bits 2 - 31 place mbox dma addr msb bits 34 - 63 */ + val |= (upper_32_bits(mbox_mem->dma) >> 2) << 2; +@@ -232,7 +400,12 @@ static int be_mbox_notify_wait(struct be_adapter *adapter) + + static int be_POST_stage_get(struct be_adapter *adapter, u16 *stage) + { +- u32 sem = ioread32(adapter->csr + MPU_EP_SEMAPHORE_OFFSET); ++ u32 sem; ++ ++ if (lancer_chip(adapter)) ++ sem = ioread32(adapter->db + MPU_EP_SEMAPHORE_IF_TYPE2_OFFSET); ++ else ++ sem = ioread32(adapter->csr + MPU_EP_SEMAPHORE_OFFSET); + + *stage = sem & EP_SEMAPHORE_POST_STAGE_MASK; + if ((sem >> EP_SEMAPHORE_POST_ERR_SHIFT) & EP_SEMAPHORE_POST_ERR_MASK) +@@ -245,30 +418,29 @@ int be_cmd_POST(struct be_adapter *adapter) + { + u16 stage; + int status, timeout = 0; ++ struct device *dev = &adapter->pdev->dev; + + do { + status = be_POST_stage_get(adapter, &stage); + if (status) { +- dev_err(&adapter->pdev->dev, "POST error; stage=0x%x\n", +- stage); ++ dev_err(dev, "POST error; stage=0x%x\n", stage); + return -1; + } else if (stage != POST_STAGE_ARMFW_RDY) { + set_current_state(TASK_INTERRUPTIBLE); +- schedule_timeout(2 * HZ); ++ if (schedule_timeout(2 * HZ)) { ++ dev_err(dev, "POST cmd aborted\n"); ++ return -EINTR; ++ } + timeout += 2; + } else { + return 0; + } +- } while (timeout < 20); ++ } while (timeout < 40); + +- dev_err(&adapter->pdev->dev, "POST timeout; stage=0x%x\n", stage); ++ dev_err(dev, "POST timeout; stage=0x%x\n", stage); + return -1; + } + +-static inline void *embedded_payload(struct be_mcc_wrb *wrb) +-{ +- return wrb->payload.embedded_payload; +-} + + static inline struct be_sge *nonembedded_sgl(struct be_mcc_wrb *wrb) + { +@@ -277,7 +449,7 @@ static inline struct be_sge *nonembedded_sgl(struct be_mcc_wrb *wrb) + + /* Don't touch the hdr after it's prepared */ + static void be_wrb_hdr_prepare(struct be_mcc_wrb *wrb, int payload_len, +- bool embedded, u8 sge_cnt) ++ bool embedded, u8 sge_cnt, u32 opcode) + { + if (embedded) + wrb->embedded |= MCC_WRB_EMBEDDED_MASK; +@@ -285,7 +457,8 @@ static void be_wrb_hdr_prepare(struct be_mcc_wrb *wrb, int payload_len, + wrb->embedded |= (sge_cnt & MCC_WRB_SGE_CNT_MASK) << + MCC_WRB_SGE_CNT_SHIFT; + wrb->payload_length = payload_len; +- be_dws_cpu_to_le(wrb, 20); ++ wrb->tag0 = opcode; ++ be_dws_cpu_to_le(wrb, 8); + } + + /* Don't touch the hdr after it's prepared */ +@@ -295,6 +468,7 @@ static void be_cmd_hdr_prepare(struct be_cmd_req_hdr *req_hdr, + req_hdr->opcode = opcode; + req_hdr->subsystem = subsystem; + req_hdr->request_length = cpu_to_le32(cmd_len - sizeof(*req_hdr)); ++ req_hdr->version = 0; + } + + static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages, +@@ -349,7 +523,11 @@ static struct be_mcc_wrb *wrb_from_mccq(struct be_adapter *adapter) + struct be_queue_info *mccq = &adapter->mcc_obj.q; + struct be_mcc_wrb *wrb; + +- BUG_ON(atomic_read(&mccq->used) >= mccq->len); ++ if (atomic_read(&mccq->used) >= mccq->len) { ++ dev_err(&adapter->pdev->dev, "Out of MCCQ wrbs\n"); ++ return NULL; ++ } ++ + wrb = queue_head_node(mccq); + queue_head_inc(mccq); + atomic_inc(&mccq->used); +@@ -357,6 +535,59 @@ static struct be_mcc_wrb *wrb_from_mccq(struct be_adapter *adapter) + return wrb; + } + ++/* Tell fw we're about to start firing cmds by writing a ++ * special pattern across the wrb hdr; uses mbox ++ */ ++int be_cmd_fw_init(struct be_adapter *adapter) ++{ ++ u8 *wrb; ++ int status; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = (u8 *)wrb_from_mbox(adapter); ++ *wrb++ = 0xFF; ++ *wrb++ = 0x12; ++ *wrb++ = 0x34; ++ *wrb++ = 0xFF; ++ *wrb++ = 0xFF; ++ *wrb++ = 0x56; ++ *wrb++ = 0x78; ++ *wrb = 0xFF; ++ ++ status = be_mbox_notify_wait(adapter); ++ ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++/* Tell fw we're done with firing cmds by writing a ++ * special pattern across the wrb hdr; uses mbox ++ */ ++int be_cmd_fw_clean(struct be_adapter *adapter) ++{ ++ u8 *wrb; ++ int status; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = (u8 *)wrb_from_mbox(adapter); ++ *wrb++ = 0xFF; ++ *wrb++ = 0xAA; ++ *wrb++ = 0xBB; ++ *wrb++ = 0xFF; ++ *wrb++ = 0xFF; ++ *wrb++ = 0xCC; ++ *wrb++ = 0xDD; ++ *wrb = 0xFF; ++ ++ status = be_mbox_notify_wait(adapter); ++ ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} + int be_cmd_eq_create(struct be_adapter *adapter, + struct be_queue_info *eq, int eq_delay) + { +@@ -365,20 +596,19 @@ int be_cmd_eq_create(struct be_adapter *adapter, + struct be_dma_mem *q_mem = &eq->dma_mem; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, OPCODE_COMMON_EQ_CREATE); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_EQ_CREATE, sizeof(*req)); + + req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); + +- AMAP_SET_BITS(struct amap_eq_context, func, req->context, +- be_pci_func(adapter)); + AMAP_SET_BITS(struct amap_eq_context, valid, req->context, 1); + /* 4byte eqe*/ + AMAP_SET_BITS(struct amap_eq_context, size, req->context, 0); +@@ -397,7 +627,7 @@ int be_cmd_eq_create(struct be_adapter *adapter, + eq->created = true; + } + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +@@ -409,12 +639,14 @@ int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr, + struct be_cmd_req_mac_query *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_MAC_QUERY); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_MAC_QUERY, sizeof(*req)); +@@ -433,13 +665,13 @@ int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr, + memcpy(mac_addr, resp->mac.addr, ETH_ALEN); + } + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + + /* Uses synchronous MCCQ */ + int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, +- u32 if_id, u32 *pmac_id) ++ u32 if_id, u32 *pmac_id, u32 domain) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_pmac_add *req; +@@ -448,13 +680,19 @@ int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_PMAC_ADD); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_PMAC_ADD, sizeof(*req)); + ++ req->hdr.domain = domain; + req->if_id = cpu_to_le32(if_id); + memcpy(req->mac_address, mac_addr, ETH_ALEN); + +@@ -464,12 +702,13 @@ int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, + *pmac_id = le32_to_cpu(resp->pmac_id); + } + ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } + + /* Uses synchronous MCCQ */ +-int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, u32 pmac_id) ++int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, u32 pmac_id, u32 dom) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_pmac_del *req; +@@ -478,20 +717,26 @@ int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, u32 pmac_id) + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_PMAC_DEL); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_PMAC_DEL, sizeof(*req)); + ++ req->hdr.domain = dom; + req->if_id = cpu_to_le32(if_id); + req->pmac_id = cpu_to_le32(pmac_id); + + status = be_mcc_notify_wait(adapter); + ++err: + spin_unlock_bh(&adapter->mcc_lock); +- + return status; + } + +@@ -506,29 +751,51 @@ int be_cmd_cq_create(struct be_adapter *adapter, + void *ctxt; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + ctxt = &req->context; + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_CQ_CREATE); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_CQ_CREATE, sizeof(*req)); + + req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); + +- AMAP_SET_BITS(struct amap_cq_context, coalescwm, ctxt, coalesce_wm); +- AMAP_SET_BITS(struct amap_cq_context, nodelay, ctxt, no_delay); +- AMAP_SET_BITS(struct amap_cq_context, count, ctxt, +- __ilog2_u32(cq->len/256)); +- AMAP_SET_BITS(struct amap_cq_context, valid, ctxt, 1); +- AMAP_SET_BITS(struct amap_cq_context, solevent, ctxt, sol_evts); +- AMAP_SET_BITS(struct amap_cq_context, eventable, ctxt, 1); +- AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); +- AMAP_SET_BITS(struct amap_cq_context, armed, ctxt, 1); +- AMAP_SET_BITS(struct amap_cq_context, func, ctxt, be_pci_func(adapter)); ++ if (lancer_chip(adapter)) { ++ req->hdr.version = 2; ++ req->page_size = 1; /* 1 for 4K */ ++ AMAP_SET_BITS(struct amap_cq_context_lancer, coalescwm, ctxt, ++ coalesce_wm); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, nodelay, ctxt, ++ no_delay); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, count, ctxt, ++ __ilog2_u32(cq->len/256)); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, eventable, ++ ctxt, 1); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, eqid, ++ ctxt, eq->id); ++ AMAP_SET_BITS(struct amap_cq_context_lancer, armed, ctxt, 1); ++ } else { ++ AMAP_SET_BITS(struct amap_cq_context_be, coalescwm, ctxt, ++ coalesce_wm); ++ AMAP_SET_BITS(struct amap_cq_context_be, nodelay, ++ ctxt, no_delay); ++ AMAP_SET_BITS(struct amap_cq_context_be, count, ctxt, ++ __ilog2_u32(cq->len/256)); ++ AMAP_SET_BITS(struct amap_cq_context_be, valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_cq_context_be, solevent, ++ ctxt, sol_evts); ++ AMAP_SET_BITS(struct amap_cq_context_be, eventable, ctxt, 1); ++ AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id); ++ AMAP_SET_BITS(struct amap_cq_context_be, armed, ctxt, 1); ++ } ++ + be_dws_cpu_to_le(ctxt, sizeof(req->context)); + + be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); +@@ -540,8 +807,7 @@ int be_cmd_cq_create(struct be_adapter *adapter, + cq->created = true; + } + +- spin_unlock(&adapter->mbox_lock); +- ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +@@ -553,7 +819,68 @@ static u32 be_encoded_q_len(int q_len) + return len_encoded; + } + +-int be_cmd_mccq_create(struct be_adapter *adapter, ++int be_cmd_mccq_ext_create(struct be_adapter *adapter, ++ struct be_queue_info *mccq, ++ struct be_queue_info *cq) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_mcc_ext_create *req; ++ struct be_dma_mem *q_mem = &mccq->dma_mem; ++ void *ctxt; ++ int status; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = wrb_from_mbox(adapter); ++ req = embedded_payload(wrb); ++ ctxt = &req->context; ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_MCC_CREATE_EXT); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_MCC_CREATE_EXT, sizeof(*req)); ++ ++ req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); ++ if (lancer_chip(adapter)) { ++ req->hdr.version = 1; ++ req->cq_id = cpu_to_le16(cq->id); ++ ++ AMAP_SET_BITS(struct amap_mcc_context_lancer, ring_size, ctxt, ++ be_encoded_q_len(mccq->len)); ++ AMAP_SET_BITS(struct amap_mcc_context_lancer, valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_mcc_context_lancer, async_cq_id, ++ ctxt, cq->id); ++ AMAP_SET_BITS(struct amap_mcc_context_lancer, async_cq_valid, ++ ctxt, 1); ++ ++ } else { ++ AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt, ++ be_encoded_q_len(mccq->len)); ++ AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id); ++ } ++ ++ /* Subscribe to Link State and Group 5 Events(bits 1 and 5 set) */ ++ req->async_event_bitmap[0] |= cpu_to_le32(0x00000022); ++ ++ be_dws_cpu_to_le(ctxt, sizeof(req->context)); ++ ++ be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); ++ ++ status = be_mbox_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_mcc_create *resp = embedded_payload(wrb); ++ mccq->id = le16_to_cpu(resp->id); ++ mccq->created = true; ++ } ++ ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++int be_cmd_mccq_org_create(struct be_adapter *adapter, + struct be_queue_info *mccq, + struct be_queue_info *cq) + { +@@ -563,24 +890,25 @@ int be_cmd_mccq_create(struct be_adapter *adapter, + void *ctxt; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + ctxt = &req->context; + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_MCC_CREATE); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_MCC_CREATE, sizeof(*req)); + +- req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); ++ req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); + +- AMAP_SET_BITS(struct amap_mcc_context, fid, ctxt, be_pci_func(adapter)); +- AMAP_SET_BITS(struct amap_mcc_context, valid, ctxt, 1); +- AMAP_SET_BITS(struct amap_mcc_context, ring_size, ctxt, +- be_encoded_q_len(mccq->len)); +- AMAP_SET_BITS(struct amap_mcc_context, cq_id, ctxt, cq->id); ++ AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt, ++ be_encoded_q_len(mccq->len)); ++ AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id); + + be_dws_cpu_to_le(ctxt, sizeof(req->context)); + +@@ -592,75 +920,93 @@ int be_cmd_mccq_create(struct be_adapter *adapter, + mccq->id = le16_to_cpu(resp->id); + mccq->created = true; + } +- spin_unlock(&adapter->mbox_lock); + ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +-int be_cmd_txq_create(struct be_adapter *adapter, +- struct be_queue_info *txq, ++int be_cmd_mccq_create(struct be_adapter *adapter, ++ struct be_queue_info *mccq, + struct be_queue_info *cq) + { ++ int status; ++ ++ status = be_cmd_mccq_ext_create(adapter, mccq, cq); ++ if (status && !lancer_chip(adapter)) { ++ dev_warn(&adapter->pdev->dev, "Upgrade to F/W ver 2.102.235.0 " ++ "or newer to avoid conflicting priorities between NIC " ++ "and FCoE traffic"); ++ status = be_cmd_mccq_org_create(adapter, mccq, cq); ++ } ++ return status; ++} ++ ++int be_cmd_txq_create(struct be_adapter *adapter, struct be_queue_info *txq, ++ struct be_queue_info *cq, u8 *tc_id) ++{ + struct be_mcc_wrb *wrb; + struct be_cmd_req_eth_tx_create *req; + struct be_dma_mem *q_mem = &txq->dma_mem; +- void *ctxt; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); +- ctxt = &req->context; +- +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); + ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, OPCODE_ETH_TX_CREATE); + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, OPCODE_ETH_TX_CREATE, + sizeof(*req)); + +- req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); ++ if (adapter->flags & BE_FLAGS_DCBX || lancer_chip(adapter)) { ++ req->hdr.version = 1; ++ req->if_id = cpu_to_le16(adapter->if_handle); ++ } ++ if (adapter->flags & BE_FLAGS_DCBX) ++ req->type = cpu_to_le16(ETX_QUEUE_TYPE_PRIORITY); ++ else ++ req->type = cpu_to_le16(ETX_QUEUE_TYPE_STANDARD); + req->ulp_num = BE_ULP1_NUM; +- req->type = BE_ETH_TX_RING_TYPE_STANDARD; +- +- AMAP_SET_BITS(struct amap_tx_context, tx_ring_size, ctxt, +- be_encoded_q_len(txq->len)); +- AMAP_SET_BITS(struct amap_tx_context, pci_func_id, ctxt, +- be_pci_func(adapter)); +- AMAP_SET_BITS(struct amap_tx_context, ctx_valid, ctxt, 1); +- AMAP_SET_BITS(struct amap_tx_context, cq_id_send, ctxt, cq->id); +- +- be_dws_cpu_to_le(ctxt, sizeof(req->context)); +- ++ req->cq_id = cpu_to_le16(cq->id); ++ req->queue_size = be_encoded_q_len(txq->len); ++ req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); + be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); + + status = be_mbox_notify_wait(adapter); + if (!status) { + struct be_cmd_resp_eth_tx_create *resp = embedded_payload(wrb); + txq->id = le16_to_cpu(resp->cid); ++ if (adapter->flags & BE_FLAGS_DCBX) ++ *tc_id = resp->tc_id; + txq->created = true; + } + +- spin_unlock(&adapter->mbox_lock); +- ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +-/* Uses mbox */ ++/* Uses MCC */ + int be_cmd_rxq_create(struct be_adapter *adapter, + struct be_queue_info *rxq, u16 cq_id, u16 frag_size, +- u16 max_frame_size, u32 if_id, u32 rss) ++ u16 max_frame_size, u32 if_id, u32 rss, u8 *rss_id) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_eth_rx_create *req; + struct be_dma_mem *q_mem = &rxq->dma_mem; + int status; + +- spin_lock(&adapter->mbox_lock); ++ spin_lock_bh(&adapter->mcc_lock); + +- wrb = wrb_from_mbox(adapter); ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_ETH_RX_CREATE); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, OPCODE_ETH_RX_CREATE, + sizeof(*req)); +@@ -673,15 +1019,16 @@ int be_cmd_rxq_create(struct be_adapter *adapter, + req->max_frame_size = cpu_to_le16(max_frame_size); + req->rss_queue = cpu_to_le32(rss); + +- status = be_mbox_notify_wait(adapter); ++ status = be_mcc_notify_wait(adapter); + if (!status) { + struct be_cmd_resp_eth_rx_create *resp = embedded_payload(wrb); + rxq->id = le16_to_cpu(resp->id); + rxq->created = true; ++ *rss_id = resp->rss_id; + } + +- spin_unlock(&adapter->mbox_lock); +- ++err: ++ spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +@@ -696,13 +1043,12 @@ int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q, + u8 subsys = 0, opcode = 0; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); +- + switch (queue_type) { + case QTYPE_EQ: + subsys = CMD_SUBSYSTEM_COMMON; +@@ -727,13 +1073,47 @@ int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q, + default: + BUG(); + } ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, opcode); ++ + be_cmd_hdr_prepare(&req->hdr, subsys, opcode, sizeof(*req)); + req->id = cpu_to_le16(q->id); + + status = be_mbox_notify_wait(adapter); ++ if (!status) ++ q->created = false; + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} + ++/* Uses MCC */ ++int be_cmd_rxq_destroy(struct be_adapter *adapter, struct be_queue_info *q) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_q_destroy *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, OPCODE_ETH_RX_DESTROY); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, OPCODE_ETH_RX_DESTROY, ++ sizeof(*req)); ++ req->id = cpu_to_le16(q->id); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) ++ q->created = false; ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +@@ -741,22 +1121,26 @@ int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q, + * Uses mbox + */ + int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, u32 en_flags, +- u8 *mac, bool pmac_invalid, u32 *if_handle, u32 *pmac_id) ++ u8 *mac, bool pmac_invalid, u32 *if_handle, u32 *pmac_id, ++ u32 domain) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_if_create *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_INTERFACE_CREATE); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_INTERFACE_CREATE, sizeof(*req)); + ++ req->hdr.domain = domain; + req->capability_flags = cpu_to_le32(cap_flags); + req->enable_flags = cpu_to_le32(en_flags); + req->pmac_invalid = pmac_invalid; +@@ -771,33 +1155,35 @@ int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, u32 en_flags, + *pmac_id = le32_to_cpu(resp->pmac_id); + } + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + + /* Uses mbox */ +-int be_cmd_if_destroy(struct be_adapter *adapter, u32 interface_id) ++int be_cmd_if_destroy(struct be_adapter *adapter, u32 interface_id, u32 domain) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_if_destroy *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_INTERFACE_DESTROY); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_INTERFACE_DESTROY, sizeof(*req)); + ++ req->hdr.domain = domain; + req->interface_id = cpu_to_le32(interface_id); + + status = be_mbox_notify_wait(adapter); + +- spin_unlock(&adapter->mbox_lock); +- ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +@@ -808,33 +1194,48 @@ int be_cmd_if_destroy(struct be_adapter *adapter, u32 interface_id) + int be_cmd_get_stats(struct be_adapter *adapter, struct be_dma_mem *nonemb_cmd) + { + struct be_mcc_wrb *wrb; +- struct be_cmd_req_get_stats *req; ++ struct be_cmd_req_hdr *hdr; + struct be_sge *sge; ++ int status = 0; ++ ++ if (MODULO(adapter->work_counter, be_get_temp_freq) == 0) ++ be_cmd_get_die_temperature(adapter); + + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); +- req = nonemb_cmd->va; ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ hdr = nonemb_cmd->va; + sge = nonembedded_sgl(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), false, 1); +- wrb->tag0 = OPCODE_ETH_GET_STATISTICS; ++ be_wrb_hdr_prepare(wrb, nonemb_cmd->size, false, 1, ++ OPCODE_ETH_GET_STATISTICS); + +- be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, +- OPCODE_ETH_GET_STATISTICS, sizeof(*req)); ++ be_cmd_hdr_prepare(hdr, CMD_SUBSYSTEM_ETH, ++ OPCODE_ETH_GET_STATISTICS, nonemb_cmd->size); ++ ++ if (adapter->generation == BE_GEN3) ++ hdr->version = 1; ++ ++ wrb->tag1 = CMD_SUBSYSTEM_ETH; + sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); + sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); + sge->len = cpu_to_le32(nonemb_cmd->size); + + be_mcc_notify(adapter); ++ adapter->stats_cmd_sent = true; + ++err: + spin_unlock_bh(&adapter->mcc_lock); +- return 0; ++ return status; + } + + /* Uses synchronous mcc */ + int be_cmd_link_status_query(struct be_adapter *adapter, +- bool *link_up) ++ int *link_status, u8 *mac_speed, u16 *link_speed, u32 dom) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_link_status *req; +@@ -843,50 +1244,216 @@ int be_cmd_link_status_query(struct be_adapter *adapter, + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- *link_up = false; ++ *link_status = LINK_DOWN; + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_LINK_STATUS_QUERY); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_LINK_STATUS_QUERY, sizeof(*req)); + ++ req->hdr.domain = dom; ++ + status = be_mcc_notify_wait(adapter); + if (!status) { + struct be_cmd_resp_link_status *resp = embedded_payload(wrb); +- if (resp->mac_speed != PHY_LINK_SPEED_ZERO) +- *link_up = true; ++ if (resp->mac_speed != PHY_LINK_SPEED_ZERO) { ++ *link_status = LINK_UP; ++ *link_speed = le16_to_cpu(resp->link_speed); ++ *mac_speed = resp->mac_speed; ++ } + } + ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +-/* Uses Mbox */ +-int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver) ++/* Uses synchronous mcc */ ++int be_cmd_get_die_temperature(struct be_adapter *adapter) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_cntl_addnl_attribs *req; ++ u16 mccq_index; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ mccq_index = adapter->mcc_obj.q.head; ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES, sizeof(*req)); ++ ++ wrb->tag1 = mccq_index; ++ ++ be_mcc_notify(adapter); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++ ++/* Uses synchronous mcc */ ++int be_cmd_get_reg_len(struct be_adapter *adapter, u32 *log_size) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_fat *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_MANAGE_FAT); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_MANAGE_FAT, sizeof(*req)); ++ req->fat_operation = cpu_to_le32(QUERY_FAT); ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_fat *resp = embedded_payload(wrb); ++ if (log_size && resp->log_size) ++ *log_size = le32_to_cpu(resp->log_size) - ++ sizeof(u32); ++ } ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++void be_cmd_get_regs(struct be_adapter *adapter, u32 buf_len, void *buf) ++{ ++ struct be_dma_mem get_fat_cmd; ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_fat *req; ++ struct be_sge *sge; ++ u32 offset = 0, total_size, buf_size, ++ log_offset = sizeof(u32), payload_len; ++ int status; ++ ++ if (buf_len == 0) ++ return; ++ ++ total_size = buf_len; ++ ++ get_fat_cmd.size = sizeof(struct be_cmd_req_get_fat) + 60*1024; ++ get_fat_cmd.va = pci_alloc_consistent(adapter->pdev, ++ get_fat_cmd.size, ++ &get_fat_cmd.dma); ++ if (!get_fat_cmd.va) { ++ status = -ENOMEM; ++ dev_err(&adapter->pdev->dev, ++ "Memory allocation failure while retrieving FAT data\n"); ++ return; ++ } ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ while (total_size) { ++ buf_size = min(total_size, (u32)60*1024); ++ total_size -= buf_size; ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = get_fat_cmd.va; ++ sge = nonembedded_sgl(wrb); ++ ++ payload_len = sizeof(struct be_cmd_req_get_fat) + buf_size; ++ be_wrb_hdr_prepare(wrb, payload_len, false, 1, ++ OPCODE_COMMON_MANAGE_FAT); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_MANAGE_FAT, payload_len); ++ ++ sge->pa_hi = cpu_to_le32(upper_32_bits(get_fat_cmd.dma)); ++ sge->pa_lo = cpu_to_le32(get_fat_cmd.dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(get_fat_cmd.size); ++ ++ req->fat_operation = cpu_to_le32(RETRIEVE_FAT); ++ req->read_log_offset = cpu_to_le32(log_offset); ++ req->read_log_length = cpu_to_le32(buf_size); ++ req->data_buffer_size = cpu_to_le32(buf_size); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_fat *resp = get_fat_cmd.va; ++ memcpy(buf + offset, ++ resp->data_buffer, ++ le32_to_cpu(resp->read_log_length)); ++ } else { ++ dev_err(&adapter->pdev->dev, "FAT Table Retrieve error\n"); ++ goto err; ++ } ++ offset += buf_size; ++ log_offset += buf_size; ++ } ++err: ++ pci_free_consistent(adapter->pdev, get_fat_cmd.size, ++ get_fat_cmd.va, ++ get_fat_cmd.dma); ++ spin_unlock_bh(&adapter->mcc_lock); ++} ++ ++/* Uses synchronous mcc */ ++int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver, ++ char *fw_on_flash) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_get_fw_version *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + +- wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_FW_VERSION); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_GET_FW_VERSION, sizeof(*req)); + +- status = be_mbox_notify_wait(adapter); ++ status = be_mcc_notify_wait(adapter); + if (!status) { + struct be_cmd_resp_get_fw_version *resp = embedded_payload(wrb); +- strncpy(fw_ver, resp->firmware_version_string, FW_VER_LEN); ++ strncpy(fw_ver, resp->firmware_version_string, FW_VER_LEN-1); ++ if (fw_on_flash) ++ strncpy(fw_on_flash, resp->fw_on_flash_version_string, ++ FW_VER_LEN-1); + } +- +- spin_unlock(&adapter->mbox_lock); ++err: ++ spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +@@ -897,13 +1464,19 @@ int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_modify_eq_delay *req; ++ int status = 0; + + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_MODIFY_EQ_DELAY); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req)); +@@ -915,8 +1488,9 @@ int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd) + + be_mcc_notify(adapter); + ++err: + spin_unlock_bh(&adapter->mcc_lock); +- return 0; ++ return status; + } + + /* Uses sycnhronous mcc */ +@@ -930,9 +1504,14 @@ int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, u16 *vtag_array, + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_VLAN_CONFIG); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_NTWK_VLAN_CONFIG, sizeof(*req)); +@@ -948,79 +1527,63 @@ int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, u16 *vtag_array, + + status = be_mcc_notify_wait(adapter); + ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +-/* Uses MCC for this command as it may be called in BH context +- * Uses synchronous mcc +- */ +-int be_cmd_promiscuous_config(struct be_adapter *adapter, u8 port_num, bool en) ++int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 value) + { + struct be_mcc_wrb *wrb; +- struct be_cmd_req_promiscuous_config *req; ++ struct be_dma_mem *mem = &adapter->rx_filter; ++ struct be_cmd_req_rx_filter *req = mem->va; ++ struct be_sge *sge; + int status; + + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); +- req = embedded_payload(wrb); +- +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); +- +- be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, +- OPCODE_ETH_PROMISCUOUS, sizeof(*req)); +- +- if (port_num) +- req->port1_promiscuous = en; +- else +- req->port0_promiscuous = en; +- +- status = be_mcc_notify_wait(adapter); +- +- spin_unlock_bh(&adapter->mcc_lock); +- return status; +-} +- +-/* +- * Uses MCC for this command as it may be called in BH context +- * (mc == NULL) => multicast promiscous +- */ +-int be_cmd_multicast_set(struct be_adapter *adapter, u32 if_id, +- struct dev_mc_list *mc_list, u32 mc_count) +-{ +-#define BE_MAX_MC 32 /* set mcast promisc if > 32 */ +- struct be_mcc_wrb *wrb; +- struct be_cmd_req_mcast_mac_config *req; +- +- spin_lock_bh(&adapter->mcc_lock); +- +- wrb = wrb_from_mccq(adapter); +- req = embedded_payload(wrb); +- +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); +- ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ sge = nonembedded_sgl(wrb); ++ sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma)); ++ sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(mem->size); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), false, 1, ++ OPCODE_COMMON_NTWK_RX_FILTER); ++ ++ memset(req, 0, sizeof(*req)); + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, +- OPCODE_COMMON_NTWK_MULTICAST_SET, sizeof(*req)); ++ OPCODE_COMMON_NTWK_RX_FILTER, sizeof(*req)); + +- req->interface_id = if_id; +- if (mc_list && mc_count <= BE_MAX_MC) { +- int i; +- struct dev_mc_list *mc; +- +- req->num_mac = cpu_to_le16(mc_count); +- +- for (mc = mc_list, i = 0; mc; mc = mc->next, i++) +- memcpy(req->mac[i].byte, mc->dmi_addr, ETH_ALEN); ++ req->if_id = cpu_to_le32(adapter->if_handle); ++ if (flags & IFF_PROMISC) { ++ req->if_flags_mask = cpu_to_le32(BE_IF_FLAGS_PROMISCUOUS | ++ BE_IF_FLAGS_VLAN_PROMISCUOUS); ++ if (value == ON) ++ req->if_flags = cpu_to_le32(BE_IF_FLAGS_PROMISCUOUS | ++ BE_IF_FLAGS_VLAN_PROMISCUOUS); ++ } else if (flags & IFF_ALLMULTI) { ++ req->if_flags_mask = req->if_flags = ++ cpu_to_le32(BE_IF_FLAGS_MCAST_PROMISCUOUS); + } else { +- req->promiscuous = 1; +- } ++ struct netdev_hw_addr *ha; ++ int i = 0; + +- be_mcc_notify_wait(adapter); ++ req->if_flags_mask = req->if_flags = ++ cpu_to_le32(BE_IF_FLAGS_MULTICAST); ++ req->mcast_num = cpu_to_le16(netdev_mc_count(adapter->netdev)); ++ netdev_for_each_mc_addr(ha, adapter->netdev) ++ memcpy(req->mcast_mac[i++].byte, ha->DMI_ADDR, ++ ETH_ALEN); ++ } ++ status = be_mcc_notify_wait(adapter); + ++err: + spin_unlock_bh(&adapter->mcc_lock); +- +- return 0; ++ return status; + } + + /* Uses synchrounous mcc */ +@@ -1033,9 +1596,14 @@ int be_cmd_set_flow_control(struct be_adapter *adapter, u32 tx_fc, u32 rx_fc) + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_SET_FLOW_CONTROL); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_SET_FLOW_CONTROL, sizeof(*req)); +@@ -1045,6 +1613,7 @@ int be_cmd_set_flow_control(struct be_adapter *adapter, u32 tx_fc, u32 rx_fc) + + status = be_mcc_notify_wait(adapter); + ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } +@@ -1059,9 +1628,14 @@ int be_cmd_get_flow_control(struct be_adapter *adapter, u32 *tx_fc, u32 *rx_fc) + spin_lock_bh(&adapter->mcc_lock); + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_FLOW_CONTROL); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_GET_FLOW_CONTROL, sizeof(*req)); +@@ -1074,23 +1648,27 @@ int be_cmd_get_flow_control(struct be_adapter *adapter, u32 *tx_fc, u32 *rx_fc) + *rx_fc = le16_to_cpu(resp->rx_flow_control); + } + ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } + + /* Uses mbox */ +-int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num, u32 *cap) ++int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num, ++ u32 *mode, u32 *function_caps) + { + struct be_mcc_wrb *wrb; + struct be_cmd_req_query_fw_cfg *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_QUERY_FIRMWARE_CONFIG); + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_QUERY_FIRMWARE_CONFIG, sizeof(*req)); +@@ -1099,10 +1677,11 @@ int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num, u32 *cap) + if (!status) { + struct be_cmd_resp_query_fw_cfg *resp = embedded_payload(wrb); + *port_num = le32_to_cpu(resp->phys_port); +- *cap = le32_to_cpu(resp->function_cap); ++ *mode = le32_to_cpu(resp->function_mode); ++ *function_caps = le32_to_cpu(resp->function_caps); + } + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); + return status; + } + +@@ -1113,19 +1692,161 @@ int be_cmd_reset_function(struct be_adapter *adapter) + struct be_cmd_req_hdr *req; + int status; + +- spin_lock(&adapter->mbox_lock); ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; + + wrb = wrb_from_mbox(adapter); + req = embedded_payload(wrb); + +- be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_FUNCTION_RESET); + + be_cmd_hdr_prepare(req, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_FUNCTION_RESET, sizeof(*req)); + + status = be_mbox_notify_wait(adapter); + +- spin_unlock(&adapter->mbox_lock); ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable, u16 table_size) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_rss_config *req; ++ u32 myhash[10] = {0x0123, 0x4567, 0x89AB, 0xCDEF, 0x01EF, ++ 0x0123, 0x4567, 0x89AB, 0xCDEF, 0x01EF}; ++ int status; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = wrb_from_mbox(adapter); ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_ETH_RSS_CONFIG); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, ++ OPCODE_ETH_RSS_CONFIG, sizeof(*req)); ++ ++ req->if_id = cpu_to_le32(adapter->if_handle); ++ req->enable_rss = cpu_to_le16(RSS_ENABLE_TCP_IPV4 | RSS_ENABLE_IPV4); ++ req->cpu_table_size_log2 = cpu_to_le16(fls(table_size) - 1); ++ memcpy(req->cpu_table, rsstable, table_size); ++ memcpy(req->hash, myhash, sizeof(myhash)); ++ be_dws_cpu_to_le(req->hash, sizeof(req->hash)); ++ ++ status = be_mbox_notify_wait(adapter); ++ ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++/* Uses sync mcc */ ++int be_cmd_set_beacon_state(struct be_adapter *adapter, u8 port_num, ++ u8 bcn, u8 sts, u8 state) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_enable_disable_beacon *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_ENABLE_DISABLE_BEACON); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_ENABLE_DISABLE_BEACON, sizeof(*req)); ++ ++ req->port_num = port_num; ++ req->beacon_state = state; ++ req->beacon_duration = bcn; ++ req->status_duration = sts; ++ ++ status = be_mcc_notify_wait(adapter); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++/* Uses sync mcc */ ++int be_cmd_get_beacon_state(struct be_adapter *adapter, u8 port_num, u32 *state) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_beacon_state *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_BEACON_STATE); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_BEACON_STATE, sizeof(*req)); ++ ++ req->port_num = port_num; ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_beacon_state *resp = ++ embedded_payload(wrb); ++ *state = resp->beacon_state; ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++/* Uses sync mcc */ ++int be_cmd_read_port_type(struct be_adapter *adapter, u32 port, ++ u8 *connector) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_port_type *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(struct be_cmd_resp_port_type), true, 0, ++ OPCODE_COMMON_READ_TRANSRECV_DATA); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_READ_TRANSRECV_DATA, sizeof(*req)); ++ ++ req->port = cpu_to_le32(port); ++ req->page_num = cpu_to_le32(TR_PAGE_A0); ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_port_type *resp = embedded_payload(wrb); ++ *connector = resp->data.connector; ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); + return status; + } + +@@ -1133,16 +1854,24 @@ int be_cmd_write_flashrom(struct be_adapter *adapter, struct be_dma_mem *cmd, + u32 flash_type, u32 flash_opcode, u32 buf_size) + { + struct be_mcc_wrb *wrb; +- struct be_cmd_write_flashrom *req = cmd->va; ++ struct be_cmd_write_flashrom *req; + struct be_sge *sge; + int status; + + spin_lock_bh(&adapter->mcc_lock); ++ adapter->flash_status = 0; + + wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err_unlock; ++ } ++ req = cmd->va; + sge = nonembedded_sgl(wrb); + +- be_wrb_hdr_prepare(wrb, cmd->size, false, 1); ++ be_wrb_hdr_prepare(wrb, cmd->size, false, 1, ++ OPCODE_COMMON_WRITE_FLASHROM); ++ wrb->tag1 = CMD_SUBSYSTEM_COMMON; + + be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, + OPCODE_COMMON_WRITE_FLASHROM, cmd->size); +@@ -1154,8 +1883,852 @@ int be_cmd_write_flashrom(struct be_adapter *adapter, struct be_dma_mem *cmd, + req->params.op_code = cpu_to_le32(flash_opcode); + req->params.data_buf_size = cpu_to_le32(buf_size); + ++ be_mcc_notify(adapter); ++ spin_unlock_bh(&adapter->mcc_lock); ++ ++ if (!wait_for_completion_timeout(&adapter->flash_compl, ++ msecs_to_jiffies(40000))) ++ status = -1; ++ else ++ status = adapter->flash_status; ++ ++ return status; ++ ++err_unlock: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc, ++ int offset) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_write_flashrom *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req)+4, true, 0, ++ OPCODE_COMMON_READ_FLASHROM); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_READ_FLASHROM, sizeof(*req)+4); ++ ++ req->params.op_type = cpu_to_le32(IMG_TYPE_REDBOOT); ++ req->params.op_code = cpu_to_le32(FLASHROM_OPER_REPORT); ++ req->params.offset = cpu_to_le32(offset); ++ req->params.data_buf_size = cpu_to_le32(0x4); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) ++ memcpy(flashed_crc, req->params.data_buf, 4); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac, ++ struct be_dma_mem *nonemb_cmd) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_acpi_wol_magic_config *req; ++ struct be_sge *sge; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = nonemb_cmd->va; ++ sge = nonembedded_sgl(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), false, 1, ++ OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, ++ OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG, sizeof(*req)); ++ memcpy(req->magic_mac, mac, ETH_ALEN); ++ ++ sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); ++ sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(nonemb_cmd->size); ++ ++ status = be_mcc_notify_wait(adapter); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num, ++ u8 loopback_type, u8 enable) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_lmode *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_LOWLEVEL_SET_LOOPBACK_MODE); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL, ++ OPCODE_LOWLEVEL_SET_LOOPBACK_MODE, ++ sizeof(*req)); ++ ++ req->src_port = port_num; ++ req->dest_port = port_num; ++ req->loopback_type = loopback_type; ++ req->loopback_state = enable; ++ ++ status = be_mcc_notify_wait(adapter); ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num, ++ u32 loopback_type, u32 pkt_size, u32 num_pkts, u64 pattern) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_loopback_test *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_LOWLEVEL_LOOPBACK_TEST); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL, ++ OPCODE_LOWLEVEL_LOOPBACK_TEST, sizeof(*req)); ++ req->hdr.timeout = cpu_to_le32(4); ++ ++ req->pattern = cpu_to_le64(pattern); ++ req->src_port = cpu_to_le32(port_num); ++ req->dest_port = cpu_to_le32(port_num); ++ req->pkt_size = cpu_to_le32(pkt_size); ++ req->num_pkts = cpu_to_le32(num_pkts); ++ req->loopback_type = cpu_to_le32(loopback_type); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_loopback_test *resp = embedded_payload(wrb); ++ status = le32_to_cpu(resp->status); ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern, ++ u32 byte_cnt, struct be_dma_mem *cmd) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_ddrdma_test *req; ++ struct be_sge *sge; ++ int status; ++ int i, j = 0; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = cmd->va; ++ sge = nonembedded_sgl(wrb); ++ be_wrb_hdr_prepare(wrb, cmd->size, false, 1, ++ OPCODE_LOWLEVEL_HOST_DDR_DMA); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL, ++ OPCODE_LOWLEVEL_HOST_DDR_DMA, cmd->size); ++ ++ sge->pa_hi = cpu_to_le32(upper_32_bits(cmd->dma)); ++ sge->pa_lo = cpu_to_le32(cmd->dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(cmd->size); ++ ++ req->pattern = cpu_to_le64(pattern); ++ req->byte_count = cpu_to_le32(byte_cnt); ++ for (i = 0; i < byte_cnt; i++) { ++ req->snd_buff[i] = (u8)(pattern >> (j*8)); ++ j++; ++ if (j > 7) ++ j = 0; ++ } ++ ++ status = be_mcc_notify_wait(adapter); ++ ++ if (!status) { ++ struct be_cmd_resp_ddrdma_test *resp; ++ resp = cmd->va; ++ if ((memcmp(resp->rcv_buff, req->snd_buff, byte_cnt) != 0) || ++ resp->snd_err) { ++ status = -1; ++ } ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_get_seeprom_data(struct be_adapter *adapter, ++ struct be_dma_mem *nonemb_cmd) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_seeprom_read *req; ++ struct be_sge *sge; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ req = nonemb_cmd->va; ++ sge = nonembedded_sgl(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), false, 1, ++ OPCODE_COMMON_SEEPROM_READ); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_SEEPROM_READ, sizeof(*req)); ++ ++ sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma)); ++ sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(nonemb_cmd->size); ++ ++ status = be_mcc_notify_wait(adapter); ++ ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_get_phy_info(struct be_adapter *adapter, ++ struct be_phy_info *phy_info) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_phy_info *req; ++ struct be_sge *sge; ++ struct be_dma_mem cmd; ++ struct be_phy_info *resp_phy_info; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ cmd.size = sizeof(struct be_cmd_req_get_phy_info); ++ cmd.va = pci_alloc_consistent(adapter->pdev, cmd.size, ++ &cmd.dma); ++ if (!cmd.va) { ++ dev_err(&adapter->pdev->dev, "Memory alloc failure\n"); ++ status = -ENOMEM; ++ goto err; ++ } ++ ++ req = cmd.va; ++ sge = nonembedded_sgl(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), false, 1, ++ OPCODE_COMMON_GET_PHY_DETAILS); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_PHY_DETAILS, ++ sizeof(*req)); ++ ++ sge->pa_hi = cpu_to_le32(upper_32_bits(cmd.dma)); ++ sge->pa_lo = cpu_to_le32(cmd.dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(cmd.size); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ resp_phy_info = cmd.va + sizeof(struct be_cmd_req_hdr); ++ phy_info->phy_type = le16_to_cpu(resp_phy_info->phy_type); ++ phy_info->interface_type = ++ le16_to_cpu(resp_phy_info->interface_type); ++ phy_info->auto_speeds_supported = ++ le16_to_cpu(resp_phy_info->auto_speeds_supported); ++ phy_info->fixed_speeds_supported = ++ le16_to_cpu(resp_phy_info->fixed_speeds_supported); ++ phy_info->misc_params = ++ le32_to_cpu(resp_phy_info->misc_params); ++ } ++ pci_free_consistent(adapter->pdev, cmd.size, ++ cmd.va, cmd.dma); ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_qos *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_SET_QOS); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_SET_QOS, sizeof(*req)); ++ ++ req->hdr.domain = domain; ++ req->valid_bits = cpu_to_le32(BE_QOS_BITS_NIC); ++ req->max_bps_nic = cpu_to_le32(bps); ++ ++ status = be_mcc_notify_wait(adapter); ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_get_cntl_attributes(struct be_adapter *adapter) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_cntl_attribs *req; ++ struct be_cmd_resp_cntl_attribs *resp; ++ struct be_sge *sge; ++ int status; ++ int payload_len = max(sizeof(*req), sizeof(*resp)); ++ struct mgmt_controller_attrib *attribs; ++ struct be_dma_mem attribs_cmd; ++ ++ memset(&attribs_cmd, 0, sizeof(struct be_dma_mem)); ++ attribs_cmd.size = sizeof(struct be_cmd_resp_cntl_attribs); ++ attribs_cmd.va = pci_alloc_consistent(adapter->pdev, attribs_cmd.size, ++ &attribs_cmd.dma); ++ if (!attribs_cmd.va) { ++ dev_err(&adapter->pdev->dev, ++ "Memory allocation failure\n"); ++ return -ENOMEM; ++ } ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = wrb_from_mbox(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = attribs_cmd.va; ++ sge = nonembedded_sgl(wrb); ++ ++ be_wrb_hdr_prepare(wrb, payload_len, false, 1, ++ OPCODE_COMMON_GET_CNTL_ATTRIBUTES); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_CNTL_ATTRIBUTES, payload_len); ++ sge->pa_hi = cpu_to_le32(upper_32_bits(attribs_cmd.dma)); ++ sge->pa_lo = cpu_to_le32(attribs_cmd.dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(attribs_cmd.size); ++ ++ status = be_mbox_notify_wait(adapter); ++ if (!status) { ++ attribs = (struct mgmt_controller_attrib *)(attribs_cmd.va + ++ sizeof(struct be_cmd_resp_hdr)); ++ adapter->hba_port_num = attribs->hba_attribs.phy_port; ++ strncpy(adapter->model_number, ++ attribs->hba_attribs.controller_model_number, 31); ++ } ++ ++err: ++ mutex_unlock(&adapter->mbox_lock); ++ pci_free_consistent(adapter->pdev, attribs_cmd.size, attribs_cmd.va, ++ attribs_cmd.dma); ++ return status; ++} ++ ++/* Uses mbox */ ++int be_cmd_req_native_mode(struct be_adapter *adapter) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_func_cap *req; ++ int status; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++ wrb = wrb_from_mbox(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_SET_DRIVER_FUNCTION_CAP); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_SET_DRIVER_FUNCTION_CAP, sizeof(*req)); ++ ++ req->valid_cap_flags = cpu_to_le32(CAPABILITY_SW_TIMESTAMPS | ++ CAPABILITY_BE3_NATIVE_ERX_API); ++ req->cap_flags = cpu_to_le32(CAPABILITY_BE3_NATIVE_ERX_API); ++ ++ status = be_mbox_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_set_func_cap *resp = embedded_payload(wrb); ++ adapter->be3_native = le32_to_cpu(resp->cap_flags) & ++ CAPABILITY_BE3_NATIVE_ERX_API; ++ } ++err: ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++static void encode_port_names(struct be_adapter *adapter) ++{ ++ switch (adapter->port_name[adapter->hba_port_num]) { ++ case '0': ++ adapter->port_name[adapter->hba_port_num] = 0; ++ break; ++ case '1': ++ adapter->port_name[adapter->hba_port_num] = 1; ++ break; ++ case '2': ++ adapter->port_name[adapter->hba_port_num] = 2; ++ break; ++ case '3': ++ adapter->port_name[adapter->hba_port_num] = 3; ++ break; ++ case '4': ++ adapter->port_name[adapter->hba_port_num] = 4; ++ break; ++ case 'A': ++ adapter->port_name[adapter->hba_port_num] = 5; ++ break; ++ case 'B': ++ adapter->port_name[adapter->hba_port_num] = 6; ++ break; ++ case 'C': ++ adapter->port_name[adapter->hba_port_num] = 7; ++ break; ++ case 'D': ++ adapter->port_name[adapter->hba_port_num] = 8; ++ break; ++ } ++} ++ ++int be_cmd_query_port_names_v0(struct be_adapter *adapter, u8 *port_name) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_port_name *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_PORT_NAME); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_PORT_NAME, sizeof(*req)); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_port_name *resp = embedded_payload(wrb); ++ port_name[0] = resp->port0_name; ++ port_name[1] = resp->port1_name; ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ ++ if(!status) ++ encode_port_names(adapter); ++ return status; ++} ++ ++int be_cmd_query_port_names_v1(struct be_adapter *adapter, u8 *port_name) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_port_name *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_PORT_NAME); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_PORT_NAME, sizeof(*req)); ++ req->hdr.version = 1; ++ + status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_port_name_v1 *resp = embedded_payload(wrb); ++ port_name[0] = resp->port0_name; ++ port_name[1] = resp->port1_name; ++ port_name[2] = resp->port2_name; ++ port_name[3] = resp->port3_name; ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ ++ if (!status) ++ encode_port_names(adapter); ++ return status; ++} ++ ++int be_cmd_req_pg_pfc(struct be_adapter *adapter, int *fw_num_txqs) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_pg *req; ++ int status, num = 0; ++ bool query = true; ++ ++ *fw_num_txqs = MAX_TX_QS; ++ ++ if (mutex_lock_interruptible(&adapter->mbox_lock)) ++ return -1; ++ ++enable_pfc: ++ wrb = wrb_from_mbox(adapter); ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_ETH_PG_FEATURE_QUERY_REQUEST); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH, ++ OPCODE_ETH_PG_FEATURE_QUERY_REQUEST, sizeof(*req)); ++ ++ if (query) ++ req->query |= cpu_to_le32(REQ_PG_QUERY); ++ req->pfc_pg |= cpu_to_le32(REQ_PG_FEAT); ++ ++ status = be_mbox_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_pg *resp = embedded_payload(wrb); ++ if (query) { ++ if (le32_to_cpu(resp->pfc_pg) & REQ_PG_FEAT) { ++ num = le32_to_cpu(resp->num_tx_rings); ++ query = false; ++ goto enable_pfc; ++ } ++ } else { ++ adapter->flags |= BE_FLAGS_DCBX; ++ *fw_num_txqs = num; ++ } ++ } ++ ++ mutex_unlock(&adapter->mbox_lock); ++ return status; ++} ++ ++/* Set privilege(s) for a function */ ++int be_cmd_set_fn_privileges(struct be_adapter *adapter, u32 mask, u32 *prev, ++ u32 domain) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_fn_privileges *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_SET_FN_PRIVILEGES); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_SET_FN_PRIVILEGES, sizeof(*req)); ++ ++ req->hdr.domain = domain; ++ req->privilege_mask = cpu_to_le32(mask); ++ ++ status = be_mcc_notify_wait(adapter); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++/* Get privilege(s) for a function */ ++int be_cmd_get_fn_privileges(struct be_adapter *adapter, u32 *privilege, ++ u32 domain) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_fn_privileges *req; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_FN_PRIVILEGES); + ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_FN_PRIVILEGES, sizeof(*req)); ++ ++ req->hdr.domain = domain; ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_fn_privileges *resp = ++ embedded_payload(wrb); ++ *privilege = le32_to_cpu(resp->privilege_mask); ++ } else ++ *privilege = 0; ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++/* Set Hyper switch config */ ++int be_cmd_set_hsw_config(struct be_adapter *adapter, u16 pvid, ++ u32 domain, u16 intf_id) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_hsw_config *req; ++ void *ctxt; ++ int status; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ctxt = &req->context; ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_SET_HSW_CONFIG); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_SET_HSW_CONFIG, sizeof(*req)); ++ ++ req->hdr.domain = domain; ++ AMAP_SET_BITS(struct amap_set_hsw_context, interface_id, ctxt, intf_id); ++ if (pvid) { ++ AMAP_SET_BITS(struct amap_set_hsw_context, pvid_valid, ctxt, 1); ++ AMAP_SET_BITS(struct amap_set_hsw_context, pvid, ctxt, pvid); ++ } ++ ++ be_dws_cpu_to_le(req->context, sizeof(req->context)); ++ status = be_mcc_notify_wait(adapter); ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++/* Get Hyper switch config */ ++int be_cmd_get_hsw_config(struct be_adapter *adapter, u16 *pvid, ++ u32 domain, u16 intf_id) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_hsw_config *req; ++ void *ctxt; ++ int status; ++ u16 vid; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ ctxt = &req->context; ++ ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_GET_HSW_CONFIG); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_GET_HSW_CONFIG, sizeof(*req)); ++ ++ req->hdr.domain = domain; ++ AMAP_SET_BITS(struct amap_get_hsw_req_context, interface_id, ctxt, ++ intf_id); ++ AMAP_SET_BITS(struct amap_get_hsw_req_context, pvid_valid, ctxt, 1); ++ be_dws_cpu_to_le(req->context, sizeof(req->context)); ++ ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_hsw_config *resp = ++ embedded_payload(wrb); ++ be_dws_le_to_cpu(&resp->context, ++ sizeof(resp->context)); ++ vid = AMAP_GET_BITS(struct amap_get_hsw_resp_context, ++ pvid, &resp->context); ++ *pvid = le16_to_cpu(vid); ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_get_port_speed(struct be_adapter *adapter, ++ u8 port_num, u16 *dac_cable_len, u16 *port_speed) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_get_port_speed *req; ++ int status = 0; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ ++ req = embedded_payload(wrb); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_GET_LINK_SPEED); ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_NTWK_GET_LINK_SPEED, ++ sizeof(*req)); ++ req->port_num = port_num; ++ status = be_mcc_notify_wait(adapter); ++ if (!status) { ++ struct be_cmd_resp_get_port_speed *resp = ++ embedded_payload(wrb); ++ *dac_cable_len = resp->dac_cable_length; ++ *port_speed = resp->mac_speed; ++ } ++ ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++int be_cmd_set_port_speed_v1(struct be_adapter *adapter, ++ u8 port_num, u16 mac_speed, ++ u16 dac_cable_len) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_cmd_req_set_port_speed_v1 *req; ++ int status = 0; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ req = embedded_payload(wrb); ++ be_wrb_hdr_prepare(wrb, sizeof(*req), true, 0, ++ OPCODE_COMMON_NTWK_SET_LINK_SPEED); ++ ++ be_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, ++ OPCODE_COMMON_NTWK_SET_LINK_SPEED, ++ sizeof(*req)); ++ req->hdr.version=1; ++ ++ req->port_num = port_num; ++ req->virt_port = port_num; ++ req->mac_speed = mac_speed; ++ req->dac_cable_length = dac_cable_len; ++ status = be_mcc_notify_wait(adapter); ++err: ++ spin_unlock_bh(&adapter->mcc_lock); ++ return status; ++} ++ ++ ++/* Uses sync mcc */ ++#ifdef CONFIG_PALAU ++int be_cmd_pass_ext_ioctl(struct be_adapter *adapter, dma_addr_t dma, ++ int req_size, void *va) ++{ ++ struct be_mcc_wrb *wrb; ++ struct be_sge *sge; ++ int status; ++ struct be_cmd_req_hdr *hdr = (struct be_cmd_req_hdr *) va; ++ ++ spin_lock_bh(&adapter->mcc_lock); ++ ++ wrb = wrb_from_mccq(adapter); ++ if (!wrb) { ++ status = -EBUSY; ++ goto err; ++ } ++ sge = nonembedded_sgl(wrb); ++ ++ be_wrb_hdr_prepare(wrb, req_size, false, 1, hdr->opcode); ++ wrb->tag1 = MCC_WRB_PASS_THRU; ++ sge->pa_hi = cpu_to_le32(upper_32_bits(dma)); ++ sge->pa_lo = cpu_to_le32(dma & 0xFFFFFFFF); ++ sge->len = cpu_to_le32(req_size); ++ ++ status = be_mcc_notify_wait(adapter); ++err: + spin_unlock_bh(&adapter->mcc_lock); + return status; + } ++#endif +diff --git a/drivers/net/benet/be_cmds.h b/drivers/net/benet/be_cmds.h +index ad33d55..35aa5c7 100644 +--- a/drivers/net/benet/be_cmds.h ++++ b/drivers/net/benet/be_cmds.h +@@ -1,20 +1,23 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + ++#ifndef BE_CMDS_H ++#define BE_CMDS_H ++ + /* + * The driver sends configuration and managements command requests to the + * firmware in the BE. These requests are communicated to the processor +@@ -29,9 +32,10 @@ struct be_sge { + u32 len; + }; + +-#define MCC_WRB_EMBEDDED_MASK 1 /* bit 0 of dword 0*/ ++#define MCC_WRB_EMBEDDED_MASK 1 /* bit 0 of dword 0*/ + #define MCC_WRB_SGE_CNT_SHIFT 3 /* bits 3 - 7 of dword 0 */ + #define MCC_WRB_SGE_CNT_MASK 0x1F /* bits 3 - 7 of dword 0 */ ++#define MCC_WRB_PASS_THRU 0xFF /* this wrb is used for pass thru cmd */ + struct be_mcc_wrb { + u32 embedded; /* dword 0 */ + u32 payload_length; /* dword 1 */ +@@ -44,24 +48,19 @@ struct be_mcc_wrb { + } payload; + }; + +-#define CQE_FLAGS_VALID_MASK (1 << 31) +-#define CQE_FLAGS_ASYNC_MASK (1 << 30) +-#define CQE_FLAGS_COMPLETED_MASK (1 << 28) +-#define CQE_FLAGS_CONSUMED_MASK (1 << 27) ++#define CQE_FLAGS_VALID_MASK (1 << 31) ++#define CQE_FLAGS_ASYNC_MASK (1 << 30) ++#define CQE_FLAGS_COMPLETED_MASK (1 << 28) ++#define CQE_FLAGS_CONSUMED_MASK (1 << 27) + + /* Completion Status */ + enum { +- MCC_STATUS_SUCCESS = 0x0, +-/* The client does not have sufficient privileges to execute the command */ +- MCC_STATUS_INSUFFICIENT_PRIVILEGES = 0x1, +-/* A parameter in the command was invalid. */ +- MCC_STATUS_INVALID_PARAMETER = 0x2, +-/* There are insufficient chip resources to execute the command */ +- MCC_STATUS_INSUFFICIENT_RESOURCES = 0x3, +-/* The command is completing because the queue was getting flushed */ +- MCC_STATUS_QUEUE_FLUSHING = 0x4, +-/* The command is completing with a DMA error */ +- MCC_STATUS_DMA_FAILED = 0x5, ++ MCC_STATUS_SUCCESS = 0, ++ MCC_STATUS_FAILED = 1, ++ MCC_STATUS_ILLEGAL_REQUEST = 2, ++ MCC_STATUS_ILLEGAL_FIELD = 3, ++ MCC_STATUS_INSUFFICIENT_BUFFER = 4, ++ MCC_STATUS_UNAUTHORIZED_REQUEST = 5, + MCC_STATUS_NOT_SUPPORTED = 66 + }; + +@@ -81,15 +80,24 @@ struct be_mcc_compl { + * mcc_compl is interpreted as follows: + */ + #define ASYNC_TRAILER_EVENT_CODE_SHIFT 8 /* bits 8 - 15 */ ++#define ASYNC_TRAILER_EVENT_TYPE_SHIFT 16 /* bits 16 - 23 */ + #define ASYNC_TRAILER_EVENT_CODE_MASK 0xFF ++#define ASYNC_TRAILER_EVENT_TYPE_MASK 0xFF + #define ASYNC_EVENT_CODE_LINK_STATE 0x1 ++#define ASYNC_EVENT_CODE_GRP_5 0x5 ++#define ASYNC_EVENT_QOS_SPEED 0x1 ++#define ASYNC_EVENT_COS_PRIORITY 0x2 ++#define ASYNC_EVENT_PVID_STATE 0x3 ++#define GRP5_TYPE_PRIO_TC_MAP 4 ++ + struct be_async_event_trailer { + u32 code; + }; + + enum { +- ASYNC_EVENT_LINK_DOWN = 0x0, +- ASYNC_EVENT_LINK_UP = 0x1 ++ ASYNC_EVENT_LINK_DOWN = 0x0, ++ ASYNC_EVENT_LINK_UP = 0x1, ++ ASYNC_EVENT_LOGICAL = 0x2 + }; + + /* When the event code of an async trailer is link-state, the mcc_compl +@@ -101,7 +109,51 @@ struct be_async_event_link_state { + u8 port_duplex; + u8 port_speed; + u8 port_fault; +- u8 rsvd0[7]; ++ u8 rsvd0; ++ u16 qos_link_speed; ++ u32 event_tag; ++ struct be_async_event_trailer trailer; ++} __packed; ++ ++/* When the event code of an async trailer is GRP-5 and event_type is QOS_SPEED ++ * the mcc_compl must be interpreted as follows ++ */ ++struct be_async_event_grp5_qos_link_speed { ++ u8 physical_port; ++ u8 rsvd[5]; ++ u16 qos_link_speed; ++ u32 event_tag; ++ struct be_async_event_trailer trailer; ++} __packed; ++ ++/* When the event code of an async trailer is GRP5 and event type is ++ * CoS-Priority, the mcc_compl must be interpreted as follows ++ */ ++struct be_async_event_grp5_cos_priority { ++ u8 physical_port; ++ u8 available_priority_bmap; ++ u8 reco_default_priority; ++ u8 valid; ++ u8 rsvd0; ++ u8 event_tag; ++ struct be_async_event_trailer trailer; ++} __packed; ++ ++/* When the event code of an async trailer is GRP5 and event type is ++ * PVID state, the mcc_compl must be interpreted as follows ++ */ ++struct be_async_event_grp5_pvid_state { ++ u8 enabled; ++ u8 rsvd0; ++ u16 tag; ++ u32 event_tag; ++ u32 rsvd1; ++ struct be_async_event_trailer trailer; ++} __packed; ++ ++/* GRP5 prio-tc-map event */ ++struct be_async_event_grp5_prio_tc_map { ++ u8 prio_tc_map[8]; /* map[prio] -> tc_id */ + struct be_async_event_trailer trailer; + } __packed; + +@@ -111,41 +163,68 @@ struct be_mcc_mailbox { + }; + + #define CMD_SUBSYSTEM_COMMON 0x1 +-#define CMD_SUBSYSTEM_ETH 0x3 ++#define CMD_SUBSYSTEM_ETH 0x3 ++#define CMD_SUBSYSTEM_LOWLEVEL 0xb + + #define OPCODE_COMMON_NTWK_MAC_QUERY 1 + #define OPCODE_COMMON_NTWK_MAC_SET 2 + #define OPCODE_COMMON_NTWK_MULTICAST_SET 3 +-#define OPCODE_COMMON_NTWK_VLAN_CONFIG 4 ++#define OPCODE_COMMON_NTWK_VLAN_CONFIG 4 + #define OPCODE_COMMON_NTWK_LINK_STATUS_QUERY 5 ++#define OPCODE_COMMON_READ_FLASHROM 6 + #define OPCODE_COMMON_WRITE_FLASHROM 7 + #define OPCODE_COMMON_CQ_CREATE 12 + #define OPCODE_COMMON_EQ_CREATE 13 +-#define OPCODE_COMMON_MCC_CREATE 21 +-#define OPCODE_COMMON_NTWK_RX_FILTER 34 ++#define OPCODE_COMMON_MCC_CREATE 21 ++#define OPCODE_COMMON_SET_QOS 28 ++#define OPCODE_COMMON_MCC_CREATE_EXT 90 ++#define OPCODE_COMMON_SEEPROM_READ 30 ++#define OPCODE_COMMON_GET_CNTL_ATTRIBUTES 32 ++#define OPCODE_COMMON_NTWK_RX_FILTER 34 + #define OPCODE_COMMON_GET_FW_VERSION 35 + #define OPCODE_COMMON_SET_FLOW_CONTROL 36 + #define OPCODE_COMMON_GET_FLOW_CONTROL 37 + #define OPCODE_COMMON_SET_FRAME_SIZE 39 + #define OPCODE_COMMON_MODIFY_EQ_DELAY 41 + #define OPCODE_COMMON_FIRMWARE_CONFIG 42 +-#define OPCODE_COMMON_NTWK_INTERFACE_CREATE 50 +-#define OPCODE_COMMON_NTWK_INTERFACE_DESTROY 51 +-#define OPCODE_COMMON_MCC_DESTROY 53 +-#define OPCODE_COMMON_CQ_DESTROY 54 +-#define OPCODE_COMMON_EQ_DESTROY 55 ++#define OPCODE_COMMON_NTWK_INTERFACE_CREATE 50 ++#define OPCODE_COMMON_NTWK_INTERFACE_DESTROY 51 ++#define OPCODE_COMMON_MCC_DESTROY 53 ++#define OPCODE_COMMON_CQ_DESTROY 54 ++#define OPCODE_COMMON_EQ_DESTROY 55 ++#define OPCODE_COMMON_NTWK_SET_LINK_SPEED 57 + #define OPCODE_COMMON_QUERY_FIRMWARE_CONFIG 58 + #define OPCODE_COMMON_NTWK_PMAC_ADD 59 + #define OPCODE_COMMON_NTWK_PMAC_DEL 60 + #define OPCODE_COMMON_FUNCTION_RESET 61 ++#define OPCODE_COMMON_MANAGE_FAT 68 ++#define OPCODE_COMMON_ENABLE_DISABLE_BEACON 69 ++#define OPCODE_COMMON_GET_BEACON_STATE 70 ++#define OPCODE_COMMON_READ_TRANSRECV_DATA 73 ++#define OPCODE_COMMON_GET_PORT_NAME 77 ++#define OPCODE_COMMON_SET_FN_PRIVILEGES 100 ++#define OPCODE_COMMON_GET_PHY_DETAILS 102 ++#define OPCODE_COMMON_SET_DRIVER_FUNCTION_CAP 103 ++#define OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES 121 ++#define OPCODE_COMMON_NTWK_GET_LINK_SPEED 134 ++#define OPCODE_COMMON_GET_HSW_CONFIG 152 ++#define OPCODE_COMMON_SET_HSW_CONFIG 153 ++#define OPCODE_COMMON_GET_FN_PRIVILEGES 170 + ++#define OPCODE_ETH_RSS_CONFIG 1 + #define OPCODE_ETH_ACPI_CONFIG 2 + #define OPCODE_ETH_PROMISCUOUS 3 + #define OPCODE_ETH_GET_STATISTICS 4 + #define OPCODE_ETH_TX_CREATE 7 +-#define OPCODE_ETH_RX_CREATE 8 +-#define OPCODE_ETH_TX_DESTROY 9 +-#define OPCODE_ETH_RX_DESTROY 10 ++#define OPCODE_ETH_RX_CREATE 8 ++#define OPCODE_ETH_TX_DESTROY 9 ++#define OPCODE_ETH_RX_DESTROY 10 ++#define OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG 12 ++#define OPCODE_ETH_PG_FEATURE_QUERY_REQUEST 23 ++ ++#define OPCODE_LOWLEVEL_HOST_DDR_DMA 17 ++#define OPCODE_LOWLEVEL_LOOPBACK_TEST 18 ++#define OPCODE_LOWLEVEL_SET_LOOPBACK_MODE 19 + + struct be_cmd_req_hdr { + u8 opcode; /* dword 0 */ +@@ -159,7 +238,7 @@ struct be_cmd_req_hdr { + }; + + #define RESP_HDR_INFO_OPCODE_SHIFT 0 /* bits 0 - 7 */ +-#define RESP_HDR_INFO_SUBSYS_SHIFT 8 /* bits 8 - 15 */ ++#define RESP_HDR_INFO_SUBSYS_SHIFT 8 /* bits 8 - 15 */ + struct be_cmd_resp_hdr { + u32 info; /* dword 0 */ + u32 status; /* dword 1 */ +@@ -265,7 +344,7 @@ struct be_cmd_req_pmac_del { + /******************** Create CQ ***************************/ + /* Pseudo amap definition in which each bit of the actual structure is defined + * as a byte: used to calculate offset/shift/mask of each field */ +-struct amap_cq_context { ++struct amap_cq_context_be { + u8 cidx[11]; /* dword 0*/ + u8 rsvd0; /* dword 0*/ + u8 coalescwm[2]; /* dword 0*/ +@@ -288,11 +367,28 @@ struct amap_cq_context { + u8 rsvd5[32]; /* dword 3*/ + } __packed; + ++struct amap_cq_context_lancer { ++ u8 rsvd0[12]; /* dword 0*/ ++ u8 coalescwm[2]; /* dword 0*/ ++ u8 nodelay; /* dword 0*/ ++ u8 rsvd1[12]; /* dword 0*/ ++ u8 count[2]; /* dword 0*/ ++ u8 valid; /* dword 0*/ ++ u8 rsvd2; /* dword 0*/ ++ u8 eventable; /* dword 0*/ ++ u8 eqid[16]; /* dword 1*/ ++ u8 rsvd3[15]; /* dword 1*/ ++ u8 armed; /* dword 1*/ ++ u8 rsvd4[32]; /* dword 2*/ ++ u8 rsvd5[32]; /* dword 3*/ ++} __packed; ++ + struct be_cmd_req_cq_create { + struct be_cmd_req_hdr hdr; + u16 num_pages; +- u16 rsvd0; +- u8 context[sizeof(struct amap_cq_context) / 8]; ++ u8 page_size; ++ u8 rsvd0; ++ u8 context[sizeof(struct amap_cq_context_be) / 8]; + struct phys_addr pages[8]; + } __packed; + +@@ -302,10 +398,28 @@ struct be_cmd_resp_cq_create { + u16 rsvd0; + } __packed; + ++struct be_cmd_req_get_fat { ++ struct be_cmd_req_hdr hdr; ++ u32 fat_operation; ++ u32 read_log_offset; ++ u32 read_log_length; ++ u32 data_buffer_size; ++ u32 data_buffer[1]; ++} __packed; ++ ++struct be_cmd_resp_get_fat { ++ struct be_cmd_resp_hdr hdr; ++ u32 log_size; ++ u32 read_log_length; ++ u32 rsvd[2]; ++ u32 data_buffer[1]; ++} __packed; ++ ++ + /******************** Create MCCQ ***************************/ + /* Pseudo amap definition in which each bit of the actual structure is defined + * as a byte: used to calculate offset/shift/mask of each field */ +-struct amap_mcc_context { ++struct amap_mcc_context_be { + u8 con_index[14]; + u8 rsvd0[2]; + u8 ring_size[4]; +@@ -320,11 +434,31 @@ struct amap_mcc_context { + u8 rsvd2[32]; + } __packed; + ++struct amap_mcc_context_lancer { ++ u8 async_cq_id[16]; ++ u8 ring_size[4]; ++ u8 rsvd0[12]; ++ u8 rsvd1[31]; ++ u8 valid; ++ u8 async_cq_valid[1]; ++ u8 rsvd2[31]; ++ u8 rsvd3[32]; ++} __packed; ++ + struct be_cmd_req_mcc_create { + struct be_cmd_req_hdr hdr; + u16 num_pages; +- u16 rsvd0; +- u8 context[sizeof(struct amap_mcc_context) / 8]; ++ u16 cq_id; ++ u8 context[sizeof(struct amap_mcc_context_be) / 8]; ++ struct phys_addr pages[8]; ++} __packed; ++ ++struct be_cmd_req_mcc_ext_create { ++ struct be_cmd_req_hdr hdr; ++ u16 num_pages; ++ u16 cq_id; ++ u32 async_event_bitmap[1]; ++ u8 context[sizeof(struct amap_mcc_context_be) / 8]; + struct phys_addr pages[8]; + } __packed; + +@@ -335,49 +469,32 @@ struct be_cmd_resp_mcc_create { + } __packed; + + /******************** Create TxQ ***************************/ +-#define BE_ETH_TX_RING_TYPE_STANDARD 2 ++#define ETX_QUEUE_TYPE_STANDARD 0x2 ++#define ETX_QUEUE_TYPE_PRIORITY 0x10 + #define BE_ULP1_NUM 1 + +-/* Pseudo amap definition in which each bit of the actual structure is defined +- * as a byte: used to calculate offset/shift/mask of each field */ +-struct amap_tx_context { +- u8 rsvd0[16]; /* dword 0 */ +- u8 tx_ring_size[4]; /* dword 0 */ +- u8 rsvd1[26]; /* dword 0 */ +- u8 pci_func_id[8]; /* dword 1 */ +- u8 rsvd2[9]; /* dword 1 */ +- u8 ctx_valid; /* dword 1 */ +- u8 cq_id_send[16]; /* dword 2 */ +- u8 rsvd3[16]; /* dword 2 */ +- u8 rsvd4[32]; /* dword 3 */ +- u8 rsvd5[32]; /* dword 4 */ +- u8 rsvd6[32]; /* dword 5 */ +- u8 rsvd7[32]; /* dword 6 */ +- u8 rsvd8[32]; /* dword 7 */ +- u8 rsvd9[32]; /* dword 8 */ +- u8 rsvd10[32]; /* dword 9 */ +- u8 rsvd11[32]; /* dword 10 */ +- u8 rsvd12[32]; /* dword 11 */ +- u8 rsvd13[32]; /* dword 12 */ +- u8 rsvd14[32]; /* dword 13 */ +- u8 rsvd15[32]; /* dword 14 */ +- u8 rsvd16[32]; /* dword 15 */ +-} __packed; +- + struct be_cmd_req_eth_tx_create { + struct be_cmd_req_hdr hdr; + u8 num_pages; + u8 ulp_num; +- u8 type; +- u8 bound_port; +- u8 context[sizeof(struct amap_tx_context) / 8]; ++ u16 type; ++ u16 if_id; ++ u8 queue_size; ++ u8 rsvd1; ++ u32 rsvd2; ++ u16 cq_id; ++ u16 rsvd3; ++ u32 rsvd4[13]; + struct phys_addr pages[8]; + } __packed; + + struct be_cmd_resp_eth_tx_create { + struct be_cmd_resp_hdr hdr; + u16 cid; +- u16 rsvd0; ++ u16 rid; ++ u32 db_offset; ++ u8 tc_id; ++ u8 rsvd0[3]; + } __packed; + + /******************** Create RxQ ***************************/ +@@ -396,7 +513,7 @@ struct be_cmd_req_eth_rx_create { + struct be_cmd_resp_eth_rx_create { + struct be_cmd_resp_hdr hdr; + u16 id; +- u8 cpu_id; ++ u8 rss_id; + u8 rsvd0; + } __packed; + +@@ -429,14 +546,15 @@ enum be_if_flags { + BE_IF_FLAGS_VLAN = 0x100, + BE_IF_FLAGS_MCAST_PROMISCUOUS = 0x200, + BE_IF_FLAGS_PASS_L2_ERRORS = 0x400, +- BE_IF_FLAGS_PASS_L3L4_ERRORS = 0x800 ++ BE_IF_FLAGS_PASS_L3L4_ERRORS = 0x800, ++ BE_IF_FLAGS_MULTICAST = 0x1000 + }; + + /* An RX interface is an object with one or more MAC addresses and + * filtering capabilities. */ + struct be_cmd_req_if_create { + struct be_cmd_req_hdr hdr; +- u32 version; /* ignore currntly */ ++ u32 version; /* ignore currently */ + u32 capability_flags; + u32 enable_flags; + u8 mac_addr[ETH_ALEN]; +@@ -458,7 +576,7 @@ struct be_cmd_req_if_destroy { + }; + + /*************** HW Stats Get **********************************/ +-struct be_port_rxf_stats { ++struct be_port_rxf_stats_v0 { + u32 rx_bytes_lsd; /* dword 0*/ + u32 rx_bytes_msd; /* dword 1*/ + u32 rx_total_frames; /* dword 2*/ +@@ -527,8 +645,8 @@ struct be_port_rxf_stats { + u32 rx_input_fifo_overflow; /* dword 65*/ + }; + +-struct be_rxf_stats { +- struct be_port_rxf_stats port[2]; ++struct be_rxf_stats_v0 { ++ struct be_port_rxf_stats_v0 port[2]; + u32 rx_drops_no_pbuf; /* dword 132*/ + u32 rx_drops_no_txpb; /* dword 133*/ + u32 rx_drops_no_erx_descr; /* dword 134*/ +@@ -545,31 +663,51 @@ struct be_rxf_stats { + u32 rx_drops_invalid_ring; /* dword 145*/ + u32 forwarded_packets; /* dword 146*/ + u32 rx_drops_mtu; /* dword 147*/ +- u32 rsvd0[15]; ++ u32 rsvd0[7]; ++ u32 port0_jabber_events; ++ u32 port1_jabber_events; ++ u32 rsvd1[6]; + }; + +-struct be_erx_stats { ++struct be_erx_stats_v0 { + u32 rx_drops_no_fragments[44]; /* dwordS 0 to 43*/ +- u32 debug_wdma_sent_hold; /* dword 44*/ +- u32 debug_wdma_pbfree_sent_hold; /* dword 45*/ +- u32 debug_wdma_zerobyte_pbfree_sent_hold; /* dword 46*/ +- u32 debug_pmem_pbuf_dealloc; /* dword 47*/ ++ u32 rsvd[4]; + }; + +-struct be_hw_stats { +- struct be_rxf_stats rxf; ++struct be_pmem_stats { ++ u32 eth_red_drops; ++ u32 rsvd[5]; ++}; ++ ++struct be_hw_stats_v0 { ++ struct be_rxf_stats_v0 rxf; + u32 rsvd[48]; +- struct be_erx_stats erx; ++ struct be_erx_stats_v0 erx; ++ struct be_pmem_stats pmem; + }; + +-struct be_cmd_req_get_stats { ++struct be_cmd_req_get_stats_v0 { + struct be_cmd_req_hdr hdr; +- u8 rsvd[sizeof(struct be_hw_stats)]; ++ u8 rsvd[sizeof(struct be_hw_stats_v0)]; + }; + +-struct be_cmd_resp_get_stats { ++struct be_cmd_resp_get_stats_v0 { + struct be_cmd_resp_hdr hdr; +- struct be_hw_stats hw_stats; ++ struct be_hw_stats_v0 hw_stats; ++}; ++ ++struct be_cmd_req_get_cntl_addnl_attribs { ++ struct be_cmd_req_hdr hdr; ++ u8 rsvd[8]; ++}; ++ ++struct be_cmd_resp_get_cntl_addnl_attribs { ++ struct be_cmd_resp_hdr hdr; ++ u16 ipl_file_number; ++ u8 ipl_file_version; ++ u8 rsvd0; ++ u8 on_die_temperature; /* in degrees centigrade*/ ++ u8 rsvd1[3]; + }; + + struct be_cmd_req_vlan_config { +@@ -581,30 +719,22 @@ struct be_cmd_req_vlan_config { + u16 normal_vlan[64]; + } __packed; + +-struct be_cmd_req_promiscuous_config { +- struct be_cmd_req_hdr hdr; +- u8 port0_promiscuous; +- u8 port1_promiscuous; +- u16 rsvd0; +-} __packed; +- ++/******************** RX FILTER ******************************/ ++#define BE_MAX_MC 64 /* set mcast promisc if > 64 */ + struct macaddr { + u8 byte[ETH_ALEN]; + }; + +-struct be_cmd_req_mcast_mac_config { ++struct be_cmd_req_rx_filter { + struct be_cmd_req_hdr hdr; +- u16 num_mac; +- u8 promiscuous; +- u8 interface_id; +- struct macaddr mac[32]; +-} __packed; +- +-static inline struct be_hw_stats * +-hw_stats_from_cmd(struct be_cmd_resp_get_stats *cmd) +-{ +- return &cmd->hw_stats; +-} ++ u32 global_flags_mask; ++ u32 global_flags; ++ u32 if_flags_mask; ++ u32 if_flags; ++ u32 if_id; ++ u32 mcast_num; ++ struct macaddr mcast_mac[BE_MAX_MC]; ++}; + + /******************** Link Status Query *******************/ + struct be_cmd_req_link_status { +@@ -619,13 +749,18 @@ enum { + }; + + enum { +- PHY_LINK_SPEED_ZERO = 0x0, /* => No link */ ++ PHY_LINK_SPEED_ZERO = 0x0, /* => No link */ + PHY_LINK_SPEED_10MBPS = 0x1, + PHY_LINK_SPEED_100MBPS = 0x2, + PHY_LINK_SPEED_1GBPS = 0x3, + PHY_LINK_SPEED_10GBPS = 0x4 + }; + ++enum { ++ LINK_DOWN = 0x0, ++ LINK_UP = 0X1 ++}; ++ + struct be_cmd_resp_link_status { + struct be_cmd_resp_hdr hdr; + u8 physical_port; +@@ -634,9 +769,47 @@ struct be_cmd_resp_link_status { + u8 mac_fault; + u8 mgmt_mac_duplex; + u8 mgmt_mac_speed; +- u16 rsvd0; ++ u16 link_speed; ++ u32 logical_link_status; + } __packed; + ++/******************** Port Identification ***************************/ ++/* Identifies the type of port attached to NIC */ ++struct be_cmd_req_port_type { ++ struct be_cmd_req_hdr hdr; ++ u32 page_num; ++ u32 port; ++}; ++ ++enum { ++ TR_PAGE_A0 = 0xa0, ++ TR_PAGE_A2 = 0xa2 ++}; ++ ++struct be_cmd_resp_port_type { ++ struct be_cmd_resp_hdr hdr; ++ u32 page_num; ++ u32 port; ++ struct data { ++ u8 identifier; ++ u8 identifier_ext; ++ u8 connector; ++ u8 transceiver[8]; ++ u8 rsvd0[3]; ++ u8 length_km; ++ u8 length_hm; ++ u8 length_om1; ++ u8 length_om2; ++ u8 length_cu; ++ u8 length_cu_m; ++ u8 vendor_name[16]; ++ u8 rsvd; ++ u8 vendor_oui[3]; ++ u8 vendor_pn[16]; ++ u8 vendor_rev[4]; ++ } data; ++}; ++ + /******************** Get FW Version *******************/ + struct be_cmd_req_get_fw_version { + struct be_cmd_req_hdr hdr; +@@ -686,9 +859,13 @@ struct be_cmd_resp_modify_eq_delay { + } __packed; + + /******************** Get FW Config *******************/ ++#define FLEX10_MODE 0x400 ++#define VNIC_MODE 0x20000 ++#define UMC_ENABLED 0x1000000 ++ + struct be_cmd_req_query_fw_cfg { + struct be_cmd_req_hdr hdr; +- u32 rsvd[30]; ++ u32 rsvd[31]; + }; + + struct be_cmd_resp_query_fw_cfg { +@@ -696,10 +873,61 @@ struct be_cmd_resp_query_fw_cfg { + u32 be_config_number; + u32 asic_revision; + u32 phys_port; +- u32 function_cap; ++ u32 function_mode; + u32 rsvd[26]; ++ u32 function_caps; + }; + ++/******************** RSS Config *******************/ ++/* RSS types */ ++#define RSS_ENABLE_NONE 0x0 ++#define RSS_ENABLE_IPV4 0x1 ++#define RSS_ENABLE_TCP_IPV4 0x2 ++#define RSS_ENABLE_IPV6 0x4 ++#define RSS_ENABLE_TCP_IPV6 0x8 ++ ++struct be_cmd_req_rss_config { ++ struct be_cmd_req_hdr hdr; ++ u32 if_id; ++ u16 enable_rss; ++ u16 cpu_table_size_log2; ++ u32 hash[10]; ++ u8 cpu_table[128]; ++ u8 flush; ++ u8 rsvd0[3]; ++}; ++ ++/******************** Port Beacon ***************************/ ++ ++#define BEACON_STATE_ENABLED 0x1 ++#define BEACON_STATE_DISABLED 0x0 ++ ++struct be_cmd_req_enable_disable_beacon { ++ struct be_cmd_req_hdr hdr; ++ u8 port_num; ++ u8 beacon_state; ++ u8 beacon_duration; ++ u8 status_duration; ++} __packed; ++ ++struct be_cmd_resp_enable_disable_beacon { ++ struct be_cmd_resp_hdr resp_hdr; ++ u32 rsvd0; ++} __packed; ++ ++struct be_cmd_req_get_beacon_state { ++ struct be_cmd_req_hdr hdr; ++ u8 port_num; ++ u8 rsvd0; ++ u16 rsvd1; ++} __packed; ++ ++struct be_cmd_resp_get_beacon_state { ++ struct be_cmd_resp_hdr resp_hdr; ++ u8 beacon_state; ++ u8 rsvd0[3]; ++} __packed; ++ + /****************** Firmware Flash ******************/ + struct flashrom_params { + u32 op_code; +@@ -714,17 +942,468 @@ struct be_cmd_write_flashrom { + struct flashrom_params params; + }; + ++/************************ WOL *******************************/ ++struct be_cmd_req_acpi_wol_magic_config { ++ struct be_cmd_req_hdr hdr; ++ u32 rsvd0[145]; ++ u8 magic_mac[6]; ++ u8 rsvd2[2]; ++} __packed; ++ ++/********************** LoopBack test *********************/ ++struct be_cmd_req_loopback_test { ++ struct be_cmd_req_hdr hdr; ++ u32 loopback_type; ++ u32 num_pkts; ++ u64 pattern; ++ u32 src_port; ++ u32 dest_port; ++ u32 pkt_size; ++}; ++ ++struct be_cmd_resp_loopback_test { ++ struct be_cmd_resp_hdr resp_hdr; ++ u32 status; ++ u32 num_txfer; ++ u32 num_rx; ++ u32 miscomp_off; ++ u32 ticks_compl; ++}; ++ ++struct be_cmd_req_set_lmode { ++ struct be_cmd_req_hdr hdr; ++ u8 src_port; ++ u8 dest_port; ++ u8 loopback_type; ++ u8 loopback_state; ++}; ++ ++struct be_cmd_resp_set_lmode { ++ struct be_cmd_resp_hdr resp_hdr; ++ u8 rsvd0[4]; ++}; ++ ++/********************** DDR DMA test *********************/ ++struct be_cmd_req_ddrdma_test { ++ struct be_cmd_req_hdr hdr; ++ u64 pattern; ++ u32 byte_count; ++ u32 rsvd0; ++ u8 snd_buff[4096]; ++ u8 rsvd1[4096]; ++}; ++ ++struct be_cmd_resp_ddrdma_test { ++ struct be_cmd_resp_hdr hdr; ++ u64 pattern; ++ u32 byte_cnt; ++ u32 snd_err; ++ u8 rsvd0[4096]; ++ u8 rcv_buff[4096]; ++}; ++ ++/*********************** SEEPROM Read ***********************/ ++ ++#define BE_READ_SEEPROM_LEN 1024 ++struct be_cmd_req_seeprom_read { ++ struct be_cmd_req_hdr hdr; ++ u8 rsvd0[BE_READ_SEEPROM_LEN]; ++}; ++ ++struct be_cmd_resp_seeprom_read { ++ struct be_cmd_req_hdr hdr; ++ u8 seeprom_data[BE_READ_SEEPROM_LEN]; ++}; ++ ++enum { ++ PHY_TYPE_CX4_10GB = 0, ++ PHY_TYPE_XFP_10GB, ++ PHY_TYPE_SFP_1GB, ++ PHY_TYPE_SFP_PLUS_10GB, ++ PHY_TYPE_KR_10GB, ++ PHY_TYPE_KX4_10GB, ++ PHY_TYPE_BASET_10GB, ++ PHY_TYPE_BASET_1GB, ++ PHY_TYPE_BASEX_1GB, ++ PHY_TYPE_SGMII, ++ PHY_TYPE_DISABLED = 255 ++}; ++ ++#define BE_AN_EN 0x2 ++#define BE_PAUSE_SYM_EN 0x80 ++ ++struct be_cmd_req_get_phy_info { ++ struct be_cmd_req_hdr hdr; ++ u8 rsvd0[24]; ++}; ++ ++struct be_phy_info { ++ u16 phy_type; ++ u16 interface_type; ++ u32 misc_params; ++ u16 ext_phy_details; ++ u16 rsvd; ++ u16 auto_speeds_supported; ++ u16 fixed_speeds_supported; ++ u32 future_use[2]; ++}; ++ ++struct be_cmd_resp_get_phy_info { ++ struct be_cmd_req_hdr hdr; ++ struct be_phy_info phy_info; ++}; ++ ++/*********************** Set QOS ***********************/ ++ ++#define BE_QOS_BITS_NIC 1 ++ ++struct be_cmd_req_set_qos { ++ struct be_cmd_req_hdr hdr; ++ u32 valid_bits; ++ u32 max_bps_nic; ++ u32 rsvd[7]; ++}; ++ ++struct be_cmd_resp_set_qos { ++ struct be_cmd_resp_hdr hdr; ++ u32 rsvd; ++}; ++ ++/*********************** Controller Attributes ***********************/ ++struct be_cmd_req_cntl_attribs { ++ struct be_cmd_req_hdr hdr; ++}; ++ ++struct be_cmd_resp_cntl_attribs { ++ struct be_cmd_resp_hdr hdr; ++ struct mgmt_controller_attrib attribs; ++}; ++ ++/******************* get port names ***************/ ++struct be_cmd_req_get_port_name { ++ struct be_cmd_req_hdr hdr; ++ u32 rsvd0; ++}; ++ ++struct be_cmd_resp_get_port_name { ++ struct be_cmd_req_hdr hdr; ++ u8 port0_name; ++ u8 port1_name; ++ u8 rsvd0[2]; ++}; ++ ++struct be_cmd_resp_get_port_name_v1 { ++ struct be_cmd_req_hdr hdr; ++ u32 pt : 2; ++ u32 rsvd0 : 30; ++ u8 port0_name; ++ u8 port1_name; ++ u8 port2_name; ++ u8 port3_name; ++}; ++ ++/*********************** Set driver function ***********************/ ++#define CAPABILITY_SW_TIMESTAMPS 2 ++#define CAPABILITY_BE3_NATIVE_ERX_API 4 ++ ++struct be_cmd_req_set_func_cap { ++ struct be_cmd_req_hdr hdr; ++ u32 valid_cap_flags; ++ u32 cap_flags; ++ u8 rsvd[212]; ++}; ++ ++struct be_cmd_resp_set_func_cap { ++ struct be_cmd_resp_hdr hdr; ++ u32 valid_cap_flags; ++ u32 cap_flags; ++ u8 rsvd[212]; ++}; ++ ++/*********************** PG Query Request ****************************/ ++#define REQ_PG_QUERY 0x1 ++#define REQ_PG_FEAT 0x1 ++struct be_cmd_req_pg { ++ struct be_cmd_req_hdr hdr; ++ u32 query; ++ u32 pfc_pg; ++}; ++ ++struct be_cmd_resp_pg { ++ struct be_cmd_resp_hdr hdr; ++ u32 pfc_pg; ++ u32 num_tx_rings; ++}; ++ ++/*********************** Function Privileges ***********************/ ++enum { ++ BE_PRIV_DEFAULT = 0x1, ++ BE_PRIV_LNKQUERY = 0x2, ++ BE_PRIV_LNKSTATS = 0x4, ++ BE_PRIV_LNKMGMT = 0x8, ++ BE_PRIV_LNKDIAG = 0x10, ++ BE_PRIV_UTILQUERY = 0x20, ++ BE_PRIV_FILTMGMT = 0x40, ++ BE_PRIV_IFACEMGMT = 0x80, ++ BE_PRIV_VHADM = 0x100, ++ BE_PRIV_DEVCFG = 0x200, ++ BE_PRIV_DEVSEC = 0x400 ++}; ++ ++struct be_cmd_req_get_fn_privileges { ++ struct be_cmd_req_hdr hdr; ++ u32 rsvd; ++}; ++ ++struct be_cmd_resp_get_fn_privileges { ++ struct be_cmd_resp_hdr hdr; ++ u32 privilege_mask; ++}; ++ ++struct be_cmd_req_set_fn_privileges { ++ struct be_cmd_req_hdr hdr; ++ u32 privilege_mask; ++}; ++ ++struct be_cmd_resp_set_fn_privileges { ++ struct be_cmd_resp_hdr hdr; ++ u32 prev_privilege_mask; ++}; ++ ++/*********************** HSW Config ***********************/ ++struct amap_set_hsw_context { ++ u8 interface_id[16]; ++ u8 rsvd0[14]; ++ u8 pvid_valid; ++ u8 rsvd1; ++ u8 rsvd2[16]; ++ u8 pvid[16]; ++ u8 rsvd3[32]; ++ u8 rsvd4[32]; ++ u8 rsvd5[32]; ++} __packed; ++ ++struct be_cmd_req_set_hsw_config { ++ struct be_cmd_req_hdr hdr; ++ u8 context[sizeof(struct amap_set_hsw_context) / 8]; ++} __packed; ++ ++struct be_cmd_resp_set_hsw_config { ++ struct be_cmd_resp_hdr hdr; ++ u32 rsvd; ++}; ++ ++struct amap_get_hsw_req_context { ++ u8 interface_id[16]; ++ u8 rsvd0[14]; ++ u8 pvid_valid; ++ u8 pport; ++} __packed; ++ ++struct amap_get_hsw_resp_context { ++ u8 rsvd1[16]; ++ u8 pvid[16]; ++ u8 rsvd2[32]; ++ u8 rsvd3[32]; ++ u8 rsvd4[32]; ++} __packed; ++ ++struct be_cmd_req_get_hsw_config { ++ struct be_cmd_req_hdr hdr; ++ u8 context[sizeof(struct amap_get_hsw_req_context) / 8]; ++} __packed; ++ ++struct be_cmd_resp_get_hsw_config { ++ struct be_cmd_resp_hdr hdr; ++ u8 context[sizeof(struct amap_get_hsw_resp_context) / 8]; ++ u32 rsvd; ++}; ++ ++/*************** Set speed ********************/ ++struct be_cmd_req_set_port_speed_v1 { ++ struct be_cmd_req_hdr hdr; ++ u8 port_num; ++ u8 virt_port; ++ u16 mac_speed; ++ u16 dac_cable_length; ++ u16 rsvd0; ++}; ++ ++struct be_cmd_resp_set_port_speed_v1 { ++ struct be_cmd_resp_hdr hdr; ++ u32 rsvd0; ++}; ++ ++/************** get port speed *******************/ ++struct be_cmd_req_get_port_speed { ++ struct be_cmd_req_hdr hdr; ++ u8 port_num; ++}; ++ ++struct be_cmd_resp_get_port_speed { ++ struct be_cmd_req_hdr hdr; ++ u16 mac_speed; ++ u16 dac_cable_length; ++}; ++ ++/*************** HW Stats Get v1 **********************************/ ++#define BE_TXP_SW_SZ 48 ++struct be_port_rxf_stats_v1 { ++ u32 rsvd0[12]; ++ u32 rx_crc_errors; ++ u32 rx_alignment_symbol_errors; ++ u32 rx_pause_frames; ++ u32 rx_priority_pause_frames; ++ u32 rx_control_frames; ++ u32 rx_in_range_errors; ++ u32 rx_out_range_errors; ++ u32 rx_frame_too_long; ++ u32 rx_address_match_errors; ++ u32 rx_dropped_too_small; ++ u32 rx_dropped_too_short; ++ u32 rx_dropped_header_too_small; ++ u32 rx_dropped_tcp_length; ++ u32 rx_dropped_runt; ++ u32 rsvd1[10]; ++ u32 rx_ip_checksum_errs; ++ u32 rx_tcp_checksum_errs; ++ u32 rx_udp_checksum_errs; ++ u32 rsvd2[7]; ++ u32 rx_switched_unicast_packets; ++ u32 rx_switched_multicast_packets; ++ u32 rx_switched_broadcast_packets; ++ u32 rsvd3[3]; ++ u32 tx_pauseframes; ++ u32 tx_priority_pauseframes; ++ u32 tx_controlframes; ++ u32 rsvd4[10]; ++ u32 rxpp_fifo_overflow_drop; ++ u32 rx_input_fifo_overflow_drop; ++ u32 pmem_fifo_overflow_drop; ++ u32 jabber_events; ++ u32 rsvd5[3]; ++}; ++ ++ ++struct be_rxf_stats_v1 { ++ struct be_port_rxf_stats_v1 port[4]; ++ u32 rsvd0[2]; ++ u32 rx_drops_no_pbuf; ++ u32 rx_drops_no_txpb; ++ u32 rx_drops_no_erx_descr; ++ u32 rx_drops_no_tpre_descr; ++ u32 rsvd1[6]; ++ u32 rx_drops_too_many_frags; ++ u32 rx_drops_invalid_ring; ++ u32 forwarded_packets; ++ u32 rx_drops_mtu; ++ u32 rsvd2[14]; ++}; ++ ++struct be_erx_stats_v1 { ++ u32 rx_drops_no_fragments[68]; /* dwordS 0 to 67*/ ++ u32 rsvd[4]; ++}; ++ ++struct be_hw_stats_v1 { ++ struct be_rxf_stats_v1 rxf; ++ u32 rsvd0[BE_TXP_SW_SZ]; ++ struct be_erx_stats_v1 erx; ++ struct be_pmem_stats pmem; ++ u32 rsvd1[3]; ++}; ++ ++struct be_cmd_req_get_stats_v1 { ++ struct be_cmd_req_hdr hdr; ++ u8 rsvd[sizeof(struct be_hw_stats_v1)]; ++}; ++ ++struct be_cmd_resp_get_stats_v1 { ++ struct be_cmd_resp_hdr hdr; ++ struct be_hw_stats_v1 hw_stats; ++}; ++ ++static inline void * ++hw_stats_from_cmd(struct be_adapter *adapter) ++{ ++ if (adapter->generation == BE_GEN3) { ++ struct be_cmd_resp_get_stats_v1 *cmd = adapter->stats_cmd.va; ++ ++ return &cmd->hw_stats; ++ } else { ++ struct be_cmd_resp_get_stats_v0 *cmd = adapter->stats_cmd.va; ++ ++ return &cmd->hw_stats; ++ } ++} ++ ++static inline void *be_port_rxf_stats_from_cmd(struct be_adapter *adapter) ++{ ++ if (adapter->generation == BE_GEN3) { ++ struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter); ++ struct be_rxf_stats_v1 *rxf_stats = &hw_stats->rxf; ++ ++ return &rxf_stats->port[adapter->port_num]; ++ } else { ++ struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter); ++ struct be_rxf_stats_v0 *rxf_stats = &hw_stats->rxf; ++ ++ return &rxf_stats->port[adapter->port_num]; ++ } ++} ++ ++static inline void *be_rxf_stats_from_cmd(struct be_adapter *adapter) ++{ ++ if (adapter->generation == BE_GEN3) { ++ struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->rxf; ++ } else { ++ struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->rxf; ++ } ++} ++ ++static inline void *be_erx_stats_from_cmd(struct be_adapter *adapter) ++{ ++ if (adapter->generation == BE_GEN3) { ++ struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->erx; ++ } else { ++ struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->erx; ++ } ++} ++ ++static inline void *be_pmem_stats_from_cmd(struct be_adapter *adapter) ++{ ++ if (adapter->generation == BE_GEN3) { ++ struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->pmem; ++ } else { ++ struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter); ++ ++ return &hw_stats->pmem; ++ } ++} ++ + extern int be_pci_fnum_get(struct be_adapter *adapter); + extern int be_cmd_POST(struct be_adapter *adapter); + extern int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr, + u8 type, bool permanent, u32 if_handle); + extern int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, +- u32 if_id, u32 *pmac_id); +-extern int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, u32 pmac_id); ++ u32 if_id, u32 *pmac_id, u32 domain); ++extern int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, u32 pmac_id, ++ u32 domain); + extern int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, + u32 en_flags, u8 *mac, bool pmac_invalid, +- u32 *if_handle, u32 *pmac_id); +-extern int be_cmd_if_destroy(struct be_adapter *adapter, u32 if_handle); ++ u32 *if_handle, u32 *pmac_id, u32 domain); ++extern int be_cmd_if_destroy(struct be_adapter *adapter, u32 if_handle, ++ u32 domain); + extern int be_cmd_eq_create(struct be_adapter *adapter, + struct be_queue_info *eq, int eq_delay); + extern int be_cmd_cq_create(struct be_adapter *adapter, +@@ -736,36 +1415,92 @@ extern int be_cmd_mccq_create(struct be_adapter *adapter, + struct be_queue_info *cq); + extern int be_cmd_txq_create(struct be_adapter *adapter, + struct be_queue_info *txq, +- struct be_queue_info *cq); ++ struct be_queue_info *cq, u8 *tc_id); + extern int be_cmd_rxq_create(struct be_adapter *adapter, + struct be_queue_info *rxq, u16 cq_id, + u16 frag_size, u16 max_frame_size, u32 if_id, +- u32 rss); ++ u32 rss, u8 *rss_id); + extern int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q, + int type); ++extern int be_cmd_rxq_destroy(struct be_adapter *adapter, ++ struct be_queue_info *q); + extern int be_cmd_link_status_query(struct be_adapter *adapter, +- bool *link_up); ++ int *link_status, u8 *mac_speed, u16 *link_speed, u32 dom); + extern int be_cmd_reset(struct be_adapter *adapter); + extern int be_cmd_get_stats(struct be_adapter *adapter, + struct be_dma_mem *nonemb_cmd); +-extern int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver); ++extern int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver, ++ char *fw_on_flash); + + extern int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd); + extern int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, + u16 *vtag_array, u32 num, bool untagged, + bool promiscuous); +-extern int be_cmd_promiscuous_config(struct be_adapter *adapter, +- u8 port_num, bool en); +-extern int be_cmd_multicast_set(struct be_adapter *adapter, u32 if_id, +- struct dev_mc_list *mc_list, u32 mc_count); ++extern int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 status); + extern int be_cmd_set_flow_control(struct be_adapter *adapter, + u32 tx_fc, u32 rx_fc); + extern int be_cmd_get_flow_control(struct be_adapter *adapter, + u32 *tx_fc, u32 *rx_fc); +-extern int be_cmd_query_fw_cfg(struct be_adapter *adapter, +- u32 *port_num, u32 *cap); ++extern int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num, ++ u32 *function_mode, u32 *functions_caps); + extern int be_cmd_reset_function(struct be_adapter *adapter); +-extern int be_process_mcc(struct be_adapter *adapter); ++extern int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable, ++ u16 table_size); ++extern int be_process_mcc(struct be_adapter *adapter, int *status); ++extern int be_cmd_set_beacon_state(struct be_adapter *adapter, ++ u8 port_num, u8 beacon, u8 status, u8 state); ++extern int be_cmd_get_beacon_state(struct be_adapter *adapter, ++ u8 port_num, u32 *state); ++extern int be_cmd_read_port_type(struct be_adapter *adapter, u32 port, ++ u8 *connector); + extern int be_cmd_write_flashrom(struct be_adapter *adapter, + struct be_dma_mem *cmd, u32 flash_oper, + u32 flash_opcode, u32 buf_size); ++int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc, ++ int offset); ++extern int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac, ++ struct be_dma_mem *nonemb_cmd); ++extern int be_cmd_fw_init(struct be_adapter *adapter); ++extern int be_cmd_fw_clean(struct be_adapter *adapter); ++extern void be_async_mcc_enable(struct be_adapter *adapter); ++extern void be_async_mcc_disable(struct be_adapter *adapter); ++extern int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num, ++ u32 loopback_type, u32 pkt_size, ++ u32 num_pkts, u64 pattern); ++extern int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern, ++ u32 byte_cnt, struct be_dma_mem *cmd); ++extern int be_cmd_get_seeprom_data(struct be_adapter *adapter, ++ struct be_dma_mem *nonemb_cmd); ++extern int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num, ++ u8 loopback_type, u8 enable); ++extern int be_cmd_get_phy_info(struct be_adapter *adapter, ++ struct be_phy_info *phy_info); ++extern int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain); ++extern void be_detect_dump_ue(struct be_adapter *adapter); ++extern int be_cmd_get_die_temperature(struct be_adapter *adapter); ++extern int be_cmd_get_cntl_attributes(struct be_adapter *adapter); ++extern int be_cmd_get_reg_len(struct be_adapter *adapter, u32 *log_size); ++extern void be_cmd_get_regs(struct be_adapter *adapter, u32 buf_len, void *buf); ++extern int be_cmd_req_native_mode(struct be_adapter *adapter); ++extern int be_cmd_query_port_names_v0(struct be_adapter *adapter, u8 *port_name); ++extern int be_cmd_query_port_names_v1(struct be_adapter *adapter, u8 *port_name); ++extern int be_cmd_req_pg_pfc(struct be_adapter *adapter, int *fw_num_txqs); ++ ++extern int be_cmd_get_fn_privileges(struct be_adapter *adapter, ++ u32 *privilege, u32 domain); ++extern int be_cmd_set_fn_privileges(struct be_adapter *adapter, ++ u32 mask, u32 *prev, u32 domain); ++extern int be_cmd_set_hsw_config(struct be_adapter *adapter, u16 pvid, ++ u32 domain, u16 intf_id); ++extern int be_cmd_get_hsw_config(struct be_adapter *adapter, u16 *pvid, ++ u32 domain, u16 intf_id); ++extern int be_cmd_set_port_speed_v1(struct be_adapter *adapter, u8 port_num, ++ u16 mac_speed, u16 dac_cable_len); ++extern int be_cmd_get_port_speed(struct be_adapter *adapter, u8 port_num, ++ u16 *dac_cable_len, u16 *port_speed); ++#ifdef CONFIG_PALAU ++int be_cmd_pass_ext_ioctl(struct be_adapter *adapter, dma_addr_t dma, ++ int req_size, void *va); ++#endif ++ ++#endif /* !BE_CMDS_H */ +diff --git a/drivers/net/benet/be_compat.c b/drivers/net/benet/be_compat.c +new file mode 100644 +index 0000000..bdd1dba +--- /dev/null ++++ b/drivers/net/benet/be_compat.c +@@ -0,0 +1,630 @@ ++/* ++ * Copyright (C) 2005 - 2011 Emulex ++ * All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License version 2 ++ * as published by the Free Software Foundation. The full GNU General ++ * Public License is included in this distribution in the file called COPYING. ++ * ++ * Contact Information: ++ * linux-drivers@emulex.com ++ * ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 ++ */ ++ ++#include "be.h" ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 28) ++void be_netdev_ops_init(struct net_device *netdev, struct net_device_ops *ops) ++{ ++ netdev->open = ops->ndo_open; ++ netdev->stop = ops->ndo_stop; ++ netdev->hard_start_xmit = ops->ndo_start_xmit; ++ netdev->set_mac_address = ops->ndo_set_mac_address; ++ netdev->get_stats = ops->ndo_get_stats; ++ netdev->set_multicast_list = ops->ndo_set_rx_mode; ++ netdev->change_mtu = ops->ndo_change_mtu; ++ netdev->vlan_rx_register = ops->ndo_vlan_rx_register; ++ netdev->vlan_rx_add_vid = ops->ndo_vlan_rx_add_vid; ++ netdev->vlan_rx_kill_vid = ops->ndo_vlan_rx_kill_vid; ++ netdev->do_ioctl = ops->ndo_do_ioctl; ++#ifdef CONFIG_NET_POLL_CONTROLLER ++ netdev->poll_controller = ops->ndo_poll_controller; ++#endif ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 27) ++ netdev->select_queue = ops->ndo_select_queue; ++#endif ++} ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 28) ++int eth_validate_addr(struct net_device *netdev) ++{ ++ return 0; ++} ++#endif ++ ++/* New NAPI backport */ ++#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 24) ++ ++int be_poll_compat(struct net_device *netdev, int *budget) ++{ ++ struct napi_struct *napi = netdev->priv; ++ u32 work_done, can_do; ++ ++ can_do = min(*budget, netdev->quota); ++ work_done = napi->poll(napi, can_do); ++ ++ *budget -= work_done; ++ netdev->quota -= work_done; ++ if (napi->rx) ++ return (work_done >= can_do); ++ return 0; ++} ++ ++ ++#endif /* New NAPI backport */ ++ ++int be_netif_napi_add(struct net_device *netdev, ++ struct napi_struct *napi, ++ int (*poll) (struct napi_struct *, int), int weight) ++{ ++#ifdef HAVE_SIMULATED_MULTI_NAPI ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct net_device *nd; ++ ++ nd = alloc_netdev(0, "", ether_setup); ++ if (!nd) ++ return -ENOMEM; ++ nd->priv = napi; ++ nd->weight = BE_NAPI_WEIGHT; ++ nd->poll = be_poll_compat; ++ set_bit(__LINK_STATE_START, &nd->state); ++ ++ if (napi == &adapter->rx_obj[0].rx_eq.napi) ++ napi->rx = true; ++ napi->poll = poll; ++ napi->dev = nd; ++#ifdef RHEL_NEW_NAPI ++ napi->napi.dev = netdev; ++#endif ++ return 0; ++#else ++ netif_napi_add(netdev, napi, poll, weight); ++ return 0; ++#endif ++} ++void be_netif_napi_del(struct net_device *netdev) ++{ ++#ifdef HAVE_SIMULATED_MULTI_NAPI ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct napi_struct *napi; ++ struct be_rx_obj *rxo; ++ int i; ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ napi = &rxo->rx_eq.napi; ++ if (napi->dev) { ++ free_netdev(napi->dev); ++ napi->dev = NULL; ++ } ++ } ++ ++ napi = &adapter->tx_eq.napi; ++ if (napi->dev) { ++ free_netdev(napi->dev); ++ napi->dev = NULL; ++ } ++#endif ++} ++/* INET_LRO backport */ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 27) ++ ++#define TCP_HDR_LEN(tcph) (tcph->doff << 2) ++#define IP_HDR_LEN(iph) (iph->ihl << 2) ++#define TCP_PAYLOAD_LENGTH(iph, tcph) (ntohs(iph->tot_len) - IP_HDR_LEN(iph) \ ++ - TCP_HDR_LEN(tcph)) ++ ++#define IPH_LEN_WO_OPTIONS 5 ++#define TCPH_LEN_WO_OPTIONS 5 ++#define TCPH_LEN_W_TIMESTAMP 8 ++ ++#define LRO_MAX_PG_HLEN 64 ++#define LRO_INC_STATS(lro_mgr, attr) { lro_mgr->stats.attr++; } ++/* ++ * Basic tcp checks whether packet is suitable for LRO ++ */ ++static int lro_tcp_ip_check(struct iphdr *iph, struct tcphdr *tcph, ++ int len, struct net_lro_desc *lro_desc) ++{ ++ /* check ip header: don't aggregate padded frames */ ++ if (ntohs(iph->tot_len) != len) ++ return -1; ++ ++ if (iph->ihl != IPH_LEN_WO_OPTIONS) ++ return -1; ++ ++ if (tcph->cwr || tcph->ece || tcph->urg || !tcph->ack ++ || tcph->rst || tcph->syn || tcph->fin) ++ return -1; ++ ++ if (INET_ECN_is_ce(ipv4_get_dsfield(iph))) ++ return -1; ++ ++ if (tcph->doff != TCPH_LEN_WO_OPTIONS ++ && tcph->doff != TCPH_LEN_W_TIMESTAMP) ++ return -1; ++ ++ /* check tcp options (only timestamp allowed) */ ++ if (tcph->doff == TCPH_LEN_W_TIMESTAMP) { ++ u32 *topt = (u32 *)(tcph + 1); ++ ++ if (*topt != htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) ++ | (TCPOPT_TIMESTAMP << 8) ++ | TCPOLEN_TIMESTAMP)) ++ return -1; ++ ++ /* timestamp should be in right order */ ++ topt++; ++ if (lro_desc && after(ntohl(lro_desc->tcp_rcv_tsval), ++ ntohl(*topt))) ++ return -1; ++ ++ /* timestamp reply should not be zero */ ++ topt++; ++ if (*topt == 0) ++ return -1; ++ } ++ ++ return 0; ++} ++ ++static void lro_update_tcp_ip_header(struct net_lro_desc *lro_desc) ++{ ++ struct iphdr *iph = lro_desc->iph; ++ struct tcphdr *tcph = lro_desc->tcph; ++ u32 *p; ++ __wsum tcp_hdr_csum; ++ ++ tcph->ack_seq = lro_desc->tcp_ack; ++ tcph->window = lro_desc->tcp_window; ++ ++ if (lro_desc->tcp_saw_tstamp) { ++ p = (u32 *)(tcph + 1); ++ *(p+2) = lro_desc->tcp_rcv_tsecr; ++ } ++ ++ iph->tot_len = htons(lro_desc->ip_tot_len); ++ ++ iph->check = 0; ++ iph->check = ip_fast_csum((u8 *)lro_desc->iph, iph->ihl); ++ ++ tcph->check = 0; ++ tcp_hdr_csum = csum_partial((u8 *)tcph, TCP_HDR_LEN(tcph), 0); ++ lro_desc->data_csum = csum_add(lro_desc->data_csum, tcp_hdr_csum); ++ tcph->check = csum_tcpudp_magic(iph->saddr, iph->daddr, ++ lro_desc->ip_tot_len - ++ IP_HDR_LEN(iph), IPPROTO_TCP, ++ lro_desc->data_csum); ++} ++ ++static __wsum lro_tcp_data_csum(struct iphdr *iph, struct tcphdr *tcph, int len) ++{ ++ __wsum tcp_csum; ++ __wsum tcp_hdr_csum; ++ __wsum tcp_ps_hdr_csum; ++ ++ tcp_csum = ~csum_unfold(tcph->check); ++ tcp_hdr_csum = csum_partial((u8 *)tcph, TCP_HDR_LEN(tcph), tcp_csum); ++ ++ tcp_ps_hdr_csum = csum_tcpudp_nofold(iph->saddr, iph->daddr, ++ len + TCP_HDR_LEN(tcph), ++ IPPROTO_TCP, 0); ++ ++ return csum_sub(csum_sub(tcp_csum, tcp_hdr_csum), ++ tcp_ps_hdr_csum); ++} ++ ++static void lro_init_desc(struct net_lro_desc *lro_desc, struct sk_buff *skb, ++ struct iphdr *iph, struct tcphdr *tcph, ++ u16 vlan_tag, struct vlan_group *vgrp) ++{ ++ int nr_frags; ++ u32 *ptr; ++ u32 tcp_data_len = TCP_PAYLOAD_LENGTH(iph, tcph); ++ ++ nr_frags = skb_shinfo(skb)->nr_frags; ++ lro_desc->parent = skb; ++ lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); ++ lro_desc->iph = iph; ++ lro_desc->tcph = tcph; ++ lro_desc->tcp_next_seq = ntohl(tcph->seq) + tcp_data_len; ++ lro_desc->tcp_ack = ntohl(tcph->ack_seq); ++ lro_desc->tcp_window = tcph->window; ++ ++ lro_desc->pkt_aggr_cnt = 1; ++ lro_desc->ip_tot_len = ntohs(iph->tot_len); ++ ++ if (tcph->doff == 8) { ++ ptr = (u32 *)(tcph+1); ++ lro_desc->tcp_saw_tstamp = 1; ++ lro_desc->tcp_rcv_tsval = *(ptr+1); ++ lro_desc->tcp_rcv_tsecr = *(ptr+2); ++ } ++ ++ lro_desc->mss = tcp_data_len; ++ lro_desc->vgrp = vgrp; ++ lro_desc->vlan_tag = vlan_tag; ++ lro_desc->active = 1; ++ ++ if (tcp_data_len) ++ lro_desc->data_csum = lro_tcp_data_csum(iph, tcph, ++ tcp_data_len); ++ ++ if (!tcp_data_len) ++ lro_desc->ack_cnt++; ++} ++ ++static inline void lro_clear_desc(struct net_lro_desc *lro_desc) ++{ ++ memset(lro_desc, 0, sizeof(struct net_lro_desc)); ++} ++ ++static void lro_add_common(struct net_lro_desc *lro_desc, struct iphdr *iph, ++ struct tcphdr *tcph, int tcp_data_len) ++{ ++ struct sk_buff *parent = lro_desc->parent; ++ u32 *topt; ++ ++ lro_desc->pkt_aggr_cnt++; ++ lro_desc->ip_tot_len += tcp_data_len; ++ lro_desc->tcp_next_seq += tcp_data_len; ++ lro_desc->tcp_window = tcph->window; ++ lro_desc->tcp_ack = tcph->ack_seq; ++ ++ /* don't update tcp_rcv_tsval, would not work with PAWS */ ++ if (lro_desc->tcp_saw_tstamp) { ++ topt = (u32 *) (tcph + 1); ++ lro_desc->tcp_rcv_tsecr = *(topt + 2); ++ } ++ ++ if (tcp_data_len) ++ lro_desc->data_csum = csum_block_add(lro_desc->data_csum, ++ lro_tcp_data_csum(iph, tcph, ++ tcp_data_len), ++ parent->len); ++ ++ parent->len += tcp_data_len; ++ parent->data_len += tcp_data_len; ++ if (tcp_data_len > lro_desc->mss) ++ lro_desc->mss = tcp_data_len; ++} ++ ++static void lro_add_frags(struct net_lro_desc *lro_desc, ++ int len, int hlen, int truesize, ++ struct skb_frag_struct *skb_frags, ++ struct iphdr *iph, struct tcphdr *tcph) ++{ ++ struct sk_buff *skb = lro_desc->parent; ++ int tcp_data_len = TCP_PAYLOAD_LENGTH(iph, tcph); ++ ++ lro_add_common(lro_desc, iph, tcph, tcp_data_len); ++ ++ skb->truesize += truesize; ++ ++ if (!tcp_data_len) { ++ put_page(skb_frags[0].page); ++ lro_desc->ack_cnt++; ++ return; ++ } ++ ++ skb_frags[0].page_offset += hlen; ++ skb_frags[0].size -= hlen; ++ ++ while (tcp_data_len > 0) { ++ *(lro_desc->next_frag) = *skb_frags; ++ tcp_data_len -= skb_frags->size; ++ lro_desc->next_frag++; ++ skb_frags++; ++ skb_shinfo(skb)->nr_frags++; ++ } ++} ++ ++static int lro_check_tcp_conn(struct net_lro_desc *lro_desc, ++ struct iphdr *iph, ++ struct tcphdr *tcph) ++{ ++ if ((lro_desc->iph->saddr != iph->saddr) ++ || (lro_desc->iph->daddr != iph->daddr) ++ || (lro_desc->tcph->source != tcph->source) ++ || (lro_desc->tcph->dest != tcph->dest)) ++ return -1; ++ return 0; ++} ++ ++static struct net_lro_desc *lro_get_desc(struct net_lro_mgr *lro_mgr, ++ struct net_lro_desc *lro_arr, ++ struct iphdr *iph, ++ struct tcphdr *tcph) ++{ ++ struct net_lro_desc *lro_desc = NULL; ++ struct net_lro_desc *tmp; ++ int max_desc = lro_mgr->max_desc; ++ int i; ++ ++ for (i = 0; i < max_desc; i++) { ++ tmp = &lro_arr[i]; ++ if (tmp->active) ++ if (!lro_check_tcp_conn(tmp, iph, tcph)) { ++ lro_desc = tmp; ++ goto out; ++ } ++ } ++ ++ for (i = 0; i < max_desc; i++) { ++ if (!lro_arr[i].active) { ++ lro_desc = &lro_arr[i]; ++ goto out; ++ } ++ } ++ ++ LRO_INC_STATS(lro_mgr, no_desc); ++out: ++ return lro_desc; ++} ++ ++static void lro_flush(struct net_lro_mgr *lro_mgr, ++ struct net_lro_desc *lro_desc) ++{ ++ struct be_adapter *adapter = netdev_priv(lro_mgr->dev); ++ ++ if (lro_desc->pkt_aggr_cnt > 1) ++ lro_update_tcp_ip_header(lro_desc); ++ ++ skb_shinfo(lro_desc->parent)->gso_size = lro_desc->mss; ++ ++ if (lro_desc->vgrp) { ++ if (test_bit(LRO_F_NAPI, &lro_mgr->features)) ++ vlan_hwaccel_receive_skb(lro_desc->parent, ++ lro_desc->vgrp, ++ lro_desc->vlan_tag); ++ else ++ vlan_hwaccel_rx(lro_desc->parent, ++ lro_desc->vgrp, ++ lro_desc->vlan_tag); ++ ++ } else { ++ if (test_bit(LRO_F_NAPI, &lro_mgr->features)) ++ netif_receive_skb(lro_desc->parent); ++ else ++ netif_rx(lro_desc->parent); ++ } ++ ++ LRO_INC_STATS(lro_mgr, flushed); ++ lro_clear_desc(lro_desc); ++} ++ ++static struct sk_buff *lro_gen_skb(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, ++ int len, int true_size, ++ void *mac_hdr, ++ int hlen, __wsum sum, ++ u32 ip_summed) ++{ ++ struct sk_buff *skb; ++ struct skb_frag_struct *skb_frags; ++ int data_len = len; ++ int hdr_len = min(len, hlen); ++ ++ skb = netdev_alloc_skb(lro_mgr->dev, hlen); ++ if (!skb) ++ return NULL; ++ ++ skb->len = len; ++ skb->data_len = len - hdr_len; ++ skb->truesize += true_size; ++ skb->tail += hdr_len; ++ ++ memcpy(skb->data, mac_hdr, hdr_len); ++ ++ if (skb->data_len) { ++ skb_frags = skb_shinfo(skb)->frags; ++ while (data_len > 0) { ++ *skb_frags = *frags; ++ data_len -= frags->size; ++ skb_frags++; ++ frags++; ++ skb_shinfo(skb)->nr_frags++; ++ } ++ skb_shinfo(skb)->frags[0].page_offset += hdr_len; ++ skb_shinfo(skb)->frags[0].size -= hdr_len; ++ } else { ++ put_page(frags[0].page); ++ } ++ ++ ++ skb->ip_summed = ip_summed; ++ skb->csum = sum; ++ skb->protocol = eth_type_trans(skb, lro_mgr->dev); ++ return skb; ++} ++ ++static struct sk_buff *__lro_proc_segment(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, ++ int len, int true_size, ++ struct vlan_group *vgrp, ++ u16 vlan_tag, void *priv, __wsum sum) ++{ ++ struct net_lro_desc *lro_desc; ++ struct iphdr *iph; ++ struct tcphdr *tcph; ++ struct sk_buff *skb; ++ u64 flags; ++ void *mac_hdr; ++ int mac_hdr_len; ++ int hdr_len = LRO_MAX_PG_HLEN; ++ int vlan_hdr_len = 0; ++ u8 pad_bytes; ++ ++ if (!lro_mgr->get_frag_header ++ || lro_mgr->get_frag_header(frags, (void *)&mac_hdr, (void *)&iph, ++ (void *)&tcph, &flags, priv)) { ++ mac_hdr = page_address(frags->page) + frags->page_offset; ++ goto out1; ++ } ++ ++ if (!(flags & LRO_IPV4) || !(flags & LRO_TCP)) ++ goto out1; ++ ++ hdr_len = (int)((void *)(tcph) + TCP_HDR_LEN(tcph) - mac_hdr); ++ mac_hdr_len = (int)((void *)(iph) - mac_hdr); ++ ++ lro_desc = lro_get_desc(lro_mgr, lro_mgr->lro_arr, iph, tcph); ++ if (!lro_desc) ++ goto out1; ++ ++ pad_bytes = len - (ntohs(iph->tot_len) + mac_hdr_len); ++ if (!TCP_PAYLOAD_LENGTH(iph, tcph) && pad_bytes) { ++ len -= pad_bytes; /* trim the packet */ ++ frags[0].size -= pad_bytes; ++ true_size -= pad_bytes; ++ } ++ ++ if (!lro_desc->active) { /* start new lro session */ ++ if (lro_tcp_ip_check(iph, tcph, len - mac_hdr_len, NULL)) ++ goto out1; ++ ++ skb = lro_gen_skb(lro_mgr, frags, len, true_size, mac_hdr, ++ hdr_len, 0, lro_mgr->ip_summed_aggr); ++ if (!skb) ++ goto out; ++ ++ if ((skb->protocol == htons(ETH_P_8021Q)) ++ && !test_bit(LRO_F_EXTRACT_VLAN_ID, &lro_mgr->features)) ++ vlan_hdr_len = VLAN_HLEN; ++ ++ iph = (void *)(skb->data + vlan_hdr_len); ++ tcph = (void *)((u8 *)skb->data + vlan_hdr_len ++ + IP_HDR_LEN(iph)); ++ ++ lro_init_desc(lro_desc, skb, iph, tcph, vlan_tag, vgrp); ++ LRO_INC_STATS(lro_mgr, aggregated); ++ return 0; ++ } ++ ++ if (lro_desc->tcp_next_seq != ntohl(tcph->seq)) ++ goto out2; ++ ++ if (lro_tcp_ip_check(iph, tcph, len - mac_hdr_len, lro_desc)) ++ goto out2; ++ ++ lro_add_frags(lro_desc, len, hdr_len, true_size, frags, iph, tcph); ++ LRO_INC_STATS(lro_mgr, aggregated); ++ ++ if ((skb_shinfo(lro_desc->parent)->nr_frags >= lro_mgr->max_aggr) || ++ lro_desc->parent->len > (0xFFFF - lro_mgr->dev->mtu)) ++ lro_flush(lro_mgr, lro_desc); ++ ++ return NULL; ++ ++out2: /* send aggregated packets to the stack */ ++ lro_flush(lro_mgr, lro_desc); ++ ++out1: /* Original packet has to be posted to the stack */ ++ skb = lro_gen_skb(lro_mgr, frags, len, true_size, mac_hdr, ++ hdr_len, sum, lro_mgr->ip_summed); ++out: ++ return skb; ++} ++ ++void lro_receive_frags_compat(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, ++ int len, int true_size, void *priv, __wsum sum) ++{ ++ struct sk_buff *skb; ++ ++ skb = __lro_proc_segment(lro_mgr, frags, len, true_size, NULL, 0, ++ priv, sum); ++ if (!skb) ++ return; ++ ++ if (test_bit(LRO_F_NAPI, &lro_mgr->features)) ++ netif_receive_skb(skb); ++ else ++ netif_rx(skb); ++} ++ ++void lro_vlan_hwaccel_receive_frags_compat(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, ++ int len, int true_size, ++ struct vlan_group *vgrp, ++ u16 vlan_tag, void *priv, __wsum sum) ++{ ++ struct sk_buff *skb; ++ ++ skb = __lro_proc_segment(lro_mgr, frags, len, true_size, vgrp, ++ vlan_tag, priv, sum); ++ if (!skb) ++ return; ++ ++ if (test_bit(LRO_F_NAPI, &lro_mgr->features)) ++ vlan_hwaccel_receive_skb(skb, vgrp, vlan_tag); ++ else ++ vlan_hwaccel_rx(skb, vgrp, vlan_tag); ++} ++ ++void lro_flush_all_compat(struct net_lro_mgr *lro_mgr) ++{ ++ int i; ++ struct net_lro_desc *lro_desc = lro_mgr->lro_arr; ++ ++ for (i = 0; i < lro_mgr->max_desc; i++) { ++ if (lro_desc[i].active) ++ lro_flush(lro_mgr, &lro_desc[i]); ++ } ++} ++#endif /* INET_LRO backport */ ++ ++#ifndef TX_MQ ++struct net_device *alloc_etherdev_mq_compat(int sizeof_priv, ++ unsigned int queue_count) ++{ ++ return alloc_etherdev(sizeof_priv); ++} ++ ++void netif_wake_subqueue_compat(struct net_device *dev, u16 queue_index) ++{ ++ netif_wake_queue(dev); ++} ++ ++void netif_stop_subqueue_compat(struct net_device *dev, u16 queue_index) ++{ ++ netif_stop_queue(dev); ++} ++ ++int __netif_subqueue_stopped_compat(const struct net_device *dev, ++ u16 queue_index) ++{ ++ return netif_queue_stopped(dev); ++} ++ ++u16 skb_get_queue_mapping_compat(const struct sk_buff *skb) ++{ ++ return 0; ++} ++ ++void netif_set_real_num_tx_queues_compat(struct net_device *dev, ++ unsigned int txq) ++{ ++ return; ++} ++ ++u16 skb_tx_hash_compat(const struct net_device *dev, ++ const struct sk_buff *skb) ++{ ++ return 0; ++} ++#endif +diff --git a/drivers/net/benet/be_compat.h b/drivers/net/benet/be_compat.h +new file mode 100644 +index 0000000..8ceecc8 +--- /dev/null ++++ b/drivers/net/benet/be_compat.h +@@ -0,0 +1,621 @@ ++/* ++ * Copyright (C) 2005 - 2011 Emulex ++ * All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License version 2 ++ * as published by the Free Software Foundation. The full GNU General ++ * Public License is included in this distribution in the file called COPYING. ++ * ++ * Contact Information: ++ * linux-drivers@emulex.com ++ * ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 ++ */ ++ ++#ifndef BE_COMPAT_H ++#define BE_COMPAT_H ++ ++/****************** RHEL5 and SLES10 backport ***************************/ ++#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 18) ++ ++#ifndef upper_32_bits ++#define upper_32_bits(n) ((u32)(((n) >> 16) >> 16)) ++#endif ++ ++#ifndef CHECKSUM_PARTIAL ++#define CHECKSUM_PARTIAL CHECKSUM_HW ++#define CHECKSUM_COMPLETE CHECKSUM_HW ++#endif ++ ++#if !defined(ip_hdr) ++#define ip_hdr(skb) (skb->nh.iph) ++#define ipv6_hdr(skb) (skb->nh.ipv6h) ++#endif ++ ++#if !defined(__packed) ++#define __packed __attribute__ ((packed)) ++#endif ++ ++#if !defined(RHEL_MINOR) ++/* Only for RH5U1 (Maui) and SLES10 NIC driver */ ++enum { ++ false = 0, ++ true = 1 ++}; ++#if LINUX_VERSION_CODE == KERNEL_VERSION(2, 6, 18) ++/* Only for RH5U1 (Maui) NIC driver */ ++static inline __attribute__((const)) ++int __ilog2_u32(u32 n) ++{ ++ return fls(n) - 1; ++} ++#endif ++#endif ++ ++#define ETH_FCS_LEN 4 ++#define bool u8 ++#ifndef PTR_ALIGN ++#define PTR_ALIGN(p, a) ((typeof(p)) \ ++ ALIGN((unsigned long)(p), (a))) ++#endif ++#define list_first_entry(ptr, type, member) \ ++ list_entry((ptr)->next, type, member) ++ ++#if (defined(RHEL_MINOR) && RHEL_MINOR < 6) || \ ++ LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 16) ++#define DEFINE_PCI_DEVICE_TABLE(_table) struct pci_device_id _table[] \ ++ __devinitdata ++#endif ++ ++/* Backport of request_irq */ ++typedef irqreturn_t(*backport_irq_handler_t) (int, void *); ++static inline int ++backport_request_irq(unsigned int irq, irqreturn_t(*handler) (int, void *), ++ unsigned long flags, const char *dev_name, void *dev_id) ++{ ++ return request_irq(irq, ++ (irqreturn_t(*) (int, void *, struct pt_regs *))handler, ++ flags, dev_name, dev_id); ++} ++#define request_irq backport_request_irq ++ ++#endif /*** RHEL5 and SLES10 backport ***/ ++ ++#if !defined(__packed) ++#define __packed __attribute__ ((packed)) ++#endif ++ ++/****************** SLES10 only backport ***************************/ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18) ++ ++#include ++ ++#define FIELD_SIZEOF(t, f) (sizeof(((t *)0)->f)) ++#define IRQF_SHARED SA_SHIRQ ++#define CHECKSUM_PARTIAL CHECKSUM_HW ++#define CHECKSUM_COMPLETE CHECKSUM_HW ++#define DIV_ROUND_UP(n, d) (((n) + (d) - 1) / (d)) ++#define NETIF_F_IPV6_CSUM NETIF_F_IP_CSUM ++#define NETIF_F_TSO6 NETIF_F_TSO ++ ++ ++static inline struct sk_buff *netdev_alloc_skb(struct net_device *dev, ++ unsigned int length) ++{ ++ /* 16 == NET_PAD_SKB */ ++ struct sk_buff *skb; ++ skb = alloc_skb(length + 16, GFP_ATOMIC); ++ if (likely(skb != NULL)) { ++ skb_reserve(skb, 16); ++ skb->dev = dev; ++ } ++ return skb; ++} ++ ++#define PCI_SAVE_STATE(x) ++ ++#else /* SLES10 only backport */ ++ ++#define PCI_SAVE_STATE(x) pci_save_state(x) ++ ++#endif /* SLES10 only backport */ ++ ++#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 31) ++#define netdev_tx_t int ++#endif ++ ++#ifndef VLAN_PRIO_MASK ++#define VLAN_PRIO_MASK 0xe000 /* Priority Code Point */ ++#define VLAN_PRIO_SHIFT 13 ++#endif ++ ++/* ++ * Backport of netdev ops struct ++ */ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 28) ++struct net_device_ops { ++ int (*ndo_init)(struct net_device *dev); ++ void (*ndo_uninit)(struct net_device *dev); ++ int (*ndo_open)(struct net_device *dev); ++ int (*ndo_stop)(struct net_device *dev); ++ int (*ndo_start_xmit) (struct sk_buff *skb, struct net_device *dev); ++ u16 (*ndo_select_queue)(struct net_device *dev, ++ struct sk_buff *skb); ++ void (*ndo_change_rx_flags)(struct net_device *dev, int flags); ++ void (*ndo_set_rx_mode)(struct net_device *dev); ++ void (*ndo_set_multicast_list)(struct net_device *dev); ++ int (*ndo_set_mac_address)(struct net_device *dev, void *addr); ++ int (*ndo_validate_addr)(struct net_device *dev); ++ int (*ndo_do_ioctl)(struct net_device *dev, ++ struct ifreq *ifr, int cmd); ++ int (*ndo_set_config)(struct net_device *dev, struct ifmap *map); ++ int (*ndo_change_mtu)(struct net_device *dev, int new_mtu); ++ int (*ndo_neigh_setup)(struct net_device *dev, ++ struct neigh_parms *); ++ void (*ndo_tx_timeout) (struct net_device *dev); ++ ++ struct net_device_stats* (*ndo_get_stats)(struct net_device *dev); ++ ++ void (*ndo_vlan_rx_register)(struct net_device *dev, ++ struct vlan_group *grp); ++ void (*ndo_vlan_rx_add_vid)(struct net_device *dev, ++ unsigned short vid); ++ void (*ndo_vlan_rx_kill_vid)(struct net_device *dev, ++ unsigned short vid); ++#ifdef CONFIG_NET_POLL_CONTROLLER ++#define HAVE_NETDEV_POLL ++ void (*ndo_poll_controller)(struct net_device *dev); ++#endif ++}; ++extern void be_netdev_ops_init(struct net_device *netdev, ++ struct net_device_ops *ops); ++extern int eth_validate_addr(struct net_device *); ++ ++#endif /* Netdev ops backport */ ++ ++#if LINUX_VERSION_CODE == KERNEL_VERSION(2, 6, 29) ++#undef NETIF_F_GRO ++#endif ++ ++#ifdef NO_GRO ++#if ((defined(RHEL_MAJOR) && (RHEL_MAJOR == 5))) ++#undef NETIF_F_GRO ++#endif ++#endif ++ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32) ++#define HAVE_ETHTOOL_FLASH ++#endif ++ ++/* ++ * Backport of NAPI ++ */ ++#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 24) ++ ++#if defined(RHEL_MINOR) && (RHEL_MINOR > 3) ++#define RHEL_NEW_NAPI ++#endif ++ ++/* We need a new struct that has some meta data beyond rhel 5.4's napi_struct ++ * to fix rhel5.4's half-baked new napi implementation. ++ * We don't want to use rhel 5.4's broken napi_complete; so ++ * define a new be_napi_complete that executes the logic only for Rx ++ */ ++ ++#ifdef RHEL_NEW_NAPI ++#define napi_complete be_napi_complete ++typedef struct napi_struct rhel_napi_struct; ++#endif ++#define napi_struct be_napi_struct ++#define napi_gro_frags(napi) napi_gro_frags((rhel_napi_struct *) napi) ++#define vlan_gro_frags(napi, vlan_grp, vid)\ ++ vlan_gro_frags((rhel_napi_struct *) napi, vlan_grp, vid) ++#define napi_get_frags(napi) napi_get_frags((rhel_napi_struct *) napi) ++ ++struct napi_struct { ++#ifdef RHEL_NEW_NAPI ++ rhel_napi_struct napi; /* must be the first member */ ++#endif ++ struct net_device *dev; ++ int (*poll) (struct napi_struct *napi, int budget); ++ bool rx; ++}; ++ ++static inline void napi_complete(struct napi_struct *napi) ++{ ++#ifdef NETIF_F_GRO ++ napi_gro_flush((rhel_napi_struct *)napi); ++#endif ++ netif_rx_complete(napi->dev); ++} ++ ++static inline void napi_schedule(struct napi_struct *napi) ++{ ++ netif_rx_schedule(napi->dev); ++} ++ ++static inline void napi_enable(struct napi_struct *napi) ++{ ++ netif_poll_enable(napi->dev); ++} ++ ++static inline void napi_disable(struct napi_struct *napi) ++{ ++ netif_poll_disable(napi->dev); ++} ++ ++#if (defined(RHEL_MINOR) && RHEL_MINOR < 6) || \ ++ LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 16) ++static inline void vlan_group_set_device(struct vlan_group *vg, ++ u16 vlan_id, ++ struct net_device *dev) ++{ ++ struct net_device **array; ++ if (!vg) ++ return; ++ array = vg->vlan_devices; ++ array[vlan_id] = dev; ++} ++#endif ++ ++#endif /* New NAPI backport */ ++ ++extern int be_netif_napi_add(struct net_device *netdev, ++ struct napi_struct *napi, ++ int (*poll) (struct napi_struct *, int), int weight); ++extern void be_netif_napi_del(struct net_device *netdev); ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 27) ++#define HAVE_SIMULATED_MULTI_NAPI ++#endif ++ ++/************** Backport of Delayed work queues interface ****************/ ++#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 19) ++#if (defined(RHEL_MINOR) && RHEL_MINOR < 6) || \ ++ LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 16) ++struct delayed_work { ++ struct work_struct work; ++}; ++#endif ++ ++#define INIT_DELAYED_WORK(_work, _func) \ ++ INIT_WORK(&(_work)->work, _func, &(_work)->work) ++ ++static inline int backport_cancel_delayed_work_sync(struct delayed_work *work) ++{ ++ cancel_rearming_delayed_work(&work->work); ++ return 0; ++} ++#define cancel_delayed_work_sync backport_cancel_delayed_work_sync ++ ++static inline int backport_schedule_delayed_work(struct delayed_work *work, ++ unsigned long delay) ++{ ++ if (unlikely(!delay)) ++ return schedule_work(&work->work); ++ else ++ return schedule_delayed_work(&work->work, delay); ++} ++#define schedule_delayed_work backport_schedule_delayed_work ++#endif /* backport delayed workqueue */ ++ ++ ++/************** Backport of INET_LRO **********************************/ ++#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 18) ++ ++#include ++ ++#else ++ ++#if LINUX_VERSION_CODE == KERNEL_VERSION(2, 6, 18) ++ ++#if defined(RHEL_MINOR) && RHEL_MINOR < 6 ++typedef __u16 __bitwise __sum16; ++typedef __u32 __bitwise __wsum; ++#endif ++ ++#if ((defined(RHEL_MAJOR) && (RHEL_MAJOR == 5) && (RHEL_MINOR <= 3)) || \ ++ (!defined(RHEL_MINOR))) ++static inline __wsum csum_unfold(__sum16 n) ++{ ++ return (__force __wsum)n; ++} ++#endif ++ ++#endif ++ ++#define lro_flush_all lro_flush_all_compat ++#define lro_vlan_hwaccel_receive_frags lro_vlan_hwaccel_receive_frags_compat ++#define lro_receive_frags lro_receive_frags_compat ++ ++struct net_lro_stats { ++ unsigned long aggregated; ++ unsigned long flushed; ++ unsigned long no_desc; ++}; ++ ++struct net_lro_desc { ++ struct sk_buff *parent; ++ struct sk_buff *last_skb; ++ struct skb_frag_struct *next_frag; ++ struct iphdr *iph; ++ struct tcphdr *tcph; ++ struct vlan_group *vgrp; ++ __wsum data_csum; ++ u32 tcp_rcv_tsecr; ++ u32 tcp_rcv_tsval; ++ u32 tcp_ack; ++ u32 tcp_next_seq; ++ u32 skb_tot_frags_len; ++ u32 ack_cnt; ++ u16 ip_tot_len; ++ u16 tcp_saw_tstamp; /* timestamps enabled */ ++ u16 tcp_window; ++ u16 vlan_tag; ++ int pkt_aggr_cnt; /* counts aggregated packets */ ++ int vlan_packet; ++ int mss; ++ int active; ++}; ++ ++struct net_lro_mgr { ++ struct net_device *dev; ++ struct net_lro_stats stats; ++ ++ /* LRO features */ ++ unsigned long features; ++#define LRO_F_NAPI 1 /* Pass packets to stack via NAPI */ ++#define LRO_F_EXTRACT_VLAN_ID 2 /* Set flag if VLAN IDs are extracted ++ from received packets and eth protocol ++ is still ETH_P_8021Q */ ++ ++ u32 ip_summed; /* Set in non generated SKBs in page mode */ ++ u32 ip_summed_aggr; /* Set in aggregated SKBs: CHECKSUM_UNNECESSARY ++ * or CHECKSUM_NONE */ ++ ++ int max_desc; /* Max number of LRO descriptors */ ++ int max_aggr; /* Max number of LRO packets to be aggregated */ ++ ++ struct net_lro_desc *lro_arr; /* Array of LRO descriptors */ ++ ++ /* Optimized driver functions ++ * get_skb_header: returns tcp and ip header for packet in SKB ++ */ ++ int (*get_skb_header)(struct sk_buff *skb, void **ip_hdr, ++ void **tcpudp_hdr, u64 *hdr_flags, void *priv); ++ ++ /* hdr_flags: */ ++#define LRO_IPV4 1 /* ip_hdr is IPv4 header */ ++#define LRO_TCP 2 /* tcpudp_hdr is TCP header */ ++ ++ /* ++ * get_frag_header: returns mac, tcp and ip header for packet in SKB ++ * ++ * @hdr_flags: Indicate what kind of LRO has to be done ++ * (IPv4/IPv6/TCP/UDP) ++ */ ++ int (*get_frag_header)(struct skb_frag_struct *frag, void **mac_hdr, ++ void **ip_hdr, void **tcpudp_hdr, u64 *hdr_flags, ++ void *priv); ++}; ++ ++extern void lro_receive_skb(struct net_lro_mgr *lro_mgr, struct sk_buff *skb, ++ void *priv); ++ ++extern void lro_vlan_hwaccel_receive_skb(struct net_lro_mgr *lro_mgr, ++ struct sk_buff *skb, struct vlan_group *vgrp, ++ u16 vlan_tag, void *priv); ++ ++/* This functions aggregate fragments and generate SKBs do pass ++ * the packets to the stack. ++ * ++ * @lro_mgr: LRO manager to use ++ * @frags: Fragment to be processed. Must contain entire header in first ++ * element. ++ * @len: Length of received data ++ * @true_size: Actual size of memory the fragment is consuming ++ * @priv: Private data that may be used by driver functions ++ * (for example get_tcp_ip_hdr) ++ */ ++extern void lro_receive_frags_compat(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, int len, int true_size, ++ void *priv, __wsum sum); ++ ++extern void lro_vlan_hwaccel_receive_frags_compat(struct net_lro_mgr *lro_mgr, ++ struct skb_frag_struct *frags, int len, int true_size, ++ struct vlan_group *vgrp, u16 vlan_tag, void *priv, ++ __wsum sum); ++ ++/* Forward all aggregated SKBs held by lro_mgr to network stack */ ++extern void lro_flush_all_compat(struct net_lro_mgr *lro_mgr); ++ ++extern void lro_flush_pkt(struct net_lro_mgr *lro_mgr, struct iphdr *iph, ++ struct tcphdr *tcph); ++#endif /* backport of inet_lro */ ++ ++#ifndef ETHTOOL_FLASH_MAX_FILENAME ++#define ETHTOOL_FLASH_MAX_FILENAME 128 ++#endif ++ ++#if defined(CONFIG_XEN) && !defined(NETIF_F_GRO) ++#define BE_INIT_FRAGS_PER_FRAME (u32) 1 ++#else ++#define BE_INIT_FRAGS_PER_FRAME (min((u32) 16, (u32) MAX_SKB_FRAGS)) ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32) ++#ifdef CONFIG_PCI_IOV ++#if (!(defined(RHEL_MAJOR) && (RHEL_MAJOR == 5) && (RHEL_MINOR == 6))) ++#undef CONFIG_PCI_IOV ++#endif ++#endif ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 20) ++#define dev_to_node(dev) -1 ++#endif ++ ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32) ++#if (!(defined(RHEL_MAJOR) && (RHEL_MAJOR == 5) && (RHEL_MINOR > 6))) ++static inline struct sk_buff *netdev_alloc_skb_ip_align(struct net_device *dev, ++ unsigned int length) ++{ ++ struct sk_buff *skb = netdev_alloc_skb(dev, length + NET_IP_ALIGN); ++ ++ if (NET_IP_ALIGN && skb) ++ skb_reserve(skb, NET_IP_ALIGN); ++ return skb; ++} ++#endif ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 26) ++#ifndef netif_set_gso_max_size ++#define netif_set_gso_max_size(netdev, size) do {} while (0) ++#endif ++#endif ++ ++#if (LINUX_VERSION_CODE == KERNEL_VERSION(2, 6, 18)) ++#if defined(RHEL_MINOR) && (RHEL_MINOR <= 4) ++static inline int skb_is_gso_v6(const struct sk_buff *skb) ++{ ++ return skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6; ++} ++#endif ++#endif ++ ++#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18)) ++static inline int skb_is_gso_v6(const struct sk_buff *skb) ++{ ++ return (ip_hdr(skb)->version == 6); ++} ++#endif ++ ++#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18)) ++#define DMA_BIT_MASK(n) (((n) == 64) ? ~0ULL : ((1ULL<<(n))-1)) ++#endif ++ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32) ++#if ((defined(RHEL_MAJOR) && (RHEL_MAJOR == 6))) ++#define HAVE_SRIOV_CONFIG ++#endif ++#endif ++ ++#ifndef NETIF_F_VLAN_SG ++#define NETIF_F_VLAN_SG NETIF_F_SG ++#endif ++ ++#ifndef NETIF_F_VLAN_CSUM ++#define NETIF_F_VLAN_CSUM NETIF_F_HW_CSUM ++#endif ++ ++#ifndef NETIF_F_VLAN_TSO ++#define NETIF_F_VLAN_TSO NETIF_F_TSO ++#endif ++ ++#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 27)) ++#define vlan_features features ++#endif ++ ++#ifndef DEFINE_DMA_UNMAP_ADDR ++#define DEFINE_DMA_UNMAP_ADDR(bus) dma_addr_t bus ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 0, 0) ++ ++#ifndef netdev_mc_count ++#define netdev_mc_count(nd) (nd->mc_count) ++#endif ++ ++#ifndef netdev_hw_addr ++#define netdev_hw_addr dev_mc_list ++#endif ++ ++#ifndef netdev_for_each_mc_addr ++#define netdev_for_each_mc_addr(ha, nd) \ ++ for (ha = (nd)->mc_list; ha; ha = ha->next) ++#endif ++ ++#define DMI_ADDR dmi_addr ++#else ++#define DMI_ADDR addr ++#endif ++ ++#ifndef VLAN_GROUP_ARRAY_LEN ++#define VLAN_GROUP_ARRAY_LEN VLAN_N_VID ++#endif ++/**************************** Multi TXQ Support ******************************/ ++ ++/* Supported only in RHEL6 and SL11.1 (barring one execption) */ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32) ++#define MQ_TX ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 27) ++#define alloc_etherdev_mq(sz, cnt) alloc_etherdev(sz) ++#define skb_get_queue_mapping(skb) 0 ++#define skb_tx_hash(dev, skb) 0 ++#define netif_set_real_num_tx_queues(dev, txq) do {} while(0) ++#define netif_wake_subqueue(dev, idx) netif_wake_queue(dev) ++#define netif_stop_subqueue(dev, idx) netif_stop_queue(dev) ++#define __netif_subqueue_stopped(dev, idx) netif_queue_stopped(dev) ++#endif /* < 2.6.27 */ ++ ++#if ((LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 27)) && \ ++ (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32))) ++#define skb_tx_hash(dev, skb) 0 ++#define netif_set_real_num_tx_queues(dev, txq) do {} while(0) ++#endif ++ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 32) ++#define netif_set_real_num_tx_queues be_set_real_num_tx_queues ++static inline void be_set_real_num_tx_queues(struct net_device *dev, ++ unsigned int txq) ++{ ++ dev->real_num_tx_queues = txq; ++} ++#endif ++ ++#include ++static inline void be_reset_skb_tx_vlan(struct sk_buff *skb) ++{ ++#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 18) ++ skb->vlan_tci = 0; ++#else ++ struct vlan_skb_tx_cookie *cookie; ++ ++ cookie = VLAN_TX_SKB_CB(skb); ++ cookie->magic = 0; ++#endif ++} ++ ++#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18)) ++static inline void skb_set_network_header(struct sk_buff *skb, const int offset) ++{ ++ skb->nh.raw = skb->data + offset; ++} ++#endif ++ ++static inline struct sk_buff *be_vlan_put_tag(struct sk_buff *skb, ++ unsigned short vlan_tag) ++{ ++ struct sk_buff *new_skb = __vlan_put_tag(skb, vlan_tag); ++ /* On kernel versions < 2.6.27 the __vlan_put_tag() function ++ * distorts the network layer hdr pointer in the skb which ++ * affects the detection of UDP/TCP packets down the line in ++ * wrb_fill_hdr().This work-around sets it right. ++ */ ++#if (LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 27)) ++ skb_set_network_header(new_skb, VLAN_ETH_HLEN); ++#endif ++ return new_skb; ++} ++ ++#ifndef ACCESS_ONCE ++#define ACCESS_ONCE(x) (*(volatile typeof(x) *)&(x)) ++#endif ++ ++#endif /* BE_COMPAT_H */ +diff --git a/drivers/net/benet/be_ethtool.c b/drivers/net/benet/be_ethtool.c +index f0fd95b..37bad99 100644 +--- a/drivers/net/benet/be_ethtool.c ++++ b/drivers/net/benet/be_ethtool.c +@@ -1,18 +1,18 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + + #include "be.h" +@@ -26,21 +26,19 @@ struct be_ethtool_stat { + int offset; + }; + +-enum {NETSTAT, PORTSTAT, MISCSTAT, DRVSTAT, ERXSTAT}; ++enum {NETSTAT, DRVSTAT_TX, DRVSTAT_RX, DRVSTAT}; + #define FIELDINFO(_struct, field) FIELD_SIZEOF(_struct, field), \ + offsetof(_struct, field) +-#define NETSTAT_INFO(field) #field, NETSTAT,\ ++#define NETSTAT_INFO(field) #field, NETSTAT,\ + FIELDINFO(struct net_device_stats,\ + field) +-#define DRVSTAT_INFO(field) #field, DRVSTAT,\ +- FIELDINFO(struct be_drvr_stats, field) +-#define MISCSTAT_INFO(field) #field, MISCSTAT,\ +- FIELDINFO(struct be_rxf_stats, field) +-#define PORTSTAT_INFO(field) #field, PORTSTAT,\ +- FIELDINFO(struct be_port_rxf_stats, \ ++#define DRVSTAT_TX_INFO(field) #field, DRVSTAT_TX,\ ++ FIELDINFO(struct be_tx_stats, field) ++#define DRVSTAT_RX_INFO(field) #field, DRVSTAT_RX,\ ++ FIELDINFO(struct be_rx_stats, field) ++#define DRVSTAT_INFO(field) #field, DRVSTAT,\ ++ FIELDINFO(struct be_drv_stats, \ + field) +-#define ERXSTAT_INFO(field) #field, ERXSTAT,\ +- FIELDINFO(struct be_erx_stats, field) + + static const struct be_ethtool_stat et_stats[] = { + {NETSTAT_INFO(rx_packets)}, +@@ -51,70 +49,131 @@ static const struct be_ethtool_stat et_stats[] = { + {NETSTAT_INFO(tx_errors)}, + {NETSTAT_INFO(rx_dropped)}, + {NETSTAT_INFO(tx_dropped)}, +- {DRVSTAT_INFO(be_tx_reqs)}, +- {DRVSTAT_INFO(be_tx_stops)}, +- {DRVSTAT_INFO(be_fwd_reqs)}, +- {DRVSTAT_INFO(be_tx_wrbs)}, +- {DRVSTAT_INFO(be_polls)}, + {DRVSTAT_INFO(be_tx_events)}, +- {DRVSTAT_INFO(be_rx_events)}, +- {DRVSTAT_INFO(be_tx_compl)}, +- {DRVSTAT_INFO(be_rx_compl)}, +- {DRVSTAT_INFO(be_ethrx_post_fail)}, +- {DRVSTAT_INFO(be_802_3_dropped_frames)}, +- {DRVSTAT_INFO(be_802_3_malformed_frames)}, +- {DRVSTAT_INFO(be_tx_rate)}, +- {DRVSTAT_INFO(be_rx_rate)}, +- {PORTSTAT_INFO(rx_unicast_frames)}, +- {PORTSTAT_INFO(rx_multicast_frames)}, +- {PORTSTAT_INFO(rx_broadcast_frames)}, +- {PORTSTAT_INFO(rx_crc_errors)}, +- {PORTSTAT_INFO(rx_alignment_symbol_errors)}, +- {PORTSTAT_INFO(rx_pause_frames)}, +- {PORTSTAT_INFO(rx_control_frames)}, +- {PORTSTAT_INFO(rx_in_range_errors)}, +- {PORTSTAT_INFO(rx_out_range_errors)}, +- {PORTSTAT_INFO(rx_frame_too_long)}, +- {PORTSTAT_INFO(rx_address_match_errors)}, +- {PORTSTAT_INFO(rx_vlan_mismatch)}, +- {PORTSTAT_INFO(rx_dropped_too_small)}, +- {PORTSTAT_INFO(rx_dropped_too_short)}, +- {PORTSTAT_INFO(rx_dropped_header_too_small)}, +- {PORTSTAT_INFO(rx_dropped_tcp_length)}, +- {PORTSTAT_INFO(rx_dropped_runt)}, +- {PORTSTAT_INFO(rx_fifo_overflow)}, +- {PORTSTAT_INFO(rx_input_fifo_overflow)}, +- {PORTSTAT_INFO(rx_ip_checksum_errs)}, +- {PORTSTAT_INFO(rx_tcp_checksum_errs)}, +- {PORTSTAT_INFO(rx_udp_checksum_errs)}, +- {PORTSTAT_INFO(rx_non_rss_packets)}, +- {PORTSTAT_INFO(rx_ipv4_packets)}, +- {PORTSTAT_INFO(rx_ipv6_packets)}, +- {PORTSTAT_INFO(tx_unicastframes)}, +- {PORTSTAT_INFO(tx_multicastframes)}, +- {PORTSTAT_INFO(tx_broadcastframes)}, +- {PORTSTAT_INFO(tx_pauseframes)}, +- {PORTSTAT_INFO(tx_controlframes)}, +- {MISCSTAT_INFO(rx_drops_no_pbuf)}, +- {MISCSTAT_INFO(rx_drops_no_txpb)}, +- {MISCSTAT_INFO(rx_drops_no_erx_descr)}, +- {MISCSTAT_INFO(rx_drops_no_tpre_descr)}, +- {MISCSTAT_INFO(rx_drops_too_many_frags)}, +- {MISCSTAT_INFO(rx_drops_invalid_ring)}, +- {MISCSTAT_INFO(forwarded_packets)}, +- {MISCSTAT_INFO(rx_drops_mtu)}, +- {ERXSTAT_INFO(rx_drops_no_fragments)}, ++ {DRVSTAT_INFO(rx_crc_errors)}, ++ {DRVSTAT_INFO(rx_alignment_symbol_errors)}, ++ {DRVSTAT_INFO(rx_pause_frames)}, ++ {DRVSTAT_INFO(rx_control_frames)}, ++ {DRVSTAT_INFO(rx_in_range_errors)}, ++ {DRVSTAT_INFO(rx_out_range_errors)}, ++ {DRVSTAT_INFO(rx_frame_too_long)}, ++ {DRVSTAT_INFO(rx_address_match_errors)}, ++ {DRVSTAT_INFO(rx_dropped_too_small)}, ++ {DRVSTAT_INFO(rx_dropped_too_short)}, ++ {DRVSTAT_INFO(rx_dropped_header_too_small)}, ++ {DRVSTAT_INFO(rx_dropped_tcp_length)}, ++ {DRVSTAT_INFO(rx_dropped_runt)}, ++ {DRVSTAT_INFO(rxpp_fifo_overflow_drop)}, ++ {DRVSTAT_INFO(rx_input_fifo_overflow_drop)}, ++ {DRVSTAT_INFO(rx_ip_checksum_errs)}, ++ {DRVSTAT_INFO(rx_tcp_checksum_errs)}, ++ {DRVSTAT_INFO(rx_udp_checksum_errs)}, ++ {DRVSTAT_INFO(rx_switched_unicast_packets)}, ++ {DRVSTAT_INFO(rx_switched_multicast_packets)}, ++ {DRVSTAT_INFO(rx_switched_broadcast_packets)}, ++ {DRVSTAT_INFO(tx_pauseframes)}, ++ {DRVSTAT_INFO(tx_controlframes)}, ++ {DRVSTAT_INFO(rx_priority_pause_frames)}, ++ {DRVSTAT_INFO(pmem_fifo_overflow_drop)}, ++ {DRVSTAT_INFO(jabber_events)}, ++ {DRVSTAT_INFO(rx_drops_no_pbuf)}, ++ {DRVSTAT_INFO(rx_drops_no_txpb)}, ++ {DRVSTAT_INFO(rx_drops_no_erx_descr)}, ++ {DRVSTAT_INFO(rx_drops_no_tpre_descr)}, ++ {DRVSTAT_INFO(rx_drops_too_many_frags)}, ++ {DRVSTAT_INFO(rx_drops_invalid_ring)}, ++ {DRVSTAT_INFO(forwarded_packets)}, ++ {DRVSTAT_INFO(rx_drops_mtu)}, ++ {DRVSTAT_INFO(eth_red_drops)}, ++ {DRVSTAT_INFO(be_on_die_temperature)} + }; + #define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats) + ++/* Stats related to multi RX queues */ ++static const struct be_ethtool_stat et_rx_stats[] = { ++ {DRVSTAT_RX_INFO(rx_bytes)}, ++ {DRVSTAT_RX_INFO(rx_pkts)}, ++ {DRVSTAT_RX_INFO(rx_rate)}, ++ {DRVSTAT_RX_INFO(rx_polls)}, ++ {DRVSTAT_RX_INFO(rx_events)}, ++ {DRVSTAT_RX_INFO(rx_compl)}, ++ {DRVSTAT_RX_INFO(rx_mcast_pkts)}, ++ {DRVSTAT_RX_INFO(rx_post_fail)}, ++ {DRVSTAT_RX_INFO(rx_drops_no_frags)} ++}; ++#define ETHTOOL_RXSTATS_NUM (ARRAY_SIZE(et_rx_stats)) ++ ++/* Stats related to multi TX queues */ ++static const struct be_ethtool_stat et_tx_stats[] = { ++ {DRVSTAT_TX_INFO(be_tx_rate)}, ++ {DRVSTAT_TX_INFO(be_tx_reqs)}, ++ {DRVSTAT_TX_INFO(be_tx_wrbs)}, ++ {DRVSTAT_TX_INFO(be_tx_stops)}, ++ {DRVSTAT_TX_INFO(be_tx_compl)}, ++ {DRVSTAT_TX_INFO(be_ipv6_ext_hdr_tx_drop)} ++}; ++#define ETHTOOL_TXSTATS_NUM (ARRAY_SIZE(et_tx_stats)) ++ ++static const char et_self_tests[][ETH_GSTRING_LEN] = { ++ "MAC Loopback test", ++ "PHY Loopback test", ++ "External Loopback test", ++ "DDR DMA test", ++ "Link test" ++}; ++ ++#define ETHTOOL_TESTS_NUM ARRAY_SIZE(et_self_tests) ++#define BE_MAC_LOOPBACK 0x0 ++#define BE_PHY_LOOPBACK 0x1 ++#define BE_ONE_PORT_EXT_LOOPBACK 0x2 ++#define BE_NO_LOOPBACK 0xff ++ ++/* MAC speed valid values */ ++#define SPEED_DEFAULT 0x0 ++#define SPEED_FORCED_10GB 0x1 ++#define SPEED_FORCED_1GB 0x2 ++#define SPEED_AUTONEG_10GB 0x3 ++#define SPEED_AUTONEG_1GB 0x4 ++#define SPEED_AUTONEG_100MB 0x5 ++#define SPEED_AUTONEG_10GB_1GB 0x6 ++#define SPEED_AUTONEG_10GB_1GB_100MB 0x7 ++#define SPEED_AUTONEG_1GB_100MB 0x8 ++#define SPEED_AUTONEG_10MB 0x9 ++#define SPEED_AUTONEG_1GB_100MB_10MB 0xa ++#define SPEED_AUTONEG_100MB_10MB 0xb ++#define SPEED_FORCED_100MB 0xc ++#define SPEED_FORCED_10MB 0xd ++ ++ ++ + static void + be_get_drvinfo(struct net_device *netdev, struct ethtool_drvinfo *drvinfo) + { + struct be_adapter *adapter = netdev_priv(netdev); ++ int len; ++ char fw_on_flash[FW_VER_LEN]; ++ ++ memset(fw_on_flash, 0 , sizeof(fw_on_flash)); ++ ++ be_cmd_get_fw_ver(adapter, adapter->fw_ver, ++ fw_on_flash); + + strcpy(drvinfo->driver, DRV_NAME); + strcpy(drvinfo->version, DRV_VER); ++ + strncpy(drvinfo->fw_version, adapter->fw_ver, FW_VER_LEN); ++ if (memcmp(adapter->fw_ver, fw_on_flash, ++ FW_VER_LEN) != 0) { ++ len = strlen(drvinfo->fw_version); ++ strncpy(drvinfo->fw_version+len, " [", ++ FW_VER_LEN-len-1); ++ len = strlen(drvinfo->fw_version); ++ strncpy(drvinfo->fw_version+len, fw_on_flash, ++ FW_VER_LEN-len-1); ++ len = strlen(drvinfo->fw_version); ++ strncpy(drvinfo->fw_version+len, "]", FW_VER_LEN-len-1); ++ } ++ + strcpy(drvinfo->bus_info, pci_name(adapter->pdev)); + drvinfo->testinfo_len = 0; + drvinfo->regdump_len = 0; +@@ -122,12 +181,37 @@ be_get_drvinfo(struct net_device *netdev, struct ethtool_drvinfo *drvinfo) + } + + static int ++be_get_reg_len(struct net_device *netdev) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ u32 log_size = 0; ++ ++ if (be_physfn(adapter)) ++ be_cmd_get_reg_len(adapter, &log_size); ++ ++ return log_size; ++} ++ ++static void ++be_get_regs(struct net_device *netdev, struct ethtool_regs *regs, void *buf) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ if (be_physfn(adapter)) { ++ memset(buf, 0, regs->len); ++ be_cmd_get_regs(adapter, regs->len, buf); ++ } ++} ++ ++static int + be_get_coalesce(struct net_device *netdev, struct ethtool_coalesce *coalesce) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_eq_obj *rx_eq = &adapter->rx_eq; ++ struct be_eq_obj *rx_eq = &adapter->rx_obj[0].rx_eq; + struct be_eq_obj *tx_eq = &adapter->tx_eq; + ++ coalesce->rx_max_coalesced_frames = adapter->max_rx_coal; ++ + coalesce->rx_coalesce_usecs = rx_eq->cur_eqd; + coalesce->rx_coalesce_usecs_high = rx_eq->max_eqd; + coalesce->rx_coalesce_usecs_low = rx_eq->min_eqd; +@@ -149,25 +233,52 @@ static int + be_set_coalesce(struct net_device *netdev, struct ethtool_coalesce *coalesce) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_eq_obj *rx_eq = &adapter->rx_eq; ++ struct be_rx_obj *rxo; ++ struct be_eq_obj *rx_eq; + struct be_eq_obj *tx_eq = &adapter->tx_eq; + u32 tx_max, tx_min, tx_cur; + u32 rx_max, rx_min, rx_cur; +- int status = 0; ++ int status = 0, i; + + if (coalesce->use_adaptive_tx_coalesce == 1) + return -EINVAL; ++ adapter->max_rx_coal = coalesce->rx_max_coalesced_frames; ++ if (adapter->max_rx_coal > BE_MAX_FRAGS_PER_FRAME) ++ adapter->max_rx_coal = BE_MAX_FRAGS_PER_FRAME; + +- /* if AIC is being turned on now, start with an EQD of 0 */ +- if (rx_eq->enable_aic == 0 && +- coalesce->use_adaptive_rx_coalesce == 1) { +- rx_eq->cur_eqd = 0; ++ for_all_rx_queues(adapter, rxo, i) { ++ rx_eq = &rxo->rx_eq; ++ ++ if (!rx_eq->enable_aic && coalesce->use_adaptive_rx_coalesce) ++ rx_eq->cur_eqd = 0; ++ rx_eq->enable_aic = coalesce->use_adaptive_rx_coalesce; ++ ++ rx_max = coalesce->rx_coalesce_usecs_high; ++ rx_min = coalesce->rx_coalesce_usecs_low; ++ rx_cur = coalesce->rx_coalesce_usecs; ++ ++ if (rx_eq->enable_aic) { ++ if (rx_max > BE_MAX_EQD) ++ rx_max = BE_MAX_EQD; ++ if (rx_min > rx_max) ++ rx_min = rx_max; ++ rx_eq->max_eqd = rx_max; ++ rx_eq->min_eqd = rx_min; ++ if (rx_eq->cur_eqd > rx_max) ++ rx_eq->cur_eqd = rx_max; ++ if (rx_eq->cur_eqd < rx_min) ++ rx_eq->cur_eqd = rx_min; ++ } else { ++ if (rx_cur > BE_MAX_EQD) ++ rx_cur = BE_MAX_EQD; ++ if (rx_eq->cur_eqd != rx_cur) { ++ status = be_cmd_modify_eqd(adapter, rx_eq->q.id, ++ rx_cur); ++ if (!status) ++ rx_eq->cur_eqd = rx_cur; ++ } ++ } + } +- rx_eq->enable_aic = coalesce->use_adaptive_rx_coalesce; +- +- rx_max = coalesce->rx_coalesce_usecs_high; +- rx_min = coalesce->rx_coalesce_usecs_low; +- rx_cur = coalesce->rx_coalesce_usecs; + + tx_max = coalesce->tx_coalesce_usecs_high; + tx_min = coalesce->tx_coalesce_usecs_low; +@@ -181,27 +292,6 @@ be_set_coalesce(struct net_device *netdev, struct ethtool_coalesce *coalesce) + tx_eq->cur_eqd = tx_cur; + } + +- if (rx_eq->enable_aic) { +- if (rx_max > BE_MAX_EQD) +- rx_max = BE_MAX_EQD; +- if (rx_min > rx_max) +- rx_min = rx_max; +- rx_eq->max_eqd = rx_max; +- rx_eq->min_eqd = rx_min; +- if (rx_eq->cur_eqd > rx_max) +- rx_eq->cur_eqd = rx_max; +- if (rx_eq->cur_eqd < rx_min) +- rx_eq->cur_eqd = rx_min; +- } else { +- if (rx_cur > BE_MAX_EQD) +- rx_cur = BE_MAX_EQD; +- if (rx_eq->cur_eqd != rx_cur) { +- status = be_cmd_modify_eqd(adapter, rx_eq->q.id, +- rx_cur); +- if (!status) +- rx_eq->cur_eqd = rx_cur; +- } +- } + return 0; + } + +@@ -229,81 +319,294 @@ be_get_ethtool_stats(struct net_device *netdev, + struct ethtool_stats *stats, uint64_t *data) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_drvr_stats *drvr_stats = &adapter->stats.drvr_stats; +- struct be_hw_stats *hw_stats = hw_stats_from_cmd(adapter->stats.cmd.va); +- struct be_rxf_stats *rxf_stats = &hw_stats->rxf; +- struct be_port_rxf_stats *port_stats = +- &rxf_stats->port[adapter->port_num]; +- struct net_device_stats *net_stats = &adapter->stats.net_stats; +- struct be_erx_stats *erx_stats = &hw_stats->erx; ++ struct be_rx_obj *rxo; ++ struct be_tx_obj *txo; + void *p = NULL; +- int i; ++ int i, j, base; + + for (i = 0; i < ETHTOOL_STATS_NUM; i++) { + switch (et_stats[i].type) { + case NETSTAT: +- p = net_stats; ++ p = &adapter->net_stats; + break; + case DRVSTAT: +- p = drvr_stats; +- break; +- case PORTSTAT: +- p = port_stats; +- break; +- case MISCSTAT: +- p = rxf_stats; +- break; +- case ERXSTAT: /* Currently only one ERX stat is provided */ +- p = (u32 *)erx_stats + adapter->rx_obj.q.id; ++ p = &adapter->drv_stats; + break; + } + + p = (u8 *)p + et_stats[i].offset; + data[i] = (et_stats[i].size == sizeof(u64)) ? +- *(u64 *)p: *(u32 *)p; ++ *(u64 *)p:(*(u32 *)p); + } + +- return; ++ base = ETHTOOL_STATS_NUM; ++ for_all_rx_queues(adapter, rxo, j) { ++ for (i = 0; i < ETHTOOL_RXSTATS_NUM; i++) { ++ p = (u8 *)&rxo->stats + et_rx_stats[i].offset; ++ data[base + j * ETHTOOL_RXSTATS_NUM + i] = ++ (et_rx_stats[i].size == sizeof(u64)) ? ++ *(u64 *)p: *(u32 *)p; ++ } ++ } ++ ++ base = ETHTOOL_STATS_NUM + adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM; ++ for_all_tx_queues(adapter, txo, j) { ++ for (i = 0; i < ETHTOOL_TXSTATS_NUM; i++) { ++ p = (u8 *)&txo->stats + et_tx_stats[i].offset; ++ data[base + j * ETHTOOL_TXSTATS_NUM + i] = ++ (et_tx_stats[i].size == sizeof(u64)) ? ++ *(u64 *)p: *(u32 *)p; ++ } ++ } + } + + static void + be_get_stat_strings(struct net_device *netdev, uint32_t stringset, + uint8_t *data) + { +- int i; ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int i, j; ++ + switch (stringset) { + case ETH_SS_STATS: + for (i = 0; i < ETHTOOL_STATS_NUM; i++) { + memcpy(data, et_stats[i].desc, ETH_GSTRING_LEN); + data += ETH_GSTRING_LEN; + } ++ for (i = 0; i < adapter->num_rx_qs; i++) { ++ for (j = 0; j < ETHTOOL_RXSTATS_NUM; j++) { ++ sprintf(data, "rxq%d: %s", i, ++ et_rx_stats[j].desc); ++ data += ETH_GSTRING_LEN; ++ } ++ } ++ for (i = 0; i < adapter->num_tx_qs; i++) { ++ for (j = 0; j < ETHTOOL_TXSTATS_NUM; j++) { ++ sprintf(data, "txq%d: %s", i, ++ et_tx_stats[j].desc); ++ data += ETH_GSTRING_LEN; ++ } ++ } ++ break; ++ case ETH_SS_TEST: ++ for (i = 0; i < ETHTOOL_TESTS_NUM; i++) { ++ memcpy(data, et_self_tests[i], ETH_GSTRING_LEN); ++ data += ETH_GSTRING_LEN; ++ } + break; + } + } + ++#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 0, 0) + static int be_get_stats_count(struct net_device *netdev) + { +- return ETHTOOL_STATS_NUM; ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ return ETHTOOL_STATS_NUM + adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM ++ + adapter->num_tx_qs * ETHTOOL_TXSTATS_NUM; + } ++static int ++be_self_test_count(struct net_device *dev) ++{ ++ return ETHTOOL_TESTS_NUM; ++} ++#else ++ ++static int be_get_sset_count(struct net_device *netdev, int stringset) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ switch (stringset) { ++ case ETH_SS_TEST: ++ return ETHTOOL_TESTS_NUM; ++ case ETH_SS_STATS: ++ return ETHTOOL_STATS_NUM + ++ adapter->num_rx_qs * ETHTOOL_RXSTATS_NUM + ++ adapter->num_tx_qs * ETHTOOL_TXSTATS_NUM; ++ default: ++ return -EINVAL; ++ } ++} ++#endif + + static int be_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd) + { +- ecmd->speed = SPEED_10000; ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_phy_info phy_info; ++ u8 mac_speed = 0; ++ u16 link_speed = 0; ++ int link_status = LINK_DOWN; ++ int status; ++ ++ if ((adapter->link_speed < 0) || (!(netdev->flags & IFF_UP))) { ++ status = be_cmd_link_status_query(adapter, &link_status, ++ &mac_speed, &link_speed, 0); ++ ++ be_link_status_update(adapter, link_status); ++ /* link_speed is in units of 10 Mbps */ ++ if (link_speed) { ++ ecmd->speed = link_speed*10; ++ } else { ++ switch (mac_speed) { ++ case PHY_LINK_SPEED_10MBPS: ++ ecmd->speed = SPEED_10; ++ break; ++ case PHY_LINK_SPEED_100MBPS: ++ ecmd->speed = SPEED_100; ++ break; ++ case PHY_LINK_SPEED_1GBPS: ++ ecmd->speed = SPEED_1000; ++ break; ++ case PHY_LINK_SPEED_10GBPS: ++ ecmd->speed = SPEED_10000; ++ break; ++ case PHY_LINK_SPEED_ZERO: ++ ecmd->speed = 0; ++ break; ++ } ++ } ++ ++ status = be_cmd_get_phy_info(adapter, &phy_info); ++ if (!status) { ++ switch (phy_info.interface_type) { ++ case PHY_TYPE_XFP_10GB: ++ case PHY_TYPE_SFP_1GB: ++ case PHY_TYPE_SFP_PLUS_10GB: ++ ecmd->port = PORT_FIBRE; ++ break; ++ default: ++ ecmd->port = PORT_TP; ++ break; ++ } ++ ++ switch (phy_info.interface_type) { ++ case PHY_TYPE_KR_10GB: ++ case PHY_TYPE_KX4_10GB: ++ ecmd->transceiver = XCVR_INTERNAL; ++ break; ++ default: ++ ecmd->transceiver = XCVR_EXTERNAL; ++ break; ++ } ++ ++ if (phy_info.auto_speeds_supported) { ++ ecmd->supported |= SUPPORTED_Autoneg; ++ ecmd->autoneg = AUTONEG_ENABLE; ++ ecmd->advertising |= ADVERTISED_Autoneg; ++ } ++ ++ if (phy_info.misc_params & BE_PAUSE_SYM_EN) { ++ ecmd->supported |= SUPPORTED_Pause; ++ ecmd->advertising |= ADVERTISED_Pause; ++ } ++ ++ } ++ ++ /* Save for future use */ ++ adapter->link_speed = ecmd->speed; ++ adapter->port_type = ecmd->port; ++ adapter->transceiver = ecmd->transceiver; ++ adapter->autoneg = ecmd->autoneg; ++ } else { ++ ecmd->speed = adapter->link_speed; ++ ecmd->port = adapter->port_type; ++ ecmd->transceiver = adapter->transceiver; ++ ecmd->autoneg = adapter->autoneg; ++ } ++ + ecmd->duplex = DUPLEX_FULL; +- ecmd->autoneg = AUTONEG_DISABLE; ++ ecmd->phy_address = (adapter->hba_port_num << 4) | ++ (adapter->port_name[adapter->hba_port_num]); ++ switch (ecmd->port) { ++ case PORT_FIBRE: ++ ecmd->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_FIBRE); ++ break; ++ case PORT_TP: ++ ecmd->supported = (SUPPORTED_10000baseT_Full | SUPPORTED_TP); ++ break; ++ } ++ ++ if (ecmd->autoneg) { ++ ecmd->supported |= SUPPORTED_1000baseT_Full; ++ ecmd->advertising |= (ADVERTISED_10000baseT_Full | ++ ADVERTISED_1000baseT_Full); ++ } ++ + return 0; + } + ++static int be_set_settings(struct net_device *netdev, ++ struct ethtool_cmd *ecmd) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_phy_info phy_info; ++ u16 mac_speed=0; ++ u16 dac_cable_len=0; ++ u16 port_speed = 0; ++ int status; ++ ++ status = be_cmd_get_phy_info(adapter, &phy_info); ++ if (status) { ++ dev_warn(&adapter->pdev->dev, "port speed set failed.\n"); ++ return status; ++ } ++ ++ if (ecmd->autoneg == AUTONEG_ENABLE) { ++ switch(phy_info.interface_type) { ++ case PHY_TYPE_SFP_1GB: ++ case PHY_TYPE_BASET_1GB: ++ case PHY_TYPE_BASEX_1GB: ++ case PHY_TYPE_SGMII: ++ mac_speed = SPEED_AUTONEG_1GB_100MB_10MB; ++ break; ++ case PHY_TYPE_SFP_PLUS_10GB: ++ dev_warn(&adapter->pdev->dev, ++ "Autoneg not supported on this module. \n"); ++ return -EINVAL; ++ case PHY_TYPE_KR_10GB: ++ case PHY_TYPE_KX4_10GB: ++ mac_speed = SPEED_AUTONEG_10GB_1GB; ++ break; ++ case PHY_TYPE_BASET_10GB: ++ mac_speed = SPEED_AUTONEG_10GB_1GB_100MB; ++ break; ++ } ++ } else if(ecmd->autoneg == AUTONEG_DISABLE) { ++ if(ecmd->speed == SPEED_10) { ++ mac_speed = SPEED_FORCED_10MB; ++ } else if(ecmd->speed == SPEED_100) { ++ mac_speed = SPEED_FORCED_100MB; ++ } else if(ecmd->speed == SPEED_1000) { ++ mac_speed = SPEED_FORCED_1GB; ++ } else if(ecmd->speed == SPEED_10000) { ++ mac_speed = SPEED_FORCED_10GB; ++ } ++ } ++ ++ status = be_cmd_get_port_speed(adapter, adapter->hba_port_num, ++ &dac_cable_len, &port_speed); ++ ++ if (!status && port_speed != mac_speed) ++ status = be_cmd_set_port_speed_v1(adapter, ++ adapter->hba_port_num, mac_speed, ++ dac_cable_len); ++ if (status) ++ dev_warn(&adapter->pdev->dev, "port speed set failed.\n"); ++ ++ return status; ++ ++} ++ + static void + be_get_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring) + { + struct be_adapter *adapter = netdev_priv(netdev); + +- ring->rx_max_pending = adapter->rx_obj.q.len; +- ring->tx_max_pending = adapter->tx_obj.q.len; ++ ring->rx_max_pending = adapter->rx_obj[0].q.len; ++ ring->tx_max_pending = adapter->tx_obj[0].q.len; + +- ring->rx_pending = atomic_read(&adapter->rx_obj.q.used); +- ring->tx_pending = atomic_read(&adapter->tx_obj.q.used); ++ ring->rx_pending = atomic_read(&adapter->rx_obj[0].q.used); ++ ring->tx_pending = atomic_read(&adapter->tx_obj[0].q.used); + } + + static void +@@ -312,7 +615,7 @@ be_get_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd) + struct be_adapter *adapter = netdev_priv(netdev); + + be_cmd_get_flow_control(adapter, &ecmd->tx_pause, &ecmd->rx_pause); +- ecmd->autoneg = 0; ++ ecmd->autoneg = adapter->autoneg; + } + + static int +@@ -334,6 +637,203 @@ be_set_pauseparam(struct net_device *netdev, struct ethtool_pauseparam *ecmd) + return status; + } + ++#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 0, 0) ++static int ++be_phys_id(struct net_device *netdev, u32 data) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int status; ++ u32 cur; ++ ++ be_cmd_get_beacon_state(adapter, adapter->hba_port_num, &cur); ++ ++ if (cur == BEACON_STATE_ENABLED) ++ return 0; ++ ++ if (data < 2) ++ data = 2; ++ ++ status = be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0, ++ BEACON_STATE_ENABLED); ++ set_current_state(TASK_INTERRUPTIBLE); ++ schedule_timeout(data*HZ); ++ ++ status = be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0, ++ BEACON_STATE_DISABLED); ++ ++ return status; ++} ++#else ++static int ++be_set_phys_id(struct net_device *netdev, ++ enum ethtool_phys_id_state state) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ switch (state) { ++ case ETHTOOL_ID_ACTIVE: ++ be_cmd_get_beacon_state(adapter, adapter->hba_port_num, ++ &adapter->beacon_state); ++ return 1; /* cycle on/off once per second */ ++ ++ case ETHTOOL_ID_ON: ++ be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0, ++ BEACON_STATE_ENABLED); ++ break; ++ ++ case ETHTOOL_ID_OFF: ++ be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0, ++ BEACON_STATE_DISABLED); ++ break; ++ ++ case ETHTOOL_ID_INACTIVE: ++ be_cmd_set_beacon_state(adapter, adapter->hba_port_num, 0, 0, ++ adapter->beacon_state); ++ } ++ ++ return 0; ++} ++#endif ++ ++static bool ++be_is_wol_supported(struct be_adapter *adapter) ++{ ++ struct pci_dev *pdev = adapter->pdev; ++ ++ if (!be_physfn(adapter)) ++ return false; ++ ++ switch (pdev->subsystem_device) { ++ case OC_SUBSYS_DEVICE_ID1: ++ case OC_SUBSYS_DEVICE_ID2: ++ case OC_SUBSYS_DEVICE_ID3: ++ case OC_SUBSYS_DEVICE_ID4: ++ return false; ++ default: ++ return true; ++ } ++} ++ ++static void ++be_get_wol(struct net_device *netdev, struct ethtool_wolinfo *wol) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ if (be_is_wol_supported(adapter)) ++ wol->supported = WAKE_MAGIC; ++ if (adapter->wol) ++ wol->wolopts = WAKE_MAGIC; ++ else ++ wol->wolopts = 0; ++ memset(&wol->sopass, 0, sizeof(wol->sopass)); ++} ++ ++static int ++be_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ if (wol->wolopts & ~WAKE_MAGIC) ++ return -EOPNOTSUPP; ++ ++ if (!be_is_wol_supported(adapter)) { ++ dev_warn(&adapter->pdev->dev, ++ "WOL not supported for this subsystemid: %x\n", ++ adapter->pdev->subsystem_device); ++ return -EOPNOTSUPP; ++ } ++ ++ if (wol->wolopts & WAKE_MAGIC) ++ adapter->wol = true; ++ else ++ adapter->wol = false; ++ ++ return 0; ++} ++ ++static int ++be_test_ddr_dma(struct be_adapter *adapter) ++{ ++ int ret, i; ++ struct be_dma_mem ddrdma_cmd; ++ u64 pattern[2] = {0x5a5a5a5a5a5a5a5aULL, 0xa5a5a5a5a5a5a5a5ULL}; ++ ++ ddrdma_cmd.size = sizeof(struct be_cmd_req_ddrdma_test); ++ ddrdma_cmd.va = pci_alloc_consistent(adapter->pdev, ddrdma_cmd.size, ++ &ddrdma_cmd.dma); ++ if (!ddrdma_cmd.va) { ++ dev_err(&adapter->pdev->dev, "Memory allocation failure\n"); ++ return -ENOMEM; ++ } ++ ++ for (i = 0; i < 2; i++) { ++ ret = be_cmd_ddr_dma_test(adapter, pattern[i], ++ 4096, &ddrdma_cmd); ++ if (ret != 0) ++ goto err; ++ } ++ ++err: ++ pci_free_consistent(adapter->pdev, ddrdma_cmd.size, ++ ddrdma_cmd.va, ddrdma_cmd.dma); ++ return ret; ++} ++ ++static u64 be_loopback_test(struct be_adapter *adapter, u8 loopback_type, ++ u64 *status) ++{ ++ be_cmd_set_loopback(adapter, adapter->hba_port_num, ++ loopback_type, 1); ++ *status = be_cmd_loopback_test(adapter, adapter->hba_port_num, ++ loopback_type, 1500, ++ 2, 0xabc); ++ be_cmd_set_loopback(adapter, adapter->hba_port_num, ++ BE_NO_LOOPBACK, 1); ++ return *status; ++} ++ ++static void ++be_self_test(struct net_device *netdev, struct ethtool_test *test, u64 *data) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int link_status; ++ u8 mac_speed = 0; ++ u16 qos_link_speed = 0; ++ ++ memset(data, 0, sizeof(u64) * ETHTOOL_TESTS_NUM); ++ ++ if (test->flags & ETH_TEST_FL_OFFLINE) { ++ if (be_loopback_test(adapter, BE_MAC_LOOPBACK, ++ &data[0]) != 0) { ++ test->flags |= ETH_TEST_FL_FAILED; ++ } ++ if (be_loopback_test(adapter, BE_PHY_LOOPBACK, ++ &data[1]) != 0) { ++ test->flags |= ETH_TEST_FL_FAILED; ++ } ++ if (be_loopback_test(adapter, BE_ONE_PORT_EXT_LOOPBACK, ++ &data[2]) != 0) { ++ test->flags |= ETH_TEST_FL_FAILED; ++ } ++ } ++ ++ if (be_test_ddr_dma(adapter) != 0) { ++ data[3] = 1; ++ test->flags |= ETH_TEST_FL_FAILED; ++ } ++ ++ if (be_cmd_link_status_query(adapter, &link_status, &mac_speed, ++ &qos_link_speed, 0) != 0) { ++ test->flags |= ETH_TEST_FL_FAILED; ++ data[4] = -1; ++ } else if (!mac_speed) { ++ test->flags |= ETH_TEST_FL_FAILED; ++ data[4] = 1; ++ } ++ ++} ++ ++#ifdef HAVE_ETHTOOL_FLASH + static int + be_do_flash(struct net_device *netdev, struct ethtool_flash *efl) + { +@@ -347,11 +847,73 @@ be_do_flash(struct net_device *netdev, struct ethtool_flash *efl) + + return be_load_fw(adapter, file_name); + } ++#endif + +-const struct ethtool_ops be_ethtool_ops = { ++static int ++be_get_eeprom_len(struct net_device *netdev) ++{ ++ return BE_READ_SEEPROM_LEN; ++} ++ ++static int ++be_read_eeprom(struct net_device *netdev, struct ethtool_eeprom *eeprom, ++ uint8_t *data) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_dma_mem eeprom_cmd; ++ struct be_cmd_resp_seeprom_read *resp; ++ int status; ++ ++ if (!eeprom->len) ++ return -EINVAL; ++ ++ eeprom->magic = BE_VENDOR_ID | (adapter->pdev->device<<16); ++ ++ memset(&eeprom_cmd, 0, sizeof(struct be_dma_mem)); ++ eeprom_cmd.size = sizeof(struct be_cmd_req_seeprom_read); ++ eeprom_cmd.va = pci_alloc_consistent(adapter->pdev, eeprom_cmd.size, ++ &eeprom_cmd.dma); ++ ++ if (!eeprom_cmd.va) { ++ dev_err(&adapter->pdev->dev, ++ "Memory allocation failure. Could not read eeprom\n"); ++ return -ENOMEM; ++ } ++ ++ status = be_cmd_get_seeprom_data(adapter, &eeprom_cmd); ++ ++ if (!status) { ++ resp = (struct be_cmd_resp_seeprom_read *) eeprom_cmd.va; ++ memcpy(data, resp->seeprom_data + eeprom->offset, eeprom->len); ++ } ++ pci_free_consistent(adapter->pdev, eeprom_cmd.size, eeprom_cmd.va, ++ eeprom_cmd.dma); ++ ++ return status; ++} ++ ++static int be_set_tso(struct net_device *netdev, uint32_t data) ++{ ++ if (data) { ++ netdev->features |= NETIF_F_TSO; ++ netdev->features |= NETIF_F_TSO6; ++ } else { ++ netdev->features &= ~NETIF_F_TSO; ++ netdev->features &= ~NETIF_F_TSO6; ++ } ++ return 0; ++} ++ ++ ++struct ethtool_ops be_ethtool_ops = { + .get_settings = be_get_settings, ++ .set_settings = be_set_settings, + .get_drvinfo = be_get_drvinfo, ++ .get_wol = be_get_wol, ++ .set_wol = be_set_wol, + .get_link = ethtool_op_get_link, ++ .get_eeprom_len = be_get_eeprom_len, ++ .get_eeprom = be_read_eeprom, + .get_coalesce = be_get_coalesce, + .set_coalesce = be_set_coalesce, + .get_ringparam = be_get_ringparam, +@@ -364,9 +926,21 @@ const struct ethtool_ops be_ethtool_ops = { + .get_sg = ethtool_op_get_sg, + .set_sg = ethtool_op_set_sg, + .get_tso = ethtool_op_get_tso, +- .set_tso = ethtool_op_set_tso, ++ .set_tso = be_set_tso, + .get_strings = be_get_stat_strings, ++#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 0, 0) ++ .phys_id = be_phys_id, + .get_stats_count = be_get_stats_count, ++ .self_test_count = be_self_test_count, ++#else ++ .set_phys_id = be_set_phys_id, ++ .get_sset_count = be_get_sset_count, ++#endif + .get_ethtool_stats = be_get_ethtool_stats, ++ .get_regs_len = be_get_reg_len, ++ .get_regs = be_get_regs, ++#ifdef HAVE_ETHTOOL_FLASH + .flash_device = be_do_flash, ++#endif ++ .self_test = be_self_test + }; +diff --git a/drivers/net/benet/be_hw.h b/drivers/net/benet/be_hw.h +index a3394b4..f871d8c 100644 +--- a/drivers/net/benet/be_hw.h ++++ b/drivers/net/benet/be_hw.h +@@ -1,18 +1,18 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + + /********* Mailbox door bell *************/ +@@ -26,24 +26,34 @@ + * queue entry. + */ + #define MPU_MAILBOX_DB_OFFSET 0x160 +-#define MPU_MAILBOX_DB_RDY_MASK 0x1 /* bit 0 */ ++#define MPU_MAILBOX_DB_RDY_MASK 0x1 /* bit 0 */ + #define MPU_MAILBOX_DB_HI_MASK 0x2 /* bit 1 */ + +-#define MPU_EP_CONTROL 0 ++#define MPU_EP_CONTROL 0 + + /********** MPU semphore ******************/ +-#define MPU_EP_SEMAPHORE_OFFSET 0xac ++#define MPU_EP_SEMAPHORE_OFFSET 0xac ++#define MPU_EP_SEMAPHORE_IF_TYPE2_OFFSET 0x400 + #define EP_SEMAPHORE_POST_STAGE_MASK 0x0000FFFF + #define EP_SEMAPHORE_POST_ERR_MASK 0x1 + #define EP_SEMAPHORE_POST_ERR_SHIFT 31 + /* MPU semphore POST stage values */ +-#define POST_STAGE_AWAITING_HOST_RDY 0x1 /* FW awaiting goahead from host */ +-#define POST_STAGE_HOST_RDY 0x2 /* Host has given go-ahed to FW */ ++#define POST_STAGE_AWAITING_HOST_RDY 0x1 /* FW awaiting goahead from host */ ++#define POST_STAGE_HOST_RDY 0x2 /* Host has given go-ahed to FW */ + #define POST_STAGE_BE_RESET 0x3 /* Host wants to reset chip */ + #define POST_STAGE_ARMFW_RDY 0xc000 /* FW is done with POST */ + ++/* Lancer SLIPORT_CONTROL SLIPORT_STATUS registers */ ++#define SLIPORT_STATUS_OFFSET 0x404 ++#define SLIPORT_CONTROL_OFFSET 0x408 ++ ++#define SLIPORT_STATUS_ERR_MASK 0x80000000 ++#define SLIPORT_STATUS_RN_MASK 0x01000000 ++#define SLIPORT_STATUS_RDY_MASK 0x00800000 ++#define SLI_PORT_CONTROL_IP_MASK 0x08000000 ++ + /********* Memory BAR register ************/ +-#define PCICFG_MEMBAR_CTRL_INT_CTRL_OFFSET 0xfc ++#define PCICFG_MEMBAR_CTRL_INT_CTRL_OFFSET 0xfc + /* Host Interrupt Enable, if set interrupts are enabled although "PCI Interrupt + * Disable" may still globally block interrupts in addition to individual + * interrupt masks; a mechanism for the device driver to block all interrupts +@@ -52,13 +62,70 @@ + */ + #define MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK (1 << 29) /* bit 29 */ + ++/********* Link Status CSR ****************/ ++#define PCICFG_PCIE_LINK_STATUS_OFFSET 0xd0 ++#define PCIE_LINK_STATUS_SPEED_MASK 0xFF /* bits 16 - 19 */ ++#define PCIE_LINK_STATUS_SPEED_SHIFT 16 ++#define PCIE_LINK_STATUS_NEG_WIDTH_MASK 0x3F /* bits 20 - 25 */ ++#define PCIE_LINK_STATUS_NEG_WIDTH_SHIFT 20 ++ ++/********* Link Capability CSR ************/ ++#define PCICFG_PCIE_LINK_CAP_OFFSET 0xcc ++#define PCIE_LINK_CAP_MAX_SPEED_MASK 0xFF /* bits 0 - 3 */ ++#define PCIE_LINK_CAP_MAX_SPEED_SHIFT 0 ++#define PCIE_LINK_CAP_MAX_WIDTH_MASK 0x3F /* bits 4 - 9 */ ++#define PCIE_LINK_CAP_MAX_WIDTH_SHIFT 4 ++ ++/********* PCI Function Capability ************/ ++#define BE_FUNCTION_CAPS_UNCLASSIFIED_STATS 0x1 ++#define BE_FUNCTION_CAPS_RSS 0x2 ++#define BE_FUNCTION_CAPS_PROMISCUOUS 0x4 ++#define BE_FUNCTION_CAPS_LEGACY_MODE 0x8 ++ ++/********* Power managment (WOL) **********/ ++#define PCICFG_PM_CONTROL_OFFSET 0x44 ++#define PCICFG_PM_CONTROL_MASK 0x108 /* bits 3 & 8 */ ++ ++/********* Online Control Registers *******/ ++#define PCICFG_ONLINE0 0xB0 ++#define PCICFG_ONLINE1 0xB4 ++ ++/********* UE Status and Mask Registers ***/ ++#define PCICFG_UE_STATUS_LOW 0xA0 ++#define PCICFG_UE_STATUS_HIGH 0xA4 ++#define PCICFG_UE_STATUS_LOW_MASK 0xA8 ++#define PCICFG_UE_STATUS_HI_MASK 0xAC ++ ++/******** SLI_INTF ***********************/ ++#define SLI_INTF_REG_OFFSET 0x58 ++#define SLI_INTF_VALID_MASK 0xE0000000 ++#define SLI_INTF_VALID 0xC0000000 ++#define SLI_INTF_HINT2_MASK 0x1F000000 ++#define SLI_INTF_HINT2_SHIFT 24 ++#define SLI_INTF_HINT1_MASK 0x00FF0000 ++#define SLI_INTF_HINT1_SHIFT 16 ++#define SLI_INTF_FAMILY_MASK 0x00000F00 ++#define SLI_INTF_FAMILY_SHIFT 8 ++#define SLI_INTF_IF_TYPE_MASK 0x0000F000 ++#define SLI_INTF_IF_TYPE_SHIFT 12 ++#define SLI_INTF_REV_MASK 0x000000F0 ++#define SLI_INTF_REV_SHIFT 4 ++#define SLI_INTF_FT_MASK 0x00000001 ++ ++/* SLI family */ ++#define BE_SLI_FAMILY 0x0 ++#define LANCER_A0_SLI_FAMILY 0xA ++ + /********* ISR0 Register offset **********/ +-#define CEV_ISR0_OFFSET 0xC18 ++#define CEV_ISR0_OFFSET 0xC18 + #define CEV_ISR_SIZE 4 + + /********* Event Q door bell *************/ + #define DB_EQ_OFFSET DB_CQ_OFFSET + #define DB_EQ_RING_ID_MASK 0x1FF /* bits 0 - 8 */ ++#define DB_EQ_RING_ID_EXT_MASK 0x3e00 /* bits 9-13 */ ++#define DB_EQ_RING_ID_EXT_MASK_SHIFT (2) /* qid bits 9-13 placing at 11-15 */ ++ + /* Clear the interrupt for this eq */ + #define DB_EQ_CLR_SHIFT (9) /* bit 9 */ + /* Must be 1 */ +@@ -69,12 +136,16 @@ + #define DB_EQ_REARM_SHIFT (29) /* bit 29 */ + + /********* Compl Q door bell *************/ +-#define DB_CQ_OFFSET 0x120 ++#define DB_CQ_OFFSET 0x120 + #define DB_CQ_RING_ID_MASK 0x3FF /* bits 0 - 9 */ ++#define DB_CQ_RING_ID_EXT_MASK 0x7C00 /* bits 10-14 */ ++#define DB_CQ_RING_ID_EXT_MASK_SHIFT (1) /* qid bits 10-14 ++ placing at 11-15 */ ++ + /* Number of event entries processed */ +-#define DB_CQ_NUM_POPPED_SHIFT (16) /* bits 16 - 28 */ ++#define DB_CQ_NUM_POPPED_SHIFT (16) /* bits 16 - 28 */ + /* Rearm bit */ +-#define DB_CQ_REARM_SHIFT (29) /* bit 29 */ ++#define DB_CQ_REARM_SHIFT (29) /* bit 29 */ + + /********** TX ULP door bell *************/ + #define DB_TXULP1_OFFSET 0x60 +@@ -84,25 +155,103 @@ + #define DB_TXULP_NUM_POSTED_MASK 0x3FFF /* bits 16 - 29 */ + + /********** RQ(erx) door bell ************/ +-#define DB_RQ_OFFSET 0x100 ++#define DB_RQ_OFFSET 0x100 + #define DB_RQ_RING_ID_MASK 0x3FF /* bits 0 - 9 */ + /* Number of rx frags posted */ + #define DB_RQ_NUM_POSTED_SHIFT (24) /* bits 24 - 31 */ + + /********** MCC door bell ************/ +-#define DB_MCCQ_OFFSET 0x140 ++#define DB_MCCQ_OFFSET 0x140 + #define DB_MCCQ_RING_ID_MASK 0x7FF /* bits 0 - 10 */ + /* Number of entries posted */ + #define DB_MCCQ_NUM_POSTED_SHIFT (16) /* bits 16 - 29 */ + ++/********** SRIOV VF PCICFG OFFSET ********/ ++#define SRIOV_VF_PCICFG_OFFSET (4096) ++ ++/********** FAT TABLE ********/ ++#define RETRIEVE_FAT 0 ++#define QUERY_FAT 1 ++ ++/* Flashrom related descriptors */ ++#define IMAGE_TYPE_FIRMWARE 160 ++#define IMAGE_TYPE_BOOTCODE 224 ++#define IMAGE_TYPE_OPTIONROM 32 ++ ++#define NUM_FLASHDIR_ENTRIES 32 ++ ++#define IMG_TYPE_ISCSI_ACTIVE 0 ++#define IMG_TYPE_REDBOOT 1 ++#define IMG_TYPE_BIOS 2 ++#define IMG_TYPE_PXE_BIOS 3 ++#define IMG_TYPE_FCOE_BIOS 8 ++#define IMG_TYPE_ISCSI_BACKUP 9 ++#define IMG_TYPE_FCOE_FW_ACTIVE 10 ++#define IMG_TYPE_FCOE_FW_BACKUP 11 ++#define IMG_TYPE_NCSI_FW 13 ++#define IMG_TYPE_PHY_FW 99 ++#define TN_8022 13 ++ ++#define ILLEGAL_IOCTL_REQ 2 ++#define FLASHROM_OPER_PHY_FLASH 9 ++#define FLASHROM_OPER_PHY_SAVE 10 ++#define FLASHROM_OPER_FLASH 1 ++#define FLASHROM_OPER_SAVE 2 ++#define FLASHROM_OPER_REPORT 4 ++ ++#define FLASH_IMAGE_MAX_SIZE_g2 (1310720) /* Max firmware image size */ ++#define FLASH_BIOS_IMAGE_MAX_SIZE_g2 (262144) /* Max OPTION ROM image sz */ ++#define FLASH_REDBOOT_IMAGE_MAX_SIZE_g2 (262144) /* Max Redboot image sz */ ++#define FLASH_IMAGE_MAX_SIZE_g3 (2097152) /* Max firmware image size */ ++#define FLASH_BIOS_IMAGE_MAX_SIZE_g3 (524288) /* Max OPTION ROM image sz */ ++#define FLASH_REDBOOT_IMAGE_MAX_SIZE_g3 (1048576) /* Max Redboot image sz */ ++#define FLASH_NCSI_IMAGE_MAX_SIZE_g3 (262144) ++#define FLASH_PHY_FW_IMAGE_MAX_SIZE_g3 (262144) ++ ++#define FLASH_NCSI_MAGIC (0x16032009) ++#define FLASH_NCSI_DISABLED (0) ++#define FLASH_NCSI_ENABLED (1) ++ ++#define FLASH_NCSI_BITFILE_HDR_OFFSET (0x600000) ++ ++/* Offsets for components on Flash. */ ++#define FLASH_iSCSI_PRIMARY_IMAGE_START_g2 (1048576) ++#define FLASH_iSCSI_BACKUP_IMAGE_START_g2 (2359296) ++#define FLASH_FCoE_PRIMARY_IMAGE_START_g2 (3670016) ++#define FLASH_FCoE_BACKUP_IMAGE_START_g2 (4980736) ++#define FLASH_iSCSI_BIOS_START_g2 (7340032) ++#define FLASH_PXE_BIOS_START_g2 (7864320) ++#define FLASH_FCoE_BIOS_START_g2 (524288) ++#define FLASH_REDBOOT_START_g2 (0) ++ ++#define FLASH_NCSI_START_g3 (15990784) ++#define FLASH_iSCSI_PRIMARY_IMAGE_START_g3 (2097152) ++#define FLASH_iSCSI_BACKUP_IMAGE_START_g3 (4194304) ++#define FLASH_FCoE_PRIMARY_IMAGE_START_g3 (6291456) ++#define FLASH_FCoE_BACKUP_IMAGE_START_g3 (8388608) ++#define FLASH_iSCSI_BIOS_START_g3 (12582912) ++#define FLASH_PXE_BIOS_START_g3 (13107200) ++#define FLASH_FCoE_BIOS_START_g3 (13631488) ++#define FLASH_REDBOOT_START_g3 (262144) ++#define FLASH_PHY_FW_START_g3 (1310720) ++ ++/************* Rx Packet Type Encoding **************/ ++#define BE_UNICAST_PACKET 0 ++#define BE_MULTICAST_PACKET 1 ++#define BE_BROADCAST_PACKET 2 ++#define BE_RSVD_PACKET 3 ++ + /* + * BE descriptors: host memory data structures whose formats + * are hardwired in BE silicon. + */ + /* Event Queue Descriptor */ +-#define EQ_ENTRY_VALID_MASK 0x1 /* bit 0 */ +-#define EQ_ENTRY_RES_ID_MASK 0xFFFF /* bits 16 - 31 */ +-#define EQ_ENTRY_RES_ID_SHIFT 16 ++#define EQ_ENTRY_VALID_MASK 0x1 /* bit 0 */ ++#define EQ_ENTRY_RES_ID_MASK 0xFFFF /* bits 16 - 31 */ ++#define EQ_ENTRY_RES_ID_SHIFT 16 ++ ++#define BE_MAC_PROMISCUOUS 62 /* Promiscuous mode */ ++ + struct be_eq_entry { + u32 evt; + }; +@@ -126,7 +275,7 @@ struct amap_eth_hdr_wrb { + u8 event; + u8 crc; + u8 forward; +- u8 ipsec; ++ u8 lso6; + u8 mgmt; + u8 ipcs; + u8 udpcs; +@@ -151,7 +300,7 @@ struct be_eth_hdr_wrb { + * offset/shift/mask of each field */ + struct amap_eth_tx_compl { + u8 wrb_index[16]; /* dword 0 */ +- u8 ct[2]; /* dword 0 */ ++ u8 ct[2]; /* dword 0 */ + u8 port[2]; /* dword 0 */ + u8 rsvd0[8]; /* dword 0 */ + u8 status[4]; /* dword 0 */ +@@ -179,10 +328,10 @@ struct be_eth_rx_d { + + /* RX Compl Queue Descriptor */ + +-/* Pseudo amap definition for eth_rx_compl in which each bit of the +- * actual structure is defined as a byte: used to calculate ++/* Pseudo amap definition for BE2 and BE3 legacy mode eth_rx_compl in which ++ * each bit of the actual structure is defined as a byte: used to calculate + * offset/shift/mask of each field */ +-struct amap_eth_rx_compl { ++struct amap_eth_rx_compl_v0 { + u8 vlan_tag[16]; /* dword 0 */ + u8 pktsize[14]; /* dword 0 */ + u8 port; /* dword 0 */ +@@ -213,39 +362,91 @@ struct amap_eth_rx_compl { + u8 rsshash[32]; /* dword 3 */ + } __packed; + ++/* Pseudo amap definition for BE3 native mode eth_rx_compl in which ++ * each bit of the actual structure is defined as a byte: used to calculate ++ * offset/shift/mask of each field */ ++struct amap_eth_rx_compl_v1 { ++ u8 vlan_tag[16]; /* dword 0 */ ++ u8 pktsize[14]; /* dword 0 */ ++ u8 vtp; /* dword 0 */ ++ u8 ip_opt; /* dword 0 */ ++ u8 err; /* dword 1 */ ++ u8 rsshp; /* dword 1 */ ++ u8 ipf; /* dword 1 */ ++ u8 tcpf; /* dword 1 */ ++ u8 udpf; /* dword 1 */ ++ u8 ipcksm; /* dword 1 */ ++ u8 l4_cksm; /* dword 1 */ ++ u8 ip_version; /* dword 1 */ ++ u8 macdst[7]; /* dword 1 */ ++ u8 rsvd0; /* dword 1 */ ++ u8 fragndx[10]; /* dword 1 */ ++ u8 ct[2]; /* dword 1 */ ++ u8 sw; /* dword 1 */ ++ u8 numfrags[3]; /* dword 1 */ ++ u8 rss_flush; /* dword 2 */ ++ u8 cast_enc[2]; /* dword 2 */ ++ u8 vtm; /* dword 2 */ ++ u8 rss_bank; /* dword 2 */ ++ u8 port[2]; /* dword 2 */ ++ u8 vntagp; /* dword 2 */ ++ u8 header_len[8]; /* dword 2 */ ++ u8 header_split[2]; /* dword 2 */ ++ u8 rsvd1[13]; /* dword 2 */ ++ u8 valid; /* dword 2 */ ++ u8 rsshash[32]; /* dword 3 */ ++} __packed; ++ + struct be_eth_rx_compl { + u32 dw[4]; + }; + +-/* Flashrom related descriptors */ +-#define IMAGE_TYPE_FIRMWARE 160 +-#define IMAGE_TYPE_BOOTCODE 224 +-#define IMAGE_TYPE_OPTIONROM 32 ++struct mgmt_hba_attribs { ++ u8 flashrom_version_string[32]; ++ u8 manufacturer_name[32]; ++ u32 supported_modes; ++ u32 rsvd0[3]; ++ u8 ncsi_ver_string[12]; ++ u32 default_extended_timeout; ++ u8 controller_model_number[32]; ++ u8 controller_description[64]; ++ u8 controller_serial_number[32]; ++ u8 ip_version_string[32]; ++ u8 firmware_version_string[32]; ++ u8 bios_version_string[32]; ++ u8 redboot_version_string[32]; ++ u8 driver_version_string[32]; ++ u8 fw_on_flash_version_string[32]; ++ u32 functionalities_supported; ++ u16 max_cdblength; ++ u8 asic_revision; ++ u8 generational_guid[16]; ++ u8 hba_port_count; ++ u16 default_link_down_timeout; ++ u8 iscsi_ver_min_max; ++ u8 multifunction_device; ++ u8 cache_valid; ++ u8 hba_status; ++ u8 max_domains_supported; ++ u8 phy_port; ++ u32 firmware_post_status; ++ u32 hba_mtu[8]; ++ u32 rsvd1[4]; ++}; + +-#define NUM_FLASHDIR_ENTRIES 32 +- +-#define FLASHROM_TYPE_ISCSI_ACTIVE 0 +-#define FLASHROM_TYPE_BIOS 2 +-#define FLASHROM_TYPE_PXE_BIOS 3 +-#define FLASHROM_TYPE_FCOE_BIOS 8 +-#define FLASHROM_TYPE_ISCSI_BACKUP 9 +-#define FLASHROM_TYPE_FCOE_FW_ACTIVE 10 +-#define FLASHROM_TYPE_FCOE_FW_BACKUP 11 +- +-#define FLASHROM_OPER_FLASH 1 +-#define FLASHROM_OPER_SAVE 2 +- +-#define FLASH_IMAGE_MAX_SIZE (1310720) /* Max firmware image size */ +-#define FLASH_BIOS_IMAGE_MAX_SIZE (262144) /* Max OPTION ROM image sz */ +- +-/* Offsets for components on Flash. */ +-#define FLASH_iSCSI_PRIMARY_IMAGE_START (1048576) +-#define FLASH_iSCSI_BACKUP_IMAGE_START (2359296) +-#define FLASH_FCoE_PRIMARY_IMAGE_START (3670016) +-#define FLASH_FCoE_BACKUP_IMAGE_START (4980736) +-#define FLASH_iSCSI_BIOS_START (7340032) +-#define FLASH_PXE_BIOS_START (7864320) +-#define FLASH_FCoE_BIOS_START (524288) ++struct mgmt_controller_attrib { ++ struct mgmt_hba_attribs hba_attribs; ++ u16 pci_vendor_id; ++ u16 pci_device_id; ++ u16 pci_sub_vendor_id; ++ u16 pci_sub_system_id; ++ u8 pci_bus_number; ++ u8 pci_device_number; ++ u8 pci_function_number; ++ u8 interface_type; ++ u64 unique_identifier; ++ u32 rsvd0[5]; ++}; + + struct controller_id { + u32 vendor; +@@ -254,7 +455,20 @@ struct controller_id { + u32 subdevice; + }; + +-struct flash_file_hdr { ++struct flash_comp { ++ unsigned long offset; ++ int optype; ++ int size; ++}; ++ ++struct image_hdr { ++ u32 imageid; ++ u32 imageoffset; ++ u32 imagelength; ++ u32 image_checksum; ++ u8 image_version[32]; ++}; ++struct flash_file_hdr_g2 { + u8 sign[32]; + u32 cksum; + u32 antidote; +@@ -266,6 +480,17 @@ struct flash_file_hdr { + u8 build[24]; + }; + ++struct flash_file_hdr_g3 { ++ u8 sign[52]; ++ u8 ufi_version[4]; ++ u32 file_len; ++ u32 cksum; ++ u32 antidote; ++ u32 num_imgs; ++ u8 build[24]; ++ u8 rsvd[32]; ++}; ++ + struct flash_section_hdr { + u32 format_rev; + u32 cksum; +@@ -299,3 +524,19 @@ struct flash_section_info { + struct flash_section_hdr fsec_hdr; + struct flash_section_entry fsec_entry[32]; + }; ++ ++struct flash_ncsi_image_hdr { ++ u32 magic; ++ u8 hdr_len; ++ u8 type; ++ u16 hdr_ver; ++ u8 rsvd0[2]; ++ u16 load_offset; ++ u32 len; ++ u32 flash_offset; ++ u8 ver[16]; ++ u8 name[24]; ++ u32 img_cksum; ++ u8 rsvd1[4]; ++ u32 hdr_cksum; ++}; +diff --git a/drivers/net/benet/be_main.c b/drivers/net/benet/be_main.c +index 000e377..f501aa3 100644 +--- a/drivers/net/benet/be_main.c ++++ b/drivers/net/benet/be_main.c +@@ -1,18 +1,18 @@ + /* +- * Copyright (C) 2005 - 2009 ServerEngines ++ * Copyright (C) 2005 - 2011 Emulex + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License version 2 +- * as published by the Free Software Foundation. The full GNU General ++ * as published by the Free Software Foundation. The full GNU General + * Public License is included in this distribution in the file called COPYING. + * + * Contact Information: +- * linux-drivers@serverengines.com ++ * linux-drivers@emulex.com + * +- * ServerEngines +- * 209 N. Fair Oaks Ave +- * Sunnyvale, CA 94085 ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 + */ + + #include "be.h" +@@ -22,23 +22,119 @@ + MODULE_VERSION(DRV_VER); + MODULE_DEVICE_TABLE(pci, be_dev_ids); + MODULE_DESCRIPTION(DRV_DESC " " DRV_VER); +-MODULE_AUTHOR("ServerEngines Corporation"); ++MODULE_AUTHOR("Emulex Corporation"); + MODULE_LICENSE("GPL"); ++MODULE_INFO(supported, "external"); + +-static unsigned int rx_frag_size = 2048; +-module_param(rx_frag_size, uint, S_IRUGO); +-MODULE_PARM_DESC(rx_frag_size, "Size of a fragment that holds rcvd data."); ++static ushort rx_frag_size = 2048; ++static unsigned int num_vfs; ++static unsigned int msix = 1; ++module_param(rx_frag_size, ushort, S_IRUGO); ++module_param(num_vfs, uint, S_IRUGO); ++module_param(msix, uint, S_IRUGO); ++MODULE_PARM_DESC(rx_frag_size, "Size of receive fragment buffer" ++ " - 2048 (default), 4096 or 8192"); ++MODULE_PARM_DESC(num_vfs, "Number of PCI VFs to initialize"); ++MODULE_PARM_DESC(msix, "Enable and disable the MSI" ++ "x (By default MSIx is enabled)"); ++static unsigned int gro = 1; ++module_param(gro, uint, S_IRUGO); ++MODULE_PARM_DESC(gro, "Enable or Disable GRO. Enabled by default"); ++ ++static unsigned int multi_rxq = true; ++module_param(multi_rxq, uint, S_IRUGO); ++MODULE_PARM_DESC(multi_rxq, "Multi Rx Queue support. Enabled by default"); + + static DEFINE_PCI_DEVICE_TABLE(be_dev_ids) = { + { PCI_DEVICE(BE_VENDOR_ID, BE_DEVICE_ID1) }, + { PCI_DEVICE(BE_VENDOR_ID, BE_DEVICE_ID2) }, + { PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID1) }, + { PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID2) }, +- { PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID3) }, ++ /* ++ * Lancer is not part of Palau 4.0 ++ * { PCI_DEVICE(EMULEX_VENDOR_ID, OC_DEVICE_ID3)}, ++ */ + { 0 } + }; + MODULE_DEVICE_TABLE(pci, be_dev_ids); + ++/* UE Status Low CSR */ ++static char *ue_status_low_desc[] = { ++ "CEV", ++ "CTX", ++ "DBUF", ++ "ERX", ++ "Host", ++ "MPU", ++ "NDMA", ++ "PTC ", ++ "RDMA ", ++ "RXF ", ++ "RXIPS ", ++ "RXULP0 ", ++ "RXULP1 ", ++ "RXULP2 ", ++ "TIM ", ++ "TPOST ", ++ "TPRE ", ++ "TXIPS ", ++ "TXULP0 ", ++ "TXULP1 ", ++ "UC ", ++ "WDMA ", ++ "TXULP2 ", ++ "HOST1 ", ++ "P0_OB_LINK ", ++ "P1_OB_LINK ", ++ "HOST_GPIO ", ++ "MBOX ", ++ "AXGMAC0", ++ "AXGMAC1", ++ "JTAG", ++ "MPU_INTPEND" ++}; ++ ++/* UE Status High CSR */ ++static char *ue_status_hi_desc[] = { ++ "LPCMEMHOST", ++ "MGMT_MAC", ++ "PCS0ONLINE", ++ "MPU_IRAM", ++ "PCS1ONLINE", ++ "PCTL0", ++ "PCTL1", ++ "PMEM", ++ "RR", ++ "TXPB", ++ "RXPP", ++ "XAUI", ++ "TXP", ++ "ARM", ++ "IPC", ++ "HOST2", ++ "HOST3", ++ "HOST4", ++ "HOST5", ++ "HOST6", ++ "HOST7", ++ "HOST8", ++ "HOST9", ++ "NETC", ++ "Unknown", ++ "Unknown", ++ "Unknown", ++ "Unknown", ++ "Unknown", ++ "Unknown", ++ "Unknown", ++ "Unknown" ++}; ++ ++static inline bool be_multi_rxq(struct be_adapter *adapter) ++{ ++ return (adapter->num_rx_qs > 1); ++} ++ + static void be_queue_free(struct be_adapter *adapter, struct be_queue_info *q) + { + struct be_dma_mem *mem = &q->dma_mem; +@@ -69,6 +165,9 @@ static void be_intr_set(struct be_adapter *adapter, bool enable) + u32 reg = ioread32(addr); + u32 enabled = reg & MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK; + ++ if (adapter->eeh_err) ++ return; ++ + if (!enabled && enable) + reg |= MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK; + else if (enabled && !enable) +@@ -84,6 +183,8 @@ static void be_rxq_notify(struct be_adapter *adapter, u16 qid, u16 posted) + u32 val = 0; + val |= qid & DB_RQ_RING_ID_MASK; + val |= posted << DB_RQ_NUM_POSTED_SHIFT; ++ ++ wmb(); + iowrite32(val, adapter->db + DB_RQ_OFFSET); + } + +@@ -92,6 +193,8 @@ static void be_txq_notify(struct be_adapter *adapter, u16 qid, u16 posted) + u32 val = 0; + val |= qid & DB_TXULP_RING_ID_MASK; + val |= (posted & DB_TXULP_NUM_POSTED_MASK) << DB_TXULP_NUM_POSTED_SHIFT; ++ ++ wmb(); + iowrite32(val, adapter->db + DB_TXULP1_OFFSET); + } + +@@ -100,6 +203,12 @@ static void be_eq_notify(struct be_adapter *adapter, u16 qid, + { + u32 val = 0; + val |= qid & DB_EQ_RING_ID_MASK; ++ val |= ((qid & DB_EQ_RING_ID_EXT_MASK) << ++ DB_EQ_RING_ID_EXT_MASK_SHIFT); ++ ++ if (adapter->eeh_err) ++ return; ++ + if (arm) + val |= 1 << DB_EQ_REARM_SHIFT; + if (clear_int) +@@ -113,6 +222,12 @@ void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm, u16 num_popped) + { + u32 val = 0; + val |= qid & DB_CQ_RING_ID_MASK; ++ val |= ((qid & DB_CQ_RING_ID_EXT_MASK) << ++ DB_CQ_RING_ID_EXT_MASK_SHIFT); ++ ++ if (adapter->eeh_err) ++ return; ++ + if (arm) + val |= 1 << DB_CQ_REARM_SHIFT; + val |= num_popped << DB_CQ_NUM_POPPED_SHIFT; +@@ -124,96 +239,250 @@ static int be_mac_addr_set(struct net_device *netdev, void *p) + struct be_adapter *adapter = netdev_priv(netdev); + struct sockaddr *addr = p; + int status = 0; ++ u8 current_mac[ETH_ALEN]; ++ u32 pmac_id = adapter->pmac_id; + +- status = be_cmd_pmac_del(adapter, adapter->if_handle, adapter->pmac_id); ++ if (!is_valid_ether_addr(addr->sa_data)) ++ return -EADDRNOTAVAIL; ++ ++ status = be_cmd_mac_addr_query(adapter, current_mac, ++ MAC_ADDRESS_TYPE_NETWORK, false, ++ adapter->if_handle); + if (status) +- return status; ++ goto err; ++ ++ if (!memcmp(addr->sa_data, current_mac, ETH_ALEN)) ++ goto done; + + status = be_cmd_pmac_add(adapter, (u8 *)addr->sa_data, +- adapter->if_handle, &adapter->pmac_id); +- if (!status) ++ adapter->if_handle, &adapter->pmac_id, 0); ++ ++ if (!status) { ++ status = be_cmd_pmac_del(adapter, adapter->if_handle, ++ pmac_id, 0); + memcpy(netdev->dev_addr, addr->sa_data, netdev->addr_len); ++ goto done; ++ } + +- return status; ++err: ++ if (status == MCC_STATUS_UNAUTHORIZED_REQUEST) ++ return -EPERM; ++ else ++ dev_err(&adapter->pdev->dev, "MAC %pM set Failed\n", ++ addr->sa_data); ++done: ++ return status; ++} ++ ++static void populate_be2_stats(struct be_adapter *adapter) ++{ ++ ++ struct be_drv_stats *drvs = &adapter->drv_stats; ++ struct be_pmem_stats *pmem_sts = be_pmem_stats_from_cmd(adapter); ++ struct be_port_rxf_stats_v0 *port_stats = ++ be_port_rxf_stats_from_cmd(adapter); ++ struct be_rxf_stats_v0 *rxf_stats = ++ be_rxf_stats_from_cmd(adapter); ++ ++ drvs->rx_pause_frames = port_stats->rx_pause_frames; ++ drvs->rx_crc_errors = port_stats->rx_crc_errors; ++ drvs->rx_control_frames = port_stats->rx_control_frames; ++ drvs->rx_in_range_errors = port_stats->rx_in_range_errors; ++ drvs->rx_frame_too_long = port_stats->rx_frame_too_long; ++ drvs->rx_dropped_runt = port_stats->rx_dropped_runt; ++ drvs->rx_ip_checksum_errs = port_stats->rx_ip_checksum_errs; ++ drvs->rx_tcp_checksum_errs = port_stats->rx_tcp_checksum_errs; ++ drvs->rx_udp_checksum_errs = port_stats->rx_udp_checksum_errs; ++ drvs->rxpp_fifo_overflow_drop = port_stats->rx_fifo_overflow; ++ drvs->rx_dropped_tcp_length = port_stats->rx_dropped_tcp_length; ++ drvs->rx_dropped_too_small = port_stats->rx_dropped_too_small; ++ drvs->rx_dropped_too_short = port_stats->rx_dropped_too_short; ++ drvs->rx_out_range_errors = port_stats->rx_out_range_errors; ++ drvs->rx_input_fifo_overflow_drop = ++ port_stats->rx_input_fifo_overflow; ++ drvs->rx_dropped_header_too_small = ++ port_stats->rx_dropped_header_too_small; ++ drvs->rx_address_match_errors = ++ port_stats->rx_address_match_errors; ++ drvs->rx_alignment_symbol_errors = ++ port_stats->rx_alignment_symbol_errors; ++ ++ drvs->tx_pauseframes = port_stats->tx_pauseframes; ++ drvs->tx_controlframes = port_stats->tx_controlframes; ++ ++ if (adapter->port_num) ++ drvs->jabber_events = ++ rxf_stats->port1_jabber_events; ++ else ++ drvs->jabber_events = ++ rxf_stats->port0_jabber_events; ++ drvs->rx_drops_no_pbuf = rxf_stats->rx_drops_no_pbuf; ++ drvs->rx_drops_no_txpb = rxf_stats->rx_drops_no_txpb; ++ drvs->rx_drops_no_erx_descr = rxf_stats->rx_drops_no_erx_descr; ++ drvs->rx_drops_invalid_ring = rxf_stats->rx_drops_invalid_ring; ++ drvs->forwarded_packets = rxf_stats->forwarded_packets; ++ drvs->rx_drops_mtu = rxf_stats->rx_drops_mtu; ++ drvs->rx_drops_no_tpre_descr = ++ rxf_stats->rx_drops_no_tpre_descr; ++ drvs->rx_drops_too_many_frags = ++ rxf_stats->rx_drops_too_many_frags; ++ adapter->drv_stats.eth_red_drops = pmem_sts->eth_red_drops; ++} ++ ++static void populate_be3_stats(struct be_adapter *adapter) ++{ ++ struct be_drv_stats *drvs = &adapter->drv_stats; ++ struct be_pmem_stats *pmem_sts = be_pmem_stats_from_cmd(adapter); ++ ++ struct be_rxf_stats_v1 *rxf_stats = ++ be_rxf_stats_from_cmd(adapter); ++ struct be_port_rxf_stats_v1 *port_stats = ++ be_port_rxf_stats_from_cmd(adapter); ++ ++ drvs->pmem_fifo_overflow_drop = port_stats->pmem_fifo_overflow_drop; ++ drvs->rx_priority_pause_frames = port_stats->rx_priority_pause_frames; ++ drvs->rx_pause_frames = port_stats->rx_pause_frames; ++ drvs->rx_crc_errors = port_stats->rx_crc_errors; ++ drvs->rx_control_frames = port_stats->rx_control_frames; ++ drvs->rx_in_range_errors = port_stats->rx_in_range_errors; ++ drvs->rx_frame_too_long = port_stats->rx_frame_too_long; ++ drvs->rx_dropped_runt = port_stats->rx_dropped_runt; ++ drvs->rx_ip_checksum_errs = port_stats->rx_ip_checksum_errs; ++ drvs->rx_tcp_checksum_errs = port_stats->rx_tcp_checksum_errs; ++ drvs->rx_udp_checksum_errs = port_stats->rx_udp_checksum_errs; ++ drvs->rx_dropped_tcp_length = port_stats->rx_dropped_tcp_length; ++ drvs->rx_dropped_too_small = port_stats->rx_dropped_too_small; ++ drvs->rx_dropped_too_short = port_stats->rx_dropped_too_short; ++ drvs->rx_out_range_errors = port_stats->rx_out_range_errors; ++ drvs->rx_dropped_header_too_small = ++ port_stats->rx_dropped_header_too_small; ++ drvs->rx_input_fifo_overflow_drop = ++ port_stats->rx_input_fifo_overflow_drop; ++ drvs->rx_address_match_errors = ++ port_stats->rx_address_match_errors; ++ drvs->rx_alignment_symbol_errors = ++ port_stats->rx_alignment_symbol_errors; ++ drvs->rxpp_fifo_overflow_drop = ++ port_stats->rxpp_fifo_overflow_drop; ++ drvs->tx_pauseframes = port_stats->tx_pauseframes; ++ drvs->tx_controlframes = port_stats->tx_controlframes; ++ drvs->jabber_events = port_stats->jabber_events; ++ drvs->rx_drops_no_pbuf = rxf_stats->rx_drops_no_pbuf; ++ drvs->rx_drops_no_txpb = rxf_stats->rx_drops_no_txpb; ++ drvs->rx_drops_no_erx_descr = rxf_stats->rx_drops_no_erx_descr; ++ drvs->rx_drops_invalid_ring = rxf_stats->rx_drops_invalid_ring; ++ drvs->forwarded_packets = rxf_stats->forwarded_packets; ++ drvs->rx_drops_mtu = rxf_stats->rx_drops_mtu; ++ drvs->rx_drops_no_tpre_descr = ++ rxf_stats->rx_drops_no_tpre_descr; ++ drvs->rx_drops_too_many_frags = ++ rxf_stats->rx_drops_too_many_frags; ++ adapter->drv_stats.eth_red_drops = pmem_sts->eth_red_drops; ++} ++ ++ ++static void accumulate_16bit_val(u32 *acc, u16 val) ++{ ++#define lo(x) (x & 0xFFFF) ++#define hi(x) (x & 0xFFFF0000) ++ bool wrapped = val < lo(*acc); ++ u32 newacc = hi(*acc) + val; ++ ++ if (wrapped) ++ newacc += 65536; ++ ACCESS_ONCE_RW(*acc) = newacc; ++} ++ ++void be_parse_stats(struct be_adapter *adapter) ++{ ++ struct be_erx_stats_v1 *erx = be_erx_stats_from_cmd(adapter); ++ struct be_rx_obj *rxo; ++ int i; ++ ++ if (adapter->generation == BE_GEN3) { ++ populate_be3_stats(adapter); ++ } else { ++ populate_be2_stats(adapter); ++ } ++ ++ /* as erx_v1 is longer than v0, ok to use v1 defn for v0 access */ ++ for_all_rx_queues(adapter, rxo, i) { ++ /* below erx HW counter can actually wrap around after ++ * 65535. Driver accumulates a 32-bit value ++ */ ++ accumulate_16bit_val(&rx_stats(rxo)->rx_drops_no_frags, ++ (u16)erx->rx_drops_no_fragments[rxo->q.id]); ++ } + } + + void netdev_stats_update(struct be_adapter *adapter) + { +- struct be_hw_stats *hw_stats = hw_stats_from_cmd(adapter->stats.cmd.va); +- struct be_rxf_stats *rxf_stats = &hw_stats->rxf; +- struct be_port_rxf_stats *port_stats = +- &rxf_stats->port[adapter->port_num]; +- struct net_device_stats *dev_stats = &adapter->stats.net_stats; +- struct be_erx_stats *erx_stats = &hw_stats->erx; ++ struct be_drv_stats *drvs = &adapter->drv_stats; ++ struct net_device_stats *dev_stats = &adapter->net_stats; ++ struct be_rx_obj *rxo; ++ struct be_tx_obj *txo; ++ unsigned long pkts = 0, bytes = 0, mcast = 0, drops = 0; ++ int i; + +- dev_stats->rx_packets = drvr_stats(adapter)->be_rx_pkts; +- dev_stats->tx_packets = drvr_stats(adapter)->be_tx_pkts; +- dev_stats->rx_bytes = drvr_stats(adapter)->be_rx_bytes; +- dev_stats->tx_bytes = drvr_stats(adapter)->be_tx_bytes; ++ for_all_rx_queues(adapter, rxo, i) { ++ pkts += rx_stats(rxo)->rx_pkts; ++ bytes += rx_stats(rxo)->rx_bytes; ++ mcast += rx_stats(rxo)->rx_mcast_pkts; ++ drops += rx_stats(rxo)->rx_drops_no_frags; ++ } ++ dev_stats->rx_packets = pkts; ++ dev_stats->rx_bytes = bytes; ++ dev_stats->multicast = mcast; ++ dev_stats->rx_dropped = drops; ++ ++ pkts = bytes = 0; ++ for_all_tx_queues(adapter, txo, i) { ++ pkts += tx_stats(txo)->be_tx_pkts; ++ bytes += tx_stats(txo)->be_tx_bytes; ++ } ++ dev_stats->tx_packets = pkts; ++ dev_stats->tx_bytes = bytes; + + /* bad pkts received */ +- dev_stats->rx_errors = port_stats->rx_crc_errors + +- port_stats->rx_alignment_symbol_errors + +- port_stats->rx_in_range_errors + +- port_stats->rx_out_range_errors + +- port_stats->rx_frame_too_long + +- port_stats->rx_dropped_too_small + +- port_stats->rx_dropped_too_short + +- port_stats->rx_dropped_header_too_small + +- port_stats->rx_dropped_tcp_length + +- port_stats->rx_dropped_runt + +- port_stats->rx_tcp_checksum_errs + +- port_stats->rx_ip_checksum_errs + +- port_stats->rx_udp_checksum_errs; +- +- /* no space in linux buffers: best possible approximation */ +- dev_stats->rx_dropped = erx_stats->rx_drops_no_fragments[0]; ++ dev_stats->rx_errors = drvs->rx_crc_errors + ++ drvs->rx_alignment_symbol_errors + ++ drvs->rx_in_range_errors + ++ drvs->rx_out_range_errors + ++ drvs->rx_frame_too_long + ++ drvs->rx_dropped_too_small + ++ drvs->rx_dropped_too_short + ++ drvs->rx_dropped_header_too_small + ++ drvs->rx_dropped_tcp_length + ++ drvs->rx_dropped_runt + ++ drvs->rx_tcp_checksum_errs + ++ drvs->rx_ip_checksum_errs + ++ drvs->rx_udp_checksum_errs; + + /* detailed rx errors */ +- dev_stats->rx_length_errors = port_stats->rx_in_range_errors + +- port_stats->rx_out_range_errors + +- port_stats->rx_frame_too_long; ++ dev_stats->rx_length_errors = drvs->rx_in_range_errors + ++ drvs->rx_out_range_errors + ++ drvs->rx_frame_too_long; + +- /* receive ring buffer overflow */ +- dev_stats->rx_over_errors = 0; +- +- dev_stats->rx_crc_errors = port_stats->rx_crc_errors; ++ dev_stats->rx_crc_errors = drvs->rx_crc_errors; + + /* frame alignment errors */ +- dev_stats->rx_frame_errors = port_stats->rx_alignment_symbol_errors; ++ dev_stats->rx_frame_errors = drvs->rx_alignment_symbol_errors; + + /* receiver fifo overrun */ + /* drops_no_pbuf is no per i/f, it's per BE card */ +- dev_stats->rx_fifo_errors = port_stats->rx_fifo_overflow + +- port_stats->rx_input_fifo_overflow + +- rxf_stats->rx_drops_no_pbuf; +- /* receiver missed packetd */ +- dev_stats->rx_missed_errors = 0; +- +- /* packet transmit problems */ +- dev_stats->tx_errors = 0; +- +- /* no space available in linux */ +- dev_stats->tx_dropped = 0; +- +- dev_stats->multicast = port_stats->rx_multicast_frames; +- dev_stats->collisions = 0; +- +- /* detailed tx_errors */ +- dev_stats->tx_aborted_errors = 0; +- dev_stats->tx_carrier_errors = 0; +- dev_stats->tx_fifo_errors = 0; +- dev_stats->tx_heartbeat_errors = 0; +- dev_stats->tx_window_errors = 0; ++ dev_stats->rx_fifo_errors = drvs->rxpp_fifo_overflow_drop + ++ drvs->rx_input_fifo_overflow_drop + ++ drvs->rx_drops_no_pbuf; + } + +-void be_link_status_update(struct be_adapter *adapter, bool link_up) ++void be_link_status_update(struct be_adapter *adapter, int link_status) + { + struct net_device *netdev = adapter->netdev; + + /* If link came up or went down */ +- if (adapter->link_up != link_up) { +- if (link_up) { ++ if (adapter->link_status != link_status) { ++ adapter->link_speed = -1; ++ if (link_status == LINK_UP) { + netif_start_queue(netdev); + netif_carrier_on(netdev); + printk(KERN_INFO "%s: Link up\n", netdev->name); +@@ -222,15 +491,15 @@ void be_link_status_update(struct be_adapter *adapter, bool link_up) + netif_carrier_off(netdev); + printk(KERN_INFO "%s: Link down\n", netdev->name); + } +- adapter->link_up = link_up; ++ adapter->link_status = link_status; + } + } + + /* Update the EQ delay n BE based on the RX frags consumed / sec */ +-static void be_rx_eqd_update(struct be_adapter *adapter) ++static void be_rx_eqd_update(struct be_adapter *adapter, struct be_rx_obj *rxo) + { +- struct be_eq_obj *rx_eq = &adapter->rx_eq; +- struct be_drvr_stats *stats = &adapter->stats.drvr_stats; ++ struct be_eq_obj *rx_eq = &rxo->rx_eq; ++ struct be_rx_stats *stats = &rxo->stats; + ulong now = jiffies; + u32 eqd; + +@@ -247,19 +516,17 @@ static void be_rx_eqd_update(struct be_adapter *adapter) + if ((now - stats->rx_fps_jiffies) < HZ) + return; + +- stats->be_rx_fps = (stats->be_rx_frags - stats->be_prev_rx_frags) / ++ stats->rx_fps = (stats->rx_frags - stats->prev_rx_frags) / + ((now - stats->rx_fps_jiffies) / HZ); + + stats->rx_fps_jiffies = now; +- stats->be_prev_rx_frags = stats->be_rx_frags; +- eqd = stats->be_rx_fps / 110000; ++ stats->prev_rx_frags = stats->rx_frags; ++ eqd = stats->rx_fps / 110000; + eqd = eqd << 3; + if (eqd > rx_eq->max_eqd) + eqd = rx_eq->max_eqd; + if (eqd < rx_eq->min_eqd) + eqd = rx_eq->min_eqd; +- if (eqd < 10) +- eqd = 0; + if (eqd != rx_eq->cur_eqd) + be_cmd_modify_eqd(adapter, rx_eq->q.id, eqd); + +@@ -270,7 +537,7 @@ static struct net_device_stats *be_get_stats(struct net_device *dev) + { + struct be_adapter *adapter = netdev_priv(dev); + +- return &adapter->stats.net_stats; ++ return &adapter->net_stats; + } + + static u32 be_calc_rate(u64 bytes, unsigned long ticks) +@@ -284,9 +551,9 @@ static u32 be_calc_rate(u64 bytes, unsigned long ticks) + return rate; + } + +-static void be_tx_rate_update(struct be_adapter *adapter) ++static void be_tx_rate_update(struct be_tx_obj *txo) + { +- struct be_drvr_stats *stats = drvr_stats(adapter); ++ struct be_tx_stats *stats = tx_stats(txo); + ulong now = jiffies; + + /* Wrapped around? */ +@@ -305,10 +572,11 @@ static void be_tx_rate_update(struct be_adapter *adapter) + } + } + +-static void be_tx_stats_update(struct be_adapter *adapter, ++static void be_tx_stats_update(struct be_tx_obj *txo, + u32 wrb_cnt, u32 copied, u32 gso_segs, bool stopped) + { +- struct be_drvr_stats *stats = drvr_stats(adapter); ++ struct be_tx_stats *stats = tx_stats(txo); ++ + stats->be_tx_reqs++; + stats->be_tx_wrbs += wrb_cnt; + stats->be_tx_bytes += copied; +@@ -318,7 +586,8 @@ static void be_tx_stats_update(struct be_adapter *adapter, + } + + /* Determine number of WRB entries needed to xmit data in an skb */ +-static u32 wrb_cnt_for_skb(struct sk_buff *skb, bool *dummy) ++static u32 wrb_cnt_for_skb(struct be_adapter *adapter, struct sk_buff *skb, ++ bool *dummy) + { + int cnt = (skb->len > skb->data_len); + +@@ -326,12 +595,13 @@ static u32 wrb_cnt_for_skb(struct sk_buff *skb, bool *dummy) + + /* to account for hdr wrb */ + cnt++; +- if (cnt & 1) { ++ if (lancer_chip(adapter) || !(cnt & 1)) { ++ *dummy = false; ++ } else { + /* add a dummy to make it an even num */ + cnt++; + *dummy = true; +- } else +- *dummy = false; ++ } + BUG_ON(cnt > BE_MAX_TX_FRAG_COUNT); + return cnt; + } +@@ -343,17 +613,31 @@ static inline void wrb_fill(struct be_eth_wrb *wrb, u64 addr, int len) + wrb->frag_len = len & ETH_WRB_FRAG_LEN_MASK; + } + +-static void wrb_fill_hdr(struct be_eth_hdr_wrb *hdr, struct sk_buff *skb, +- bool vlan, u32 wrb_cnt, u32 len) ++static void wrb_fill_hdr(struct be_adapter *adapter, struct be_eth_hdr_wrb *hdr, ++ struct sk_buff *skb, u32 wrb_cnt, u32 len) + { ++ u16 vlan_tag = 0; ++ + memset(hdr, 0, sizeof(*hdr)); + + AMAP_SET_BITS(struct amap_eth_hdr_wrb, crc, hdr, 1); + +- if (skb_shinfo(skb)->gso_segs > 1 && skb_shinfo(skb)->gso_size) { ++ if (skb_is_gso(skb)) { + AMAP_SET_BITS(struct amap_eth_hdr_wrb, lso, hdr, 1); + AMAP_SET_BITS(struct amap_eth_hdr_wrb, lso_mss, + hdr, skb_shinfo(skb)->gso_size); ++ if (skb_is_gso_v6(skb) && !lancer_chip(adapter)) ++ AMAP_SET_BITS(struct amap_eth_hdr_wrb, lso6, hdr, 1); ++ ++ if (lancer_A0_chip(adapter)) { ++ AMAP_SET_BITS(struct amap_eth_hdr_wrb, ipcs, hdr, 1); ++ if (is_tcp_pkt(skb)) ++ AMAP_SET_BITS(struct amap_eth_hdr_wrb, ++ tcpcs, hdr, 1); ++ else if (is_udp_pkt(skb)) ++ AMAP_SET_BITS(struct amap_eth_hdr_wrb, ++ udpcs, hdr, 1); ++ } + } else if (skb->ip_summed == CHECKSUM_PARTIAL) { + if (is_tcp_pkt(skb)) + AMAP_SET_BITS(struct amap_eth_hdr_wrb, tcpcs, hdr, 1); +@@ -361,10 +645,10 @@ static void wrb_fill_hdr(struct be_eth_hdr_wrb *hdr, struct sk_buff *skb, + AMAP_SET_BITS(struct amap_eth_hdr_wrb, udpcs, hdr, 1); + } + +- if (vlan && vlan_tx_tag_present(skb)) { ++ if (adapter->vlan_grp && vlan_tx_tag_present(skb)) { + AMAP_SET_BITS(struct amap_eth_hdr_wrb, vlan, hdr, 1); +- AMAP_SET_BITS(struct amap_eth_hdr_wrb, vlan_tag, +- hdr, vlan_tx_tag_get(skb)); ++ vlan_tag = be_get_tx_vlan_tag(adapter, skb); ++ AMAP_SET_BITS(struct amap_eth_hdr_wrb, vlan_tag, hdr, vlan_tag); + } + + AMAP_SET_BITS(struct amap_eth_hdr_wrb, event, hdr, 1); +@@ -374,14 +658,13 @@ static void wrb_fill_hdr(struct be_eth_hdr_wrb *hdr, struct sk_buff *skb, + } + + +-static int make_tx_wrbs(struct be_adapter *adapter, ++static int make_tx_wrbs(struct be_adapter *adapter, struct be_queue_info *txq, + struct sk_buff *skb, u32 wrb_cnt, bool dummy_wrb) + { +- u64 busaddr; +- u32 i, copied = 0; ++ dma_addr_t busaddr; ++ int i, copied = 0; + struct pci_dev *pdev = adapter->pdev; + struct sk_buff *first_skb = skb; +- struct be_queue_info *txq = &adapter->tx_obj.q; + struct be_eth_wrb *wrb; + struct be_eth_hdr_wrb *hdr; + +@@ -389,15 +672,11 @@ static int make_tx_wrbs(struct be_adapter *adapter, + atomic_add(wrb_cnt, &txq->used); + queue_head_inc(txq); + +- if (skb_dma_map(&pdev->dev, skb, DMA_TO_DEVICE)) { +- dev_err(&pdev->dev, "TX DMA mapping failed\n"); +- return 0; +- } +- + if (skb->len > skb->data_len) { +- int len = skb->len - skb->data_len; ++ int len = skb_headlen(skb); ++ busaddr = pci_map_single(pdev, skb->data, len, ++ PCI_DMA_TODEVICE); + wrb = queue_head_node(txq); +- busaddr = skb_shinfo(skb)->dma_head; + wrb_fill(wrb, busaddr, len); + be_dws_cpu_to_le(wrb, sizeof(*wrb)); + queue_head_inc(txq); +@@ -407,8 +686,9 @@ static int make_tx_wrbs(struct be_adapter *adapter, + for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { + struct skb_frag_struct *frag = + &skb_shinfo(skb)->frags[i]; +- +- busaddr = skb_shinfo(skb)->dma_maps[i]; ++ busaddr = pci_map_page(pdev, frag->page, ++ frag->page_offset, ++ frag->size, PCI_DMA_TODEVICE); + wrb = queue_head_node(txq); + wrb_fill(wrb, busaddr, frag->size); + be_dws_cpu_to_le(wrb, sizeof(*wrb)); +@@ -423,8 +703,7 @@ static int make_tx_wrbs(struct be_adapter *adapter, + queue_head_inc(txq); + } + +- wrb_fill_hdr(hdr, first_skb, adapter->vlan_grp ? true : false, +- wrb_cnt, copied); ++ wrb_fill_hdr(adapter, hdr, first_skb, wrb_cnt, copied); + be_dws_cpu_to_le(hdr, sizeof(*hdr)); + + return copied; +@@ -434,19 +713,70 @@ static netdev_tx_t be_xmit(struct sk_buff *skb, + struct net_device *netdev) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_tx_obj *tx_obj = &adapter->tx_obj; +- struct be_queue_info *txq = &tx_obj->q; ++ struct be_tx_obj *txo = &adapter->tx_obj[skb_get_queue_mapping(skb)]; ++ struct be_queue_info *txq = &txo->q; + u32 wrb_cnt = 0, copied = 0; + u32 start = txq->head; + bool dummy_wrb, stopped = false; + +- wrb_cnt = wrb_cnt_for_skb(skb, &dummy_wrb); ++ if (unlikely((skb_shinfo(skb)->gso_segs > 1) && ++ skb_shinfo(skb)->gso_size && is_ipv6_ext_hdr(skb))) { ++ tx_stats(txo)->be_ipv6_ext_hdr_tx_drop++; ++ goto tx_drop; ++ } + +- copied = make_tx_wrbs(adapter, skb, wrb_cnt, dummy_wrb); ++ /* If the skb is a large pkt forwarded to this interface ++ * after being LRO'd on another interface, drop the pkt. ++ * HW cannot handle such pkts. LRO must be disabled when ++ * using the server as a router. ++ */ ++ if (!skb_is_gso(skb)) { ++ int eth_hdr_len = ntohs(skb->protocol) == ETH_P_8021Q ? ++ VLAN_ETH_HLEN : ETH_HLEN; ++ ++ if ((skb->len - eth_hdr_len) > adapter->netdev->mtu) ++ goto tx_drop; ++ } ++ ++ /* The ASIC is calculating checksum for Vlan tagged pkts ++ * though CSO is disabled. ++ * To work around this, insert the Vlan tag in the driver ++ * and donot set the vlan bit, cso bit in the Tx WRB. ++ */ ++ if (unlikely(vlan_tx_tag_present(skb) && ++ ((skb->ip_summed != CHECKSUM_PARTIAL) || (skb->len <= 60)))) { ++ /* Bug 28694: Don't embed the host VLAN tag in SKB ++ * when UMC mode enabled on that interface ++ */ ++ if (!(adapter->function_mode & UMC_ENABLED)) { ++ skb = skb_share_check(skb, GFP_ATOMIC); ++ if (unlikely(!skb)) ++ goto tx_drop; ++ ++ skb = be_vlan_put_tag(skb, ++ be_get_tx_vlan_tag(adapter, skb)); ++ if (unlikely(!skb)) ++ goto tx_drop; ++ ++ be_reset_skb_tx_vlan(skb); ++ } ++ } ++ ++ /* Bug 12422: the stack can send us skbs with length more than 65535 ++ * BE cannot handle such requests. Hack the extra data out and drop it. ++ */ ++ if (skb->len > 65535) { ++ int err = __pskb_trim(skb, 65535); ++ BUG_ON(err); ++ } ++ ++ wrb_cnt = wrb_cnt_for_skb(adapter, skb, &dummy_wrb); ++ ++ copied = make_tx_wrbs(adapter, txq, skb, wrb_cnt, dummy_wrb); + if (copied) { + /* record the sent skb in the sent_skb table */ +- BUG_ON(tx_obj->sent_skb_list[start]); +- tx_obj->sent_skb_list[start] = skb; ++ BUG_ON(txo->sent_skb_list[start]); ++ txo->sent_skb_list[start] = skb; + + /* Ensure txq has space for the next skb; Else stop the queue + * *BEFORE* ringing the tx doorbell, so that we serialze the +@@ -454,16 +784,21 @@ static netdev_tx_t be_xmit(struct sk_buff *skb, + */ + if ((BE_MAX_TX_FRAG_COUNT + atomic_read(&txq->used)) >= + txq->len) { +- netif_stop_queue(netdev); ++ netif_stop_subqueue(netdev, skb_get_queue_mapping(skb)); + stopped = true; + } + + be_txq_notify(adapter, txq->id, wrb_cnt); + +- be_tx_stats_update(adapter, wrb_cnt, copied, ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32) ++ netdev->trans_start = jiffies; ++#endif ++ ++ be_tx_stats_update(txo, wrb_cnt, copied, + skb_shinfo(skb)->gso_segs, stopped); + } else { + txq->head = start; ++tx_drop: + dev_kfree_skb_any(skb); + } + return NETDEV_TX_OK; +@@ -473,10 +808,12 @@ static int be_change_mtu(struct net_device *netdev, int new_mtu) + { + struct be_adapter *adapter = netdev_priv(netdev); + if (new_mtu < BE_MIN_MTU || +- new_mtu > BE_MAX_JUMBO_FRAME_SIZE) { ++ new_mtu > (BE_MAX_JUMBO_FRAME_SIZE - ++ (ETH_HLEN + ETH_FCS_LEN))) { + dev_info(&adapter->pdev->dev, + "MTU must be between %d and %d bytes\n", +- BE_MIN_MTU, BE_MAX_JUMBO_FRAME_SIZE); ++ BE_MIN_MTU, ++ (BE_MAX_JUMBO_FRAME_SIZE - (ETH_HLEN + ETH_FCS_LEN))); + return -EINVAL; + } + dev_info(&adapter->pdev->dev, "MTU changed from %d to %d bytes\n", +@@ -486,17 +823,19 @@ static int be_change_mtu(struct net_device *netdev, int new_mtu) + } + + /* +- * if there are BE_NUM_VLANS_SUPPORTED or lesser number of VLANS configured, +- * program them in BE. If more than BE_NUM_VLANS_SUPPORTED are configured, +- * set the BE in promiscuous VLAN mode. ++ * A max of 64 (BE_NUM_VLANS_SUPPORTED) vlans can be configured in BE. ++ * If the user configures more, place BE in vlan promiscuous mode. + */ +-static int be_vid_config(struct be_adapter *adapter) ++static int be_vid_config(struct be_adapter *adapter, bool vf, u32 vf_num) + { + u16 vtag[BE_NUM_VLANS_SUPPORTED]; + u16 ntags = 0, i; +- int status; ++ int status = 0; + +- if (adapter->num_vlans <= BE_NUM_VLANS_SUPPORTED) { ++ /* No need to change the VLAN state if the I/F is in promiscous */ ++ if (adapter->promiscuous) ++ return 0; ++ if (adapter->vlans_added <= adapter->max_vlans) { + /* Construct VLAN Table to give to HW */ + for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { + if (adapter->vlan_tag[i]) { +@@ -504,47 +843,46 @@ static int be_vid_config(struct be_adapter *adapter) + ntags++; + } + } +- status = be_cmd_vlan_config(adapter, adapter->if_handle, +- vtag, ntags, 1, 0); ++ /* Send command only if there is something to be programmed */ ++ if (ntags) ++ status = be_cmd_vlan_config(adapter, adapter->if_handle, ++ vtag, ntags, 1, 0); + } else { + status = be_cmd_vlan_config(adapter, adapter->if_handle, +- NULL, 0, 1, 1); ++ NULL, 0, 1, 1); + } ++ + return status; + } + + static void be_vlan_register(struct net_device *netdev, struct vlan_group *grp) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_eq_obj *rx_eq = &adapter->rx_eq; +- struct be_eq_obj *tx_eq = &adapter->tx_eq; + +- be_eq_notify(adapter, rx_eq->q.id, false, false, 0); +- be_eq_notify(adapter, tx_eq->q.id, false, false, 0); + adapter->vlan_grp = grp; +- be_eq_notify(adapter, rx_eq->q.id, true, false, 0); +- be_eq_notify(adapter, tx_eq->q.id, true, false, 0); + } + + static void be_vlan_add_vid(struct net_device *netdev, u16 vid) + { + struct be_adapter *adapter = netdev_priv(netdev); + +- adapter->num_vlans++; ++ adapter->vlans_added++; ++ + adapter->vlan_tag[vid] = 1; +- +- be_vid_config(adapter); ++ if (adapter->vlans_added <= (adapter->max_vlans + 1)) ++ be_vid_config(adapter, false, 0); + } + + static void be_vlan_rem_vid(struct net_device *netdev, u16 vid) + { + struct be_adapter *adapter = netdev_priv(netdev); + +- adapter->num_vlans--; +- adapter->vlan_tag[vid] = 0; +- ++ adapter->vlans_added--; + vlan_group_set_device(adapter->vlan_grp, vid, NULL); +- be_vid_config(adapter); ++ ++ adapter->vlan_tag[vid] = 0; ++ if (adapter->vlans_added <= adapter->max_vlans) ++ be_vid_config(adapter, false, 0); + } + + static void be_set_multicast_list(struct net_device *netdev) +@@ -552,7 +890,7 @@ static void be_set_multicast_list(struct net_device *netdev) + struct be_adapter *adapter = netdev_priv(netdev); + + if (netdev->flags & IFF_PROMISC) { +- be_cmd_promiscuous_config(adapter, adapter->port_num, 1); ++ be_cmd_rx_filter(adapter, IFF_PROMISC, ON); + adapter->promiscuous = true; + goto done; + } +@@ -560,81 +898,244 @@ static void be_set_multicast_list(struct net_device *netdev) + /* BE was previously in promiscous mode; disable it */ + if (adapter->promiscuous) { + adapter->promiscuous = false; +- be_cmd_promiscuous_config(adapter, adapter->port_num, 0); ++ be_cmd_rx_filter(adapter, IFF_PROMISC, OFF); ++ ++ if (adapter->vlans_added) ++ be_vid_config(adapter, false, 0); + } + +- if (netdev->flags & IFF_ALLMULTI) { +- be_cmd_multicast_set(adapter, adapter->if_handle, NULL, 0); ++ /* Enable multicast promisc if num configured exceeds what we support */ ++ if (netdev->flags & IFF_ALLMULTI || ++ netdev_mc_count(netdev) > BE_MAX_MC) { ++ be_cmd_rx_filter(adapter, IFF_ALLMULTI, ON); + goto done; + } + +- be_cmd_multicast_set(adapter, adapter->if_handle, netdev->mc_list, +- netdev->mc_count); ++ be_cmd_rx_filter(adapter, IFF_MULTICAST, ON); + done: + return; + } + +-static void be_rx_rate_update(struct be_adapter *adapter) ++#ifdef HAVE_SRIOV_CONFIG ++static int be_set_vf_mac(struct net_device *netdev, int vf, u8 *mac) + { +- struct be_drvr_stats *stats = drvr_stats(adapter); ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int status; ++ ++ if (adapter->num_vfs == 0) ++ return -EPERM; ++ ++ if (!is_valid_ether_addr(mac) || (vf >= adapter->num_vfs)) ++ return -EINVAL; ++ ++ if (adapter->vf_cfg[vf].vf_pmac_id != BE_INVALID_PMAC_ID) ++ status = be_cmd_pmac_del(adapter, ++ adapter->vf_cfg[vf].vf_if_handle, ++ adapter->vf_cfg[vf].vf_pmac_id, vf + 1); ++ ++ status = be_cmd_pmac_add(adapter, mac, ++ adapter->vf_cfg[vf].vf_if_handle, ++ &adapter->vf_cfg[vf].vf_pmac_id, vf + 1); ++ ++ if (status) ++ dev_err(&adapter->pdev->dev, "MAC %pM set on VF %d Failed\n", ++ mac, vf); ++ else ++ memcpy(adapter->vf_cfg[vf].vf_mac_addr, mac, ETH_ALEN); ++ ++ return status; ++} ++ ++static int be_get_vf_config(struct net_device *netdev, int vf, ++ struct ifla_vf_info *vi) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ ++ if (adapter->num_vfs == 0) ++ return -EPERM; ++ ++ if (vf >= adapter->num_vfs) ++ return -EINVAL; ++ ++ vi->vf = vf; ++ vi->tx_rate = adapter->vf_cfg[vf].vf_tx_rate; ++ vi->vlan = adapter->vf_cfg[vf].vf_vlan_tag & VLAN_VID_MASK; ++ vi->qos = adapter->vf_cfg[vf].vf_vlan_tag >> VLAN_PRIO_SHIFT; ++ memcpy(&vi->mac, adapter->vf_cfg[vf].vf_mac_addr, ETH_ALEN); ++ ++ return 0; ++} ++ ++/* ++ * Entry point to configure vlan behavior for a VF. ++ * 1. By default a VF is vlan Challenged. ++ * 2. It may or may not have Transparent Tagging enabled. ++ * 3. Vlan privilege for a VF can be toggled using special VID 4095. ++ * 4. When removing the Vlan privilege for a VF there is no need set default vid ++ * 5. Transparent Tagging configured for a VF resets its Vlan privilege ++ * 6. To disable the current Transparet Tagging for a VF: ++ * 6a. run the last iproute command with vlan set to 0. ++ * 6b. programing the default vid will disable Transparent Tagging in ARM/ASIC ++ */ ++static int be_set_vf_vlan(struct net_device *netdev, ++ int vf, u16 vlan, u8 qos) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int status = 0; ++ u32 en = 0; ++ ++ if (adapter->num_vfs == 0) ++ return -EPERM; ++ ++ if ((vf >= adapter->num_vfs) || (vlan > 4095) || (qos > 7)) ++ return -EINVAL; ++ ++ status = be_cmd_get_fn_privileges(adapter, &en, vf + 1); ++ if (status) ++ goto sts; ++ ++ if (vlan == 4095) { ++ if (en & BE_PRIV_FILTMGMT) { ++ /* Knock off filtering privileges */ ++ en &= ~BE_PRIV_FILTMGMT; ++ } else { ++ en |= BE_PRIV_FILTMGMT; ++ /* Transparent Tagging is currently enabled, Reset it */ ++ if (adapter->vf_cfg[vf].vf_vlan_tag) { ++ adapter->vf_cfg[vf].vf_vlan_tag = 0; ++ vlan = adapter->vf_cfg[vf].vf_def_vid; ++ be_cmd_set_hsw_config(adapter, vlan, vf + 1, ++ adapter->vf_cfg[vf].vf_if_handle); ++ } ++ } ++ ++ adapter->vf_cfg[vf].vf_vlan_tag = 0; ++ status = be_cmd_set_fn_privileges(adapter, en, NULL, vf + 1); ++ ++ goto sts; ++ } ++ ++ if (vlan || qos) { ++ if (en & BE_PRIV_FILTMGMT) { ++ /* Check privilege and reset it to default */ ++ en &= ~BE_PRIV_FILTMGMT; ++ be_cmd_set_fn_privileges(adapter, en, NULL, vf + 1); ++ } ++ ++ vlan |= qos << VLAN_PRIO_SHIFT; ++ if (adapter->vf_cfg[vf].vf_vlan_tag != vlan) { ++ /* If this is new value, program it. Else skip. */ ++ adapter->vf_cfg[vf].vf_vlan_tag = vlan; ++ ++ status = be_cmd_set_hsw_config(adapter, vlan, ++ vf + 1, adapter->vf_cfg[vf].vf_if_handle); ++ } ++ ++ } else { ++ /* Reset Transparent Vlan Tagging. */ ++ adapter->vf_cfg[vf].vf_vlan_tag = 0; ++ vlan = adapter->vf_cfg[vf].vf_def_vid; ++ status = be_cmd_set_hsw_config(adapter, vlan, vf + 1, ++ adapter->vf_cfg[vf].vf_if_handle); ++ } ++ ++sts: ++ if (status) ++ dev_info(&adapter->pdev->dev, ++ "VLAN %d config on VF %d failed\n", vlan, vf); ++ return status; ++} ++ ++static int be_set_vf_tx_rate(struct net_device *netdev, ++ int vf, int rate) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ int status = 0; ++ ++ if (adapter->num_vfs == 0) ++ return -EPERM; ++ ++ if ((vf >= adapter->num_vfs) || (rate < 0)) ++ return -EINVAL; ++ ++ if (rate > 10000) ++ rate = 10000; ++ ++ adapter->vf_cfg[vf].vf_tx_rate = rate; ++ status = be_cmd_set_qos(adapter, rate / 10, vf + 1); ++ ++ if (status) ++ dev_info(&adapter->pdev->dev, ++ "tx rate %d on VF %d failed\n", rate, vf); ++ return status; ++} ++#endif /* HAVE_SRIOV_CONFIG */ ++ ++static void be_rx_rate_update(struct be_rx_obj *rxo) ++{ ++ struct be_rx_stats *stats = &rxo->stats; + ulong now = jiffies; + + /* Wrapped around */ +- if (time_before(now, stats->be_rx_jiffies)) { +- stats->be_rx_jiffies = now; ++ if (time_before(now, stats->rx_jiffies)) { ++ stats->rx_jiffies = now; + return; + } + + /* Update the rate once in two seconds */ +- if ((now - stats->be_rx_jiffies) < 2 * HZ) ++ if ((now - stats->rx_jiffies) < 2 * HZ) + return; + +- stats->be_rx_rate = be_calc_rate(stats->be_rx_bytes +- - stats->be_rx_bytes_prev, +- now - stats->be_rx_jiffies); +- stats->be_rx_jiffies = now; +- stats->be_rx_bytes_prev = stats->be_rx_bytes; ++ stats->rx_rate = be_calc_rate(stats->rx_bytes - stats->rx_bytes_prev, ++ now - stats->rx_jiffies); ++ stats->rx_jiffies = now; ++ stats->rx_bytes_prev = stats->rx_bytes; + } + +-static void be_rx_stats_update(struct be_adapter *adapter, +- u32 pktsize, u16 numfrags) ++static void be_rx_stats_update(struct be_rx_obj *rxo, ++ struct be_rx_compl_info *rxcp) + { +- struct be_drvr_stats *stats = drvr_stats(adapter); ++ struct be_rx_stats *stats = &rxo->stats; + +- stats->be_rx_compl++; +- stats->be_rx_frags += numfrags; +- stats->be_rx_bytes += pktsize; +- stats->be_rx_pkts++; ++ stats->rx_compl++; ++ stats->rx_frags += rxcp->num_rcvd; ++ stats->rx_bytes += rxcp->pkt_size; ++ stats->rx_pkts++; ++ if (rxcp->pkt_type == BE_MULTICAST_PACKET) ++ stats->rx_mcast_pkts++; ++ if (rxcp->err) ++ stats->rxcp_err++; + } + +-static inline bool do_pkt_csum(struct be_eth_rx_compl *rxcp, bool cso) ++static inline bool csum_passed(struct be_rx_compl_info *rxcp) + { +- u8 l4_cksm, ip_version, ipcksm, tcpf = 0, udpf = 0, ipv6_chk; +- +- l4_cksm = AMAP_GET_BITS(struct amap_eth_rx_compl, l4_cksm, rxcp); +- ipcksm = AMAP_GET_BITS(struct amap_eth_rx_compl, ipcksm, rxcp); +- ip_version = AMAP_GET_BITS(struct amap_eth_rx_compl, ip_version, rxcp); +- if (ip_version) { +- tcpf = AMAP_GET_BITS(struct amap_eth_rx_compl, tcpf, rxcp); +- udpf = AMAP_GET_BITS(struct amap_eth_rx_compl, udpf, rxcp); +- } +- ipv6_chk = (ip_version && (tcpf || udpf)); +- +- return ((l4_cksm && ipv6_chk && ipcksm) && cso) ? false : true; ++ /* L4 checksum is not reliable for non TCP/UDP packets. ++ * Also ignore ipcksm for ipv6 pkts */ ++ return (rxcp->tcpf || rxcp->udpf) && rxcp->l4_csum && ++ (rxcp->ip_csum || rxcp->ipv6); + } + + static struct be_rx_page_info * +-get_rx_page_info(struct be_adapter *adapter, u16 frag_idx) ++get_rx_page_info(struct be_adapter *adapter, struct be_rx_obj *rxo, ++ u16 frag_idx) + { + struct be_rx_page_info *rx_page_info; +- struct be_queue_info *rxq = &adapter->rx_obj.q; ++ struct be_queue_info *rxq = &rxo->q; + +- rx_page_info = &adapter->rx_obj.page_info_tbl[frag_idx]; +- BUG_ON(!rx_page_info->page); ++ rx_page_info = &rxo->page_info_tbl[frag_idx]; ++ if (!rx_page_info->page) { ++ printk(KERN_EMERG "curr_idx=%d prev_dix=%d rxq->head=%d\n", ++ frag_idx, rxo->prev_frag_idx, rxq->head); ++ BUG_ON(!rx_page_info->page); ++ } + +- if (rx_page_info->last_page_user) ++ if (rx_page_info->last_page_user) { + pci_unmap_page(adapter->pdev, pci_unmap_addr(rx_page_info, bus), + adapter->big_page_size, PCI_DMA_FROMDEVICE); ++ rx_page_info->last_page_user = false; ++ } ++ ++ rxo->prev_frag_idx = frag_idx; + + atomic_dec(&rxq->used); + return rx_page_info; +@@ -642,20 +1143,26 @@ get_rx_page_info(struct be_adapter *adapter, u16 frag_idx) + + /* Throwaway the data in the Rx completion */ + static void be_rx_compl_discard(struct be_adapter *adapter, +- struct be_eth_rx_compl *rxcp) ++ struct be_rx_obj *rxo, ++ struct be_rx_compl_info *rxcp) + { +- struct be_queue_info *rxq = &adapter->rx_obj.q; ++ struct be_queue_info *rxq = &rxo->q; + struct be_rx_page_info *page_info; +- u16 rxq_idx, i, num_rcvd; ++ u16 i; ++ bool oob_error; ++ u16 num_rcvd = rxcp->num_rcvd; + +- rxq_idx = AMAP_GET_BITS(struct amap_eth_rx_compl, fragndx, rxcp); +- num_rcvd = AMAP_GET_BITS(struct amap_eth_rx_compl, numfrags, rxcp); ++ oob_error = lancer_A0_chip(adapter) && rxcp->err; ++ ++ /* In case of OOB error num_rcvd will be 1 more than actual */ ++ if (oob_error && num_rcvd) ++ num_rcvd -= 1; + + for (i = 0; i < num_rcvd; i++) { +- page_info = get_rx_page_info(adapter, rxq_idx); ++ page_info = get_rx_page_info(adapter, rxo, rxcp->rxq_idx); + put_page(page_info->page); + memset(page_info, 0, sizeof(*page_info)); +- index_inc(&rxq_idx, rxq->len); ++ index_inc(&rxcp->rxq_idx, rxq->len); + } + } + +@@ -663,29 +1170,24 @@ static void be_rx_compl_discard(struct be_adapter *adapter, + * skb_fill_rx_data forms a complete skb for an ether frame + * indicated by rxcp. + */ +-static void skb_fill_rx_data(struct be_adapter *adapter, +- struct sk_buff *skb, struct be_eth_rx_compl *rxcp) ++static void skb_fill_rx_data(struct be_adapter *adapter, struct be_rx_obj *rxo, ++ struct sk_buff *skb, struct be_rx_compl_info *rxcp) + { +- struct be_queue_info *rxq = &adapter->rx_obj.q; ++ struct be_queue_info *rxq = &rxo->q; + struct be_rx_page_info *page_info; +- u16 rxq_idx, i, num_rcvd, j; +- u32 pktsize, hdr_len, curr_frag_len, size; ++ u16 i, j; ++ u16 hdr_len, curr_frag_len, remaining; + u8 *start; + +- rxq_idx = AMAP_GET_BITS(struct amap_eth_rx_compl, fragndx, rxcp); +- pktsize = AMAP_GET_BITS(struct amap_eth_rx_compl, pktsize, rxcp); +- num_rcvd = AMAP_GET_BITS(struct amap_eth_rx_compl, numfrags, rxcp); +- +- page_info = get_rx_page_info(adapter, rxq_idx); +- ++ page_info = get_rx_page_info(adapter, rxo, rxcp->rxq_idx); + start = page_address(page_info->page) + page_info->page_offset; + prefetch(start); + + /* Copy data in the first descriptor of this completion */ +- curr_frag_len = min(pktsize, rx_frag_size); ++ curr_frag_len = min(rxcp->pkt_size, rx_frag_size); + + /* Copy the header portion into skb_data */ +- hdr_len = min((u32)BE_HDR_LEN, curr_frag_len); ++ hdr_len = min(BE_HDR_LEN, curr_frag_len); + memcpy(skb->data, start, hdr_len); + skb->len = curr_frag_len; + if (curr_frag_len <= BE_HDR_LEN) { /* tiny packet */ +@@ -702,21 +1204,19 @@ static void skb_fill_rx_data(struct be_adapter *adapter, + skb->data_len = curr_frag_len - hdr_len; + skb->tail += hdr_len; + } +- memset(page_info, 0, sizeof(*page_info)); ++ page_info->page = NULL; + +- if (pktsize <= rx_frag_size) { +- BUG_ON(num_rcvd != 1); +- goto done; ++ if (rxcp->pkt_size <= rx_frag_size) { ++ BUG_ON(rxcp->num_rcvd != 1); ++ return; + } + + /* More frags present for this completion */ +- size = pktsize; +- for (i = 1, j = 0; i < num_rcvd; i++) { +- size -= curr_frag_len; +- index_inc(&rxq_idx, rxq->len); +- page_info = get_rx_page_info(adapter, rxq_idx); +- +- curr_frag_len = min(size, rx_frag_size); ++ index_inc(&rxcp->rxq_idx, rxq->len); ++ remaining = rxcp->pkt_size - curr_frag_len; ++ for (i = 1, j = 0; i < rxcp->num_rcvd; i++) { ++ page_info = get_rx_page_info(adapter, rxo, rxcp->rxq_idx); ++ curr_frag_len = min(remaining, rx_frag_size); + + /* Coalesce all frags from the same physical page in one slot */ + if (page_info->page_offset == 0) { +@@ -735,99 +1235,122 @@ static void skb_fill_rx_data(struct be_adapter *adapter, + skb->len += curr_frag_len; + skb->data_len += curr_frag_len; + +- memset(page_info, 0, sizeof(*page_info)); ++ remaining -= curr_frag_len; ++ index_inc(&rxcp->rxq_idx, rxq->len); ++ page_info->page = NULL; + } + BUG_ON(j > MAX_SKB_FRAGS); +- +-done: +- be_rx_stats_update(adapter, pktsize, num_rcvd); +- return; + } + +-/* Process the RX completion indicated by rxcp when GRO is disabled */ ++/* Process the RX completion indicated by rxcp when LRO is disabled */ + static void be_rx_compl_process(struct be_adapter *adapter, +- struct be_eth_rx_compl *rxcp) ++ struct be_rx_obj *rxo, ++ struct be_rx_compl_info *rxcp) + { + struct sk_buff *skb; +- u32 vlanf, vid; +- u8 vtm; + +- vlanf = AMAP_GET_BITS(struct amap_eth_rx_compl, vtp, rxcp); +- vtm = AMAP_GET_BITS(struct amap_eth_rx_compl, vtm, rxcp); +- +- /* vlanf could be wrongly set in some cards. +- * ignore if vtm is not set */ +- if ((adapter->cap == 0x400) && !vtm) +- vlanf = 0; +- +- skb = netdev_alloc_skb(adapter->netdev, BE_HDR_LEN + NET_IP_ALIGN); +- if (!skb) { ++ skb = netdev_alloc_skb_ip_align(adapter->netdev, BE_HDR_LEN); ++ if (unlikely(!skb)) { + if (net_ratelimit()) + dev_warn(&adapter->pdev->dev, "skb alloc failed\n"); +- be_rx_compl_discard(adapter, rxcp); ++ be_rx_compl_discard(adapter, rxo, rxcp); + return; + } + +- skb_reserve(skb, NET_IP_ALIGN); ++ skb_fill_rx_data(adapter, rxo, skb, rxcp); + +- skb_fill_rx_data(adapter, skb, rxcp); +- +- if (do_pkt_csum(rxcp, adapter->rx_csum)) +- skb->ip_summed = CHECKSUM_NONE; +- else ++ if (likely(adapter->rx_csum && csum_passed(rxcp))) + skb->ip_summed = CHECKSUM_UNNECESSARY; ++ else ++ skb->ip_summed = CHECKSUM_NONE; + + skb->truesize = skb->len + sizeof(struct sk_buff); ++ if (unlikely(rxcp->vlanf) && ++ unlikely(!vlan_configured(adapter))) { ++ __vlan_put_tag(skb, rxcp->vlan_tag); ++ } + skb->protocol = eth_type_trans(skb, adapter->netdev); + skb->dev = adapter->netdev; + +- if (vlanf) { +- if (!adapter->vlan_grp || adapter->num_vlans == 0) { +- kfree_skb(skb); +- return; +- } +- vid = AMAP_GET_BITS(struct amap_eth_rx_compl, vlan_tag, rxcp); +- vid = be16_to_cpu(vid); +- vlan_hwaccel_receive_skb(skb, adapter->vlan_grp, vid); +- } else { ++ if (unlikely(rxcp->vlanf) && ++ vlan_configured(adapter)) ++ vlan_hwaccel_receive_skb(skb, adapter->vlan_grp, ++ rxcp->vlan_tag); ++ else + netif_receive_skb(skb); ++ ++ return; ++} ++ ++/* Process the RX completion indicated by rxcp when LRO is enabled */ ++static void be_rx_compl_process_lro(struct be_adapter *adapter, ++ struct be_rx_obj *rxo, ++ struct be_rx_compl_info *rxcp) ++{ ++ struct be_rx_page_info *page_info; ++ struct skb_frag_struct rx_frags[BE_MAX_FRAGS_PER_FRAME]; ++ struct be_queue_info *rxq = &rxo->q; ++ u16 remaining, curr_frag_len; ++ u16 i, j; ++ ++ remaining = rxcp->pkt_size; ++ for (i = 0, j = -1; i < rxcp->num_rcvd; i++) { ++ page_info = get_rx_page_info(adapter, rxo, rxcp->rxq_idx); ++ ++ curr_frag_len = min(remaining, rx_frag_size); ++ ++ /* Coalesce all frags from the same physical page in one slot */ ++ if (i == 0 || page_info->page_offset == 0) { ++ /* First frag or Fresh page */ ++ j++; ++ rx_frags[j].page = page_info->page; ++ rx_frags[j].page_offset = page_info->page_offset; ++ rx_frags[j].size = 0; ++ } else { ++ put_page(page_info->page); ++ } ++ rx_frags[j].size += curr_frag_len; ++ ++ remaining -= curr_frag_len; ++ index_inc(&rxcp->rxq_idx, rxq->len); ++ memset(page_info, 0, sizeof(*page_info)); ++ } ++ BUG_ON(j > MAX_SKB_FRAGS); ++ ++ if (likely(!rxcp->vlanf)) { ++ lro_receive_frags(&rxo->lro_mgr, rx_frags, rxcp->pkt_size, ++ rxcp->pkt_size, NULL, 0); ++ } else { ++ lro_vlan_hwaccel_receive_frags(&rxo->lro_mgr, rx_frags, ++ rxcp->pkt_size, rxcp->pkt_size, adapter->vlan_grp, ++ rxcp->vlan_tag, NULL, 0); + } + + return; + } + + /* Process the RX completion indicated by rxcp when GRO is enabled */ +-static void be_rx_compl_process_gro(struct be_adapter *adapter, +- struct be_eth_rx_compl *rxcp) ++void be_rx_compl_process_gro(struct be_adapter *adapter, ++ struct be_rx_obj *rxo, ++ struct be_rx_compl_info *rxcp) + { ++#ifdef NETIF_F_GRO + struct be_rx_page_info *page_info; + struct sk_buff *skb = NULL; +- struct be_queue_info *rxq = &adapter->rx_obj.q; +- struct be_eq_obj *eq_obj = &adapter->rx_eq; +- u32 num_rcvd, pkt_size, remaining, vlanf, curr_frag_len; +- u16 i, rxq_idx = 0, vid, j; +- u8 vtm; +- +- num_rcvd = AMAP_GET_BITS(struct amap_eth_rx_compl, numfrags, rxcp); +- pkt_size = AMAP_GET_BITS(struct amap_eth_rx_compl, pktsize, rxcp); +- vlanf = AMAP_GET_BITS(struct amap_eth_rx_compl, vtp, rxcp); +- rxq_idx = AMAP_GET_BITS(struct amap_eth_rx_compl, fragndx, rxcp); +- vtm = AMAP_GET_BITS(struct amap_eth_rx_compl, vtm, rxcp); +- +- /* vlanf could be wrongly set in some cards. +- * ignore if vtm is not set */ +- if ((adapter->cap == 0x400) && !vtm) +- vlanf = 0; ++ struct be_queue_info *rxq = &rxo->q; ++ struct be_eq_obj *eq_obj = &rxo->rx_eq; ++ u16 remaining, curr_frag_len; ++ u16 i, j; + + skb = napi_get_frags(&eq_obj->napi); + if (!skb) { +- be_rx_compl_discard(adapter, rxcp); ++ be_rx_compl_discard(adapter, rxo, rxcp); + return; + } + +- remaining = pkt_size; +- for (i = 0, j = -1; i < num_rcvd; i++) { +- page_info = get_rx_page_info(adapter, rxq_idx); ++ remaining = rxcp->pkt_size; ++ for (i = 0, j = -1; i < rxcp->num_rcvd; i++) { ++ page_info = get_rx_page_info(adapter, rxo, rxcp->rxq_idx); + + curr_frag_len = min(remaining, rx_frag_size); + +@@ -845,55 +1368,129 @@ static void be_rx_compl_process_gro(struct be_adapter *adapter, + skb_shinfo(skb)->frags[j].size += curr_frag_len; + + remaining -= curr_frag_len; +- index_inc(&rxq_idx, rxq->len); ++ index_inc(&rxcp->rxq_idx, rxq->len); + memset(page_info, 0, sizeof(*page_info)); + } + BUG_ON(j > MAX_SKB_FRAGS); + + skb_shinfo(skb)->nr_frags = j + 1; +- skb->len = pkt_size; +- skb->data_len = pkt_size; +- skb->truesize += pkt_size; ++ skb->len = rxcp->pkt_size; ++ skb->data_len = rxcp->pkt_size; ++ skb->truesize += rxcp->pkt_size; + skb->ip_summed = CHECKSUM_UNNECESSARY; + +- if (likely(!vlanf)) { ++ if (likely(!rxcp->vlanf)) + napi_gro_frags(&eq_obj->napi); +- } else { +- vid = AMAP_GET_BITS(struct amap_eth_rx_compl, vlan_tag, rxcp); +- vid = be16_to_cpu(vid); ++ else ++ vlan_gro_frags(&eq_obj->napi, ++ adapter->vlan_grp, rxcp->vlan_tag); ++#endif + +- if (!adapter->vlan_grp || adapter->num_vlans == 0) +- return; +- +- vlan_gro_frags(&eq_obj->napi, adapter->vlan_grp, vid); +- } +- +- be_rx_stats_update(adapter, pkt_size, num_rcvd); + return; + } + +-static struct be_eth_rx_compl *be_rx_compl_get(struct be_adapter *adapter) ++static void be_parse_rx_compl_v1(struct be_adapter *adapter, ++ struct be_eth_rx_compl *compl, ++ struct be_rx_compl_info *rxcp) + { +- struct be_eth_rx_compl *rxcp = queue_tail_node(&adapter->rx_obj.cq); ++ rxcp->pkt_size = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, pktsize, compl); ++ rxcp->vlanf = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, vtp, compl); ++ rxcp->err = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, err, compl); ++ rxcp->tcpf = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, tcpf, compl); ++ rxcp->udpf = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, udpf, compl); ++ rxcp->ip_csum = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, ipcksm, compl); ++ rxcp->l4_csum = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, l4_cksm, compl); ++ rxcp->ipv6 = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, ip_version, compl); ++ rxcp->rxq_idx = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, fragndx, compl); ++ rxcp->num_rcvd = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, numfrags, compl); ++ rxcp->pkt_type = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v1, cast_enc, compl); ++ if (rxcp->vlanf) { ++ rxcp->vtm = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, vtm, ++ compl); ++ rxcp->vlan_tag = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, ++ vlan_tag, compl); ++ } ++ rxcp->port = AMAP_GET_BITS(struct amap_eth_rx_compl_v1, port, compl); ++} + +- if (rxcp->dw[offsetof(struct amap_eth_rx_compl, valid) / 32] == 0) ++static void be_parse_rx_compl_v0(struct be_adapter *adapter, ++ struct be_eth_rx_compl *compl, ++ struct be_rx_compl_info *rxcp) ++{ ++ rxcp->pkt_size = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, pktsize, compl); ++ rxcp->vlanf = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, vtp, compl); ++ rxcp->err = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, err, compl); ++ rxcp->tcpf = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, tcpf, compl); ++ rxcp->udpf = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, udpf, compl); ++ rxcp->ip_csum = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, ipcksm, compl); ++ rxcp->l4_csum = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, l4_cksm, compl); ++ rxcp->ipv6 = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, ip_version, compl); ++ rxcp->rxq_idx = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, fragndx, compl); ++ rxcp->num_rcvd = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, numfrags, compl); ++ rxcp->pkt_type = ++ AMAP_GET_BITS(struct amap_eth_rx_compl_v0, cast_enc, compl); ++ if (rxcp->vlanf) { ++ rxcp->vtm = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, vtm, ++ compl); ++ rxcp->vlan_tag = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, ++ vlan_tag, compl); ++ } ++ rxcp->port = AMAP_GET_BITS(struct amap_eth_rx_compl_v0, port, compl); ++} ++ ++static struct be_rx_compl_info *be_rx_compl_get(struct be_rx_obj *rxo) ++{ ++ struct be_eth_rx_compl *compl = queue_tail_node(&rxo->cq); ++ struct be_rx_compl_info *rxcp = &rxo->rxcp; ++ struct be_adapter *adapter = rxo->adapter; ++ ++ /* For checking the valid bit it is Ok to use either definition as the ++ * valid bit is at the same position in both v0 and v1 Rx compl */ ++ if (compl->dw[offsetof(struct amap_eth_rx_compl_v1, valid) / 32] == 0) + return NULL; + +- be_dws_le_to_cpu(rxcp, sizeof(*rxcp)); ++ rmb(); ++ be_dws_le_to_cpu(compl, sizeof(*compl)); + +- queue_tail_inc(&adapter->rx_obj.cq); ++ if (adapter->be3_native) ++ be_parse_rx_compl_v1(adapter, compl, rxcp); ++ else ++ be_parse_rx_compl_v0(adapter, compl, rxcp); ++ ++ if (rxcp->vlanf) { ++ /* vlanf could be wrongly set in some cards. ++ * ignore if vtm is not set */ ++ if ((adapter->function_mode & FLEX10_MODE) && !rxcp->vtm) ++ rxcp->vlanf = 0; ++ ++ if (!lancer_chip(adapter)) ++ rxcp->vlan_tag = swab16(rxcp->vlan_tag); ++ ++ if ((adapter->pvid == (rxcp->vlan_tag & VLAN_VID_MASK)) && ++ !adapter->vlan_tag[rxcp->vlan_tag]) ++ rxcp->vlanf = 0; ++ } ++ ++ /* As the compl has been parsed, reset it; we wont touch it again */ ++ compl->dw[offsetof(struct amap_eth_rx_compl_v1, valid) / 32] = 0; ++ ++ queue_tail_inc(&rxo->cq); + return rxcp; + } + +-/* To reset the valid bit, we need to reset the whole word as +- * when walking the queue the valid entries are little-endian +- * and invalid entries are host endian +- */ +-static inline void be_rx_compl_reset(struct be_eth_rx_compl *rxcp) +-{ +- rxcp->dw[offsetof(struct amap_eth_rx_compl, valid) / 32] = 0; +-} +- + static inline struct page *be_alloc_pages(u32 size) + { + gfp_t alloc_flags = GFP_ATOMIC; +@@ -907,11 +1504,12 @@ static inline struct page *be_alloc_pages(u32 size) + * Allocate a page, split it to fragments of size rx_frag_size and post as + * receive buffers to BE + */ +-static void be_post_rx_frags(struct be_adapter *adapter) ++static void be_post_rx_frags(struct be_rx_obj *rxo) + { +- struct be_rx_page_info *page_info_tbl = adapter->rx_obj.page_info_tbl; +- struct be_rx_page_info *page_info = NULL; +- struct be_queue_info *rxq = &adapter->rx_obj.q; ++ struct be_adapter *adapter = rxo->adapter; ++ struct be_rx_page_info *page_info_tbl = rxo->page_info_tbl; ++ struct be_rx_page_info *page_info = NULL, *prev_page_info = NULL; ++ struct be_queue_info *rxq = &rxo->q; + struct page *pagep = NULL; + struct be_eth_rx_d *rxd; + u64 page_dmaaddr = 0, frag_dmaaddr; +@@ -922,7 +1520,7 @@ static void be_post_rx_frags(struct be_adapter *adapter) + if (!pagep) { + pagep = be_alloc_pages(adapter->big_page_size); + if (unlikely(!pagep)) { +- drvr_stats(adapter)->be_ethrx_post_fail++; ++ rxo->stats.rx_post_fail++; + break; + } + page_dmaaddr = pci_map_page(adapter->pdev, pagep, 0, +@@ -941,7 +1539,6 @@ static void be_post_rx_frags(struct be_adapter *adapter) + rxd = queue_head_node(rxq); + rxd->fragpa_lo = cpu_to_le32(frag_dmaaddr & 0xFFFFFFFF); + rxd->fragpa_hi = cpu_to_le32(upper_32_bits(frag_dmaaddr)); +- queue_head_inc(rxq); + + /* Any space left in the current big page for another frag? */ + if ((page_offset + rx_frag_size + rx_frag_size) > +@@ -949,17 +1546,24 @@ static void be_post_rx_frags(struct be_adapter *adapter) + pagep = NULL; + page_info->last_page_user = true; + } ++ ++ prev_page_info = page_info; ++ queue_head_inc(rxq); + page_info = &page_info_tbl[rxq->head]; + } + if (pagep) +- page_info->last_page_user = true; ++ prev_page_info->last_page_user = true; + ++ /* Ensure that posting buffers is the last thing done by this ++ * routine to avoid racing between rx bottom-half and ++ * be_worker (process) contexts. ++ */ + if (posted) { + atomic_add(posted, &rxq->used); + be_rxq_notify(adapter, rxq->id, posted); + } else if (atomic_read(&rxq->used) == 0) { + /* Let be_worker replenish when memory is available */ +- adapter->rx_post_starved = true; ++ rxo->rx_post_starved = true; + } + + return; +@@ -972,6 +1576,7 @@ static struct be_eth_tx_compl *be_tx_compl_get(struct be_queue_info *tx_cq) + if (txcp->dw[offsetof(struct amap_eth_tx_compl, valid) / 32] == 0) + return NULL; + ++ rmb(); + be_dws_le_to_cpu(txcp, sizeof(*txcp)); + + txcp->dw[offsetof(struct amap_eth_tx_compl, valid) / 32] = 0; +@@ -980,11 +1585,14 @@ static struct be_eth_tx_compl *be_tx_compl_get(struct be_queue_info *tx_cq) + return txcp; + } + +-static void be_tx_compl_process(struct be_adapter *adapter, u16 last_index) ++static u16 be_tx_compl_process(struct be_adapter *adapter, ++ struct be_tx_obj *txo, u16 last_index) + { +- struct be_queue_info *txq = &adapter->tx_obj.q; +- struct sk_buff **sent_skbs = adapter->tx_obj.sent_skb_list; ++ struct be_queue_info *txq = &txo->q; ++ struct be_eth_wrb *wrb; ++ struct sk_buff **sent_skbs = txo->sent_skb_list; + struct sk_buff *sent_skb; ++ u64 busaddr; + u16 cur_index, num_wrbs = 0; + + cur_index = txq->tail; +@@ -992,15 +1600,31 @@ static void be_tx_compl_process(struct be_adapter *adapter, u16 last_index) + BUG_ON(!sent_skb); + sent_skbs[cur_index] = NULL; + +- do { ++ wrb = queue_tail_node(txq); ++ be_dws_le_to_cpu(wrb, sizeof(*wrb)); ++ busaddr = ((u64)wrb->frag_pa_hi << 32) | (u64)wrb->frag_pa_lo; ++ if (busaddr != 0) { ++ pci_unmap_single(adapter->pdev, busaddr, ++ wrb->frag_len, PCI_DMA_TODEVICE); ++ } ++ num_wrbs++; ++ queue_tail_inc(txq); ++ ++ while (cur_index != last_index) { + cur_index = txq->tail; ++ wrb = queue_tail_node(txq); ++ be_dws_le_to_cpu(wrb, sizeof(*wrb)); ++ busaddr = ((u64)wrb->frag_pa_hi << 32) | (u64)wrb->frag_pa_lo; ++ if (busaddr != 0) { ++ pci_unmap_page(adapter->pdev, busaddr, ++ wrb->frag_len, PCI_DMA_TODEVICE); ++ } + num_wrbs++; + queue_tail_inc(txq); +- } while (cur_index != last_index); ++ } + +- atomic_sub(num_wrbs, &txq->used); +- skb_dma_unmap(&adapter->pdev->dev, sent_skb, DMA_TO_DEVICE); + kfree_skb(sent_skb); ++ return num_wrbs; + } + + static inline struct be_eq_entry *event_get(struct be_eq_obj *eq_obj) +@@ -1010,13 +1634,15 @@ static inline struct be_eq_entry *event_get(struct be_eq_obj *eq_obj) + if (!eqe->evt) + return NULL; + ++ rmb(); + eqe->evt = le32_to_cpu(eqe->evt); + queue_tail_inc(&eq_obj->q); + return eqe; + } + + static int event_handle(struct be_adapter *adapter, +- struct be_eq_obj *eq_obj) ++ struct be_eq_obj *eq_obj, ++ bool rearm) + { + struct be_eq_entry *eqe; + u16 num = 0; +@@ -1029,7 +1655,10 @@ static int event_handle(struct be_adapter *adapter, + /* Deal with any spurious interrupts that come + * without events + */ +- be_eq_notify(adapter, eq_obj->q.id, true, true, num); ++ if (!num) ++ rearm = true; ++ ++ be_eq_notify(adapter, eq_obj->q.id, rearm, true, num); + if (num) + napi_schedule(&eq_obj->napi); + +@@ -1053,49 +1682,55 @@ static void be_eq_clean(struct be_adapter *adapter, + be_eq_notify(adapter, eq_obj->q.id, false, true, num); + } + +-static void be_rx_q_clean(struct be_adapter *adapter) ++static void be_rx_q_clean(struct be_adapter *adapter, struct be_rx_obj *rxo) + { + struct be_rx_page_info *page_info; +- struct be_queue_info *rxq = &adapter->rx_obj.q; +- struct be_queue_info *rx_cq = &adapter->rx_obj.cq; +- struct be_eth_rx_compl *rxcp; ++ struct be_queue_info *rxq = &rxo->q; ++ struct be_queue_info *rx_cq = &rxo->cq; ++ struct be_rx_compl_info *rxcp; + u16 tail; + + /* First cleanup pending rx completions */ +- while ((rxcp = be_rx_compl_get(adapter)) != NULL) { +- be_rx_compl_discard(adapter, rxcp); +- be_rx_compl_reset(rxcp); ++ while ((rxcp = be_rx_compl_get(rxo)) != NULL) { ++ be_rx_compl_discard(adapter, rxo, rxcp); + be_cq_notify(adapter, rx_cq->id, true, 1); + } + + /* Then free posted rx buffer that were not used */ + tail = (rxq->head + rxq->len - atomic_read(&rxq->used)) % rxq->len; + for (; atomic_read(&rxq->used) > 0; index_inc(&tail, rxq->len)) { +- page_info = get_rx_page_info(adapter, tail); ++ page_info = get_rx_page_info(adapter, rxo, tail); + put_page(page_info->page); + memset(page_info, 0, sizeof(*page_info)); + } + BUG_ON(atomic_read(&rxq->used)); ++ rxq->tail = rxq->head = 0; + } + +-static void be_tx_compl_clean(struct be_adapter *adapter) ++static void be_tx_compl_clean(struct be_adapter *adapter, ++ struct be_tx_obj *txo) + { +- struct be_queue_info *tx_cq = &adapter->tx_obj.cq; +- struct be_queue_info *txq = &adapter->tx_obj.q; ++ struct be_queue_info *tx_cq = &txo->cq; ++ struct be_queue_info *txq = &txo->q; + struct be_eth_tx_compl *txcp; +- u16 end_idx, cmpl = 0, timeo = 0; ++ u16 end_idx, cmpl = 0, timeo = 0, num_wrbs = 0; ++ struct sk_buff **sent_skbs = txo->sent_skb_list; ++ struct sk_buff *sent_skb; ++ bool dummy_wrb; + + /* Wait for a max of 200ms for all the tx-completions to arrive. */ + do { + while ((txcp = be_tx_compl_get(tx_cq))) { + end_idx = AMAP_GET_BITS(struct amap_eth_tx_compl, + wrb_index, txcp); +- be_tx_compl_process(adapter, end_idx); ++ num_wrbs += be_tx_compl_process(adapter, txo, end_idx); + cmpl++; + } + if (cmpl) { + be_cq_notify(adapter, tx_cq->id, false, cmpl); ++ atomic_sub(num_wrbs, &txq->used); + cmpl = 0; ++ num_wrbs = 0; + } + + if (atomic_read(&txq->used) == 0 || ++timeo > 200) +@@ -1107,6 +1742,17 @@ static void be_tx_compl_clean(struct be_adapter *adapter) + if (atomic_read(&txq->used)) + dev_err(&adapter->pdev->dev, "%d pending tx-completions\n", + atomic_read(&txq->used)); ++ ++ /* free posted tx for which compls will never arrive */ ++ while (atomic_read(&txq->used)) { ++ sent_skb = sent_skbs[txq->tail]; ++ end_idx = txq->tail; ++ index_adv(&end_idx, ++ wrb_cnt_for_skb(adapter, sent_skb, &dummy_wrb) - 1, ++ txq->len); ++ num_wrbs = be_tx_compl_process(adapter, txo, end_idx); ++ atomic_sub(num_wrbs, &txq->used); ++ } + } + + static void be_mcc_queues_destroy(struct be_adapter *adapter) +@@ -1145,8 +1791,9 @@ static int be_mcc_queues_create(struct be_adapter *adapter) + goto mcc_cq_destroy; + + /* Ask BE to create MCC queue */ +- if (be_cmd_mccq_create(adapter, q, cq)) ++ if (be_cmd_mccq_create(adapter, q, cq)) { + goto mcc_q_free; ++ } + + return 0; + +@@ -1163,16 +1810,20 @@ err: + static void be_tx_queues_destroy(struct be_adapter *adapter) + { + struct be_queue_info *q; ++ struct be_tx_obj *txo; ++ u8 i; + +- q = &adapter->tx_obj.q; +- if (q->created) +- be_cmd_q_destroy(adapter, q, QTYPE_TXQ); +- be_queue_free(adapter, q); ++ for_all_tx_queues(adapter, txo, i) { ++ q = &txo->q; ++ if (q->created) ++ be_cmd_q_destroy(adapter, q, QTYPE_TXQ); ++ be_queue_free(adapter, q); + +- q = &adapter->tx_obj.cq; +- if (q->created) +- be_cmd_q_destroy(adapter, q, QTYPE_CQ); +- be_queue_free(adapter, q); ++ q = &txo->cq; ++ if (q->created) ++ be_cmd_q_destroy(adapter, q, QTYPE_CQ); ++ be_queue_free(adapter, q); ++ } + + /* Clear any residual events */ + be_eq_clean(adapter, &adapter->tx_eq); +@@ -1183,168 +1834,210 @@ static void be_tx_queues_destroy(struct be_adapter *adapter) + be_queue_free(adapter, q); + } + ++/* One TX event queue is shared by all TX compl qs */ + static int be_tx_queues_create(struct be_adapter *adapter) + { + struct be_queue_info *eq, *q, *cq; ++ struct be_tx_obj *txo; ++ u8 i, tc_id; + + adapter->tx_eq.max_eqd = 0; + adapter->tx_eq.min_eqd = 0; + adapter->tx_eq.cur_eqd = 96; + adapter->tx_eq.enable_aic = false; +- /* Alloc Tx Event queue */ ++ + eq = &adapter->tx_eq.q; +- if (be_queue_alloc(adapter, eq, EVNT_Q_LEN, sizeof(struct be_eq_entry))) ++ if (be_queue_alloc(adapter, eq, EVNT_Q_LEN, ++ sizeof(struct be_eq_entry))) + return -1; + +- /* Ask BE to create Tx Event queue */ + if (be_cmd_eq_create(adapter, eq, adapter->tx_eq.cur_eqd)) +- goto tx_eq_free; +- /* Alloc TX eth compl queue */ +- cq = &adapter->tx_obj.cq; +- if (be_queue_alloc(adapter, cq, TX_CQ_LEN, ++ goto err; ++ adapter->tx_eq.eq_idx = adapter->eq_next_idx++; ++ ++ for_all_tx_queues(adapter, txo, i) { ++ cq = &txo->cq; ++ if (be_queue_alloc(adapter, cq, TX_CQ_LEN, + sizeof(struct be_eth_tx_compl))) +- goto tx_eq_destroy; ++ goto err; + +- /* Ask BE to create Tx eth compl queue */ +- if (be_cmd_cq_create(adapter, cq, eq, false, false, 3)) +- goto tx_cq_free; ++ if (be_cmd_cq_create(adapter, cq, eq, false, false, 3)) ++ goto err; + +- /* Alloc TX eth queue */ +- q = &adapter->tx_obj.q; +- if (be_queue_alloc(adapter, q, TX_Q_LEN, sizeof(struct be_eth_wrb))) +- goto tx_cq_destroy; ++ q = &txo->q; ++ if (be_queue_alloc(adapter, q, TX_Q_LEN, ++ sizeof(struct be_eth_wrb))) ++ goto err; + +- /* Ask BE to create Tx eth queue */ +- if (be_cmd_txq_create(adapter, q, cq)) +- goto tx_q_free; ++ if (be_cmd_txq_create(adapter, q, cq, &tc_id)) ++ goto err; ++ ++ if (adapter->flags & BE_FLAGS_DCBX) ++ adapter->tc_txq_map[tc_id] = i; ++ } + return 0; + +-tx_q_free: +- be_queue_free(adapter, q); +-tx_cq_destroy: +- be_cmd_q_destroy(adapter, cq, QTYPE_CQ); +-tx_cq_free: +- be_queue_free(adapter, cq); +-tx_eq_destroy: +- be_cmd_q_destroy(adapter, eq, QTYPE_EQ); +-tx_eq_free: +- be_queue_free(adapter, eq); ++err: ++ be_tx_queues_destroy(adapter); + return -1; + } + + static void be_rx_queues_destroy(struct be_adapter *adapter) + { + struct be_queue_info *q; ++ struct be_rx_obj *rxo; ++ int i; + +- q = &adapter->rx_obj.q; +- if (q->created) { +- be_cmd_q_destroy(adapter, q, QTYPE_RXQ); +- be_rx_q_clean(adapter); +- } +- be_queue_free(adapter, q); ++ for_all_rx_queues(adapter, rxo, i) { ++ be_queue_free(adapter, &rxo->q); ++ ++ q = &rxo->cq; ++ if (q->created) ++ be_cmd_q_destroy(adapter, q, QTYPE_CQ); ++ be_queue_free(adapter, q); + +- q = &adapter->rx_obj.cq; +- if (q->created) +- be_cmd_q_destroy(adapter, q, QTYPE_CQ); +- be_queue_free(adapter, q); ++ q = &rxo->rx_eq.q; ++ if (q->created) ++ be_cmd_q_destroy(adapter, q, QTYPE_EQ); ++ be_queue_free(adapter, q); + +- /* Clear any residual events */ +- be_eq_clean(adapter, &adapter->rx_eq); ++ kfree(rxo->page_info_tbl); ++ } ++} + +- q = &adapter->rx_eq.q; +- if (q->created) +- be_cmd_q_destroy(adapter, q, QTYPE_EQ); +- be_queue_free(adapter, q); ++/* Is BE in a multi-channel mode */ ++static inline bool be_is_mc(struct be_adapter *adapter) { ++ return (adapter->function_mode & FLEX10_MODE || ++ adapter->function_mode & VNIC_MODE || ++ adapter->function_mode & UMC_ENABLED); ++} ++ ++static u32 be_num_rxqs_want(struct be_adapter *adapter) ++{ ++ if (multi_rxq && (adapter->function_caps & BE_FUNCTION_CAPS_RSS) && ++ adapter->num_vfs == 0 && be_physfn(adapter) && ++ !be_is_mc(adapter)) { ++ return 1 + MAX_RSS_QS; /* one default non-RSS queue */ ++ } else { ++ dev_warn(&adapter->pdev->dev, ++ "No support for multiple RX queues\n"); ++ return 1; ++ } + } + + static int be_rx_queues_create(struct be_adapter *adapter) + { + struct be_queue_info *eq, *q, *cq; +- int rc; ++ struct be_rx_obj *rxo; ++ int rc, i; + ++ adapter->num_rx_qs = min(be_num_rxqs_want(adapter), ++ msix_enabled(adapter) ? ++ adapter->num_msix_vec - 1 : 1); ++ if (adapter->num_rx_qs != MAX_RX_QS) ++ dev_warn(&adapter->pdev->dev, ++ "Could create only %d receive queues", ++ adapter->num_rx_qs); ++ ++ adapter->max_rx_coal = gro ? BE_INIT_FRAGS_PER_FRAME : 1; + adapter->big_page_size = (1 << get_order(rx_frag_size)) * PAGE_SIZE; +- adapter->rx_eq.max_eqd = BE_MAX_EQD; +- adapter->rx_eq.min_eqd = 0; +- adapter->rx_eq.cur_eqd = 0; +- adapter->rx_eq.enable_aic = true; ++ for_all_rx_queues(adapter, rxo, i) { ++ rxo->adapter = adapter; ++ rxo->rx_eq.max_eqd = BE_MAX_EQD; ++ rxo->rx_eq.enable_aic = true; + +- /* Alloc Rx Event queue */ +- eq = &adapter->rx_eq.q; +- rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN, +- sizeof(struct be_eq_entry)); +- if (rc) +- return rc; ++ /* EQ */ ++ eq = &rxo->rx_eq.q; ++ rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN, ++ sizeof(struct be_eq_entry)); ++ if (rc) ++ goto err; + +- /* Ask BE to create Rx Event queue */ +- rc = be_cmd_eq_create(adapter, eq, adapter->rx_eq.cur_eqd); +- if (rc) +- goto rx_eq_free; ++ rc = be_cmd_eq_create(adapter, eq, rxo->rx_eq.cur_eqd); ++ if (rc) ++ goto err; + +- /* Alloc RX eth compl queue */ +- cq = &adapter->rx_obj.cq; +- rc = be_queue_alloc(adapter, cq, RX_CQ_LEN, +- sizeof(struct be_eth_rx_compl)); +- if (rc) +- goto rx_eq_destroy; ++ rxo->rx_eq.eq_idx = adapter->eq_next_idx++; + +- /* Ask BE to create Rx eth compl queue */ +- rc = be_cmd_cq_create(adapter, cq, eq, false, false, 3); +- if (rc) +- goto rx_cq_free; ++ /* CQ */ ++ cq = &rxo->cq; ++ rc = be_queue_alloc(adapter, cq, RX_CQ_LEN, ++ sizeof(struct be_eth_rx_compl)); ++ if (rc) ++ goto err; + +- /* Alloc RX eth queue */ +- q = &adapter->rx_obj.q; +- rc = be_queue_alloc(adapter, q, RX_Q_LEN, sizeof(struct be_eth_rx_d)); +- if (rc) +- goto rx_cq_destroy; ++ rc = be_cmd_cq_create(adapter, cq, eq, false, false, 3); ++ if (rc) ++ goto err; + +- /* Ask BE to create Rx eth queue */ +- rc = be_cmd_rxq_create(adapter, q, cq->id, rx_frag_size, +- BE_MAX_JUMBO_FRAME_SIZE, adapter->if_handle, false); +- if (rc) +- goto rx_q_free; ++ /* Rx Q - will be created in be_open() */ ++ q = &rxo->q; ++ rc = be_queue_alloc(adapter, q, RX_Q_LEN, ++ sizeof(struct be_eth_rx_d)); ++ if (rc) ++ goto err; ++ ++ rxo->page_info_tbl = kzalloc(sizeof(struct be_rx_page_info) * ++ RX_Q_LEN, GFP_KERNEL); ++ if (!rxo->page_info_tbl) ++ goto err; ++ } + + return 0; +-rx_q_free: +- be_queue_free(adapter, q); +-rx_cq_destroy: +- be_cmd_q_destroy(adapter, cq, QTYPE_CQ); +-rx_cq_free: +- be_queue_free(adapter, cq); +-rx_eq_destroy: +- be_cmd_q_destroy(adapter, eq, QTYPE_EQ); +-rx_eq_free: +- be_queue_free(adapter, eq); +- return rc; ++err: ++ be_rx_queues_destroy(adapter); ++ return -1; + } + +-/* There are 8 evt ids per func. Retruns the evt id's bit number */ +-static inline int be_evt_bit_get(struct be_adapter *adapter, u32 eq_id) ++static bool event_peek(struct be_eq_obj *eq_obj) + { +- return eq_id - 8 * be_pci_func(adapter); ++ struct be_eq_entry *eqe = queue_tail_node(&eq_obj->q); ++ if (!eqe->evt) ++ return false; ++ else ++ return true; + } + + static irqreturn_t be_intx(int irq, void *dev) + { + struct be_adapter *adapter = dev; +- int isr; ++ struct be_rx_obj *rxo; ++ int isr, i, tx = 0 , rx = 0; + +- isr = ioread32(adapter->csr + CEV_ISR0_OFFSET + +- be_pci_func(adapter) * CEV_ISR_SIZE); +- if (!isr) +- return IRQ_NONE; ++ if (lancer_chip(adapter)) { ++ if (event_peek(&adapter->tx_eq)) ++ tx = event_handle(adapter, &adapter->tx_eq, false); ++ for_all_rx_queues(adapter, rxo, i) { ++ if (event_peek(&rxo->rx_eq)) ++ rx |= event_handle(adapter, &rxo->rx_eq, true); ++ } + +- event_handle(adapter, &adapter->tx_eq); +- event_handle(adapter, &adapter->rx_eq); ++ if (!(tx || rx)) ++ return IRQ_NONE; ++ } else { ++ isr = ioread32(adapter->csr + CEV_ISR0_OFFSET + ++ (adapter->tx_eq.q.id / 8) * CEV_ISR_SIZE); ++ if (!isr) ++ return IRQ_NONE; ++ ++ if ((1 << adapter->tx_eq.eq_idx & isr)) ++ event_handle(adapter, &adapter->tx_eq, false); ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ if ((1 << rxo->rx_eq.eq_idx & isr)) ++ event_handle(adapter, &rxo->rx_eq, true); ++ } ++ } + + return IRQ_HANDLED; + } + + static irqreturn_t be_msix_rx(int irq, void *dev) + { +- struct be_adapter *adapter = dev; ++ struct be_rx_obj *rxo = dev; ++ struct be_adapter *adapter = rxo->adapter; + +- event_handle(adapter, &adapter->rx_eq); ++ event_handle(adapter, &rxo->rx_eq, true); + + return IRQ_HANDLED; + } +@@ -1353,48 +2046,72 @@ static irqreturn_t be_msix_tx_mcc(int irq, void *dev) + { + struct be_adapter *adapter = dev; + +- event_handle(adapter, &adapter->tx_eq); ++ event_handle(adapter, &adapter->tx_eq, false); + + return IRQ_HANDLED; + } + + static inline bool do_gro(struct be_adapter *adapter, +- struct be_eth_rx_compl *rxcp) ++ struct be_rx_compl_info *rxcp) + { +- int err = AMAP_GET_BITS(struct amap_eth_rx_compl, err, rxcp); +- int tcp_frame = AMAP_GET_BITS(struct amap_eth_rx_compl, tcpf, rxcp); +- +- if (err) +- drvr_stats(adapter)->be_rxcp_err++; +- +- return (tcp_frame && !err) ? true : false; ++ return (!rxcp->tcpf || rxcp->err || adapter->max_rx_coal <= 1 || ++ (rxcp->vlanf && !vlan_configured(adapter))) ? ++ false : true; + } + + int be_poll_rx(struct napi_struct *napi, int budget) + { + struct be_eq_obj *rx_eq = container_of(napi, struct be_eq_obj, napi); +- struct be_adapter *adapter = +- container_of(rx_eq, struct be_adapter, rx_eq); +- struct be_queue_info *rx_cq = &adapter->rx_obj.cq; +- struct be_eth_rx_compl *rxcp; ++ struct be_rx_obj *rxo = container_of(rx_eq, struct be_rx_obj, rx_eq); ++ struct be_adapter *adapter = rxo->adapter; ++ struct be_queue_info *rx_cq = &rxo->cq; ++ struct be_rx_compl_info *rxcp; + u32 work_done; ++ bool flush_lro = false; + ++ rxo->stats.rx_polls++; + for (work_done = 0; work_done < budget; work_done++) { +- rxcp = be_rx_compl_get(adapter); ++ rxcp = be_rx_compl_get(rxo); + if (!rxcp) + break; + +- if (do_gro(adapter, rxcp)) +- be_rx_compl_process_gro(adapter, rxcp); +- else +- be_rx_compl_process(adapter, rxcp); ++ /* Is it a flush compl that has no data */ ++ if (unlikely(rxcp->num_rcvd == 0)) ++ continue; + +- be_rx_compl_reset(rxcp); ++ if (unlikely(rxcp->port != adapter->port_num)) { ++ be_rx_compl_discard(adapter, rxo, rxcp); ++ be_rx_stats_update(rxo, rxcp); ++ continue; ++ } ++ ++ if (likely((lancer_A0_chip(adapter) && !rxcp->err) || ++ !lancer_A0_chip(adapter))) { ++ if (do_gro(adapter, rxcp)) { ++ if (adapter->gro_supported) { ++ be_rx_compl_process_gro(adapter, rxo, ++ rxcp); ++ } else { ++ be_rx_compl_process_lro(adapter, rxo, ++ rxcp); ++ flush_lro = true; ++ } ++ } else { ++ be_rx_compl_process(adapter, rxo, rxcp); ++ } ++ } else if (lancer_A0_chip(adapter) && rxcp->err) { ++ be_rx_compl_discard(adapter, rxo, rxcp); ++ } ++ ++ be_rx_stats_update(rxo, rxcp); + } + ++ if (flush_lro) ++ lro_flush_all(&rxo->lro_mgr); ++ + /* Refill the queue */ +- if (atomic_read(&adapter->rx_obj.q.used) < RX_FRAGS_REFILL_WM) +- be_post_rx_frags(adapter); ++ if (work_done && atomic_read(&rxo->q.used) < RX_FRAGS_REFILL_WM) ++ be_post_rx_frags(rxo); + + /* All consumed */ + if (work_done < budget) { +@@ -1404,40 +2121,13 @@ int be_poll_rx(struct napi_struct *napi, int budget) + /* More to be consumed; continue with interrupts disabled */ + be_cq_notify(adapter, rx_cq->id, false, work_done); + } ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 32) ++ adapter->netdev->last_rx = jiffies; ++#endif + return work_done; + } + +-void be_process_tx(struct be_adapter *adapter) +-{ +- struct be_queue_info *txq = &adapter->tx_obj.q; +- struct be_queue_info *tx_cq = &adapter->tx_obj.cq; +- struct be_eth_tx_compl *txcp; +- u32 num_cmpl = 0; +- u16 end_idx; +- +- while ((txcp = be_tx_compl_get(tx_cq))) { +- end_idx = AMAP_GET_BITS(struct amap_eth_tx_compl, +- wrb_index, txcp); +- be_tx_compl_process(adapter, end_idx); +- num_cmpl++; +- } +- +- if (num_cmpl) { +- be_cq_notify(adapter, tx_cq->id, true, num_cmpl); +- +- /* As Tx wrbs have been freed up, wake up netdev queue if +- * it was stopped due to lack of tx wrbs. +- */ +- if (netif_queue_stopped(adapter->netdev) && +- atomic_read(&txq->used) < txq->len / 2) { +- netif_wake_queue(adapter->netdev); +- } +- +- drvr_stats(adapter)->be_tx_events++; +- drvr_stats(adapter)->be_tx_compl += num_cmpl; +- } +-} +- + /* As TX and MCC share the same EQ check for both TX and MCC completions. + * For TX/MCC we don't honour budget; consume everything + */ +@@ -1446,96 +2136,264 @@ static int be_poll_tx_mcc(struct napi_struct *napi, int budget) + struct be_eq_obj *tx_eq = container_of(napi, struct be_eq_obj, napi); + struct be_adapter *adapter = + container_of(tx_eq, struct be_adapter, tx_eq); ++ struct be_tx_obj *txo; ++ struct be_eth_tx_compl *txcp; ++ int tx_compl, mcc_compl, status = 0; ++ u8 i; ++ u16 num_wrbs; ++ ++ for_all_tx_queues(adapter, txo, i) { ++ tx_compl = 0; ++ num_wrbs = 0; ++ while ((txcp = be_tx_compl_get(&txo->cq))) { ++ num_wrbs += be_tx_compl_process(adapter, txo, ++ AMAP_GET_BITS(struct amap_eth_tx_compl, ++ wrb_index, txcp)); ++ tx_compl++; ++ } ++ if (tx_compl) { ++ be_cq_notify(adapter, txo->cq.id, true, tx_compl); ++ ++ atomic_sub(num_wrbs, &txo->q.used); ++ ++ /* As Tx wrbs have been freed up, wake up netdev queue ++ * if it was stopped due to lack of tx wrbs. */ ++ if (__netif_subqueue_stopped(adapter->netdev, i) && ++ atomic_read(&txo->q.used) < txo->q.len / 2) { ++ netif_wake_subqueue(adapter->netdev, i); ++ } ++ ++ adapter->drv_stats.be_tx_events++; ++ txo->stats.be_tx_compl += tx_compl; ++ } ++ } ++ ++ mcc_compl = be_process_mcc(adapter, &status); ++ ++ if (mcc_compl) { ++ struct be_mcc_obj *mcc_obj = &adapter->mcc_obj; ++ be_cq_notify(adapter, mcc_obj->cq.id, true, mcc_compl); ++ } + + napi_complete(napi); + +- be_process_tx(adapter); +- +- be_process_mcc(adapter); +- ++ be_eq_notify(adapter, tx_eq->q.id, true, false, 0); + return 1; + } + ++void be_detect_dump_ue(struct be_adapter *adapter) ++{ ++ u32 ue_status_lo, ue_status_hi, ue_status_lo_mask, ue_status_hi_mask; ++ u32 i; ++ ++ pci_read_config_dword(adapter->pdev, ++ PCICFG_UE_STATUS_LOW, &ue_status_lo); ++ pci_read_config_dword(adapter->pdev, ++ PCICFG_UE_STATUS_HIGH, &ue_status_hi); ++ pci_read_config_dword(adapter->pdev, ++ PCICFG_UE_STATUS_LOW_MASK, &ue_status_lo_mask); ++ pci_read_config_dword(adapter->pdev, ++ PCICFG_UE_STATUS_HI_MASK, &ue_status_hi_mask); ++ ++ ue_status_lo = (ue_status_lo & (~ue_status_lo_mask)); ++ ue_status_hi = (ue_status_hi & (~ue_status_hi_mask)); ++ ++ if (ue_status_lo || ue_status_hi) { ++ adapter->ue_detected = true; ++ adapter->eeh_err = true; ++ dev_err(&adapter->pdev->dev, "UE Detected!!\n"); ++ } ++ ++ if (ue_status_lo) { ++ for (i = 0; ue_status_lo; ue_status_lo >>= 1, i++) { ++ if (ue_status_lo & 1) ++ dev_err(&adapter->pdev->dev, ++ "UE: %s bit set\n", ue_status_low_desc[i]); ++ } ++ } ++ if (ue_status_hi) { ++ for (i = 0; ue_status_hi; ue_status_hi >>= 1, i++) { ++ if (ue_status_hi & 1) ++ dev_err(&adapter->pdev->dev, ++ "UE: %s bit set\n", ue_status_hi_desc[i]); ++ } ++ } ++ ++} ++ + static void be_worker(struct work_struct *work) + { + struct be_adapter *adapter = + container_of(work, struct be_adapter, work.work); ++ struct be_rx_obj *rxo; ++ struct be_tx_obj *txo; ++ int i; + +- be_cmd_get_stats(adapter, &adapter->stats.cmd); ++ if (!adapter->ue_detected && !lancer_chip(adapter)) ++ be_detect_dump_ue(adapter); + +- /* Set EQ delay */ +- be_rx_eqd_update(adapter); ++ /* when interrupts are not yet enabled, just reap any pending ++ * mcc completions */ ++ if (!netif_running(adapter->netdev)) { ++ int mcc_compl, status = 0; + +- be_tx_rate_update(adapter); +- be_rx_rate_update(adapter); ++ mcc_compl = be_process_mcc(adapter, &status); + +- if (adapter->rx_post_starved) { +- adapter->rx_post_starved = false; +- be_post_rx_frags(adapter); ++ if (mcc_compl) { ++ struct be_mcc_obj *mcc_obj = &adapter->mcc_obj; ++ be_cq_notify(adapter, mcc_obj->cq.id, false, mcc_compl); ++ } ++ ++ goto reschedule; ++ } ++ ++ if (!adapter->stats_cmd_sent) ++ be_cmd_get_stats(adapter, &adapter->stats_cmd); ++ ++ for_all_tx_queues(adapter, txo, i) ++ be_tx_rate_update(txo); ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ be_rx_rate_update(rxo); ++ be_rx_eqd_update(adapter, rxo); ++ ++ if (rxo->rx_post_starved) { ++ rxo->rx_post_starved = false; ++ be_post_rx_frags(rxo); ++ } + } + ++reschedule: ++ adapter->work_counter++; + schedule_delayed_work(&adapter->work, msecs_to_jiffies(1000)); + } + ++static void be_msix_disable(struct be_adapter *adapter) ++{ ++ if (msix_enabled(adapter)) { ++ pci_disable_msix(adapter->pdev); ++ adapter->num_msix_vec = 0; ++ } ++} ++ + static void be_msix_enable(struct be_adapter *adapter) + { +- int i, status; ++#define BE_MIN_MSIX_VECTORS (1 + 1) /* Rx + Tx */ ++ int i, status, num_vec; + +- for (i = 0; i < BE_NUM_MSIX_VECTORS; i++) ++ num_vec = be_num_rxqs_want(adapter) + 1; ++ ++ for (i = 0; i < num_vec; i++) + adapter->msix_entries[i].entry = i; + +- status = pci_enable_msix(adapter->pdev, adapter->msix_entries, +- BE_NUM_MSIX_VECTORS); +- if (status == 0) +- adapter->msix_enabled = true; ++ status = pci_enable_msix(adapter->pdev, adapter->msix_entries, num_vec); ++ if (status == 0) { ++ goto done; ++ } else if (status >= BE_MIN_MSIX_VECTORS) { ++ num_vec = status; ++ if (pci_enable_msix(adapter->pdev, adapter->msix_entries, ++ num_vec) == 0) ++ goto done; ++ } + return; ++done: ++ adapter->num_msix_vec = num_vec; ++ return; ++} ++ ++static void be_sriov_enable(struct be_adapter *adapter) ++{ ++ be_check_sriov_fn_type(adapter); ++#ifdef CONFIG_PCI_IOV ++ if (be_physfn(adapter) && num_vfs) { ++ int status, pos; ++ u16 nvfs; ++ ++ pos = pci_find_ext_capability(adapter->pdev, ++ PCI_EXT_CAP_ID_SRIOV); ++ pci_read_config_word(adapter->pdev, ++ pos + PCI_SRIOV_TOTAL_VF, &nvfs); ++ adapter->num_vfs = num_vfs; ++ if (num_vfs > nvfs) { ++ dev_info(&adapter->pdev->dev, ++ "Device supports %d VFs and not %d\n", ++ nvfs, num_vfs); ++ adapter->num_vfs = nvfs; ++ } ++ ++ status = pci_enable_sriov(adapter->pdev, adapter->num_vfs); ++ if (status) ++ adapter->num_vfs = 0; ++ } ++#endif ++} ++ ++static void be_sriov_disable(struct be_adapter *adapter) ++{ ++#ifdef CONFIG_PCI_IOV ++ if (adapter->num_vfs > 0) { ++ pci_disable_sriov(adapter->pdev); ++ adapter->num_vfs = 0; ++ } ++#endif + } + +-static inline int be_msix_vec_get(struct be_adapter *adapter, u32 eq_id) ++static inline int be_msix_vec_get(struct be_adapter *adapter, ++ struct be_eq_obj *eq_obj) + { +- return adapter->msix_entries[ +- be_evt_bit_get(adapter, eq_id)].vector; ++ return adapter->msix_entries[eq_obj->eq_idx].vector; + } + + static int be_request_irq(struct be_adapter *adapter, + struct be_eq_obj *eq_obj, +- void *handler, char *desc) ++ void *handler, char *desc, void *context) + { + struct net_device *netdev = adapter->netdev; + int vec; + + sprintf(eq_obj->desc, "%s-%s", netdev->name, desc); +- vec = be_msix_vec_get(adapter, eq_obj->q.id); +- return request_irq(vec, handler, 0, eq_obj->desc, adapter); ++ vec = be_msix_vec_get(adapter, eq_obj); ++ return request_irq(vec, handler, 0, eq_obj->desc, context); + } + +-static void be_free_irq(struct be_adapter *adapter, struct be_eq_obj *eq_obj) ++static void be_free_irq(struct be_adapter *adapter, struct be_eq_obj *eq_obj, ++ void *context) + { +- int vec = be_msix_vec_get(adapter, eq_obj->q.id); +- free_irq(vec, adapter); ++ int vec = be_msix_vec_get(adapter, eq_obj); ++ free_irq(vec, context); + } + + static int be_msix_register(struct be_adapter *adapter) + { +- int status; ++ struct be_rx_obj *rxo; ++ int status, i; ++ char qname[10]; + +- status = be_request_irq(adapter, &adapter->tx_eq, be_msix_tx_mcc, "tx"); ++ status = be_request_irq(adapter, &adapter->tx_eq, be_msix_tx_mcc, "tx", ++ adapter); + if (status) + goto err; + +- status = be_request_irq(adapter, &adapter->rx_eq, be_msix_rx, "rx"); +- if (status) +- goto free_tx_irq; ++ for_all_rx_queues(adapter, rxo, i) { ++ sprintf(qname, "rxq%d", i); ++ status = be_request_irq(adapter, &rxo->rx_eq, be_msix_rx, ++ qname, rxo); ++ if (status) ++ goto err_msix; ++ } + + return 0; + +-free_tx_irq: +- be_free_irq(adapter, &adapter->tx_eq); ++err_msix: ++ be_free_irq(adapter, &adapter->tx_eq, adapter); ++ ++ for (i--, rxo = &adapter->rx_obj[i]; i >= 0; i--, rxo--) ++ be_free_irq(adapter, &rxo->rx_eq, rxo); ++ + err: + dev_warn(&adapter->pdev->dev, + "MSIX Request IRQ failed - err %d\n", status); +- pci_disable_msix(adapter->pdev); +- adapter->msix_enabled = false; ++ be_msix_disable(adapter); + return status; + } + +@@ -1544,10 +2402,13 @@ static int be_irq_register(struct be_adapter *adapter) + struct net_device *netdev = adapter->netdev; + int status; + +- if (adapter->msix_enabled) { ++ if (msix_enabled(adapter)) { + status = be_msix_register(adapter); + if (status == 0) + goto done; ++ /* INTx is not supported for VF */ ++ if (!be_physfn(adapter)) ++ return status; + } + + /* INTx */ +@@ -1567,87 +2428,363 @@ done: + static void be_irq_unregister(struct be_adapter *adapter) + { + struct net_device *netdev = adapter->netdev; ++ struct be_rx_obj *rxo; ++ int i; + + if (!adapter->isr_registered) + return; + + /* INTx */ +- if (!adapter->msix_enabled) { ++ if (!msix_enabled(adapter)) { + free_irq(netdev->irq, adapter); + goto done; + } + + /* MSIx */ +- be_free_irq(adapter, &adapter->tx_eq); +- be_free_irq(adapter, &adapter->rx_eq); ++ be_free_irq(adapter, &adapter->tx_eq, adapter); ++ ++ for_all_rx_queues(adapter, rxo, i) ++ be_free_irq(adapter, &rxo->rx_eq, rxo); ++ + done: + adapter->isr_registered = false; +- return; + } + +-static int be_open(struct net_device *netdev) ++static u16 be_select_queue(struct net_device *netdev, ++ struct sk_buff *skb) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_eq_obj *rx_eq = &adapter->rx_eq; ++ u8 prio; ++ ++ if (adapter->num_tx_qs == 1) ++ return 0; ++ ++ prio = (vlan_tx_tag_get(skb) & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT; ++ return adapter->tc_txq_map[adapter->prio_tc_map[prio]]; ++} ++ ++static void be_rx_queues_clear(struct be_adapter *adapter) ++{ ++ struct be_queue_info *q; ++ struct be_rx_obj *rxo; ++ int i; ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ q = &rxo->q; ++ if (q->created) { ++ be_cmd_rxq_destroy(adapter, q); ++ /* After the rxq is invalidated, wait for a grace time ++ * of 1ms for all dma to end and the flush compl to ++ * arrive ++ */ ++ mdelay(1); ++ be_rx_q_clean(adapter, rxo); ++ } ++ ++ /* Clear any residual events */ ++ q = &rxo->rx_eq.q; ++ if (q->created) ++ be_eq_clean(adapter, &rxo->rx_eq); ++ } ++} ++ ++static int be_close(struct net_device *netdev) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_rx_obj *rxo; ++ struct be_tx_obj *txo; + struct be_eq_obj *tx_eq = &adapter->tx_eq; +- bool link_up; +- int status; ++ int vec, i; ++ ++ be_async_mcc_disable(adapter); ++ ++ netif_stop_queue(netdev); ++ netif_carrier_off(netdev); ++ adapter->link_status = LINK_DOWN; ++ ++ if (!lancer_chip(adapter)) ++ be_intr_set(adapter, false); ++ ++ for_all_rx_queues(adapter, rxo, i) ++ napi_disable(&rxo->rx_eq.napi); ++ ++ napi_disable(&tx_eq->napi); ++ ++ if (lancer_chip(adapter)) { ++ be_cq_notify(adapter, adapter->mcc_obj.cq.id, false, 0); ++ for_all_rx_queues(adapter, rxo, i) ++ be_cq_notify(adapter, rxo->cq.id, false, 0); ++ for_all_tx_queues(adapter, txo, i) ++ be_cq_notify(adapter, txo->cq.id, false, 0); ++ } ++ ++ if (msix_enabled(adapter)) { ++ vec = be_msix_vec_get(adapter, tx_eq); ++ synchronize_irq(vec); ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ vec = be_msix_vec_get(adapter, &rxo->rx_eq); ++ synchronize_irq(vec); ++ } ++ } else { ++ synchronize_irq(netdev->irq); ++ } ++ be_irq_unregister(adapter); ++ ++ /* Wait for all pending tx completions to arrive so that ++ * all tx skbs are freed. ++ */ ++ for_all_tx_queues(adapter, txo, i) ++ be_tx_compl_clean(adapter, txo); ++ ++ be_rx_queues_clear(adapter); ++ return 0; ++} ++ ++static int be_rx_queues_setup(struct be_adapter *adapter) ++{ ++ struct be_rx_obj *rxo; ++ int rc, i; ++ u8 rsstable[MAX_RSS_QS]; ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ rc = be_cmd_rxq_create(adapter, &rxo->q, rxo->cq.id, ++ rx_frag_size, BE_MAX_JUMBO_FRAME_SIZE, ++ adapter->if_handle, ++ (i > 0) ? 1 : 0/* rss enable */, &rxo->rss_id); ++ if (rc) ++ return rc; ++ } ++ ++ if (be_multi_rxq(adapter)) { ++ for_all_rss_queues(adapter, rxo, i) ++ rsstable[i] = rxo->rss_id; ++ ++ rc = be_cmd_rss_config(adapter, rsstable, ++ adapter->num_rx_qs - 1); ++ if (rc) ++ return rc; ++ } + + /* First time posting */ +- be_post_rx_frags(adapter); ++ for_all_rx_queues(adapter, rxo, i) { ++ be_post_rx_frags(rxo); ++ napi_enable(&rxo->rx_eq.napi); ++ } ++ return 0; ++} ++ ++static int be_open(struct net_device *netdev) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_eq_obj *tx_eq = &adapter->tx_eq; ++ struct be_rx_obj *rxo; ++ int link_status; ++ int status, i; ++ u8 mac_speed; ++ u16 link_speed; ++ ++ status = be_rx_queues_setup(adapter); ++ if (status) ++ goto err; + +- napi_enable(&rx_eq->napi); + napi_enable(&tx_eq->napi); + + be_irq_register(adapter); + +- be_intr_set(adapter, true); ++ if (!lancer_chip(adapter)) ++ be_intr_set(adapter, true); + + /* The evt queues are created in unarmed state; arm them */ +- be_eq_notify(adapter, rx_eq->q.id, true, false, 0); ++ for_all_rx_queues(adapter, rxo, i) { ++ be_eq_notify(adapter, rxo->rx_eq.q.id, true, false, 0); ++ be_cq_notify(adapter, rxo->cq.id, true, 0); ++ } + be_eq_notify(adapter, tx_eq->q.id, true, false, 0); + +- /* Rx compl queue may be in unarmed state; rearm it */ +- be_cq_notify(adapter, adapter->rx_obj.cq.id, true, 0); ++ /* Now that interrupts are on we can process async mcc */ ++ be_async_mcc_enable(adapter); + +- status = be_cmd_link_status_query(adapter, &link_up); ++ status = be_cmd_link_status_query(adapter, &link_status, &mac_speed, ++ &link_speed, 0); + if (status) +- goto ret_sts; +- be_link_status_update(adapter, link_up); ++ goto err; ++ be_link_status_update(adapter, link_status); + +- status = be_vid_config(adapter); ++ status = be_vid_config(adapter, false, 0); + if (status) +- goto ret_sts; ++ goto err; + +- status = be_cmd_set_flow_control(adapter, +- adapter->tx_fc, adapter->rx_fc); +- if (status) +- goto ret_sts; ++ if (be_physfn(adapter)) { ++ status = be_cmd_set_flow_control(adapter, ++ adapter->tx_fc, adapter->rx_fc); ++ if (status) ++ goto err; ++ } ++ ++ return 0; ++err: ++ be_close(adapter->netdev); ++ return -EIO; ++} ++ ++static int be_setup_wol(struct be_adapter *adapter, bool enable) ++{ ++ struct be_dma_mem cmd; ++ int status = 0; ++ u8 mac[ETH_ALEN]; ++ ++ memset(mac, 0, ETH_ALEN); ++ ++ cmd.size = sizeof(struct be_cmd_req_acpi_wol_magic_config); ++ cmd.va = pci_alloc_consistent(adapter->pdev, cmd.size, &cmd.dma); ++ if (cmd.va == NULL) ++ return -1; ++ memset(cmd.va, 0, cmd.size); ++ ++ if (enable) { ++ status = pci_write_config_dword(adapter->pdev, ++ PCICFG_PM_CONTROL_OFFSET, PCICFG_PM_CONTROL_MASK); ++ if (status) { ++ dev_err(&adapter->pdev->dev, ++ "Could not enable Wake-on-lan\n"); ++ pci_free_consistent(adapter->pdev, cmd.size, cmd.va, ++ cmd.dma); ++ return status; ++ } ++ status = be_cmd_enable_magic_wol(adapter, ++ adapter->netdev->dev_addr, &cmd); ++ pci_enable_wake(adapter->pdev, PCI_D3hot, 1); ++ pci_enable_wake(adapter->pdev, PCI_D3cold, 1); ++ } else { ++ status = be_cmd_enable_magic_wol(adapter, mac, &cmd); ++ pci_enable_wake(adapter->pdev, PCI_D3hot, 0); ++ pci_enable_wake(adapter->pdev, PCI_D3cold, 0); ++ } ++ ++ pci_free_consistent(adapter->pdev, cmd.size, cmd.va, cmd.dma); ++ return status; ++} ++ ++/* ++ * Generate a seed MAC address from the PF MAC Address using jhash. ++ * MAC Address for VFs are assigned incrementally starting from the seed. ++ * These addresses are programmed in the ASIC by the PF and the VF driver ++ * queries for the MAC address during its probe. ++ */ ++static inline int be_vf_eth_addr_config(struct be_adapter *adapter) ++{ ++ u32 vf = 0; ++ int status = 0; ++ u8 mac[ETH_ALEN]; ++ ++ be_vf_eth_addr_generate(adapter, mac); ++ ++ for (vf = 0; vf < adapter->num_vfs; vf++) { ++ status = be_cmd_pmac_add(adapter, mac, ++ adapter->vf_cfg[vf].vf_if_handle, ++ &adapter->vf_cfg[vf].vf_pmac_id, ++ vf + 1); ++ if (status) ++ dev_err(&adapter->pdev->dev, ++ "Mac address add failed for VF %d\n", vf); ++ else ++ memcpy(adapter->vf_cfg[vf].vf_mac_addr, mac, ETH_ALEN); + +- schedule_delayed_work(&adapter->work, msecs_to_jiffies(100)); +-ret_sts: ++ mac[5] += 1; ++ } + return status; + } + ++static inline void be_vf_eth_addr_rem(struct be_adapter *adapter) ++{ ++ u32 vf; ++ ++ for (vf = 0; vf < adapter->num_vfs; vf++) { ++ if (adapter->vf_cfg[vf].vf_pmac_id != BE_INVALID_PMAC_ID) ++ be_cmd_pmac_del(adapter, ++ adapter->vf_cfg[vf].vf_if_handle, ++ adapter->vf_cfg[vf].vf_pmac_id, vf + 1); ++ } ++} ++ ++static int be_num_txqs_want(struct be_adapter *adapter) ++{ ++ if (adapter->num_vfs > 0 || be_is_mc(adapter) || ++ lancer_chip(adapter) || !be_physfn(adapter) || ++ adapter->generation == BE_GEN2) ++ return 1; ++ else ++ return MAX_TX_QS; ++} ++ + static int be_setup(struct be_adapter *adapter) + { + struct net_device *netdev = adapter->netdev; +- u32 cap_flags, en_flags; +- int status; +- +- cap_flags = BE_IF_FLAGS_UNTAGGED | BE_IF_FLAGS_BROADCAST | +- BE_IF_FLAGS_MCAST_PROMISCUOUS | +- BE_IF_FLAGS_PROMISCUOUS | +- BE_IF_FLAGS_PASS_L3L4_ERRORS; +- en_flags = BE_IF_FLAGS_UNTAGGED | BE_IF_FLAGS_BROADCAST | +- BE_IF_FLAGS_PASS_L3L4_ERRORS; ++ int status, fw_num_txqs, num_txqs; ++ u32 cap_flags, en_flags, vf = 0; ++ u8 mac[ETH_ALEN]; ++ ++ num_txqs = be_num_txqs_want(adapter); ++ if (num_txqs > 1) { ++ be_cmd_req_pg_pfc(adapter, &fw_num_txqs); ++ num_txqs = min(num_txqs, fw_num_txqs); ++ } ++ adapter->num_tx_qs = num_txqs; ++ if (adapter->num_tx_qs != MAX_TX_QS) ++ netif_set_real_num_tx_queues(adapter->netdev, ++ adapter->num_tx_qs); ++ ++ be_cmd_req_native_mode(adapter); ++ ++ cap_flags = en_flags = BE_IF_FLAGS_UNTAGGED | ++ BE_IF_FLAGS_BROADCAST | ++ BE_IF_FLAGS_MULTICAST; ++ ++ if (be_physfn(adapter)) { ++ if (adapter->function_caps & BE_FUNCTION_CAPS_RSS) { ++ cap_flags |= BE_IF_FLAGS_RSS; ++ en_flags |= BE_IF_FLAGS_RSS; ++ } ++ cap_flags |= BE_IF_FLAGS_MCAST_PROMISCUOUS | ++ BE_IF_FLAGS_PROMISCUOUS; ++ if (!lancer_A0_chip(adapter)) { ++ cap_flags |= BE_IF_FLAGS_PASS_L3L4_ERRORS; ++ en_flags |= BE_IF_FLAGS_PASS_L3L4_ERRORS; ++ } ++ } + + status = be_cmd_if_create(adapter, cap_flags, en_flags, + netdev->dev_addr, false/* pmac_invalid */, +- &adapter->if_handle, &adapter->pmac_id); ++ &adapter->if_handle, &adapter->pmac_id, 0); + if (status != 0) + goto do_none; + ++ if (be_physfn(adapter)) { ++ while (vf < adapter->num_vfs) { ++ cap_flags = en_flags = BE_IF_FLAGS_UNTAGGED | ++ BE_IF_FLAGS_BROADCAST; ++ status = be_cmd_if_create(adapter, cap_flags, ++ en_flags, mac, true, ++ &adapter->vf_cfg[vf].vf_if_handle, ++ NULL, vf+1); ++ if (status) { ++ dev_err(&adapter->pdev->dev, ++ "Interface Create failed for VF %d\n", vf); ++ goto if_destroy; ++ } ++ adapter->vf_cfg[vf].vf_pmac_id = BE_INVALID_PMAC_ID; ++ vf++; ++ } ++ } else { ++ status = be_cmd_mac_addr_query(adapter, mac, ++ MAC_ADDRESS_TYPE_NETWORK, false, adapter->if_handle); ++ if (!status) { ++ memcpy(adapter->netdev->dev_addr, mac, ETH_ALEN); ++ memcpy(adapter->netdev->perm_addr, mac, ETH_ALEN); ++ } ++ } ++ + status = be_tx_queues_create(adapter); + if (status != 0) + goto if_destroy; +@@ -1656,10 +2793,15 @@ static int be_setup(struct be_adapter *adapter) + if (status != 0) + goto tx_qs_destroy; + ++ /* Allow all priorities by default. A GRP5 evt may modify this */ ++ adapter->vlan_prio_bmap = 0xff; ++ + status = be_mcc_queues_create(adapter); + if (status != 0) + goto rx_qs_destroy; + ++ adapter->link_speed = -1; ++ + return 0; + + rx_qs_destroy: +@@ -1667,158 +2809,392 @@ rx_qs_destroy: + tx_qs_destroy: + be_tx_queues_destroy(adapter); + if_destroy: +- be_cmd_if_destroy(adapter, adapter->if_handle); ++ if (be_physfn(adapter)) { ++ for (vf = 0; vf < adapter->num_vfs; vf++) ++ if (adapter->vf_cfg[vf].vf_if_handle) ++ be_cmd_if_destroy(adapter, ++ adapter->vf_cfg[vf].vf_if_handle, ++ vf + 1); ++ } ++ be_cmd_if_destroy(adapter, adapter->if_handle, 0); + do_none: + return status; + } + + static int be_clear(struct be_adapter *adapter) + { ++ int vf; ++ ++ if (be_physfn(adapter) && adapter->num_vfs) ++ be_vf_eth_addr_rem(adapter); ++ + be_mcc_queues_destroy(adapter); + be_rx_queues_destroy(adapter); + be_tx_queues_destroy(adapter); ++ adapter->eq_next_idx = 0; + +- be_cmd_if_destroy(adapter, adapter->if_handle); ++ if (be_physfn(adapter)) { ++ for (vf = 0; vf < adapter->num_vfs; vf++) ++ if (adapter->vf_cfg[vf].vf_if_handle) ++ be_cmd_if_destroy(adapter, ++ adapter->vf_cfg[vf].vf_if_handle, vf + 1); ++ } ++ be_cmd_if_destroy(adapter, adapter->if_handle, 0); + ++ /* tell fw we're done with firing cmds */ ++ be_cmd_fw_clean(adapter); + return 0; + } + +-static int be_close(struct net_device *netdev) ++static void be_cpy_drv_ver(struct be_adapter *adapter, void *va) ++{ ++ struct mgmt_controller_attrib *attrib = ++ (struct mgmt_controller_attrib *) ((u8*) va + ++ sizeof(struct be_cmd_resp_hdr)); ++ ++ memcpy(attrib->hba_attribs.driver_version_string, ++ DRV_VER, sizeof(DRV_VER)); ++ attrib->pci_bus_number = adapter->pdev->bus->number; ++ attrib->pci_device_number = PCI_SLOT(adapter->pdev->devfn); ++ return; ++} ++ ++#define IOCTL_COOKIE "SERVERENGINES CORP" ++static int be_do_ioctl(struct net_device *netdev, ++ struct ifreq *ifr, int cmd) + { + struct be_adapter *adapter = netdev_priv(netdev); +- struct be_eq_obj *rx_eq = &adapter->rx_eq; +- struct be_eq_obj *tx_eq = &adapter->tx_eq; +- int vec; ++ struct be_cmd_req_hdr req; ++ struct be_cmd_resp_hdr *resp; ++ void *data = ifr->ifr_data; ++ void *ioctl_ptr; ++ void *va; ++ dma_addr_t dma; ++ u32 req_size; ++ int status, ret = 0; ++ u8 cookie[32]; ++ ++ switch (cmd) { ++ case SIOCDEVPRIVATE: ++ if (copy_from_user(cookie, data, strlen(IOCTL_COOKIE))) ++ return -EFAULT; ++ ++ if (memcmp(cookie, IOCTL_COOKIE, strlen(IOCTL_COOKIE))) ++ return -EINVAL; + +- cancel_delayed_work_sync(&adapter->work); ++ ioctl_ptr = (u8 *)data + strlen(IOCTL_COOKIE); ++ if (copy_from_user(&req, ioctl_ptr, ++ sizeof(struct be_cmd_req_hdr))) ++ return -EFAULT; + +- netif_stop_queue(netdev); +- netif_carrier_off(netdev); +- adapter->link_up = false; ++ req_size = le32_to_cpu(req.request_length); ++ if (req_size > 65536) ++ return -EINVAL; + +- be_intr_set(adapter, false); ++ req_size += sizeof(struct be_cmd_req_hdr); ++ va = pci_alloc_consistent(adapter->pdev, req_size, &dma); ++ if (!va) ++ return -ENOMEM; ++ if (copy_from_user(va, ioctl_ptr, req_size)) { ++ ret = -EFAULT; ++ break; ++ } + +- if (adapter->msix_enabled) { +- vec = be_msix_vec_get(adapter, tx_eq->q.id); +- synchronize_irq(vec); +- vec = be_msix_vec_get(adapter, rx_eq->q.id); +- synchronize_irq(vec); +- } else { +- synchronize_irq(netdev->irq); ++ status = be_cmd_pass_ext_ioctl(adapter, dma, req_size, va); ++ if (status == -1) { ++ ret = -EIO; ++ break; ++ } ++ ++ resp = (struct be_cmd_resp_hdr *) va; ++ if (!status) { ++ if (req.opcode == OPCODE_COMMON_GET_CNTL_ATTRIBUTES) ++ be_cpy_drv_ver(adapter, va); ++ } ++ ++ if (copy_to_user(ioctl_ptr, va, req_size)) { ++ ret = -EFAULT; ++ break; ++ } ++ break; ++ default: ++ return -EOPNOTSUPP; + } +- be_irq_unregister(adapter); + +- napi_disable(&rx_eq->napi); +- napi_disable(&tx_eq->napi); ++ if (va) ++ pci_free_consistent(adapter->pdev, req_size, va, dma); ++ ++ return ret; ++} ++ ++#ifdef CONFIG_NET_POLL_CONTROLLER ++static void be_netpoll(struct net_device *netdev) ++{ ++ struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_rx_obj *rxo; ++ int i; + +- /* Wait for all pending tx completions to arrive so that +- * all tx skbs are freed. +- */ +- be_tx_compl_clean(adapter); ++ event_handle(adapter, &adapter->tx_eq, false); ++ for_all_rx_queues(adapter, rxo, i) ++ event_handle(adapter, &rxo->rx_eq, true); ++ ++ return; ++} ++#endif ++ ++static int be_get_frag_header(struct skb_frag_struct *frag, void **mac_hdr, ++ void **ip_hdr, void **tcpudp_hdr, ++ u64 *hdr_flags, void *priv) ++{ ++ struct ethhdr *eh; ++ struct vlan_ethhdr *veh; ++ struct iphdr *iph; ++ u8 *va = page_address(frag->page) + frag->page_offset; ++ unsigned long ll_hlen; ++ ++ prefetch(va); ++ eh = (struct ethhdr *)va; ++ *mac_hdr = eh; ++ ll_hlen = ETH_HLEN; ++ if (eh->h_proto != htons(ETH_P_IP)) { ++ if (eh->h_proto == htons(ETH_P_8021Q)) { ++ veh = (struct vlan_ethhdr *)va; ++ if (veh->h_vlan_encapsulated_proto != htons(ETH_P_IP)) ++ return -1; ++ ++ ll_hlen += VLAN_HLEN; ++ } else { ++ return -1; ++ } ++ } ++ *hdr_flags = LRO_IPV4; ++ iph = (struct iphdr *)(va + ll_hlen); ++ *ip_hdr = iph; ++ if (iph->protocol != IPPROTO_TCP) ++ return -1; ++ *hdr_flags |= LRO_TCP; ++ *tcpudp_hdr = (u8 *) (*ip_hdr) + (iph->ihl << 2); + + return 0; + } + +-#define FW_FILE_HDR_SIGN "ServerEngines Corp. " ++static void be_lro_init(struct be_adapter *adapter, struct net_device *netdev) ++{ ++ struct net_lro_mgr *lro_mgr; ++ struct be_rx_obj *rxo; ++ int i; ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ lro_mgr = &rxo->lro_mgr; ++ lro_mgr->dev = netdev; ++ lro_mgr->features = LRO_F_NAPI; ++ lro_mgr->ip_summed = CHECKSUM_UNNECESSARY; ++ lro_mgr->ip_summed_aggr = CHECKSUM_UNNECESSARY; ++ lro_mgr->max_desc = BE_MAX_LRO_DESCRIPTORS; ++ lro_mgr->lro_arr = rxo->lro_desc; ++ lro_mgr->get_frag_header = be_get_frag_header; ++ lro_mgr->max_aggr = BE_MAX_FRAGS_PER_FRAME; ++ } ++ ++#ifdef NETIF_F_GRO ++ netdev->features |= NETIF_F_GRO; ++ adapter->gro_supported = true; ++#endif ++} ++ ++#define FW_FILE_HDR_SIGN "ServerEngines Corp. " + char flash_cookie[2][16] = {"*** SE FLAS", + "H DIRECTORY *** "}; +-static int be_flash_image(struct be_adapter *adapter, ++ ++static bool be_flash_redboot(struct be_adapter *adapter, ++ const u8 *p, u32 img_start, int image_size, ++ int hdr_size) ++{ ++ u32 crc_offset; ++ u8 flashed_crc[4]; ++ int status; ++ ++ crc_offset = hdr_size + img_start + image_size - 4; ++ ++ p += crc_offset; ++ ++ status = be_cmd_get_flash_crc(adapter, flashed_crc, ++ (image_size - 4)); ++ if (status) { ++ dev_err(&adapter->pdev->dev, ++ "could not get crc from flash, not flashing redboot\n"); ++ return false; ++ } ++ ++ /*update redboot only if crc does not match*/ ++ if (!memcmp(flashed_crc, p, 4)) ++ return false; ++ else ++ return true; ++} ++ ++static bool phy_flashing_required(struct be_adapter *adapter) ++{ ++ int status = 0; ++ struct be_phy_info phy_info; ++ ++ status = be_cmd_get_phy_info(adapter, &phy_info); ++ if (status) ++ return false; ++ if ((phy_info.phy_type == TN_8022) && ++ (phy_info.interface_type == PHY_TYPE_BASET_10GB)) { ++ return true; ++ } ++ return false; ++} ++ ++static int be_flash_data(struct be_adapter *adapter, + const struct firmware *fw, +- struct be_dma_mem *flash_cmd, u32 flash_type) ++ struct be_dma_mem *flash_cmd, int num_of_images) ++ + { +- int status; +- u32 flash_op, image_offset = 0, total_bytes, image_size = 0; ++ int status = 0, i, filehdr_size = 0; ++ u32 total_bytes = 0, flash_op; + int num_bytes; + const u8 *p = fw->data; + struct be_cmd_write_flashrom *req = flash_cmd->va; ++ struct flash_comp *pflashcomp; ++ int num_comp; + +- switch (flash_type) { +- case FLASHROM_TYPE_ISCSI_ACTIVE: +- image_offset = FLASH_iSCSI_PRIMARY_IMAGE_START; +- image_size = FLASH_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_ISCSI_BACKUP: +- image_offset = FLASH_iSCSI_BACKUP_IMAGE_START; +- image_size = FLASH_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_FCOE_FW_ACTIVE: +- image_offset = FLASH_FCoE_PRIMARY_IMAGE_START; +- image_size = FLASH_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_FCOE_FW_BACKUP: +- image_offset = FLASH_FCoE_BACKUP_IMAGE_START; +- image_size = FLASH_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_BIOS: +- image_offset = FLASH_iSCSI_BIOS_START; +- image_size = FLASH_BIOS_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_FCOE_BIOS: +- image_offset = FLASH_FCoE_BIOS_START; +- image_size = FLASH_BIOS_IMAGE_MAX_SIZE; +- break; +- case FLASHROM_TYPE_PXE_BIOS: +- image_offset = FLASH_PXE_BIOS_START; +- image_size = FLASH_BIOS_IMAGE_MAX_SIZE; +- break; +- default: +- return 0; ++ struct flash_comp gen3_flash_types[10] = { ++ { FLASH_iSCSI_PRIMARY_IMAGE_START_g3, IMG_TYPE_ISCSI_ACTIVE, ++ FLASH_IMAGE_MAX_SIZE_g3}, ++ { FLASH_REDBOOT_START_g3, IMG_TYPE_REDBOOT, ++ FLASH_REDBOOT_IMAGE_MAX_SIZE_g3}, ++ { FLASH_iSCSI_BIOS_START_g3, IMG_TYPE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g3}, ++ { FLASH_PXE_BIOS_START_g3, IMG_TYPE_PXE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g3}, ++ { FLASH_FCoE_BIOS_START_g3, IMG_TYPE_FCOE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g3}, ++ { FLASH_iSCSI_BACKUP_IMAGE_START_g3, IMG_TYPE_ISCSI_BACKUP, ++ FLASH_IMAGE_MAX_SIZE_g3}, ++ { FLASH_FCoE_PRIMARY_IMAGE_START_g3, IMG_TYPE_FCOE_FW_ACTIVE, ++ FLASH_IMAGE_MAX_SIZE_g3}, ++ { FLASH_FCoE_BACKUP_IMAGE_START_g3, IMG_TYPE_FCOE_FW_BACKUP, ++ FLASH_IMAGE_MAX_SIZE_g3}, ++ { FLASH_NCSI_START_g3, IMG_TYPE_NCSI_FW, ++ FLASH_NCSI_IMAGE_MAX_SIZE_g3}, ++ { FLASH_PHY_FW_START_g3, IMG_TYPE_PHY_FW, ++ FLASH_PHY_FW_IMAGE_MAX_SIZE_g3} ++ }; ++ struct flash_comp gen2_flash_types[8] = { ++ { FLASH_iSCSI_PRIMARY_IMAGE_START_g2, IMG_TYPE_ISCSI_ACTIVE, ++ FLASH_IMAGE_MAX_SIZE_g2}, ++ { FLASH_REDBOOT_START_g2, IMG_TYPE_REDBOOT, ++ FLASH_REDBOOT_IMAGE_MAX_SIZE_g2}, ++ { FLASH_iSCSI_BIOS_START_g2, IMG_TYPE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g2}, ++ { FLASH_PXE_BIOS_START_g2, IMG_TYPE_PXE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g2}, ++ { FLASH_FCoE_BIOS_START_g2, IMG_TYPE_FCOE_BIOS, ++ FLASH_BIOS_IMAGE_MAX_SIZE_g2}, ++ { FLASH_iSCSI_BACKUP_IMAGE_START_g2, IMG_TYPE_ISCSI_BACKUP, ++ FLASH_IMAGE_MAX_SIZE_g2}, ++ { FLASH_FCoE_PRIMARY_IMAGE_START_g2, IMG_TYPE_FCOE_FW_ACTIVE, ++ FLASH_IMAGE_MAX_SIZE_g2}, ++ { FLASH_FCoE_BACKUP_IMAGE_START_g2, IMG_TYPE_FCOE_FW_BACKUP, ++ FLASH_IMAGE_MAX_SIZE_g2} ++ }; ++ if (adapter->generation == BE_GEN3) { ++ pflashcomp = gen3_flash_types; ++ filehdr_size = sizeof(struct flash_file_hdr_g3); ++ num_comp = ARRAY_SIZE(gen3_flash_types); ++ } else { ++ pflashcomp = gen2_flash_types; ++ filehdr_size = sizeof(struct flash_file_hdr_g2); ++ num_comp = ARRAY_SIZE(gen2_flash_types); + } ++ for (i = 0; i < num_comp; i++) { ++ if ((pflashcomp[i].optype == IMG_TYPE_NCSI_FW) && ++ memcmp(adapter->fw_ver, "3.102.148.0", 11) < 0) ++ continue; ++ if (pflashcomp[i].optype == IMG_TYPE_PHY_FW) { ++ if (!phy_flashing_required(adapter)) ++ continue; ++ } ++ if ((pflashcomp[i].optype == IMG_TYPE_REDBOOT) && ++ (!be_flash_redboot(adapter, fw->data, ++ pflashcomp[i].offset, pflashcomp[i].size, filehdr_size + ++ (num_of_images * sizeof(struct image_hdr))))) ++ continue; + +- p += sizeof(struct flash_file_hdr) + image_offset; +- if (p + image_size > fw->data + fw->size) +- return -1; +- +- total_bytes = image_size; +- +- while (total_bytes) { +- if (total_bytes > 32*1024) +- num_bytes = 32*1024; +- else +- num_bytes = total_bytes; +- total_bytes -= num_bytes; +- +- if (!total_bytes) +- flash_op = FLASHROM_OPER_FLASH; +- else +- flash_op = FLASHROM_OPER_SAVE; +- memcpy(req->params.data_buf, p, num_bytes); +- p += num_bytes; +- status = be_cmd_write_flashrom(adapter, flash_cmd, +- flash_type, flash_op, num_bytes); +- if (status) { +- dev_err(&adapter->pdev->dev, +- "cmd to write to flash rom failed. type/op %d/%d\n", +- flash_type, flash_op); ++ p = fw->data; ++ p += filehdr_size + pflashcomp[i].offset ++ + (num_of_images * sizeof(struct image_hdr)); ++ if (p + pflashcomp[i].size > fw->data + fw->size) + return -1; ++ total_bytes = pflashcomp[i].size; ++ while (total_bytes) { ++ if (total_bytes > 32*1024) ++ num_bytes = 32*1024; ++ else ++ num_bytes = total_bytes; ++ total_bytes -= num_bytes; ++ if (!total_bytes) { ++ if (pflashcomp[i].optype == IMG_TYPE_PHY_FW) ++ flash_op = FLASHROM_OPER_PHY_FLASH; ++ else ++ flash_op = FLASHROM_OPER_FLASH; ++ } else { ++ if (pflashcomp[i].optype == IMG_TYPE_PHY_FW) ++ flash_op = FLASHROM_OPER_PHY_SAVE; ++ else ++ flash_op = FLASHROM_OPER_SAVE; ++ } ++ memcpy(req->params.data_buf, p, num_bytes); ++ p += num_bytes; ++ status = be_cmd_write_flashrom(adapter, flash_cmd, ++ pflashcomp[i].optype, flash_op, num_bytes); ++ if (status) { ++ if ((status == ILLEGAL_IOCTL_REQ) && ++ (pflashcomp[i].optype == ++ IMG_TYPE_PHY_FW)) ++ break; ++ dev_err(&adapter->pdev->dev, ++ "cmd to write to flash rom failed.\n"); ++ return -1; ++ } ++ yield(); + } +- yield(); + } +- + return 0; + } + ++static int get_ufigen_type(struct flash_file_hdr_g2 *fhdr) ++{ ++ if (fhdr == NULL) ++ return 0; ++ if (fhdr->build[0] == '3') ++ return BE_GEN3; ++ else if (fhdr->build[0] == '2') ++ return BE_GEN2; ++ else ++ return 0; ++} ++ + int be_load_fw(struct be_adapter *adapter, u8 *func) + { + char fw_file[ETHTOOL_FLASH_MAX_FILENAME]; + const struct firmware *fw; +- struct flash_file_hdr *fhdr; +- struct flash_section_info *fsec = NULL; ++ struct flash_file_hdr_g2 *fhdr; ++ struct flash_file_hdr_g3 *fhdr3; ++ struct image_hdr *img_hdr_ptr = NULL; + struct be_dma_mem flash_cmd; +- int status; ++ int status, i = 0, num_imgs = 0; + const u8 *p; +- bool entry_found = false; +- int flash_type; +- char fw_ver[FW_VER_LEN]; +- char fw_cfg; + +- status = be_cmd_get_fw_ver(adapter, fw_ver); +- if (status) +- return status; ++ if (!netif_running(adapter->netdev)) { ++ dev_err(&adapter->pdev->dev, ++ "Firmware load not allowed (interface is down)\n"); ++ return -1; ++ } + +- fw_cfg = *(fw_ver + 2); +- if (fw_cfg == '0') +- fw_cfg = '1'; + strcpy(fw_file, func); + + status = request_firmware(&fw, fw_file, &adapter->pdev->dev); +@@ -1826,34 +3202,9 @@ int be_load_fw(struct be_adapter *adapter, u8 *func) + goto fw_exit; + + p = fw->data; +- fhdr = (struct flash_file_hdr *) p; +- if (memcmp(fhdr->sign, FW_FILE_HDR_SIGN, strlen(FW_FILE_HDR_SIGN))) { +- dev_err(&adapter->pdev->dev, +- "Firmware(%s) load error (signature did not match)\n", +- fw_file); +- status = -1; +- goto fw_exit; +- } +- ++ fhdr = (struct flash_file_hdr_g2 *) p; + dev_info(&adapter->pdev->dev, "Flashing firmware file %s\n", fw_file); + +- p += sizeof(struct flash_file_hdr); +- while (p < (fw->data + fw->size)) { +- fsec = (struct flash_section_info *)p; +- if (!memcmp(flash_cookie, fsec->cookie, sizeof(flash_cookie))) { +- entry_found = true; +- break; +- } +- p += 32; +- } +- +- if (!entry_found) { +- status = -1; +- dev_err(&adapter->pdev->dev, +- "Flash cookie not found in firmware image\n"); +- goto fw_exit; +- } +- + flash_cmd.size = sizeof(struct be_cmd_write_flashrom) + 32*1024; + flash_cmd.va = pci_alloc_consistent(adapter->pdev, flash_cmd.size, + &flash_cmd.dma); +@@ -1864,12 +3215,25 @@ int be_load_fw(struct be_adapter *adapter, u8 *func) + goto fw_exit; + } + +- for (flash_type = FLASHROM_TYPE_ISCSI_ACTIVE; +- flash_type <= FLASHROM_TYPE_FCOE_FW_BACKUP; flash_type++) { +- status = be_flash_image(adapter, fw, &flash_cmd, +- flash_type); +- if (status) +- break; ++ if ((adapter->generation == BE_GEN3) && ++ (get_ufigen_type(fhdr) == BE_GEN3)) { ++ fhdr3 = (struct flash_file_hdr_g3 *) fw->data; ++ num_imgs = le32_to_cpu(fhdr3->num_imgs); ++ for (i = 0; i < num_imgs; i++) { ++ img_hdr_ptr = (struct image_hdr *) (fw->data + ++ (sizeof(struct flash_file_hdr_g3) + ++ i * sizeof(struct image_hdr))); ++ if (le32_to_cpu(img_hdr_ptr->imageid) == 1) ++ status = be_flash_data(adapter, fw, &flash_cmd, ++ num_imgs); ++ } ++ } else if ((adapter->generation == BE_GEN2) && ++ (get_ufigen_type(fhdr) == BE_GEN2)) { ++ status = be_flash_data(adapter, fw, &flash_cmd, 0); ++ } else { ++ dev_err(&adapter->pdev->dev, ++ "UFI and Interface are not compatible for flashing\n"); ++ status = -1; + } + + pci_free_consistent(adapter->pdev, flash_cmd.size, flash_cmd.va, +@@ -1879,14 +3243,14 @@ int be_load_fw(struct be_adapter *adapter, u8 *func) + goto fw_exit; + } + +- dev_info(&adapter->pdev->dev, "Firmware flashed succesfully\n"); ++ dev_info(&adapter->pdev->dev, "Firmware flashed successfully\n"); + + fw_exit: + release_firmware(fw); + return status; + } + +-static struct net_device_ops be_netdev_ops = { ++static net_device_ops_no_const be_netdev_ops = { + .ndo_open = be_open, + .ndo_stop = be_close, + .ndo_start_xmit = be_xmit, +@@ -1898,15 +3262,32 @@ static struct net_device_ops be_netdev_ops = { + .ndo_vlan_rx_register = be_vlan_register, + .ndo_vlan_rx_add_vid = be_vlan_add_vid, + .ndo_vlan_rx_kill_vid = be_vlan_rem_vid, ++#ifdef HAVE_SRIOV_CONFIG ++ .ndo_set_vf_mac = be_set_vf_mac, ++ .ndo_set_vf_vlan = be_set_vf_vlan, ++ .ndo_set_vf_tx_rate = be_set_vf_tx_rate, ++ .ndo_get_vf_config = be_get_vf_config, ++#endif ++ .ndo_do_ioctl = be_do_ioctl, ++#ifdef CONFIG_NET_POLL_CONTROLLER ++ .ndo_poll_controller = be_netpoll, ++#endif + }; + +-static void be_netdev_init(struct net_device *netdev) ++static int be_netdev_init(struct net_device *netdev) + { + struct be_adapter *adapter = netdev_priv(netdev); ++ struct be_rx_obj *rxo; ++ int i, status = 0; + + netdev->features |= NETIF_F_SG | NETIF_F_HW_VLAN_RX | NETIF_F_TSO | +- NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_FILTER | NETIF_F_HW_CSUM | +- NETIF_F_GRO; ++ NETIF_F_HW_VLAN_TX | NETIF_F_HW_CSUM | NETIF_F_TSO6; ++ ++ netdev->vlan_features |= NETIF_F_SG | NETIF_F_TSO | NETIF_F_TSO6 | ++ NETIF_F_HW_CSUM; ++ ++ netdev->features |= NETIF_F_VLAN_SG | NETIF_F_VLAN_TSO | ++ NETIF_F_VLAN_CSUM; + + netdev->flags |= IFF_MULTICAST; + +@@ -1918,17 +3299,30 @@ static void be_netdev_init(struct net_device *netdev) + + netif_set_gso_max_size(netdev, 65535); + ++ if (adapter->flags & BE_FLAGS_DCBX) ++ be_netdev_ops.ndo_select_queue = be_select_queue; + BE_SET_NETDEV_OPS(netdev, &be_netdev_ops); +- ++ + SET_ETHTOOL_OPS(netdev, &be_ethtool_ops); + +- netif_napi_add(netdev, &adapter->rx_eq.napi, be_poll_rx, +- BE_NAPI_WEIGHT); +- netif_napi_add(netdev, &adapter->tx_eq.napi, be_poll_tx_mcc, ++ be_lro_init(adapter, netdev); ++ ++ for_all_rx_queues(adapter, rxo, i) { ++ status = be_netif_napi_add(netdev, &rxo->rx_eq.napi, be_poll_rx, ++ BE_NAPI_WEIGHT); ++ if (status) { ++ dev_err(&adapter->pdev->dev, "dummy netdev alloc fail" ++ "for rxo:%d\n", i); ++ return status; ++ } ++ } ++ status = be_netif_napi_add(netdev, &adapter->tx_eq.napi, be_poll_tx_mcc, + BE_NAPI_WEIGHT); ++ if (status) ++ dev_err(&adapter->pdev->dev, "dummy netdev alloc fail" ++ "for tx\n"); + +- netif_carrier_off(netdev); +- netif_stop_queue(netdev); ++ return status; + } + + static void be_unmap_pci_bars(struct be_adapter *adapter) +@@ -1937,37 +3331,62 @@ static void be_unmap_pci_bars(struct be_adapter *adapter) + iounmap(adapter->csr); + if (adapter->db) + iounmap(adapter->db); +- if (adapter->pcicfg) ++ if (adapter->pcicfg && be_physfn(adapter)) + iounmap(adapter->pcicfg); + } + + static int be_map_pci_bars(struct be_adapter *adapter) + { ++ struct pci_dev *pdev = adapter->pdev; + u8 __iomem *addr; +- int pcicfg_reg; ++ int pcicfg_reg, db_reg; + +- addr = ioremap_nocache(pci_resource_start(adapter->pdev, 2), +- pci_resource_len(adapter->pdev, 2)); +- if (addr == NULL) +- return -ENOMEM; +- adapter->csr = addr; ++ if (lancer_chip(adapter)) { ++ addr = ioremap_nocache(pci_resource_start(pdev, 0), ++ pci_resource_len(adapter->pdev, 0)); ++ if (addr == NULL) ++ return -ENOMEM; ++ adapter->db = addr; ++ return 0; ++ } + +- addr = ioremap_nocache(pci_resource_start(adapter->pdev, 4), +- 128 * 1024); +- if (addr == NULL) +- goto pci_map_err; +- adapter->db = addr; ++ if (be_physfn(adapter)) { ++ addr = ioremap_nocache(pci_resource_start(pdev, 2), ++ pci_resource_len(pdev, 2)); ++ if (addr == NULL) ++ return -ENOMEM; ++ adapter->csr = addr; ++ adapter->netdev->mem_start = pci_resource_start(pdev, 2); ++ adapter->netdev->mem_end = pci_resource_start(pdev, 2) + ++ pci_resource_len(pdev, 2); ++ } + +- if (adapter->generation == BE_GEN2) ++ if (adapter->generation == BE_GEN2) { + pcicfg_reg = 1; +- else ++ db_reg = 4; ++ } else { + pcicfg_reg = 0; ++ if (be_physfn(adapter)) ++ db_reg = 4; ++ else ++ db_reg = 0; ++ } + +- addr = ioremap_nocache(pci_resource_start(adapter->pdev, pcicfg_reg), +- pci_resource_len(adapter->pdev, pcicfg_reg)); ++ addr = ioremap_nocache(pci_resource_start(pdev, db_reg), ++ pci_resource_len(pdev, db_reg)); + if (addr == NULL) + goto pci_map_err; +- adapter->pcicfg = addr; ++ adapter->db = addr; ++ ++ if (be_physfn(adapter)) { ++ addr = ioremap_nocache( ++ pci_resource_start(pdev, pcicfg_reg), ++ pci_resource_len(pdev, pcicfg_reg)); ++ if (addr == NULL) ++ goto pci_map_err; ++ adapter->pcicfg = addr; ++ } else ++ adapter->pcicfg = adapter->db + SRIOV_VF_PCICFG_OFFSET; + + return 0; + pci_map_err: +@@ -1985,40 +3404,69 @@ static void be_ctrl_cleanup(struct be_adapter *adapter) + if (mem->va) + pci_free_consistent(adapter->pdev, mem->size, + mem->va, mem->dma); ++ ++ mem = &adapter->rx_filter; ++ if (mem->va) ++ pci_free_consistent(adapter->pdev, mem->size, ++ mem->va, mem->dma); + } + + static int be_ctrl_init(struct be_adapter *adapter) + { + struct be_dma_mem *mbox_mem_alloc = &adapter->mbox_mem_alloced; + struct be_dma_mem *mbox_mem_align = &adapter->mbox_mem; ++ struct be_dma_mem *rx_filter = &adapter->rx_filter; + int status; + + status = be_map_pci_bars(adapter); + if (status) +- return status; ++ goto done; + + mbox_mem_alloc->size = sizeof(struct be_mcc_mailbox) + 16; + mbox_mem_alloc->va = pci_alloc_consistent(adapter->pdev, + mbox_mem_alloc->size, &mbox_mem_alloc->dma); + if (!mbox_mem_alloc->va) { +- be_unmap_pci_bars(adapter); +- return -1; ++ status = -ENOMEM; ++ goto unmap_pci_bars; + } ++ + mbox_mem_align->size = sizeof(struct be_mcc_mailbox); + mbox_mem_align->va = PTR_ALIGN(mbox_mem_alloc->va, 16); + mbox_mem_align->dma = PTR_ALIGN(mbox_mem_alloc->dma, 16); + memset(mbox_mem_align->va, 0, sizeof(struct be_mcc_mailbox)); +- spin_lock_init(&adapter->mbox_lock); ++ ++ rx_filter->size = sizeof(struct be_cmd_req_rx_filter); ++ rx_filter->va = pci_alloc_consistent(adapter->pdev, rx_filter->size, ++ &rx_filter->dma); ++ if (rx_filter->va == NULL) { ++ status = -ENOMEM; ++ goto free_mbox; ++ } ++ memset(rx_filter->va, 0, rx_filter->size); ++ ++ mutex_init(&adapter->mbox_lock); + spin_lock_init(&adapter->mcc_lock); + spin_lock_init(&adapter->mcc_cq_lock); + ++ init_completion(&adapter->flash_compl); ++ ++ PCI_SAVE_STATE(adapter->pdev); + return 0; ++ ++free_mbox: ++ pci_free_consistent(adapter->pdev, mbox_mem_alloc->size, ++ mbox_mem_alloc->va, mbox_mem_alloc->dma); ++ ++unmap_pci_bars: ++ be_unmap_pci_bars(adapter); ++ ++done: ++ return status; + } + + static void be_stats_cleanup(struct be_adapter *adapter) + { +- struct be_stats_obj *stats = &adapter->stats; +- struct be_dma_mem *cmd = &stats->cmd; ++ struct be_dma_mem *cmd = &adapter->stats_cmd; + + if (cmd->va) + pci_free_consistent(adapter->pdev, cmd->size, +@@ -2027,10 +3475,12 @@ static void be_stats_cleanup(struct be_adapter *adapter) + + static int be_stats_init(struct be_adapter *adapter) + { +- struct be_stats_obj *stats = &adapter->stats; +- struct be_dma_mem *cmd = &stats->cmd; ++ struct be_dma_mem *cmd = &adapter->stats_cmd; + +- cmd->size = sizeof(struct be_cmd_req_get_stats); ++ if (adapter->generation == BE_GEN2) ++ cmd->size = sizeof(struct be_cmd_req_get_stats_v0); ++ else ++ cmd->size = sizeof(struct be_cmd_req_get_stats_v1); + cmd->va = pci_alloc_consistent(adapter->pdev, cmd->size, &cmd->dma); + if (cmd->va == NULL) + return -1; +@@ -2041,9 +3491,17 @@ static int be_stats_init(struct be_adapter *adapter) + static void __devexit be_remove(struct pci_dev *pdev) + { + struct be_adapter *adapter = pci_get_drvdata(pdev); ++ + if (!adapter) + return; + ++ cancel_delayed_work_sync(&adapter->work); ++ ++#ifdef CONFIG_PALAU ++ be_sysfs_remove_group(adapter); ++#endif ++ ++ /* be_close() gets called if the device is open by unregister */ + unregister_netdev(adapter->netdev); + + be_clear(adapter); +@@ -2052,36 +3510,203 @@ static void __devexit be_remove(struct pci_dev *pdev) + + be_ctrl_cleanup(adapter); + +- if (adapter->msix_enabled) { +- pci_disable_msix(adapter->pdev); +- adapter->msix_enabled = false; +- } ++ kfree(adapter->vf_cfg); ++ be_sriov_disable(adapter); ++ ++ be_msix_disable(adapter); + + pci_set_drvdata(pdev, NULL); + pci_release_regions(pdev); + pci_disable_device(pdev); +- ++ be_netif_napi_del(adapter->netdev); + free_netdev(adapter->netdev); + } + +-static int be_hw_up(struct be_adapter *adapter) ++static void be_pcie_slot_check(struct be_adapter *adapter) ++{ ++ u32 curr, max, width, max_wd, speed, max_sp; ++ ++ pci_read_config_dword(adapter->pdev, PCICFG_PCIE_LINK_STATUS_OFFSET, ++ &curr); ++ width = (curr >> PCIE_LINK_STATUS_NEG_WIDTH_SHIFT) & ++ PCIE_LINK_STATUS_NEG_WIDTH_MASK; ++ speed = (curr >> PCIE_LINK_STATUS_SPEED_SHIFT) & ++ PCIE_LINK_STATUS_SPEED_MASK; ++ ++ pci_read_config_dword(adapter->pdev, PCICFG_PCIE_LINK_CAP_OFFSET, ++ &max); ++ max_wd = (max >> PCIE_LINK_CAP_MAX_WIDTH_SHIFT) & ++ PCIE_LINK_CAP_MAX_WIDTH_MASK; ++ max_sp = (max >> PCIE_LINK_CAP_MAX_SPEED_SHIFT) & ++ PCIE_LINK_CAP_MAX_SPEED_MASK; ++ ++ if (width < max_wd || speed < max_sp) ++ dev_warn(&adapter->pdev->dev, ++ "Found network device in a Gen%s x%d PCIe slot. It " ++ "should be in a Gen2 x%d slot for best performance\n", ++ speed < max_sp ? "1" : "2", width, max_wd); ++} ++ ++static int be_get_ioctl_version(char *fw_version) { ++ char *str[4]; ++ int i; ++ int val[4]; ++ char *endptr; ++ ++ if(!fw_version) ++ return 0; ++ for(i=0; i<3; i++) { ++ str[i] = strsep(&fw_version, "."); ++ val[i] = simple_strtol(str[i], &endptr, 10); ++ } ++ ++ if (val[0]>4 || (val[0]>3 && val[2]>143)) ++ return 1; ++ return 0; ++} ++ ++static int be_get_port_names(struct be_adapter *adapter) + { + int status; ++ int ver; + +- status = be_cmd_POST(adapter); ++ status = be_cmd_get_fw_ver(adapter, ++ adapter->fw_ver, NULL); + if (status) + return status; ++ ver = be_get_ioctl_version(adapter->fw_ver); ++ if (ver && (adapter->generation == BE_GEN3)) ++ status = be_cmd_query_port_names_v1(adapter, ++ adapter->port_name); ++ else ++ status = be_cmd_query_port_names_v0(adapter, ++ adapter->port_name); ++ return status; ++} + +- status = be_cmd_reset_function(adapter); ++static int be_get_config(struct be_adapter *adapter) ++{ ++ int status; ++ u8 mac[ETH_ALEN]; ++ ++ status = be_cmd_query_fw_cfg(adapter, &adapter->port_num, ++ &adapter->function_mode, ++ &adapter->function_caps); + if (status) + return status; + +- status = be_cmd_get_fw_ver(adapter, adapter->fw_ver); ++ status = be_cmd_get_cntl_attributes(adapter); + if (status) + return status; + +- status = be_cmd_query_fw_cfg(adapter, +- &adapter->port_num, &adapter->cap); ++ memset(mac, 0, ETH_ALEN); ++ be_pcie_slot_check(adapter); ++ ++ if (be_physfn(adapter)) { ++ status = be_cmd_mac_addr_query(adapter, mac, ++ MAC_ADDRESS_TYPE_NETWORK, true /*permanent */, 0); ++ ++ if (status) ++ return status; ++ ++ if (!is_valid_ether_addr(mac)) ++ return -EADDRNOTAVAIL; ++ ++ memcpy(adapter->netdev->dev_addr, mac, ETH_ALEN); ++ memcpy(adapter->netdev->perm_addr, mac, ETH_ALEN); ++ } ++ ++ if (adapter->function_mode & FLEX10_MODE) ++ adapter->max_vlans = BE_NUM_VLANS_SUPPORTED/8; ++ else ++ adapter->max_vlans = BE_NUM_VLANS_SUPPORTED; ++ ++ return 0; ++} ++ ++static int be_dev_family_check(struct be_adapter *adapter) ++{ ++ struct pci_dev *pdev = adapter->pdev; ++ u32 sli_intf = 0, if_type; ++ ++ switch (pdev->device) { ++ case BE_DEVICE_ID1: ++ case OC_DEVICE_ID1: ++ adapter->generation = BE_GEN2; ++ break; ++ case BE_DEVICE_ID2: ++ case OC_DEVICE_ID2: ++ adapter->generation = BE_GEN3; ++ break; ++ case OC_DEVICE_ID3: ++ pci_read_config_dword(pdev, SLI_INTF_REG_OFFSET, &sli_intf); ++ if_type = (sli_intf & SLI_INTF_IF_TYPE_MASK) >> ++ SLI_INTF_IF_TYPE_SHIFT; ++ ++ if (((sli_intf & SLI_INTF_VALID_MASK) != SLI_INTF_VALID) || ++ if_type != 0x02) { ++ dev_err(&pdev->dev, "SLI_INTF reg val is not valid\n"); ++ return -EINVAL; ++ } ++ if (num_vfs > 0) { ++ dev_err(&pdev->dev, "VFs not supported\n"); ++ return -EINVAL; ++ } ++ adapter->sli_family = ((sli_intf & SLI_INTF_FAMILY_MASK) >> ++ SLI_INTF_FAMILY_SHIFT); ++ adapter->generation = BE_GEN3; ++ break; ++ default: ++ adapter->generation = 0; ++ } ++ return 0; ++} ++ ++static int lancer_wait_ready(struct be_adapter *adapter) ++{ ++#define SLIPORT_READY_TIMEOUT 500 ++ u32 sliport_status; ++ int status = 0, i; ++ ++ for (i = 0; i < SLIPORT_READY_TIMEOUT; i++) { ++ sliport_status = ioread32(adapter->db + SLIPORT_STATUS_OFFSET); ++ if (sliport_status & SLIPORT_STATUS_RDY_MASK) ++ break; ++ ++ msleep(20); ++ } ++ ++ if (i == SLIPORT_READY_TIMEOUT) ++ status = -1; ++ ++ return status; ++} ++ ++static int lancer_test_and_set_rdy_state(struct be_adapter *adapter) ++{ ++ int status; ++ u32 sliport_status, err, reset_needed; ++ status = lancer_wait_ready(adapter); ++ if (!status) { ++ sliport_status = ioread32(adapter->db + SLIPORT_STATUS_OFFSET); ++ err = sliport_status & SLIPORT_STATUS_ERR_MASK; ++ reset_needed = sliport_status & SLIPORT_STATUS_RN_MASK; ++ if (err && reset_needed) { ++ iowrite32(SLI_PORT_CONTROL_IP_MASK, ++ adapter->db + SLIPORT_CONTROL_OFFSET); ++ ++ /* check adapter has corrected the error */ ++ status = lancer_wait_ready(adapter); ++ sliport_status = ioread32(adapter->db + ++ SLIPORT_STATUS_OFFSET); ++ sliport_status &= (SLIPORT_STATUS_ERR_MASK | ++ SLIPORT_STATUS_RN_MASK); ++ if (status || sliport_status) ++ status = -1; ++ } else if (err || reset_needed) { ++ status = -1; ++ } ++ } + return status; + } + +@@ -2091,7 +3716,7 @@ static int __devinit be_probe(struct pci_dev *pdev, + int status = 0; + struct be_adapter *adapter; + struct net_device *netdev; +- u8 mac[ETH_ALEN]; ++ u32 en; + + status = pci_enable_device(pdev); + if (status) +@@ -2102,31 +3727,22 @@ static int __devinit be_probe(struct pci_dev *pdev, + goto disable_dev; + pci_set_master(pdev); + +- netdev = alloc_etherdev(sizeof(struct be_adapter)); ++ netdev = alloc_etherdev_mq(sizeof(struct be_adapter), MAX_TX_QS); + if (netdev == NULL) { + status = -ENOMEM; + goto rel_reg; + } + adapter = netdev_priv(netdev); + +- switch (pdev->device) { +- case BE_DEVICE_ID1: +- case OC_DEVICE_ID1: +- adapter->generation = BE_GEN2; +- break; +- case BE_DEVICE_ID2: +- case OC_DEVICE_ID2: +- adapter->generation = BE_GEN3; +- break; +- default: +- adapter->generation = 0; +- } +- + adapter->pdev = pdev; ++ ++ status = be_dev_family_check(adapter); ++ if (status) ++ goto free_netdev; ++ + pci_set_drvdata(pdev, adapter); + adapter->netdev = netdev; +- +- be_msix_enable(adapter); ++ SET_NETDEV_DEV(netdev, &pdev->dev); + + status = pci_set_dma_mask(pdev, DMA_BIT_MASK(64)); + if (!status) { +@@ -2139,46 +3755,150 @@ static int __devinit be_probe(struct pci_dev *pdev, + } + } + ++ be_sriov_enable(adapter); ++ if (adapter->num_vfs > 0) { ++ adapter->vf_cfg = kcalloc(adapter->num_vfs, ++ sizeof(struct be_vf_cfg), GFP_KERNEL); ++ ++ if (!adapter->vf_cfg) ++ goto free_netdev; ++ } ++ + status = be_ctrl_init(adapter); + if (status) +- goto free_netdev; ++ goto free_vf_cfg; ++ ++ if (lancer_chip(adapter)) { ++ status = lancer_test_and_set_rdy_state(adapter); ++ if (status) { ++ dev_err(&pdev->dev, "Adapter in non recoverable error\n"); ++ goto ctrl_clean; ++ } ++ } ++ ++ /* sync up with fw's ready state */ ++ if (be_physfn(adapter)) { ++ status = be_cmd_POST(adapter); ++ if (status) ++ goto ctrl_clean; ++ } ++ ++ /* tell fw we're ready to fire cmds */ ++ status = be_cmd_fw_init(adapter); ++ if (status) ++ goto ctrl_clean; ++ ++ status = be_cmd_reset_function(adapter); ++ if (status) ++ goto ctrl_clean; + + status = be_stats_init(adapter); + if (status) + goto ctrl_clean; + +- status = be_hw_up(adapter); ++ status = be_get_config(adapter); + if (status) + goto stats_clean; + +- status = be_cmd_mac_addr_query(adapter, mac, MAC_ADDRESS_TYPE_NETWORK, +- true /* permanent */, 0); +- if (status) +- goto stats_clean; +- memcpy(netdev->dev_addr, mac, ETH_ALEN); ++ /* This bit is zero in normal boot case, but in crash kernel case this ++ is not cleared. clear this bit here, until we are ready with the irqs ++ i.e in be_open call.*/ ++ if (!lancer_chip(adapter)) ++ be_intr_set(adapter, false); ++ ++ if (msix) ++ be_msix_enable(adapter); + + INIT_DELAYED_WORK(&adapter->work, be_worker); +- be_netdev_init(netdev); +- SET_NETDEV_DEV(netdev, &adapter->pdev->dev); + + status = be_setup(adapter); + if (status) +- goto stats_clean; ++ goto msix_disable; ++ ++ /* Initilize the link status to -1 */ ++ adapter->link_status = -1; ++ ++ status = be_netdev_init(netdev); ++ if (status) ++ goto unsetup; ++ + status = register_netdev(netdev); + if (status != 0) + goto unsetup; + +- dev_info(&pdev->dev, "%s port %d\n", nic_name(pdev), adapter->port_num); ++ be_cmd_get_fw_ver(adapter, adapter->fw_ver, NULL); ++ ++ if (be_physfn(adapter) && adapter->num_vfs) { ++ u8 mac_speed; ++ int link_status; ++ u16 def_vlan, vf, lnk_speed; ++ ++ status = be_vf_eth_addr_config(adapter); ++ if (status) ++ goto unreg_netdev; ++ ++ for (vf = 0; vf < adapter->num_vfs; vf++) { ++ status = be_cmd_get_hsw_config(adapter, &def_vlan, ++ vf + 1, adapter->vf_cfg[vf].vf_if_handle); ++ if (!status) ++ adapter->vf_cfg[vf].vf_def_vid = def_vlan; ++ else ++ goto unreg_netdev; ++ ++ status = be_cmd_link_status_query(adapter, &link_status, ++ &mac_speed, &lnk_speed, vf + 1); ++ if (!status) ++ adapter->vf_cfg[vf].vf_tx_rate = lnk_speed * 10; ++ else ++ goto unreg_netdev; ++ } ++ } ++ if (be_physfn(adapter)) { ++ /* Temp fix ofr bug# 23034. Till ARM ++ * f/w fixes privilege lvl */ ++ be_get_port_names(adapter); ++ } ++ ++ /* Enable Vlan capability based on privileges. ++ * PF will have Vlan capability anyway. */ ++ be_cmd_get_fn_privileges(adapter, &en, 0); ++ ++ if ((en & (BE_PRIV_FILTMGMT | BE_PRIV_VHADM | BE_PRIV_DEVCFG)) || ++ be_physfn(adapter)) ++ netdev->features |= NETIF_F_HW_VLAN_FILTER; ++ else ++ netdev->features |= NETIF_F_VLAN_CHALLENGED; ++ ++ dev_info(&pdev->dev, "%s: numa node %d\n", netdev->name, ++ dev_to_node(&pdev->dev)); ++ dev_info(&pdev->dev, "%s %s \"%s\" port %d\n", nic_name(pdev), ++ (adapter->port_num > 1 ? "1Gbps NIC" : "10Gbps NIC"), ++ adapter->model_number, adapter->hba_port_num); ++ ++ ++#ifdef CONFIG_PALAU ++ be_sysfs_create_group(adapter); ++#endif ++ schedule_delayed_work(&adapter->work, msecs_to_jiffies(100)); + return 0; + ++unreg_netdev: ++ unregister_netdev(netdev); + unsetup: + be_clear(adapter); ++msix_disable: ++ be_msix_disable(adapter); + stats_clean: + be_stats_cleanup(adapter); + ctrl_clean: + be_ctrl_cleanup(adapter); ++free_vf_cfg: ++ kfree(adapter->vf_cfg); + free_netdev: +- free_netdev(adapter->netdev); ++ be_sriov_disable(adapter); ++ be_netif_napi_del(netdev); ++ free_netdev(netdev); ++ pci_set_drvdata(pdev, NULL); + rel_reg: + pci_release_regions(pdev); + disable_dev: +@@ -2193,6 +3913,10 @@ static int be_suspend(struct pci_dev *pdev, pm_message_t state) + struct be_adapter *adapter = pci_get_drvdata(pdev); + struct net_device *netdev = adapter->netdev; + ++ cancel_delayed_work_sync(&adapter->work); ++ if (adapter->wol) ++ be_setup_wol(adapter, true); ++ + netif_device_detach(netdev); + if (netif_running(netdev)) { + rtnl_lock(); +@@ -2202,6 +3926,7 @@ static int be_suspend(struct pci_dev *pdev, pm_message_t state) + be_cmd_get_flow_control(adapter, &adapter->tx_fc, &adapter->rx_fc); + be_clear(adapter); + ++ be_msix_disable(adapter); + pci_save_state(pdev); + pci_disable_device(pdev); + pci_set_power_state(pdev, pci_choose_state(pdev, state)); +@@ -2223,6 +3948,12 @@ static int be_resume(struct pci_dev *pdev) + pci_set_power_state(pdev, 0); + pci_restore_state(pdev); + ++ be_msix_enable(adapter); ++ /* tell fw we're ready to fire cmds */ ++ status = be_cmd_fw_init(adapter); ++ if (status) ++ return status; ++ + be_setup(adapter); + if (netif_running(netdev)) { + rtnl_lock(); +@@ -2230,28 +3961,152 @@ static int be_resume(struct pci_dev *pdev) + rtnl_unlock(); + } + netif_device_attach(netdev); ++ ++ if (adapter->wol) ++ be_setup_wol(adapter, false); ++ ++ schedule_delayed_work(&adapter->work, msecs_to_jiffies(100)); + return 0; + } + ++/* ++ * An FLR will stop BE from DMAing any data. ++ */ ++static void be_shutdown(struct pci_dev *pdev) ++{ ++ struct be_adapter *adapter = pci_get_drvdata(pdev); ++ ++ if (!adapter) ++ return; ++ ++ cancel_delayed_work_sync(&adapter->work); ++ ++ netif_device_detach(adapter->netdev); ++ ++ if (adapter->wol) ++ be_setup_wol(adapter, true); ++ ++ be_cmd_reset_function(adapter); ++ ++ pci_disable_device(pdev); ++} ++ ++static pci_ers_result_t be_eeh_err_detected(struct pci_dev *pdev, ++ pci_channel_state_t state) ++{ ++ struct be_adapter *adapter = pci_get_drvdata(pdev); ++ struct net_device *netdev = adapter->netdev; ++ ++ dev_err(&adapter->pdev->dev, "EEH error detected\n"); ++ ++ adapter->eeh_err = true; ++ ++ netif_device_detach(netdev); ++ ++ if (netif_running(netdev)) { ++ rtnl_lock(); ++ be_close(netdev); ++ rtnl_unlock(); ++ } ++ be_clear(adapter); ++ ++ if (state == pci_channel_io_perm_failure) ++ return PCI_ERS_RESULT_DISCONNECT; ++ ++ pci_disable_device(pdev); ++ ++ return PCI_ERS_RESULT_NEED_RESET; ++} ++ ++static pci_ers_result_t be_eeh_reset(struct pci_dev *pdev) ++{ ++ struct be_adapter *adapter = pci_get_drvdata(pdev); ++ int status; ++ ++ dev_info(&adapter->pdev->dev, "EEH reset\n"); ++ adapter->eeh_err = false; ++ ++ status = pci_enable_device(pdev); ++ if (status) ++ return PCI_ERS_RESULT_DISCONNECT; ++ ++ pci_set_master(pdev); ++ pci_set_power_state(pdev, 0); ++ pci_restore_state(pdev); ++ ++ /* Check if card is ok and fw is ready */ ++ status = be_cmd_POST(adapter); ++ if (status) ++ return PCI_ERS_RESULT_DISCONNECT; ++ ++ return PCI_ERS_RESULT_RECOVERED; ++} ++ ++static void be_eeh_resume(struct pci_dev *pdev) ++{ ++ int status = 0; ++ struct be_adapter *adapter = pci_get_drvdata(pdev); ++ struct net_device *netdev = adapter->netdev; ++ ++ dev_info(&adapter->pdev->dev, "EEH resume\n"); ++ ++ pci_save_state(pdev); ++ ++ /* tell fw we're ready to fire cmds */ ++ status = be_cmd_fw_init(adapter); ++ if (status) ++ goto err; ++ ++ status = be_setup(adapter); ++ if (status) ++ goto err; ++ ++ if (netif_running(netdev)) { ++ status = be_open(netdev); ++ if (status) ++ goto err; ++ } ++ netif_device_attach(netdev); ++ return; ++err: ++ dev_err(&adapter->pdev->dev, "EEH resume failed\n"); ++ return; ++} ++ ++static struct pci_error_handlers be_eeh_handlers = { ++ .error_detected = be_eeh_err_detected, ++ .slot_reset = be_eeh_reset, ++ .resume = be_eeh_resume, ++}; ++ + static struct pci_driver be_driver = { + .name = DRV_NAME, + .id_table = be_dev_ids, + .probe = be_probe, + .remove = be_remove, + .suspend = be_suspend, +- .resume = be_resume ++ .resume = be_resume, ++ .shutdown = be_shutdown, ++ .err_handler = &be_eeh_handlers + }; + + static int __init be_init_module(void) + { +- if (rx_frag_size != 8192 && rx_frag_size != 4096 +- && rx_frag_size != 2048) { ++ if (rx_frag_size != 8192 && rx_frag_size != 4096 && ++ rx_frag_size != 2048) { + printk(KERN_WARNING DRV_NAME + " : Module param rx_frag_size must be 2048/4096/8192." + " Using 2048\n"); + rx_frag_size = 2048; + } + ++ if (!msix && num_vfs > 0) { ++ printk(KERN_WARNING DRV_NAME ++ " : MSIx required for num_vfs > 0. Ignoring msix=0\n"); ++ msix = 1; ++ } ++ ++ + return pci_register_driver(&be_driver); + } + module_init(be_init_module); +diff --git a/drivers/net/benet/be_misc.c b/drivers/net/benet/be_misc.c +new file mode 100644 +index 0000000..4ab499f +--- /dev/null ++++ b/drivers/net/benet/be_misc.c +@@ -0,0 +1,106 @@ ++/* ++ * Copyright (C) 2005 - 2011 Emulex ++ * All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License version 2 ++ * as published by the Free Software Foundation. The full GNU General ++ * Public License is included in this distribution in the file called COPYING. ++ * ++ * Contact Information: ++ * linux-drivers@emulex.com ++ * ++ * Emulex ++ * 3333 Susan Street ++ * Costa Mesa, CA 92626 ++ */ ++#include "be.h" ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 24) ++static ssize_t ++flash_fw_store(struct class_device *cd, const char *buf, size_t len) ++{ ++ struct be_adapter *adapter = ++ netdev_priv(container_of(cd, struct net_device, class_dev)); ++ char file_name[ETHTOOL_FLASH_MAX_FILENAME]; ++ int status; ++ ++ if (!capable(CAP_NET_ADMIN)) ++ return -EPERM; ++ ++ file_name[ETHTOOL_FLASH_MAX_FILENAME - 1] = 0; ++ strncpy(file_name, buf, (ETHTOOL_FLASH_MAX_FILENAME - 1)); ++ ++ /* Removing new-line char given by sysfs */ ++ file_name[strlen(file_name) - 1] = '\0'; ++ ++ status = be_load_fw(adapter, file_name); ++ if (!status) ++ return len; ++ else ++ return status; ++} ++ ++static CLASS_DEVICE_ATTR(flash_fw, S_IWUSR, NULL, flash_fw_store); ++ ++static struct attribute *benet_attrs[] = { ++ &class_device_attr_flash_fw.attr, ++ NULL, ++}; ++#else ++ ++static ssize_t ++flash_fw_store(struct device *dev, struct device_attribute *attr, ++ const char *buf, size_t len) ++{ ++ struct be_adapter *adapter = ++ netdev_priv(container_of(dev, struct net_device, dev)); ++ char file_name[ETHTOOL_FLASH_MAX_FILENAME]; ++ int status; ++ ++ if (!capable(CAP_NET_ADMIN)) ++ return -EPERM; ++ ++ file_name[ETHTOOL_FLASH_MAX_FILENAME - 1] = 0; ++ strncpy(file_name, buf, (ETHTOOL_FLASH_MAX_FILENAME - 1)); ++ ++ /* Removing new-line char given by sysfs */ ++ file_name[strlen(file_name) - 1] = '\0'; ++ ++ status = be_load_fw(adapter, file_name); ++ if (!status) ++ return len; ++ else ++ return status; ++} ++ ++static DEVICE_ATTR(flash_fw, S_IWUSR, NULL, flash_fw_store); ++ ++static struct attribute *benet_attrs[] = { ++ &dev_attr_flash_fw.attr, ++ NULL, ++}; ++#endif ++ ++#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 24) ++#define CLASS_DEV class_dev ++#else ++#define CLASS_DEV dev ++#endif ++ ++static struct attribute_group benet_attr_group = {.attrs = benet_attrs }; ++ ++void be_sysfs_create_group(struct be_adapter *adapter) ++{ ++ int status; ++ ++ status = sysfs_create_group(&adapter->netdev->CLASS_DEV.kobj, ++ &benet_attr_group); ++ if (status) ++ dev_err(&adapter->pdev->dev, "Could not create sysfs group\n"); ++} ++ ++void be_sysfs_remove_group(struct be_adapter *adapter) ++{ ++ sysfs_remove_group(&adapter->netdev->CLASS_DEV.kobj, &benet_attr_group); ++} +diff --git a/drivers/net/benet/be_proc.c b/drivers/net/benet/be_proc.c +new file mode 100644 +index 0000000..0bfdb3b +--- /dev/null ++++ b/drivers/net/benet/be_proc.c +@@ -0,0 +1,513 @@ ++/* ++ * Copyright (C) 2005 - 2011 ServerEngines ++ * All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License version 2 ++ * as published by the Free Software Foundation. The full GNU General ++ * Public License is included in this distribution in the file called COPYING. ++ * ++ * Contact Information: ++ * linux-drivers@serverengines.com ++ * ++ * ServerEngines ++ * 209 N. Fair Oaks Ave ++ * Sunnyvale, CA 94085 ++ */ ++#include ++#include "be.h" ++ ++char *be_adpt_name[] = { ++ "driver/be2net0", ++ "driver/be2net1", ++ "driver/be2net2", ++ "driver/be2net3", ++ "driver/be2net4", ++ "driver/be2net5", ++ "driver/be2net6", ++ "driver/be2net7" ++}; ++ ++#define MAX_BE_DEVICES 8 ++struct proc_dir_entry *be_proc_dir[MAX_BE_DEVICES]; ++ ++/*File to read Eth Ring Information */ ++#define BE_ETH_RING_FILE "eth_ring" ++#define BE_DRVR_STAT_FILE "drvr_stat" ++ ++/* ++ * this file enables user to read a 32 bit CSR register. ++ * to read 32 bit value of a register at offset 0x1234, ++ * first write the offset 0x1234 (echo "0x1234") in ++ * the file and then read the value from this file. ++ * the written offset is latched until another value is written ++ */ ++#define BE_CSR_R_FILE "csrr" ++/* ++ * this file enables user to write to a 32 bit CSR register. ++ * to write a value 0xdeadbeef to a register at offset 0x1234, ++ * write 0x1234 0xdeadbeef (echo "0x1234 0xdeadbeeb") to ++ * the file. ++ */ ++#define BE_CSR_W_FILE "csrw" ++ ++#define BE_PROC_MODE 0600 ++ ++static char read_eth_ring_buf[4096]; ++static int read_eth_ring_count; ++ ++/* ++ * Get Various Eth Ring Properties ++ */ ++static int proc_eth_read_ring(char *page, char **start, ++ off_t off, int count, int *eof, void *data) ++{ ++ int i, n; ++ char *p = read_eth_ring_buf; ++ struct be_adapter *adapter = (struct be_adapter *) data; ++ ++ if (off == 0) { ++ /* Reset read_eth_ring_count */ ++ read_eth_ring_count = 0; ++ ++ n = sprintf(p, " PhyAddr VirtAddr Size TotalEntries ProducerIndex ConsumerIndex NumUsed\n"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, " ------- -------- ---- ------------ ------------- ------------- -------\n"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, "%s", "EthSendRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, " %7lx %8p %4u %12u %13u %13u %7u \n", ++ (long) adapter->tx_obj.q.dma_mem.dma, ++ (void *)adapter->tx_obj.q.dma_mem.va, ++ (u32) (adapter->tx_obj.q.len * ++ sizeof(struct be_eth_wrb)), ++ adapter->tx_obj.q.len, adapter->tx_obj.q.head, ++ adapter->tx_obj.q.tail, ++ atomic_read(&adapter->tx_obj.q.used)); ++ ++ p += n; ++ read_eth_ring_count += n; ++ ++ /* Get Eth Send Compl Queue Details */ ++ n = sprintf(p, "%s", "EthSendCmplRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, " %7lx %8p %4u %12u %13s %13u %7s\n", ++ (long)adapter->tx_obj.cq.dma_mem.dma, ++ (void *)adapter->tx_obj.cq.dma_mem.va, ++ (u32) (adapter->tx_obj.cq.len * ++ sizeof(struct be_eth_tx_compl)), ++ adapter->tx_obj.cq.len, "NA", ++ adapter->tx_obj.cq.tail, "NA"); ++ ++ p += n; ++ read_eth_ring_count += n; ++ /* Get Eth Rx Queue Details */ ++ n = sprintf(p, "%s", "EthRxRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, " %7lx %8p %4u %12u %13u %13s %7u \n", ++ (long)adapter->rx_obj.q.dma_mem.dma, ++ (void *)adapter->rx_obj.q.dma_mem.va, ++ (u32) (adapter->rx_obj.q.len * ++ sizeof(struct be_eth_rx_d)), ++ adapter->rx_obj.q.len, adapter->rx_obj.q.head,"NA", ++ atomic_read(&adapter->rx_obj.q.used)); ++ p += n; ++ read_eth_ring_count += n; ++ ++ /* Get Eth Unicast Rx Compl Queue Details */ ++ n = sprintf(p, "%s", "EthRxCmplRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, " %7lx %8p %4u %12u %13s %13u %7s\n", ++ (long)adapter->rx_obj.cq.dma_mem.dma, ++ (void *)adapter->rx_obj.cq.dma_mem.va, ++ (u32) (adapter->rx_obj.cq.len * ++ sizeof(struct be_eth_rx_compl)), ++ adapter->rx_obj.cq.len, "NA", ++ adapter->rx_obj.cq.tail, "NA"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ /* Get Eth Event Queue Details */ ++ n = sprintf(p, "%s", "EthTxEventRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, ++ " %7lx %8p %4u %12u %13s %13u %7s\n", ++ (long) adapter->tx_eq.q.dma_mem.dma, ++ (void *)adapter->tx_eq.q.dma_mem.va, ++ (u32) (adapter->tx_eq.q.len * ++ sizeof(struct be_eq_entry)), ++ adapter->tx_eq.q.len, "NA", ++ adapter->tx_eq.q.tail, "NA"); ++ ++ p += n; ++ read_eth_ring_count += n; ++ ++ /* Get Eth Event Queue Details */ ++ n = sprintf(p, "%s", "EthRxEventRing"); ++ p += n; ++ read_eth_ring_count += n; ++ ++ n = sprintf(p, ++ " %7lx %8p %4u %12u %13s %13u %7s\n", ++ (long) adapter->rx_eq.q.dma_mem.dma, ++ (void *)adapter->rx_eq.q.dma_mem.va, ++ (u32) (adapter->rx_eq.q.len * ++ sizeof(struct be_eq_entry)), ++ adapter->rx_eq.q.len, "NA", ++ adapter->rx_eq.q.tail, "NA"); ++ ++ p += n; ++ read_eth_ring_count += n; ++ } ++ ++ *start = page; ++ /* copy whatever we can */ ++ if (count < (read_eth_ring_count - off)) { ++ i = count; ++ *eof = 0; /* More bytes left */ ++ } else { ++ i = read_eth_ring_count - off; ++ *eof = 1; /* Nothing left. indicate EOF */ ++ } ++ ++ memcpy(page, read_eth_ring_buf + off, i); ++ return (i); ++} ++ ++static int proc_eth_write_ring(struct file *file, ++ const char *buffer, unsigned long count, ++ void *data) ++{ ++ return (count); /* we do not support write */ ++} ++ ++/* ++ * read the driver stats. ++ */ ++static int proc_read_drvr_stat(char *page, char **start, ++ off_t off, int count, int *eof, void *data) ++{ ++ int n, lro_cp; ++ char *p = page; ++ struct be_adapter *adapter = (struct be_adapter *) data; ++ struct net_device *netdev = adapter->netdev; ++ ++ if (off == 0) { ++ n = sprintf(p, "interface = %s\n", netdev->name); ++ p += n; ++ n = sprintf(p, "tx_reqs = %d\n", ++ drvr_stats(adapter)->be_tx_reqs); ++ p += n; ++ n = sprintf(p, "tx_stops = %d\n", ++ drvr_stats(adapter)->be_tx_stops); ++ p += n; ++ n = sprintf(p, "fwd_reqs = %d\n", ++ drvr_stats(adapter)->be_fwd_reqs); ++ p += n; ++ n = sprintf(p, "tx_wrbs = %d\n", ++ drvr_stats(adapter)->be_tx_wrbs); ++ p += n; ++ n = sprintf(p, "rx_poll = %d\n", drvr_stats(adapter)->be_rx_polls); ++ p += n; ++ n = sprintf(p, "tx_events = %d\n", ++ drvr_stats(adapter)->be_tx_events); ++ p += n; ++ n = sprintf(p, "rx_events = %d\n", ++ drvr_stats(adapter)->be_rx_events); ++ p += n; ++ n = sprintf(p, "tx_compl = %d\n", ++ drvr_stats(adapter)->be_tx_compl); ++ p += n; ++ n = sprintf(p, "rx_compl = %d\n", ++ drvr_stats(adapter)->be_rx_compl); ++ p += n; ++ n = sprintf(p, "ethrx_post_fail = %d\n", ++ drvr_stats(adapter)->be_ethrx_post_fail); ++ p += n; ++ n = sprintf(p, "802.3_dropped_frames = %d\n", ++ drvr_stats(adapter)->be_802_3_dropped_frames); ++ p += n; ++ n = sprintf(p, "802.3_malformed_frames = %d\n", ++ drvr_stats(adapter)->be_802_3_malformed_frames); ++ p += n; ++ n = sprintf(p, "eth_tx_rate = %d\n", ++ drvr_stats(adapter)->be_tx_rate); ++ p += n; ++ n = sprintf(p, "eth_rx_rate = %d\n", ++ drvr_stats(adapter)->be_rx_rate); ++ p += n; ++ ++ lro_cp = (drvr_stats(adapter)->be_lro_hgram_data[0] + ++ drvr_stats(adapter)->be_lro_hgram_data[1] + ++ drvr_stats(adapter)->be_lro_hgram_data[2] + ++ drvr_stats(adapter)->be_lro_hgram_data[3] + ++ drvr_stats(adapter)->be_lro_hgram_data[4] + ++ drvr_stats(adapter)->be_lro_hgram_data[5] + ++ drvr_stats(adapter)->be_lro_hgram_data[6] + ++ drvr_stats(adapter)->be_lro_hgram_data[7])/100; ++ lro_cp = (lro_cp == 0) ? 1 : lro_cp; /* avoid divide by 0 */ ++ n = sprintf(p, ++ "LRO data count %% histogram (1, 2-3, 4-5,..,>=16) = " ++ "%d, %d, %d, %d - %d, %d, %d, %d\n", ++ drvr_stats(adapter)->be_lro_hgram_data[0]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[1]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[2]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[3]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[4]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[5]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[6]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_data[7]/lro_cp); ++ p += n; ++ ++ lro_cp = (drvr_stats(adapter)->be_lro_hgram_ack[0] + ++ drvr_stats(adapter)->be_lro_hgram_ack[1] + ++ drvr_stats(adapter)->be_lro_hgram_ack[2] + ++ drvr_stats(adapter)->be_lro_hgram_ack[3] + ++ drvr_stats(adapter)->be_lro_hgram_ack[4] + ++ drvr_stats(adapter)->be_lro_hgram_ack[5] + ++ drvr_stats(adapter)->be_lro_hgram_ack[6] + ++ drvr_stats(adapter)->be_lro_hgram_ack[7])/100; ++ lro_cp = (lro_cp == 0) ? 1 : lro_cp; /* avoid divide by 0 */ ++ n = sprintf(p, ++ "LRO ack count %% histogram (1, 2-3, 4-5,..,>=16) = " ++ "%d, %d, %d, %d - %d, %d, %d, %d\n", ++ drvr_stats(adapter)->be_lro_hgram_ack[0]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[1]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[2]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[3]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[4]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[5]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[6]/lro_cp, ++ drvr_stats(adapter)->be_lro_hgram_ack[7]/lro_cp); ++ p += n; ++ n = sprintf(p, "rx_eq_delay = %d \n", adapter->rx_eq.cur_eqd); ++ p += n; ++ n = sprintf(p, "rx frags per sec=%d \n", ++ drvr_stats(adapter)->be_rx_fps); ++ p += n; ++ ++ } ++ *eof = 1; ++ return (p - page); ++} ++ ++static int proc_write_drvr_stat(struct file *file, ++ const char *buffer, unsigned long count, ++ void *data) ++{ ++ struct be_adapter *adapter = (struct be_adapter *) data; ++ ++ memset(&(adapter->stats.drvr_stats), 0, ++ sizeof(adapter->stats.drvr_stats)); ++ return (count); /* we do not support write */ ++} ++ ++#if 0 ++/* the following are some of the functions that are needed here ++ * until all initializations are done by MPU. ++ */ ++ ++u32 ++CsrReadDr(void* BaseAddress, u32 Offset) ++{ ++ u32 *rp; ++ ++ rp = (u32 *) (((u8 *) BaseAddress) + Offset); ++ return (*rp); ++} ++ ++/*! ++ ++@brief ++ This routine writes to a register located within the CSR ++ space for a given function object. ++ ++@param ++ FuncObj - Pointer to the function object to read from. ++ ++@param ++ Offset - The Offset (in bytes) to write to within the function's CSR space. ++ ++@param ++ Value - The value to write to the register. ++ ++@return ++ ++@note ++ IRQL: any ++ ++*/ ++void ++CsrWriteDr(void* BaseAddress, u32 Offset, u32 Value) ++{ ++ u32 *Register; ++ ++ Register = (u32 *) (((u8 *) BaseAddress) + Offset); ++ ++ //TRACE(DL_INFO, "CsrWrite[ %X ] <= %X", Register, Value); ++ *Register = Value; ++} ++u32 be_proc_csrr_offset = -1; /* to latch the offset of next CSR Read req. */ ++ ++/* ++ * read the csr_r file. return the 32 bit register value from ++ * CSR space at offset latched in the global location ++ * be_proc_csrr_offset ++ */ ++static int proc_read_csr_r(char *page, char **start, ++ off_t off, int count, int *eof, void *data) ++{ ++ struct be_adapter * adapter = (struct be_adapter *)data; ++ u32 val; ++ int n = 0; ++ if (be_proc_csrr_offset == -1) ++ return -EINVAL; ++ ++ if (off == 0) { ++ /* read the CSR at offset be_proc_csrr_offset and return */ ++ val = CsrReadDr(adapter->csr_va, be_proc_csrr_offset); ++ n = sprintf(page, "0x%x\n", val); ++ } ++ *eof = 1; ++ return n; ++} ++ ++/* ++ * save the written value in be_proc_csrr_offset for next ++ * read from the file ++ */ ++static int proc_write_csr_r(struct file *file, ++ const char *buffer, unsigned long count, void *data) ++{ ++ char buf[64]; ++ u32 n; ++ ++ if (count > sizeof(buf) + 1) ++ return -EINVAL; ++ if (copy_from_user(buf, buffer, count)) ++ return -EFAULT; ++ buf[count] = '\0'; ++ ++ n = simple_strtoul(buf, NULL, 16); ++ if (n < 0x50000) ++ be_proc_csrr_offset = n; ++ return (count); ++} ++ ++/* ++ * return the latched offset for reading the csr_r file. ++ */ ++static int proc_read_csr_w(char *page, char **start, ++ off_t off, int count, int *eof, void *data) ++{ ++ ++ *eof = 1; ++ return sprintf(page, "0x%x\n", be_proc_csrr_offset); ++} ++ ++/* ++ * the incoming string is of the form " " ++ * where the offset is the offset of the register to be written ++ * and value is the value to be written. ++ */ ++static int proc_write_csr_w(struct file *file, ++ const char *buffer, unsigned long count, ++ void *data) ++{ ++ char buf[64]; ++ char *p; ++ u32 n, val; ++ struct be_adapter * adapter = (struct be_adapter *)data; ++ ++ if (count > sizeof(buf) + 1) ++ return -EINVAL; ++ if (copy_from_user(buf, buffer, count)) ++ return -EFAULT; ++ buf[count] = '\0'; ++ ++ n = simple_strtoul(buf, &p, 16); ++ if (n > 0x50000) ++ return -EINVAL; ++ ++ /* now get the actual value to be written */ ++ while (*p == ' ' || *p == '\t') ++ p++; ++ val = simple_strtoul(p, NULL, 16); ++ CsrWriteDr(adapter->csr_va, n, val); ++ return (count); ++} ++#endif ++ ++void be_init_procfs(struct be_adapter *adapter, int adapt_num) ++{ ++ static struct proc_dir_entry *pde; ++ ++ if (adapt_num > MAX_BE_DEVICES - 1) ++ return; ++ ++ /* create directory */ ++ be_proc_dir[adapt_num] = ++ proc_mkdir(be_adpt_name[adapt_num], NULL); ++ if (be_proc_dir[adapt_num]) { ++ (be_proc_dir[adapt_num])->owner = THIS_MODULE; ++ } ++ ++ pde = create_proc_entry(BE_ETH_RING_FILE, BE_PROC_MODE, ++ be_proc_dir[adapt_num]); ++ if (pde) { ++ pde->read_proc = proc_eth_read_ring; ++ pde->write_proc = proc_eth_write_ring; ++ pde->data = adapter; ++ pde->owner = THIS_MODULE; ++ } ++ ++ pde = create_proc_entry(BE_DRVR_STAT_FILE, BE_PROC_MODE, ++ be_proc_dir[adapt_num]); ++ if (pde) { ++ pde->read_proc = proc_read_drvr_stat; ++ pde->write_proc = proc_write_drvr_stat; ++ pde->data = adapter; ++ pde->owner = THIS_MODULE; ++ } ++ ++#if 0 ++ if ((pde = create_proc_entry(BE_CSR_R_FILE, BE_PROC_MODE, be_proc_dir[adapt_num]))) { ++ pde->read_proc = proc_read_csr_r; ++ pde->write_proc = proc_write_csr_r; ++ pde->data = adapter; ++ pde->owner = THIS_MODULE; ++ } ++ ++ if ((pde = create_proc_entry(BE_CSR_W_FILE, BE_PROC_MODE, be_proc_dir[adapt_num]))) { ++ pde->read_proc = proc_read_csr_w; ++ pde->write_proc = proc_write_csr_w; ++ pde->data = adapter; ++ pde->owner = THIS_MODULE; ++ } ++#endif ++} ++ ++void be_cleanup_procfs(struct be_adapter *adapter, int adapt_num) ++{ ++ if (adapt_num > MAX_BE_DEVICES - 1) ++ return; ++ remove_proc_entry(BE_ETH_RING_FILE, be_proc_dir[adapt_num]); ++ remove_proc_entry(BE_DRVR_STAT_FILE, be_proc_dir[adapt_num]); ++ remove_proc_entry(BE_CSR_R_FILE, be_proc_dir[adapt_num]); ++ remove_proc_entry(BE_CSR_W_FILE, be_proc_dir[adapt_num]); ++ remove_proc_entry(be_adpt_name[adapt_num], NULL); ++} +diff --git a/drivers/net/benet/version.h b/drivers/net/benet/version.h +new file mode 100644 +index 0000000..c7ed692 +--- /dev/null ++++ b/drivers/net/benet/version.h +@@ -0,0 +1,51 @@ ++#define STR_BE_BRANCH "0" ++#define STR_BE_BUILD "479" ++#define STR_BE_DOT "0" ++#define STR_BE_MINOR "0" ++#define STR_BE_MAJOR "4" ++ ++#define BE_BRANCH 0 ++#define BE_BUILD 479 ++#define BE_DOT 0 ++#define BE_MINOR 0 ++#define BE_MAJOR 4 ++ ++#define MGMT_BRANCH 0 ++#define MGMT_BUILDNUM 479 ++#define MGMT_MINOR 0 ++#define MGMT_MAJOR 4 ++ ++#define BE_REDBOOT_VERSION "2.0.5.0" ++ ++//start-auto ++#define BUILD_MONTH "12" ++#define BUILD_MONTH_NAME "December" ++#define BUILD_DAY "6" ++#define BUILD_YEAR "2011" ++#define BUILD_24HOUR "21" ++#define BUILD_12HOUR "9" ++#define BUILD_AM_PM "PM" ++#define BUILD_MIN "48" ++#define BUILD_SEC "05" ++#define BUILD_MONTH_NUMBER 12 ++#define BUILD_DAY_NUMBER 6 ++#define BUILD_YEAR_NUMBER 2011 ++#define BUILD_24HOUR_NUMBER 21 ++#define BUILD_12HOUR_NUMBER 9 ++#define BUILD_MIN_NUMBER 48 ++#define BUILD_SEC_NUMBER 5 ++#undef MAJOR_BUILD ++#undef MINOR_BUILD ++#undef DOT_BUILD ++#define NUMBERED_BUILD ++#undef BRANCH_BUILD ++//end-auto ++ ++#define ELX_FCOE_XROM_BIOS_VER "7.03a1" ++#define ELX_FCoE_X86_VER "4.02a1" ++#define ELX_FCoE_EFI_VER "5.01a1" ++#define ELX_FCoE_FCODE_VER "4.01a0" ++#define ELX_PXE_BIOS_VER "3.00a5" ++#define ELX_UEFI_NIC_VER "2.10A10" ++#define ELX_UEFI_FCODE_VER "1.10A0" ++#define ELX_ISCSI_BIOS_VER "1.00A8" +diff --git a/drivers/net/bnx2.c b/drivers/net/bnx2.c +index 4874b2b..67f8526 100644 +--- a/drivers/net/bnx2.c ++++ b/drivers/net/bnx2.c +@@ -5809,6 +5809,8 @@ bnx2_test_nvram(struct bnx2 *bp) + int rc = 0; + u32 magic, csum; + ++ pax_track_stack(); ++ + if ((rc = bnx2_nvram_read(bp, 0, data, 4)) != 0) + goto test_nvram_done; + +diff --git a/drivers/net/cxgb3/l2t.h b/drivers/net/cxgb3/l2t.h +index fd3eb07..8a6978d 100644 +--- a/drivers/net/cxgb3/l2t.h ++++ b/drivers/net/cxgb3/l2t.h +@@ -86,7 +86,7 @@ typedef void (*arp_failure_handler_func)(struct t3cdev * dev, + */ + struct l2t_skb_cb { + arp_failure_handler_func arp_failure_handler; +-}; ++} __no_const; + + #define L2T_SKB_CB(skb) ((struct l2t_skb_cb *)(skb)->cb) + +diff --git a/drivers/net/cxgb3/t3_hw.c b/drivers/net/cxgb3/t3_hw.c +index 032cfe0..411af379 100644 +--- a/drivers/net/cxgb3/t3_hw.c ++++ b/drivers/net/cxgb3/t3_hw.c +@@ -699,6 +699,8 @@ static int get_vpd_params(struct adapter *adapter, struct vpd_params *p) + int i, addr, ret; + struct t3_vpd vpd; + ++ pax_track_stack(); ++ + /* + * Card information is normally at VPD_BASE but some early cards had + * it at 0. +diff --git a/drivers/net/e1000e/82571.c b/drivers/net/e1000e/82571.c +index d1e0563..b9e129c 100644 +--- a/drivers/net/e1000e/82571.c ++++ b/drivers/net/e1000e/82571.c +@@ -212,7 +212,7 @@ static s32 e1000_init_mac_params_82571(struct e1000_adapter *adapter) + { + struct e1000_hw *hw = &adapter->hw; + struct e1000_mac_info *mac = &hw->mac; +- struct e1000_mac_operations *func = &mac->ops; ++ e1000_mac_operations_no_const *func = &mac->ops; + u32 swsm = 0; + u32 swsm2 = 0; + bool force_clear_smbi = false; +@@ -1656,7 +1656,7 @@ static void e1000_clear_hw_cntrs_82571(struct e1000_hw *hw) + temp = er32(ICRXDMTC); + } + +-static struct e1000_mac_operations e82571_mac_ops = { ++static const struct e1000_mac_operations e82571_mac_ops = { + /* .check_mng_mode: mac type dependent */ + /* .check_for_link: media type dependent */ + .id_led_init = e1000e_id_led_init, +@@ -1674,7 +1674,7 @@ static struct e1000_mac_operations e82571_mac_ops = { + .setup_led = e1000e_setup_led_generic, + }; + +-static struct e1000_phy_operations e82_phy_ops_igp = { ++static const struct e1000_phy_operations e82_phy_ops_igp = { + .acquire_phy = e1000_get_hw_semaphore_82571, + .check_reset_block = e1000e_check_reset_block_generic, + .commit_phy = NULL, +@@ -1691,7 +1691,7 @@ static struct e1000_phy_operations e82_phy_ops_igp = { + .cfg_on_link_up = NULL, + }; + +-static struct e1000_phy_operations e82_phy_ops_m88 = { ++static const struct e1000_phy_operations e82_phy_ops_m88 = { + .acquire_phy = e1000_get_hw_semaphore_82571, + .check_reset_block = e1000e_check_reset_block_generic, + .commit_phy = e1000e_phy_sw_reset, +@@ -1708,7 +1708,7 @@ static struct e1000_phy_operations e82_phy_ops_m88 = { + .cfg_on_link_up = NULL, + }; + +-static struct e1000_phy_operations e82_phy_ops_bm = { ++static const struct e1000_phy_operations e82_phy_ops_bm = { + .acquire_phy = e1000_get_hw_semaphore_82571, + .check_reset_block = e1000e_check_reset_block_generic, + .commit_phy = e1000e_phy_sw_reset, +@@ -1725,7 +1725,7 @@ static struct e1000_phy_operations e82_phy_ops_bm = { + .cfg_on_link_up = NULL, + }; + +-static struct e1000_nvm_operations e82571_nvm_ops = { ++static const struct e1000_nvm_operations e82571_nvm_ops = { + .acquire_nvm = e1000_acquire_nvm_82571, + .read_nvm = e1000e_read_nvm_eerd, + .release_nvm = e1000_release_nvm_82571, +diff --git a/drivers/net/e1000e/e1000.h b/drivers/net/e1000e/e1000.h +index 47db9bd..fa58ccd 100644 +--- a/drivers/net/e1000e/e1000.h ++++ b/drivers/net/e1000e/e1000.h +@@ -375,9 +375,9 @@ struct e1000_info { + u32 pba; + u32 max_hw_frame_size; + s32 (*get_variants)(struct e1000_adapter *); +- struct e1000_mac_operations *mac_ops; +- struct e1000_phy_operations *phy_ops; +- struct e1000_nvm_operations *nvm_ops; ++ const struct e1000_mac_operations *mac_ops; ++ const struct e1000_phy_operations *phy_ops; ++ const struct e1000_nvm_operations *nvm_ops; + }; + + /* hardware capability, feature, and workaround flags */ +diff --git a/drivers/net/e1000e/es2lan.c b/drivers/net/e1000e/es2lan.c +index ae5d736..e9a93a1 100644 +--- a/drivers/net/e1000e/es2lan.c ++++ b/drivers/net/e1000e/es2lan.c +@@ -207,7 +207,7 @@ static s32 e1000_init_mac_params_80003es2lan(struct e1000_adapter *adapter) + { + struct e1000_hw *hw = &adapter->hw; + struct e1000_mac_info *mac = &hw->mac; +- struct e1000_mac_operations *func = &mac->ops; ++ e1000_mac_operations_no_const *func = &mac->ops; + + /* Set media type */ + switch (adapter->pdev->device) { +@@ -1365,7 +1365,7 @@ static void e1000_clear_hw_cntrs_80003es2lan(struct e1000_hw *hw) + temp = er32(ICRXDMTC); + } + +-static struct e1000_mac_operations es2_mac_ops = { ++static const struct e1000_mac_operations es2_mac_ops = { + .id_led_init = e1000e_id_led_init, + .check_mng_mode = e1000e_check_mng_mode_generic, + /* check_for_link dependent on media type */ +@@ -1383,7 +1383,7 @@ static struct e1000_mac_operations es2_mac_ops = { + .setup_led = e1000e_setup_led_generic, + }; + +-static struct e1000_phy_operations es2_phy_ops = { ++static const struct e1000_phy_operations es2_phy_ops = { + .acquire_phy = e1000_acquire_phy_80003es2lan, + .check_reset_block = e1000e_check_reset_block_generic, + .commit_phy = e1000e_phy_sw_reset, +@@ -1400,7 +1400,7 @@ static struct e1000_phy_operations es2_phy_ops = { + .cfg_on_link_up = e1000_cfg_on_link_up_80003es2lan, + }; + +-static struct e1000_nvm_operations es2_nvm_ops = { ++static const struct e1000_nvm_operations es2_nvm_ops = { + .acquire_nvm = e1000_acquire_nvm_80003es2lan, + .read_nvm = e1000e_read_nvm_eerd, + .release_nvm = e1000_release_nvm_80003es2lan, +diff --git a/drivers/net/e1000e/hw.h b/drivers/net/e1000e/hw.h +index 11f3b7c..6381887 100644 +--- a/drivers/net/e1000e/hw.h ++++ b/drivers/net/e1000e/hw.h +@@ -753,6 +753,7 @@ struct e1000_mac_operations { + s32 (*setup_physical_interface)(struct e1000_hw *); + s32 (*setup_led)(struct e1000_hw *); + }; ++typedef struct e1000_mac_operations __no_const e1000_mac_operations_no_const; + + /* Function pointers for the PHY. */ + struct e1000_phy_operations { +@@ -774,6 +775,7 @@ struct e1000_phy_operations { + s32 (*write_phy_reg_locked)(struct e1000_hw *, u32, u16); + s32 (*cfg_on_link_up)(struct e1000_hw *); + }; ++typedef struct e1000_phy_operations __no_const e1000_phy_operations_no_const; + + /* Function pointers for the NVM. */ + struct e1000_nvm_operations { +@@ -785,9 +787,10 @@ struct e1000_nvm_operations { + s32 (*validate_nvm)(struct e1000_hw *); + s32 (*write_nvm)(struct e1000_hw *, u16, u16, u16 *); + }; ++typedef struct e1000_nvm_operations __no_const e1000_nvm_operations_no_const; + + struct e1000_mac_info { +- struct e1000_mac_operations ops; ++ e1000_mac_operations_no_const ops; + + u8 addr[6]; + u8 perm_addr[6]; +@@ -823,7 +826,7 @@ struct e1000_mac_info { + }; + + struct e1000_phy_info { +- struct e1000_phy_operations ops; ++ e1000_phy_operations_no_const ops; + + enum e1000_phy_type type; + +@@ -857,7 +860,7 @@ struct e1000_phy_info { + }; + + struct e1000_nvm_info { +- struct e1000_nvm_operations ops; ++ e1000_nvm_operations_no_const ops; + + enum e1000_nvm_type type; + enum e1000_nvm_override override; +diff --git a/drivers/net/e1000e/ich8lan.c b/drivers/net/e1000e/ich8lan.c +index de39f9a..e28d3e0 100644 +--- a/drivers/net/e1000e/ich8lan.c ++++ b/drivers/net/e1000e/ich8lan.c +@@ -3463,7 +3463,7 @@ static void e1000_clear_hw_cntrs_ich8lan(struct e1000_hw *hw) + } + } + +-static struct e1000_mac_operations ich8_mac_ops = { ++static const struct e1000_mac_operations ich8_mac_ops = { + .id_led_init = e1000e_id_led_init, + .check_mng_mode = e1000_check_mng_mode_ich8lan, + .check_for_link = e1000_check_for_copper_link_ich8lan, +@@ -3481,7 +3481,7 @@ static struct e1000_mac_operations ich8_mac_ops = { + /* id_led_init dependent on mac type */ + }; + +-static struct e1000_phy_operations ich8_phy_ops = { ++static const struct e1000_phy_operations ich8_phy_ops = { + .acquire_phy = e1000_acquire_swflag_ich8lan, + .check_reset_block = e1000_check_reset_block_ich8lan, + .commit_phy = NULL, +@@ -3497,7 +3497,7 @@ static struct e1000_phy_operations ich8_phy_ops = { + .write_phy_reg = e1000e_write_phy_reg_igp, + }; + +-static struct e1000_nvm_operations ich8_nvm_ops = { ++static const struct e1000_nvm_operations ich8_nvm_ops = { + .acquire_nvm = e1000_acquire_nvm_ich8lan, + .read_nvm = e1000_read_nvm_ich8lan, + .release_nvm = e1000_release_nvm_ich8lan, +diff --git a/drivers/net/fealnx.c b/drivers/net/fealnx.c +index 18d5fbb..542d96d 100644 +--- a/drivers/net/fealnx.c ++++ b/drivers/net/fealnx.c +@@ -151,7 +151,7 @@ struct chip_info { + int flags; + }; + +-static const struct chip_info skel_netdrv_tbl[] __devinitdata = { ++static const struct chip_info skel_netdrv_tbl[] __devinitconst = { + { "100/10M Ethernet PCI Adapter", HAS_MII_XCVR }, + { "100/10M Ethernet PCI Adapter", HAS_CHIP_XCVR }, + { "1000/100/10M Ethernet PCI Adapter", HAS_MII_XCVR }, +diff --git a/drivers/net/hamradio/6pack.c b/drivers/net/hamradio/6pack.c +index 0e5b54b..b503f82 100644 +--- a/drivers/net/hamradio/6pack.c ++++ b/drivers/net/hamradio/6pack.c +@@ -461,6 +461,8 @@ static void sixpack_receive_buf(struct tty_struct *tty, + unsigned char buf[512]; + int count1; + ++ pax_track_stack(); ++ + if (!count) + return; + +diff --git a/drivers/net/ibmveth.c b/drivers/net/ibmveth.c +index 5862282..7cce8cb 100644 +--- a/drivers/net/ibmveth.c ++++ b/drivers/net/ibmveth.c +@@ -1577,7 +1577,7 @@ static struct attribute * veth_pool_attrs[] = { + NULL, + }; + +-static struct sysfs_ops veth_pool_ops = { ++static const struct sysfs_ops veth_pool_ops = { + .show = veth_pool_show, + .store = veth_pool_store, + }; +diff --git a/drivers/net/igb/e1000_82575.c b/drivers/net/igb/e1000_82575.c +index d617f2d..57b5309 100644 +--- a/drivers/net/igb/e1000_82575.c ++++ b/drivers/net/igb/e1000_82575.c +@@ -1411,7 +1411,7 @@ void igb_vmdq_set_replication_pf(struct e1000_hw *hw, bool enable) + wr32(E1000_VT_CTL, vt_ctl); + } + +-static struct e1000_mac_operations e1000_mac_ops_82575 = { ++static const struct e1000_mac_operations e1000_mac_ops_82575 = { + .reset_hw = igb_reset_hw_82575, + .init_hw = igb_init_hw_82575, + .check_for_link = igb_check_for_link_82575, +@@ -1420,13 +1420,13 @@ static struct e1000_mac_operations e1000_mac_ops_82575 = { + .get_speed_and_duplex = igb_get_speed_and_duplex_copper, + }; + +-static struct e1000_phy_operations e1000_phy_ops_82575 = { ++static const struct e1000_phy_operations e1000_phy_ops_82575 = { + .acquire = igb_acquire_phy_82575, + .get_cfg_done = igb_get_cfg_done_82575, + .release = igb_release_phy_82575, + }; + +-static struct e1000_nvm_operations e1000_nvm_ops_82575 = { ++static const struct e1000_nvm_operations e1000_nvm_ops_82575 = { + .acquire = igb_acquire_nvm_82575, + .read = igb_read_nvm_eerd, + .release = igb_release_nvm_82575, +diff --git a/drivers/net/igb/e1000_hw.h b/drivers/net/igb/e1000_hw.h +index 72081df..d855cf5 100644 +--- a/drivers/net/igb/e1000_hw.h ++++ b/drivers/net/igb/e1000_hw.h +@@ -288,6 +288,7 @@ struct e1000_mac_operations { + s32 (*read_mac_addr)(struct e1000_hw *); + s32 (*get_speed_and_duplex)(struct e1000_hw *, u16 *, u16 *); + }; ++typedef struct e1000_mac_operations __no_const e1000_mac_operations_no_const; + + struct e1000_phy_operations { + s32 (*acquire)(struct e1000_hw *); +@@ -303,6 +304,7 @@ struct e1000_phy_operations { + s32 (*set_d3_lplu_state)(struct e1000_hw *, bool); + s32 (*write_reg)(struct e1000_hw *, u32, u16); + }; ++typedef struct e1000_phy_operations __no_const e1000_phy_operations_no_const; + + struct e1000_nvm_operations { + s32 (*acquire)(struct e1000_hw *); +@@ -310,6 +312,7 @@ struct e1000_nvm_operations { + void (*release)(struct e1000_hw *); + s32 (*write)(struct e1000_hw *, u16, u16, u16 *); + }; ++typedef struct e1000_nvm_operations __no_const e1000_nvm_operations_no_const; + + struct e1000_info { + s32 (*get_invariants)(struct e1000_hw *); +@@ -321,7 +324,7 @@ struct e1000_info { + extern const struct e1000_info e1000_82575_info; + + struct e1000_mac_info { +- struct e1000_mac_operations ops; ++ e1000_mac_operations_no_const ops; + + u8 addr[6]; + u8 perm_addr[6]; +@@ -365,7 +368,7 @@ struct e1000_mac_info { + }; + + struct e1000_phy_info { +- struct e1000_phy_operations ops; ++ e1000_phy_operations_no_const ops; + + enum e1000_phy_type type; + +@@ -400,7 +403,7 @@ struct e1000_phy_info { + }; + + struct e1000_nvm_info { +- struct e1000_nvm_operations ops; ++ e1000_nvm_operations_no_const ops; + + enum e1000_nvm_type type; + enum e1000_nvm_override override; +@@ -446,6 +449,7 @@ struct e1000_mbx_operations { + s32 (*check_for_ack)(struct e1000_hw *, u16); + s32 (*check_for_rst)(struct e1000_hw *, u16); + }; ++typedef struct e1000_mbx_operations __no_const e1000_mbx_operations_no_const; + + struct e1000_mbx_stats { + u32 msgs_tx; +@@ -457,7 +461,7 @@ struct e1000_mbx_stats { + }; + + struct e1000_mbx_info { +- struct e1000_mbx_operations ops; ++ e1000_mbx_operations_no_const ops; + struct e1000_mbx_stats stats; + u32 timeout; + u32 usec_delay; +diff --git a/drivers/net/igbvf/vf.h b/drivers/net/igbvf/vf.h +index 1e8ce37..549c453 100644 +--- a/drivers/net/igbvf/vf.h ++++ b/drivers/net/igbvf/vf.h +@@ -187,9 +187,10 @@ struct e1000_mac_operations { + s32 (*read_mac_addr)(struct e1000_hw *); + s32 (*set_vfta)(struct e1000_hw *, u16, bool); + }; ++typedef struct e1000_mac_operations __no_const e1000_mac_operations_no_const; + + struct e1000_mac_info { +- struct e1000_mac_operations ops; ++ e1000_mac_operations_no_const ops; + u8 addr[6]; + u8 perm_addr[6]; + +@@ -211,6 +212,7 @@ struct e1000_mbx_operations { + s32 (*check_for_ack)(struct e1000_hw *); + s32 (*check_for_rst)(struct e1000_hw *); + }; ++typedef struct e1000_mbx_operations __no_const e1000_mbx_operations_no_const; + + struct e1000_mbx_stats { + u32 msgs_tx; +@@ -222,7 +224,7 @@ struct e1000_mbx_stats { + }; + + struct e1000_mbx_info { +- struct e1000_mbx_operations ops; ++ e1000_mbx_operations_no_const ops; + struct e1000_mbx_stats stats; + u32 timeout; + u32 usec_delay; +diff --git a/drivers/net/iseries_veth.c b/drivers/net/iseries_veth.c +index aa7286b..a61394f 100644 +--- a/drivers/net/iseries_veth.c ++++ b/drivers/net/iseries_veth.c +@@ -384,7 +384,7 @@ static struct attribute *veth_cnx_default_attrs[] = { + NULL + }; + +-static struct sysfs_ops veth_cnx_sysfs_ops = { ++static const struct sysfs_ops veth_cnx_sysfs_ops = { + .show = veth_cnx_attribute_show + }; + +@@ -441,7 +441,7 @@ static struct attribute *veth_port_default_attrs[] = { + NULL + }; + +-static struct sysfs_ops veth_port_sysfs_ops = { ++static const struct sysfs_ops veth_port_sysfs_ops = { + .show = veth_port_attribute_show + }; + +diff --git a/drivers/net/ixgb/ixgb_main.c b/drivers/net/ixgb/ixgb_main.c +index 8aa44dc..fa1e797 100644 +--- a/drivers/net/ixgb/ixgb_main.c ++++ b/drivers/net/ixgb/ixgb_main.c +@@ -1052,6 +1052,8 @@ ixgb_set_multi(struct net_device *netdev) + u32 rctl; + int i; + ++ pax_track_stack(); ++ + /* Check for Promiscuous and All Multicast modes */ + + rctl = IXGB_READ_REG(hw, RCTL); +diff --git a/drivers/net/ixgb/ixgb_param.c b/drivers/net/ixgb/ixgb_param.c +index af35e1d..8781785 100644 +--- a/drivers/net/ixgb/ixgb_param.c ++++ b/drivers/net/ixgb/ixgb_param.c +@@ -260,6 +260,9 @@ void __devinit + ixgb_check_options(struct ixgb_adapter *adapter) + { + int bd = adapter->bd_number; ++ ++ pax_track_stack(); ++ + if (bd >= IXGB_MAX_NIC) { + printk(KERN_NOTICE + "Warning: no configuration for board #%i\n", bd); +diff --git a/drivers/net/ixgbe/ixgbe_type.h b/drivers/net/ixgbe/ixgbe_type.h +index b17aa73..ed74540 100644 +--- a/drivers/net/ixgbe/ixgbe_type.h ++++ b/drivers/net/ixgbe/ixgbe_type.h +@@ -2327,6 +2327,7 @@ struct ixgbe_eeprom_operations { + s32 (*validate_checksum)(struct ixgbe_hw *, u16 *); + s32 (*update_checksum)(struct ixgbe_hw *); + }; ++typedef struct ixgbe_eeprom_operations __no_const ixgbe_eeprom_operations_no_const; + + struct ixgbe_mac_operations { + s32 (*init_hw)(struct ixgbe_hw *); +@@ -2376,6 +2377,7 @@ struct ixgbe_mac_operations { + /* Flow Control */ + s32 (*fc_enable)(struct ixgbe_hw *, s32); + }; ++typedef struct ixgbe_mac_operations __no_const ixgbe_mac_operations_no_const; + + struct ixgbe_phy_operations { + s32 (*identify)(struct ixgbe_hw *); +@@ -2394,9 +2396,10 @@ struct ixgbe_phy_operations { + s32 (*read_i2c_eeprom)(struct ixgbe_hw *, u8 , u8 *); + s32 (*write_i2c_eeprom)(struct ixgbe_hw *, u8, u8); + }; ++typedef struct ixgbe_phy_operations __no_const ixgbe_phy_operations_no_const; + + struct ixgbe_eeprom_info { +- struct ixgbe_eeprom_operations ops; ++ ixgbe_eeprom_operations_no_const ops; + enum ixgbe_eeprom_type type; + u32 semaphore_delay; + u16 word_size; +@@ -2404,7 +2407,7 @@ struct ixgbe_eeprom_info { + }; + + struct ixgbe_mac_info { +- struct ixgbe_mac_operations ops; ++ ixgbe_mac_operations_no_const ops; + enum ixgbe_mac_type type; + u8 addr[IXGBE_ETH_LENGTH_OF_ADDRESS]; + u8 perm_addr[IXGBE_ETH_LENGTH_OF_ADDRESS]; +@@ -2423,7 +2426,7 @@ struct ixgbe_mac_info { + }; + + struct ixgbe_phy_info { +- struct ixgbe_phy_operations ops; ++ ixgbe_phy_operations_no_const ops; + struct mdio_if_info mdio; + enum ixgbe_phy_type type; + u32 id; +diff --git a/drivers/net/mlx4/main.c b/drivers/net/mlx4/main.c +index 291a505..2543756 100644 +--- a/drivers/net/mlx4/main.c ++++ b/drivers/net/mlx4/main.c +@@ -38,6 +38,7 @@ + #include + #include + #include ++#include + + #include + #include +@@ -730,6 +731,8 @@ static int mlx4_init_hca(struct mlx4_dev *dev) + u64 icm_size; + int err; + ++ pax_track_stack(); ++ + err = mlx4_QUERY_FW(dev); + if (err) { + if (err == -EACCES) +diff --git a/drivers/net/niu.c b/drivers/net/niu.c +index 2dce134..fa5ce75 100644 +--- a/drivers/net/niu.c ++++ b/drivers/net/niu.c +@@ -9128,6 +9128,8 @@ static void __devinit niu_try_msix(struct niu *np, u8 *ldg_num_map) + int i, num_irqs, err; + u8 first_ldg; + ++ pax_track_stack(); ++ + first_ldg = (NIU_NUM_LDG / parent->num_ports) * np->port; + for (i = 0; i < (NIU_NUM_LDG / parent->num_ports); i++) + ldg_num_map[i] = first_ldg + i; +diff --git a/drivers/net/pcnet32.c b/drivers/net/pcnet32.c +index c1b3f09..97cd8c4 100644 +--- a/drivers/net/pcnet32.c ++++ b/drivers/net/pcnet32.c +@@ -79,7 +79,7 @@ static int cards_found; + /* + * VLB I/O addresses + */ +-static unsigned int pcnet32_portlist[] __initdata = ++static unsigned int pcnet32_portlist[] __devinitdata = + { 0x300, 0x320, 0x340, 0x360, 0 }; + + static int pcnet32_debug = 0; +@@ -267,7 +267,7 @@ struct pcnet32_private { + struct sk_buff **rx_skbuff; + dma_addr_t *tx_dma_addr; + dma_addr_t *rx_dma_addr; +- struct pcnet32_access a; ++ struct pcnet32_access *a; + spinlock_t lock; /* Guard lock */ + unsigned int cur_rx, cur_tx; /* The next free ring entry */ + unsigned int rx_ring_size; /* current rx ring size */ +@@ -457,9 +457,9 @@ static void pcnet32_netif_start(struct net_device *dev) + u16 val; + + netif_wake_queue(dev); +- val = lp->a.read_csr(ioaddr, CSR3); ++ val = lp->a->read_csr(ioaddr, CSR3); + val &= 0x00ff; +- lp->a.write_csr(ioaddr, CSR3, val); ++ lp->a->write_csr(ioaddr, CSR3, val); + napi_enable(&lp->napi); + } + +@@ -744,7 +744,7 @@ static u32 pcnet32_get_link(struct net_device *dev) + r = mii_link_ok(&lp->mii_if); + } else if (lp->chip_version >= PCNET32_79C970A) { + ulong ioaddr = dev->base_addr; /* card base I/O address */ +- r = (lp->a.read_bcr(ioaddr, 4) != 0xc0); ++ r = (lp->a->read_bcr(ioaddr, 4) != 0xc0); + } else { /* can not detect link on really old chips */ + r = 1; + } +@@ -806,7 +806,7 @@ static int pcnet32_set_ringparam(struct net_device *dev, + pcnet32_netif_stop(dev); + + spin_lock_irqsave(&lp->lock, flags); +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); /* stop the chip */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); /* stop the chip */ + + size = min(ering->tx_pending, (unsigned int)TX_MAX_RING_SIZE); + +@@ -886,7 +886,7 @@ static void pcnet32_ethtool_test(struct net_device *dev, + static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + { + struct pcnet32_private *lp = netdev_priv(dev); +- struct pcnet32_access *a = &lp->a; /* access to registers */ ++ struct pcnet32_access *a = lp->a; /* access to registers */ + ulong ioaddr = dev->base_addr; /* card base I/O address */ + struct sk_buff *skb; /* sk buff */ + int x, i; /* counters */ +@@ -906,21 +906,21 @@ static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + pcnet32_netif_stop(dev); + + spin_lock_irqsave(&lp->lock, flags); +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); /* stop the chip */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); /* stop the chip */ + + numbuffs = min(numbuffs, (int)min(lp->rx_ring_size, lp->tx_ring_size)); + + /* Reset the PCNET32 */ +- lp->a.reset(ioaddr); +- lp->a.write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ ++ lp->a->reset(ioaddr); ++ lp->a->write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ + + /* switch pcnet32 to 32bit mode */ +- lp->a.write_bcr(ioaddr, 20, 2); ++ lp->a->write_bcr(ioaddr, 20, 2); + + /* purge & init rings but don't actually restart */ + pcnet32_restart(dev, 0x0000); + +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); /* Set STOP bit */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); /* Set STOP bit */ + + /* Initialize Transmit buffers. */ + size = data_len + 15; +@@ -966,10 +966,10 @@ static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + + /* set int loopback in CSR15 */ + x = a->read_csr(ioaddr, CSR15) & 0xfffc; +- lp->a.write_csr(ioaddr, CSR15, x | 0x0044); ++ lp->a->write_csr(ioaddr, CSR15, x | 0x0044); + + teststatus = cpu_to_le16(0x8000); +- lp->a.write_csr(ioaddr, CSR0, CSR0_START); /* Set STRT bit */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_START); /* Set STRT bit */ + + /* Check status of descriptors */ + for (x = 0; x < numbuffs; x++) { +@@ -990,7 +990,7 @@ static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + } + } + +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); /* Set STOP bit */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); /* Set STOP bit */ + wmb(); + if (netif_msg_hw(lp) && netif_msg_pktdata(lp)) { + printk(KERN_DEBUG "%s: RX loopback packets:\n", dev->name); +@@ -1039,7 +1039,7 @@ static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + pcnet32_restart(dev, CSR0_NORMAL); + } else { + pcnet32_purge_rx_ring(dev); +- lp->a.write_bcr(ioaddr, 20, 4); /* return to 16bit mode */ ++ lp->a->write_bcr(ioaddr, 20, 4); /* return to 16bit mode */ + } + spin_unlock_irqrestore(&lp->lock, flags); + +@@ -1049,7 +1049,7 @@ static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1) + static void pcnet32_led_blink_callback(struct net_device *dev) + { + struct pcnet32_private *lp = netdev_priv(dev); +- struct pcnet32_access *a = &lp->a; ++ struct pcnet32_access *a = lp->a; + ulong ioaddr = dev->base_addr; + unsigned long flags; + int i; +@@ -1066,7 +1066,7 @@ static void pcnet32_led_blink_callback(struct net_device *dev) + static int pcnet32_phys_id(struct net_device *dev, u32 data) + { + struct pcnet32_private *lp = netdev_priv(dev); +- struct pcnet32_access *a = &lp->a; ++ struct pcnet32_access *a = lp->a; + ulong ioaddr = dev->base_addr; + unsigned long flags; + int i, regs[4]; +@@ -1112,7 +1112,7 @@ static int pcnet32_suspend(struct net_device *dev, unsigned long *flags, + { + int csr5; + struct pcnet32_private *lp = netdev_priv(dev); +- struct pcnet32_access *a = &lp->a; ++ struct pcnet32_access *a = lp->a; + ulong ioaddr = dev->base_addr; + int ticks; + +@@ -1388,8 +1388,8 @@ static int pcnet32_poll(struct napi_struct *napi, int budget) + spin_lock_irqsave(&lp->lock, flags); + if (pcnet32_tx(dev)) { + /* reset the chip to clear the error condition, then restart */ +- lp->a.reset(ioaddr); +- lp->a.write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ ++ lp->a->reset(ioaddr); ++ lp->a->write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ + pcnet32_restart(dev, CSR0_START); + netif_wake_queue(dev); + } +@@ -1401,12 +1401,12 @@ static int pcnet32_poll(struct napi_struct *napi, int budget) + __napi_complete(napi); + + /* clear interrupt masks */ +- val = lp->a.read_csr(ioaddr, CSR3); ++ val = lp->a->read_csr(ioaddr, CSR3); + val &= 0x00ff; +- lp->a.write_csr(ioaddr, CSR3, val); ++ lp->a->write_csr(ioaddr, CSR3, val); + + /* Set interrupt enable. */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_INTEN); + + spin_unlock_irqrestore(&lp->lock, flags); + } +@@ -1429,7 +1429,7 @@ static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs, + int i, csr0; + u16 *buff = ptr; + struct pcnet32_private *lp = netdev_priv(dev); +- struct pcnet32_access *a = &lp->a; ++ struct pcnet32_access *a = lp->a; + ulong ioaddr = dev->base_addr; + unsigned long flags; + +@@ -1466,9 +1466,9 @@ static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs, + for (j = 0; j < PCNET32_MAX_PHYS; j++) { + if (lp->phymask & (1 << j)) { + for (i = 0; i < PCNET32_REGS_PER_PHY; i++) { +- lp->a.write_bcr(ioaddr, 33, ++ lp->a->write_bcr(ioaddr, 33, + (j << 5) | i); +- *buff++ = lp->a.read_bcr(ioaddr, 34); ++ *buff++ = lp->a->read_bcr(ioaddr, 34); + } + } + } +@@ -1858,7 +1858,7 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) + ((cards_found >= MAX_UNITS) || full_duplex[cards_found])) + lp->options |= PCNET32_PORT_FD; + +- lp->a = *a; ++ lp->a = a; + + /* prior to register_netdev, dev->name is not yet correct */ + if (pcnet32_alloc_ring(dev, pci_name(lp->pci_dev))) { +@@ -1917,7 +1917,7 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) + if (lp->mii) { + /* lp->phycount and lp->phymask are set to 0 by memset above */ + +- lp->mii_if.phy_id = ((lp->a.read_bcr(ioaddr, 33)) >> 5) & 0x1f; ++ lp->mii_if.phy_id = ((lp->a->read_bcr(ioaddr, 33)) >> 5) & 0x1f; + /* scan for PHYs */ + for (i = 0; i < PCNET32_MAX_PHYS; i++) { + unsigned short id1, id2; +@@ -1938,7 +1938,7 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) + "Found PHY %04x:%04x at address %d.\n", + id1, id2, i); + } +- lp->a.write_bcr(ioaddr, 33, (lp->mii_if.phy_id) << 5); ++ lp->a->write_bcr(ioaddr, 33, (lp->mii_if.phy_id) << 5); + if (lp->phycount > 1) { + lp->options |= PCNET32_PORT_MII; + } +@@ -2109,10 +2109,10 @@ static int pcnet32_open(struct net_device *dev) + } + + /* Reset the PCNET32 */ +- lp->a.reset(ioaddr); ++ lp->a->reset(ioaddr); + + /* switch pcnet32 to 32bit mode */ +- lp->a.write_bcr(ioaddr, 20, 2); ++ lp->a->write_bcr(ioaddr, 20, 2); + + if (netif_msg_ifup(lp)) + printk(KERN_DEBUG +@@ -2122,14 +2122,14 @@ static int pcnet32_open(struct net_device *dev) + (u32) (lp->init_dma_addr)); + + /* set/reset autoselect bit */ +- val = lp->a.read_bcr(ioaddr, 2) & ~2; ++ val = lp->a->read_bcr(ioaddr, 2) & ~2; + if (lp->options & PCNET32_PORT_ASEL) + val |= 2; +- lp->a.write_bcr(ioaddr, 2, val); ++ lp->a->write_bcr(ioaddr, 2, val); + + /* handle full duplex setting */ + if (lp->mii_if.full_duplex) { +- val = lp->a.read_bcr(ioaddr, 9) & ~3; ++ val = lp->a->read_bcr(ioaddr, 9) & ~3; + if (lp->options & PCNET32_PORT_FD) { + val |= 1; + if (lp->options == (PCNET32_PORT_FD | PCNET32_PORT_AUI)) +@@ -2139,14 +2139,14 @@ static int pcnet32_open(struct net_device *dev) + if (lp->chip_version == 0x2627) + val |= 3; + } +- lp->a.write_bcr(ioaddr, 9, val); ++ lp->a->write_bcr(ioaddr, 9, val); + } + + /* set/reset GPSI bit in test register */ +- val = lp->a.read_csr(ioaddr, 124) & ~0x10; ++ val = lp->a->read_csr(ioaddr, 124) & ~0x10; + if ((lp->options & PCNET32_PORT_PORTSEL) == PCNET32_PORT_GPSI) + val |= 0x10; +- lp->a.write_csr(ioaddr, 124, val); ++ lp->a->write_csr(ioaddr, 124, val); + + /* Allied Telesyn AT 2700/2701 FX are 100Mbit only and do not negotiate */ + if (pdev && pdev->subsystem_vendor == PCI_VENDOR_ID_AT && +@@ -2167,24 +2167,24 @@ static int pcnet32_open(struct net_device *dev) + * duplex, and/or enable auto negotiation, and clear DANAS + */ + if (lp->mii && !(lp->options & PCNET32_PORT_ASEL)) { +- lp->a.write_bcr(ioaddr, 32, +- lp->a.read_bcr(ioaddr, 32) | 0x0080); ++ lp->a->write_bcr(ioaddr, 32, ++ lp->a->read_bcr(ioaddr, 32) | 0x0080); + /* disable Auto Negotiation, set 10Mpbs, HD */ +- val = lp->a.read_bcr(ioaddr, 32) & ~0xb8; ++ val = lp->a->read_bcr(ioaddr, 32) & ~0xb8; + if (lp->options & PCNET32_PORT_FD) + val |= 0x10; + if (lp->options & PCNET32_PORT_100) + val |= 0x08; +- lp->a.write_bcr(ioaddr, 32, val); ++ lp->a->write_bcr(ioaddr, 32, val); + } else { + if (lp->options & PCNET32_PORT_ASEL) { +- lp->a.write_bcr(ioaddr, 32, +- lp->a.read_bcr(ioaddr, ++ lp->a->write_bcr(ioaddr, 32, ++ lp->a->read_bcr(ioaddr, + 32) | 0x0080); + /* enable auto negotiate, setup, disable fd */ +- val = lp->a.read_bcr(ioaddr, 32) & ~0x98; ++ val = lp->a->read_bcr(ioaddr, 32) & ~0x98; + val |= 0x20; +- lp->a.write_bcr(ioaddr, 32, val); ++ lp->a->write_bcr(ioaddr, 32, val); + } + } + } else { +@@ -2197,10 +2197,10 @@ static int pcnet32_open(struct net_device *dev) + * There is really no good other way to handle multiple PHYs + * other than turning off all automatics + */ +- val = lp->a.read_bcr(ioaddr, 2); +- lp->a.write_bcr(ioaddr, 2, val & ~2); +- val = lp->a.read_bcr(ioaddr, 32); +- lp->a.write_bcr(ioaddr, 32, val & ~(1 << 7)); /* stop MII manager */ ++ val = lp->a->read_bcr(ioaddr, 2); ++ lp->a->write_bcr(ioaddr, 2, val & ~2); ++ val = lp->a->read_bcr(ioaddr, 32); ++ lp->a->write_bcr(ioaddr, 32, val & ~(1 << 7)); /* stop MII manager */ + + if (!(lp->options & PCNET32_PORT_ASEL)) { + /* setup ecmd */ +@@ -2210,7 +2210,7 @@ static int pcnet32_open(struct net_device *dev) + ecmd.speed = + lp-> + options & PCNET32_PORT_100 ? SPEED_100 : SPEED_10; +- bcr9 = lp->a.read_bcr(ioaddr, 9); ++ bcr9 = lp->a->read_bcr(ioaddr, 9); + + if (lp->options & PCNET32_PORT_FD) { + ecmd.duplex = DUPLEX_FULL; +@@ -2219,7 +2219,7 @@ static int pcnet32_open(struct net_device *dev) + ecmd.duplex = DUPLEX_HALF; + bcr9 |= ~(1 << 0); + } +- lp->a.write_bcr(ioaddr, 9, bcr9); ++ lp->a->write_bcr(ioaddr, 9, bcr9); + } + + for (i = 0; i < PCNET32_MAX_PHYS; i++) { +@@ -2252,9 +2252,9 @@ static int pcnet32_open(struct net_device *dev) + + #ifdef DO_DXSUFLO + if (lp->dxsuflo) { /* Disable transmit stop on underflow */ +- val = lp->a.read_csr(ioaddr, CSR3); ++ val = lp->a->read_csr(ioaddr, CSR3); + val |= 0x40; +- lp->a.write_csr(ioaddr, CSR3, val); ++ lp->a->write_csr(ioaddr, CSR3, val); + } + #endif + +@@ -2270,11 +2270,11 @@ static int pcnet32_open(struct net_device *dev) + napi_enable(&lp->napi); + + /* Re-initialize the PCNET32, and start it when done. */ +- lp->a.write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff)); +- lp->a.write_csr(ioaddr, 2, (lp->init_dma_addr >> 16)); ++ lp->a->write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff)); ++ lp->a->write_csr(ioaddr, 2, (lp->init_dma_addr >> 16)); + +- lp->a.write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_INIT); ++ lp->a->write_csr(ioaddr, CSR4, 0x0915); /* auto tx pad */ ++ lp->a->write_csr(ioaddr, CSR0, CSR0_INIT); + + netif_start_queue(dev); + +@@ -2286,20 +2286,20 @@ static int pcnet32_open(struct net_device *dev) + + i = 0; + while (i++ < 100) +- if (lp->a.read_csr(ioaddr, CSR0) & CSR0_IDON) ++ if (lp->a->read_csr(ioaddr, CSR0) & CSR0_IDON) + break; + /* + * We used to clear the InitDone bit, 0x0100, here but Mark Stockton + * reports that doing so triggers a bug in the '974. + */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_NORMAL); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_NORMAL); + + if (netif_msg_ifup(lp)) + printk(KERN_DEBUG + "%s: pcnet32 open after %d ticks, init block %#x csr0 %4.4x.\n", + dev->name, i, + (u32) (lp->init_dma_addr), +- lp->a.read_csr(ioaddr, CSR0)); ++ lp->a->read_csr(ioaddr, CSR0)); + + spin_unlock_irqrestore(&lp->lock, flags); + +@@ -2313,7 +2313,7 @@ static int pcnet32_open(struct net_device *dev) + * Switch back to 16bit mode to avoid problems with dumb + * DOS packet driver after a warm reboot + */ +- lp->a.write_bcr(ioaddr, 20, 4); ++ lp->a->write_bcr(ioaddr, 20, 4); + + err_free_irq: + spin_unlock_irqrestore(&lp->lock, flags); +@@ -2420,7 +2420,7 @@ static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits) + + /* wait for stop */ + for (i = 0; i < 100; i++) +- if (lp->a.read_csr(ioaddr, CSR0) & CSR0_STOP) ++ if (lp->a->read_csr(ioaddr, CSR0) & CSR0_STOP) + break; + + if (i >= 100 && netif_msg_drv(lp)) +@@ -2433,13 +2433,13 @@ static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits) + return; + + /* ReInit Ring */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_INIT); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_INIT); + i = 0; + while (i++ < 1000) +- if (lp->a.read_csr(ioaddr, CSR0) & CSR0_IDON) ++ if (lp->a->read_csr(ioaddr, CSR0) & CSR0_IDON) + break; + +- lp->a.write_csr(ioaddr, CSR0, csr0_bits); ++ lp->a->write_csr(ioaddr, CSR0, csr0_bits); + } + + static void pcnet32_tx_timeout(struct net_device *dev) +@@ -2452,8 +2452,8 @@ static void pcnet32_tx_timeout(struct net_device *dev) + if (pcnet32_debug & NETIF_MSG_DRV) + printk(KERN_ERR + "%s: transmit timed out, status %4.4x, resetting.\n", +- dev->name, lp->a.read_csr(ioaddr, CSR0)); +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); ++ dev->name, lp->a->read_csr(ioaddr, CSR0)); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); + dev->stats.tx_errors++; + if (netif_msg_tx_err(lp)) { + int i; +@@ -2497,7 +2497,7 @@ static netdev_tx_t pcnet32_start_xmit(struct sk_buff *skb, + if (netif_msg_tx_queued(lp)) { + printk(KERN_DEBUG + "%s: pcnet32_start_xmit() called, csr0 %4.4x.\n", +- dev->name, lp->a.read_csr(ioaddr, CSR0)); ++ dev->name, lp->a->read_csr(ioaddr, CSR0)); + } + + /* Default status -- will not enable Successful-TxDone +@@ -2528,7 +2528,7 @@ static netdev_tx_t pcnet32_start_xmit(struct sk_buff *skb, + dev->stats.tx_bytes += skb->len; + + /* Trigger an immediate send poll. */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN | CSR0_TXPOLL); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_INTEN | CSR0_TXPOLL); + + dev->trans_start = jiffies; + +@@ -2555,18 +2555,18 @@ pcnet32_interrupt(int irq, void *dev_id) + + spin_lock(&lp->lock); + +- csr0 = lp->a.read_csr(ioaddr, CSR0); ++ csr0 = lp->a->read_csr(ioaddr, CSR0); + while ((csr0 & 0x8f00) && --boguscnt >= 0) { + if (csr0 == 0xffff) { + break; /* PCMCIA remove happened */ + } + /* Acknowledge all of the current interrupt sources ASAP. */ +- lp->a.write_csr(ioaddr, CSR0, csr0 & ~0x004f); ++ lp->a->write_csr(ioaddr, CSR0, csr0 & ~0x004f); + + if (netif_msg_intr(lp)) + printk(KERN_DEBUG + "%s: interrupt csr0=%#2.2x new csr=%#2.2x.\n", +- dev->name, csr0, lp->a.read_csr(ioaddr, CSR0)); ++ dev->name, csr0, lp->a->read_csr(ioaddr, CSR0)); + + /* Log misc errors. */ + if (csr0 & 0x4000) +@@ -2595,19 +2595,19 @@ pcnet32_interrupt(int irq, void *dev_id) + if (napi_schedule_prep(&lp->napi)) { + u16 val; + /* set interrupt masks */ +- val = lp->a.read_csr(ioaddr, CSR3); ++ val = lp->a->read_csr(ioaddr, CSR3); + val |= 0x5f00; +- lp->a.write_csr(ioaddr, CSR3, val); ++ lp->a->write_csr(ioaddr, CSR3, val); + + __napi_schedule(&lp->napi); + break; + } +- csr0 = lp->a.read_csr(ioaddr, CSR0); ++ csr0 = lp->a->read_csr(ioaddr, CSR0); + } + + if (netif_msg_intr(lp)) + printk(KERN_DEBUG "%s: exiting interrupt, csr0=%#4.4x.\n", +- dev->name, lp->a.read_csr(ioaddr, CSR0)); ++ dev->name, lp->a->read_csr(ioaddr, CSR0)); + + spin_unlock(&lp->lock); + +@@ -2627,21 +2627,21 @@ static int pcnet32_close(struct net_device *dev) + + spin_lock_irqsave(&lp->lock, flags); + +- dev->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112); ++ dev->stats.rx_missed_errors = lp->a->read_csr(ioaddr, 112); + + if (netif_msg_ifdown(lp)) + printk(KERN_DEBUG + "%s: Shutting down ethercard, status was %2.2x.\n", +- dev->name, lp->a.read_csr(ioaddr, CSR0)); ++ dev->name, lp->a->read_csr(ioaddr, CSR0)); + + /* We stop the PCNET32 here -- it occasionally polls memory if we don't. */ +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); + + /* + * Switch back to 16bit mode to avoid problems with dumb + * DOS packet driver after a warm reboot + */ +- lp->a.write_bcr(ioaddr, 20, 4); ++ lp->a->write_bcr(ioaddr, 20, 4); + + spin_unlock_irqrestore(&lp->lock, flags); + +@@ -2664,7 +2664,7 @@ static struct net_device_stats *pcnet32_get_stats(struct net_device *dev) + unsigned long flags; + + spin_lock_irqsave(&lp->lock, flags); +- dev->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112); ++ dev->stats.rx_missed_errors = lp->a->read_csr(ioaddr, 112); + spin_unlock_irqrestore(&lp->lock, flags); + + return &dev->stats; +@@ -2686,10 +2686,10 @@ static void pcnet32_load_multicast(struct net_device *dev) + if (dev->flags & IFF_ALLMULTI) { + ib->filter[0] = cpu_to_le32(~0U); + ib->filter[1] = cpu_to_le32(~0U); +- lp->a.write_csr(ioaddr, PCNET32_MC_FILTER, 0xffff); +- lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+1, 0xffff); +- lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+2, 0xffff); +- lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+3, 0xffff); ++ lp->a->write_csr(ioaddr, PCNET32_MC_FILTER, 0xffff); ++ lp->a->write_csr(ioaddr, PCNET32_MC_FILTER+1, 0xffff); ++ lp->a->write_csr(ioaddr, PCNET32_MC_FILTER+2, 0xffff); ++ lp->a->write_csr(ioaddr, PCNET32_MC_FILTER+3, 0xffff); + return; + } + /* clear the multicast filter */ +@@ -2710,7 +2710,7 @@ static void pcnet32_load_multicast(struct net_device *dev) + mcast_table[crc >> 4] |= cpu_to_le16(1 << (crc & 0xf)); + } + for (i = 0; i < 4; i++) +- lp->a.write_csr(ioaddr, PCNET32_MC_FILTER + i, ++ lp->a->write_csr(ioaddr, PCNET32_MC_FILTER + i, + le16_to_cpu(mcast_table[i])); + return; + } +@@ -2726,7 +2726,7 @@ static void pcnet32_set_multicast_list(struct net_device *dev) + + spin_lock_irqsave(&lp->lock, flags); + suspended = pcnet32_suspend(dev, &flags, 0); +- csr15 = lp->a.read_csr(ioaddr, CSR15); ++ csr15 = lp->a->read_csr(ioaddr, CSR15); + if (dev->flags & IFF_PROMISC) { + /* Log any net taps. */ + if (netif_msg_hw(lp)) +@@ -2735,21 +2735,21 @@ static void pcnet32_set_multicast_list(struct net_device *dev) + lp->init_block->mode = + cpu_to_le16(0x8000 | (lp->options & PCNET32_PORT_PORTSEL) << + 7); +- lp->a.write_csr(ioaddr, CSR15, csr15 | 0x8000); ++ lp->a->write_csr(ioaddr, CSR15, csr15 | 0x8000); + } else { + lp->init_block->mode = + cpu_to_le16((lp->options & PCNET32_PORT_PORTSEL) << 7); +- lp->a.write_csr(ioaddr, CSR15, csr15 & 0x7fff); ++ lp->a->write_csr(ioaddr, CSR15, csr15 & 0x7fff); + pcnet32_load_multicast(dev); + } + + if (suspended) { + int csr5; + /* clear SUSPEND (SPND) - CSR5 bit 0 */ +- csr5 = lp->a.read_csr(ioaddr, CSR5); +- lp->a.write_csr(ioaddr, CSR5, csr5 & (~CSR5_SUSPEND)); ++ csr5 = lp->a->read_csr(ioaddr, CSR5); ++ lp->a->write_csr(ioaddr, CSR5, csr5 & (~CSR5_SUSPEND)); + } else { +- lp->a.write_csr(ioaddr, CSR0, CSR0_STOP); ++ lp->a->write_csr(ioaddr, CSR0, CSR0_STOP); + pcnet32_restart(dev, CSR0_NORMAL); + netif_wake_queue(dev); + } +@@ -2767,8 +2767,8 @@ static int mdio_read(struct net_device *dev, int phy_id, int reg_num) + if (!lp->mii) + return 0; + +- lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); +- val_out = lp->a.read_bcr(ioaddr, 34); ++ lp->a->write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); ++ val_out = lp->a->read_bcr(ioaddr, 34); + + return val_out; + } +@@ -2782,8 +2782,8 @@ static void mdio_write(struct net_device *dev, int phy_id, int reg_num, int val) + if (!lp->mii) + return; + +- lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); +- lp->a.write_bcr(ioaddr, 34, val); ++ lp->a->write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); ++ lp->a->write_bcr(ioaddr, 34, val); + } + + static int pcnet32_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) +@@ -2862,7 +2862,7 @@ static void pcnet32_check_media(struct net_device *dev, int verbose) + curr_link = mii_link_ok(&lp->mii_if); + } else { + ulong ioaddr = dev->base_addr; /* card base I/O address */ +- curr_link = (lp->a.read_bcr(ioaddr, 4) != 0xc0); ++ curr_link = (lp->a->read_bcr(ioaddr, 4) != 0xc0); + } + if (!curr_link) { + if (prev_link || verbose) { +@@ -2887,13 +2887,13 @@ static void pcnet32_check_media(struct net_device *dev, int verbose) + (ecmd.duplex == + DUPLEX_FULL) ? "full" : "half"); + } +- bcr9 = lp->a.read_bcr(dev->base_addr, 9); ++ bcr9 = lp->a->read_bcr(dev->base_addr, 9); + if ((bcr9 & (1 << 0)) != lp->mii_if.full_duplex) { + if (lp->mii_if.full_duplex) + bcr9 |= (1 << 0); + else + bcr9 &= ~(1 << 0); +- lp->a.write_bcr(dev->base_addr, 9, bcr9); ++ lp->a->write_bcr(dev->base_addr, 9, bcr9); + } + } else { + if (netif_msg_link(lp)) +diff --git a/drivers/net/sis190.c b/drivers/net/sis190.c +index 7cc9898..6eb50d3 100644 +--- a/drivers/net/sis190.c ++++ b/drivers/net/sis190.c +@@ -1598,7 +1598,7 @@ static int __devinit sis190_get_mac_addr_from_eeprom(struct pci_dev *pdev, + static int __devinit sis190_get_mac_addr_from_apc(struct pci_dev *pdev, + struct net_device *dev) + { +- static const u16 __devinitdata ids[] = { 0x0965, 0x0966, 0x0968 }; ++ static const u16 __devinitconst ids[] = { 0x0965, 0x0966, 0x0968 }; + struct sis190_private *tp = netdev_priv(dev); + struct pci_dev *isa_bridge; + u8 reg, tmp8; +diff --git a/drivers/net/sundance.c b/drivers/net/sundance.c +index e13685a..60c948c 100644 +--- a/drivers/net/sundance.c ++++ b/drivers/net/sundance.c +@@ -225,7 +225,7 @@ enum { + struct pci_id_info { + const char *name; + }; +-static const struct pci_id_info pci_id_tbl[] __devinitdata = { ++static const struct pci_id_info pci_id_tbl[] __devinitconst = { + {"D-Link DFE-550TX FAST Ethernet Adapter"}, + {"D-Link DFE-550FX 100Mbps Fiber-optics Adapter"}, + {"D-Link DFE-580TX 4 port Server Adapter"}, +diff --git a/drivers/net/tg3.h b/drivers/net/tg3.h +index 529f55a..cccaa18 100644 +--- a/drivers/net/tg3.h ++++ b/drivers/net/tg3.h +@@ -95,6 +95,7 @@ + #define CHIPREV_ID_5750_A0 0x4000 + #define CHIPREV_ID_5750_A1 0x4001 + #define CHIPREV_ID_5750_A3 0x4003 ++#define CHIPREV_ID_5750_C1 0x4201 + #define CHIPREV_ID_5750_C2 0x4202 + #define CHIPREV_ID_5752_A0_HW 0x5000 + #define CHIPREV_ID_5752_A0 0x6000 +diff --git a/drivers/net/tokenring/abyss.c b/drivers/net/tokenring/abyss.c +index b9db1b5..720f9ce 100644 +--- a/drivers/net/tokenring/abyss.c ++++ b/drivers/net/tokenring/abyss.c +@@ -451,10 +451,12 @@ static struct pci_driver abyss_driver = { + + static int __init abyss_init (void) + { +- abyss_netdev_ops = tms380tr_netdev_ops; ++ pax_open_kernel(); ++ memcpy((void *)&abyss_netdev_ops, &tms380tr_netdev_ops, sizeof(tms380tr_netdev_ops)); + +- abyss_netdev_ops.ndo_open = abyss_open; +- abyss_netdev_ops.ndo_stop = abyss_close; ++ *(void **)&abyss_netdev_ops.ndo_open = abyss_open; ++ *(void **)&abyss_netdev_ops.ndo_stop = abyss_close; ++ pax_close_kernel(); + + return pci_register_driver(&abyss_driver); + } +diff --git a/drivers/net/tokenring/madgemc.c b/drivers/net/tokenring/madgemc.c +index 456f8bf..373e56d 100644 +--- a/drivers/net/tokenring/madgemc.c ++++ b/drivers/net/tokenring/madgemc.c +@@ -755,9 +755,11 @@ static struct mca_driver madgemc_driver = { + + static int __init madgemc_init (void) + { +- madgemc_netdev_ops = tms380tr_netdev_ops; +- madgemc_netdev_ops.ndo_open = madgemc_open; +- madgemc_netdev_ops.ndo_stop = madgemc_close; ++ pax_open_kernel(); ++ memcpy((void *)&madgemc_netdev_ops, &tms380tr_netdev_ops, sizeof(tms380tr_netdev_ops)); ++ *(void **)&madgemc_netdev_ops.ndo_open = madgemc_open; ++ *(void **)&madgemc_netdev_ops.ndo_stop = madgemc_close; ++ pax_close_kernel(); + + return mca_register_driver (&madgemc_driver); + } +diff --git a/drivers/net/tokenring/proteon.c b/drivers/net/tokenring/proteon.c +index 16e8783..925bd49 100644 +--- a/drivers/net/tokenring/proteon.c ++++ b/drivers/net/tokenring/proteon.c +@@ -353,9 +353,11 @@ static int __init proteon_init(void) + struct platform_device *pdev; + int i, num = 0, err = 0; + +- proteon_netdev_ops = tms380tr_netdev_ops; +- proteon_netdev_ops.ndo_open = proteon_open; +- proteon_netdev_ops.ndo_stop = tms380tr_close; ++ pax_open_kernel(); ++ memcpy((void *)&proteon_netdev_ops, &tms380tr_netdev_ops, sizeof(tms380tr_netdev_ops)); ++ *(void **)&proteon_netdev_ops.ndo_open = proteon_open; ++ *(void **)&proteon_netdev_ops.ndo_stop = tms380tr_close; ++ pax_close_kernel(); + + err = platform_driver_register(&proteon_driver); + if (err) +diff --git a/drivers/net/tokenring/skisa.c b/drivers/net/tokenring/skisa.c +index 46db5c5..37c1536 100644 +--- a/drivers/net/tokenring/skisa.c ++++ b/drivers/net/tokenring/skisa.c +@@ -363,9 +363,11 @@ static int __init sk_isa_init(void) + struct platform_device *pdev; + int i, num = 0, err = 0; + +- sk_isa_netdev_ops = tms380tr_netdev_ops; +- sk_isa_netdev_ops.ndo_open = sk_isa_open; +- sk_isa_netdev_ops.ndo_stop = tms380tr_close; ++ pax_open_kernel(); ++ memcpy((void *)&sk_isa_netdev_ops, &tms380tr_netdev_ops, sizeof(tms380tr_netdev_ops)); ++ *(void **)&sk_isa_netdev_ops.ndo_open = sk_isa_open; ++ *(void **)&sk_isa_netdev_ops.ndo_stop = tms380tr_close; ++ pax_close_kernel(); + + err = platform_driver_register(&sk_isa_driver); + if (err) +diff --git a/drivers/net/tulip/de2104x.c b/drivers/net/tulip/de2104x.c +index 74e5ba4..5cf6bc9 100644 +--- a/drivers/net/tulip/de2104x.c ++++ b/drivers/net/tulip/de2104x.c +@@ -1785,6 +1785,8 @@ static void __devinit de21041_get_srom_info (struct de_private *de) + struct de_srom_info_leaf *il; + void *bufp; + ++ pax_track_stack(); ++ + /* download entire eeprom */ + for (i = 0; i < DE_EEPROM_WORDS; i++) + ((__le16 *)ee_data)[i] = +diff --git a/drivers/net/tulip/de4x5.c b/drivers/net/tulip/de4x5.c +index a8349b7..90f9dfe 100644 +--- a/drivers/net/tulip/de4x5.c ++++ b/drivers/net/tulip/de4x5.c +@@ -5472,7 +5472,7 @@ de4x5_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) + for (i=0; idev_addr[i]; + } +- if (copy_to_user(ioc->data, tmp.addr, ioc->len)) return -EFAULT; ++ if (ioc->len > sizeof tmp.addr || copy_to_user(ioc->data, tmp.addr, ioc->len)) return -EFAULT; + break; + + case DE4X5_SET_HWADDR: /* Set the hardware address */ +@@ -5512,7 +5512,7 @@ de4x5_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) + spin_lock_irqsave(&lp->lock, flags); + memcpy(&statbuf, &lp->pktStats, ioc->len); + spin_unlock_irqrestore(&lp->lock, flags); +- if (copy_to_user(ioc->data, &statbuf, ioc->len)) ++ if (ioc->len > sizeof statbuf || copy_to_user(ioc->data, &statbuf, ioc->len)) + return -EFAULT; + break; + } +diff --git a/drivers/net/tulip/eeprom.c b/drivers/net/tulip/eeprom.c +index 391acd3..56d11cd 100644 +--- a/drivers/net/tulip/eeprom.c ++++ b/drivers/net/tulip/eeprom.c +@@ -80,7 +80,7 @@ static struct eeprom_fixup eeprom_fixups[] __devinitdata = { + {NULL}}; + + +-static const char *block_name[] __devinitdata = { ++static const char *block_name[] __devinitconst = { + "21140 non-MII", + "21140 MII PHY", + "21142 Serial PHY", +diff --git a/drivers/net/tulip/winbond-840.c b/drivers/net/tulip/winbond-840.c +index b38d3b7..b1cff23 100644 +--- a/drivers/net/tulip/winbond-840.c ++++ b/drivers/net/tulip/winbond-840.c +@@ -235,7 +235,7 @@ struct pci_id_info { + int drv_flags; /* Driver use, intended as capability flags. */ + }; + +-static const struct pci_id_info pci_id_tbl[] __devinitdata = { ++static const struct pci_id_info pci_id_tbl[] __devinitconst = { + { /* Sometime a Level-One switch card. */ + "Winbond W89c840", CanHaveMII | HasBrokenTx | FDXOnNoMII}, + { "Winbond W89c840", CanHaveMII | HasBrokenTx}, +diff --git a/drivers/net/usb/hso.c b/drivers/net/usb/hso.c +index f450bc9..2b747c8 100644 +--- a/drivers/net/usb/hso.c ++++ b/drivers/net/usb/hso.c +@@ -71,7 +71,7 @@ + #include + #include + #include +- ++#include + + #define DRIVER_VERSION "1.2" + #define MOD_AUTHOR "Option Wireless" +@@ -258,7 +258,7 @@ struct hso_serial { + + /* from usb_serial_port */ + struct tty_struct *tty; +- int open_count; ++ local_t open_count; + spinlock_t serial_lock; + + int (*write_data) (struct hso_serial *serial); +@@ -1180,7 +1180,7 @@ static void put_rxbuf_data_and_resubmit_ctrl_urb(struct hso_serial *serial) + struct urb *urb; + + urb = serial->rx_urb[0]; +- if (serial->open_count > 0) { ++ if (local_read(&serial->open_count) > 0) { + count = put_rxbuf_data(urb, serial); + if (count == -1) + return; +@@ -1216,7 +1216,7 @@ static void hso_std_serial_read_bulk_callback(struct urb *urb) + DUMP1(urb->transfer_buffer, urb->actual_length); + + /* Anyone listening? */ +- if (serial->open_count == 0) ++ if (local_read(&serial->open_count) == 0) + return; + + if (status == 0) { +@@ -1311,8 +1311,7 @@ static int hso_serial_open(struct tty_struct *tty, struct file *filp) + spin_unlock_irq(&serial->serial_lock); + + /* check for port already opened, if not set the termios */ +- serial->open_count++; +- if (serial->open_count == 1) { ++ if (local_inc_return(&serial->open_count) == 1) { + tty->low_latency = 1; + serial->rx_state = RX_IDLE; + /* Force default termio settings */ +@@ -1325,7 +1324,7 @@ static int hso_serial_open(struct tty_struct *tty, struct file *filp) + result = hso_start_serial_device(serial->parent, GFP_KERNEL); + if (result) { + hso_stop_serial_device(serial->parent); +- serial->open_count--; ++ local_dec(&serial->open_count); + kref_put(&serial->parent->ref, hso_serial_ref_free); + } + } else { +@@ -1362,10 +1361,10 @@ static void hso_serial_close(struct tty_struct *tty, struct file *filp) + + /* reset the rts and dtr */ + /* do the actual close */ +- serial->open_count--; ++ local_dec(&serial->open_count); + +- if (serial->open_count <= 0) { +- serial->open_count = 0; ++ if (local_read(&serial->open_count) <= 0) { ++ local_set(&serial->open_count, 0); + spin_lock_irq(&serial->serial_lock); + if (serial->tty == tty) { + serial->tty->driver_data = NULL; +@@ -1447,7 +1446,7 @@ static void hso_serial_set_termios(struct tty_struct *tty, struct ktermios *old) + + /* the actual setup */ + spin_lock_irqsave(&serial->serial_lock, flags); +- if (serial->open_count) ++ if (local_read(&serial->open_count)) + _hso_serial_set_termios(tty, old); + else + tty->termios = old; +@@ -3097,7 +3096,7 @@ static int hso_resume(struct usb_interface *iface) + /* Start all serial ports */ + for (i = 0; i < HSO_SERIAL_TTY_MINORS; i++) { + if (serial_table[i] && (serial_table[i]->interface == iface)) { +- if (dev2ser(serial_table[i])->open_count) { ++ if (local_read(&dev2ser(serial_table[i])->open_count)) { + result = + hso_start_serial_device(serial_table[i], GFP_NOIO); + hso_kick_transmit(dev2ser(serial_table[i])); +diff --git a/drivers/net/vxge/vxge-config.h b/drivers/net/vxge/vxge-config.h +index 3e94f0c..ffdd926 100644 +--- a/drivers/net/vxge/vxge-config.h ++++ b/drivers/net/vxge/vxge-config.h +@@ -474,7 +474,7 @@ struct vxge_hw_uld_cbs { + void (*link_down)(struct __vxge_hw_device *devh); + void (*crit_err)(struct __vxge_hw_device *devh, + enum vxge_hw_event type, u64 ext_data); +-}; ++} __no_const; + + /* + * struct __vxge_hw_blockpool_entry - Block private data structure +diff --git a/drivers/net/vxge/vxge-main.c b/drivers/net/vxge/vxge-main.c +index 068d7a9..35293de 100644 +--- a/drivers/net/vxge/vxge-main.c ++++ b/drivers/net/vxge/vxge-main.c +@@ -93,6 +93,8 @@ static inline void VXGE_COMPLETE_VPATH_TX(struct vxge_fifo *fifo) + struct sk_buff *completed[NR_SKB_COMPLETED]; + int more; + ++ pax_track_stack(); ++ + do { + more = 0; + skb_ptr = completed; +@@ -1779,6 +1781,8 @@ static enum vxge_hw_status vxge_rth_configure(struct vxgedev *vdev) + u8 mtable[256] = {0}; /* CPU to vpath mapping */ + int index; + ++ pax_track_stack(); ++ + /* + * Filling + * - itable with bucket numbers +diff --git a/drivers/net/vxge/vxge-traffic.h b/drivers/net/vxge/vxge-traffic.h +index 461742b..81be42e 100644 +--- a/drivers/net/vxge/vxge-traffic.h ++++ b/drivers/net/vxge/vxge-traffic.h +@@ -2123,7 +2123,7 @@ struct vxge_hw_mempool_cbs { + struct vxge_hw_mempool_dma *dma_object, + u32 index, + u32 is_last); +-}; ++} __no_const; + + void + __vxge_hw_mempool_destroy( +diff --git a/drivers/net/wan/cycx_x25.c b/drivers/net/wan/cycx_x25.c +index cd8cb95..4153b79 100644 +--- a/drivers/net/wan/cycx_x25.c ++++ b/drivers/net/wan/cycx_x25.c +@@ -1017,6 +1017,8 @@ static void hex_dump(char *msg, unsigned char *p, int len) + unsigned char hex[1024], + * phex = hex; + ++ pax_track_stack(); ++ + if (len >= (sizeof(hex) / 2)) + len = (sizeof(hex) / 2) - 1; + +diff --git a/drivers/net/wan/hdlc_x25.c b/drivers/net/wan/hdlc_x25.c +index aa9248f..a4e3c3b 100644 +--- a/drivers/net/wan/hdlc_x25.c ++++ b/drivers/net/wan/hdlc_x25.c +@@ -136,16 +136,16 @@ static netdev_tx_t x25_xmit(struct sk_buff *skb, struct net_device *dev) + + static int x25_open(struct net_device *dev) + { +- struct lapb_register_struct cb; ++ static struct lapb_register_struct cb = { ++ .connect_confirmation = x25_connected, ++ .connect_indication = x25_connected, ++ .disconnect_confirmation = x25_disconnected, ++ .disconnect_indication = x25_disconnected, ++ .data_indication = x25_data_indication, ++ .data_transmit = x25_data_transmit ++ }; + int result; + +- cb.connect_confirmation = x25_connected; +- cb.connect_indication = x25_connected; +- cb.disconnect_confirmation = x25_disconnected; +- cb.disconnect_indication = x25_disconnected; +- cb.data_indication = x25_data_indication; +- cb.data_transmit = x25_data_transmit; +- + result = lapb_register(dev, &cb); + if (result != LAPB_OK) + return result; +diff --git a/drivers/net/wimax/i2400m/usb-fw.c b/drivers/net/wimax/i2400m/usb-fw.c +index 5ad287c..783b020 100644 +--- a/drivers/net/wimax/i2400m/usb-fw.c ++++ b/drivers/net/wimax/i2400m/usb-fw.c +@@ -263,6 +263,8 @@ ssize_t i2400mu_bus_bm_wait_for_ack(struct i2400m *i2400m, + int do_autopm = 1; + DECLARE_COMPLETION_ONSTACK(notif_completion); + ++ pax_track_stack(); ++ + d_fnstart(8, dev, "(i2400m %p ack %p size %zu)\n", + i2400m, ack, ack_size); + BUG_ON(_ack == i2400m->bm_ack_buf); +diff --git a/drivers/net/wireless/airo.c b/drivers/net/wireless/airo.c +index 6c26840..62c97c3 100644 +--- a/drivers/net/wireless/airo.c ++++ b/drivers/net/wireless/airo.c +@@ -3003,6 +3003,8 @@ static void airo_process_scan_results (struct airo_info *ai) { + BSSListElement * loop_net; + BSSListElement * tmp_net; + ++ pax_track_stack(); ++ + /* Blow away current list of scan results */ + list_for_each_entry_safe (loop_net, tmp_net, &ai->network_list, list) { + list_move_tail (&loop_net->list, &ai->network_free_list); +@@ -3783,6 +3785,8 @@ static u16 setup_card(struct airo_info *ai, u8 *mac, int lock) + WepKeyRid wkr; + int rc; + ++ pax_track_stack(); ++ + memset( &mySsid, 0, sizeof( mySsid ) ); + kfree (ai->flash); + ai->flash = NULL; +@@ -4758,6 +4762,8 @@ static int proc_stats_rid_open( struct inode *inode, + __le32 *vals = stats.vals; + int len; + ++ pax_track_stack(); ++ + if ((file->private_data = kzalloc(sizeof(struct proc_data ), GFP_KERNEL)) == NULL) + return -ENOMEM; + data = (struct proc_data *)file->private_data; +@@ -5487,6 +5493,8 @@ static int proc_BSSList_open( struct inode *inode, struct file *file ) { + /* If doLoseSync is not 1, we won't do a Lose Sync */ + int doLoseSync = -1; + ++ pax_track_stack(); ++ + if ((file->private_data = kzalloc(sizeof(struct proc_data ), GFP_KERNEL)) == NULL) + return -ENOMEM; + data = (struct proc_data *)file->private_data; +@@ -7193,6 +7201,8 @@ static int airo_get_aplist(struct net_device *dev, + int i; + int loseSync = capable(CAP_NET_ADMIN) ? 1: -1; + ++ pax_track_stack(); ++ + qual = kmalloc(IW_MAX_AP * sizeof(*qual), GFP_KERNEL); + if (!qual) + return -ENOMEM; +@@ -7753,6 +7763,8 @@ static void airo_read_wireless_stats(struct airo_info *local) + CapabilityRid cap_rid; + __le32 *vals = stats_rid.vals; + ++ pax_track_stack(); ++ + /* Get stats out of the card */ + clear_bit(JOB_WSTATS, &local->jobs); + if (local->power.event) { +diff --git a/drivers/net/wireless/ath/ath5k/debug.c b/drivers/net/wireless/ath/ath5k/debug.c +index 747508c..82e965d 100644 +--- a/drivers/net/wireless/ath/ath5k/debug.c ++++ b/drivers/net/wireless/ath/ath5k/debug.c +@@ -205,6 +205,8 @@ static ssize_t read_file_beacon(struct file *file, char __user *user_buf, + unsigned int v; + u64 tsf; + ++ pax_track_stack(); ++ + v = ath5k_hw_reg_read(sc->ah, AR5K_BEACON); + len += snprintf(buf+len, sizeof(buf)-len, + "%-24s0x%08x\tintval: %d\tTIM: 0x%x\n", +@@ -318,6 +320,8 @@ static ssize_t read_file_debug(struct file *file, char __user *user_buf, + unsigned int len = 0; + unsigned int i; + ++ pax_track_stack(); ++ + len += snprintf(buf+len, sizeof(buf)-len, + "DEBUG LEVEL: 0x%08x\n\n", sc->debug.level); + +diff --git a/drivers/net/wireless/ath/ath9k/debug.c b/drivers/net/wireless/ath/ath9k/debug.c +index 2be4c22..593b1eb 100644 +--- a/drivers/net/wireless/ath/ath9k/debug.c ++++ b/drivers/net/wireless/ath/ath9k/debug.c +@@ -220,6 +220,8 @@ static ssize_t read_file_interrupt(struct file *file, char __user *user_buf, + char buf[512]; + unsigned int len = 0; + ++ pax_track_stack(); ++ + len += snprintf(buf + len, sizeof(buf) - len, + "%8s: %10u\n", "RX", sc->debug.stats.istats.rxok); + len += snprintf(buf + len, sizeof(buf) - len, +@@ -360,6 +362,8 @@ static ssize_t read_file_wiphy(struct file *file, char __user *user_buf, + int i; + u8 addr[ETH_ALEN]; + ++ pax_track_stack(); ++ + len += snprintf(buf + len, sizeof(buf) - len, + "primary: %s (%s chan=%d ht=%d)\n", + wiphy_name(sc->pri_wiphy->hw->wiphy), +diff --git a/drivers/net/wireless/b43/debugfs.c b/drivers/net/wireless/b43/debugfs.c +index 80b19a4..dab3a45 100644 +--- a/drivers/net/wireless/b43/debugfs.c ++++ b/drivers/net/wireless/b43/debugfs.c +@@ -43,7 +43,7 @@ static struct dentry *rootdir; + struct b43_debugfs_fops { + ssize_t (*read)(struct b43_wldev *dev, char *buf, size_t bufsize); + int (*write)(struct b43_wldev *dev, const char *buf, size_t count); +- struct file_operations fops; ++ const struct file_operations fops; + /* Offset of struct b43_dfs_file in struct b43_dfsentry */ + size_t file_struct_offset; + }; +diff --git a/drivers/net/wireless/b43legacy/debugfs.c b/drivers/net/wireless/b43legacy/debugfs.c +index 1f85ac5..c99b4b4 100644 +--- a/drivers/net/wireless/b43legacy/debugfs.c ++++ b/drivers/net/wireless/b43legacy/debugfs.c +@@ -44,7 +44,7 @@ static struct dentry *rootdir; + struct b43legacy_debugfs_fops { + ssize_t (*read)(struct b43legacy_wldev *dev, char *buf, size_t bufsize); + int (*write)(struct b43legacy_wldev *dev, const char *buf, size_t count); +- struct file_operations fops; ++ const struct file_operations fops; + /* Offset of struct b43legacy_dfs_file in struct b43legacy_dfsentry */ + size_t file_struct_offset; + /* Take wl->irq_lock before calling read/write? */ +diff --git a/drivers/net/wireless/ipw2x00/ipw2100.c b/drivers/net/wireless/ipw2x00/ipw2100.c +index 43102bf..3b569c3 100644 +--- a/drivers/net/wireless/ipw2x00/ipw2100.c ++++ b/drivers/net/wireless/ipw2x00/ipw2100.c +@@ -2014,6 +2014,8 @@ static int ipw2100_set_essid(struct ipw2100_priv *priv, char *essid, + int err; + DECLARE_SSID_BUF(ssid); + ++ pax_track_stack(); ++ + IPW_DEBUG_HC("SSID: '%s'\n", print_ssid(ssid, essid, ssid_len)); + + if (ssid_len) +@@ -5380,6 +5382,8 @@ static int ipw2100_set_key(struct ipw2100_priv *priv, + struct ipw2100_wep_key *wep_key = (void *)cmd.host_command_parameters; + int err; + ++ pax_track_stack(); ++ + IPW_DEBUG_HC("WEP_KEY_INFO: index = %d, len = %d/%d\n", + idx, keylen, len); + +diff --git a/drivers/net/wireless/ipw2x00/libipw_rx.c b/drivers/net/wireless/ipw2x00/libipw_rx.c +index 282b1f7..169f0cf 100644 +--- a/drivers/net/wireless/ipw2x00/libipw_rx.c ++++ b/drivers/net/wireless/ipw2x00/libipw_rx.c +@@ -1566,6 +1566,8 @@ static void libipw_process_probe_response(struct libipw_device + unsigned long flags; + DECLARE_SSID_BUF(ssid); + ++ pax_track_stack(); ++ + LIBIPW_DEBUG_SCAN("'%s' (%pM" + "): %c%c%c%c %c%c%c%c-%c%c%c%c %c%c%c%c\n", + print_ssid(ssid, info_element->data, info_element->len), +diff --git a/drivers/net/wireless/iwlwifi/iwl-1000.c b/drivers/net/wireless/iwlwifi/iwl-1000.c +index 950267a..80d5fd2 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-1000.c ++++ b/drivers/net/wireless/iwlwifi/iwl-1000.c +@@ -137,7 +137,7 @@ static struct iwl_lib_ops iwl1000_lib = { + }, + }; + +-static struct iwl_ops iwl1000_ops = { ++static const struct iwl_ops iwl1000_ops = { + .ucode = &iwl5000_ucode, + .lib = &iwl1000_lib, + .hcmd = &iwl5000_hcmd, +diff --git a/drivers/net/wireless/iwlwifi/iwl-3945.c b/drivers/net/wireless/iwlwifi/iwl-3945.c +index 56bfcc3..b348020 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-3945.c ++++ b/drivers/net/wireless/iwlwifi/iwl-3945.c +@@ -2874,7 +2874,7 @@ static struct iwl_hcmd_utils_ops iwl3945_hcmd_utils = { + .build_addsta_hcmd = iwl3945_build_addsta_hcmd, + }; + +-static struct iwl_ops iwl3945_ops = { ++static const struct iwl_ops iwl3945_ops = { + .ucode = &iwl3945_ucode, + .lib = &iwl3945_lib, + .hcmd = &iwl3945_hcmd, +diff --git a/drivers/net/wireless/iwlwifi/iwl-4965.c b/drivers/net/wireless/iwlwifi/iwl-4965.c +index 585b8d4..e142963 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-4965.c ++++ b/drivers/net/wireless/iwlwifi/iwl-4965.c +@@ -2345,7 +2345,7 @@ static struct iwl_lib_ops iwl4965_lib = { + }, + }; + +-static struct iwl_ops iwl4965_ops = { ++static const struct iwl_ops iwl4965_ops = { + .ucode = &iwl4965_ucode, + .lib = &iwl4965_lib, + .hcmd = &iwl4965_hcmd, +diff --git a/drivers/net/wireless/iwlwifi/iwl-5000.c b/drivers/net/wireless/iwlwifi/iwl-5000.c +index 1f423f2..e37c192 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-5000.c ++++ b/drivers/net/wireless/iwlwifi/iwl-5000.c +@@ -1633,14 +1633,14 @@ static struct iwl_lib_ops iwl5150_lib = { + }, + }; + +-struct iwl_ops iwl5000_ops = { ++const struct iwl_ops iwl5000_ops = { + .ucode = &iwl5000_ucode, + .lib = &iwl5000_lib, + .hcmd = &iwl5000_hcmd, + .utils = &iwl5000_hcmd_utils, + }; + +-static struct iwl_ops iwl5150_ops = { ++static const struct iwl_ops iwl5150_ops = { + .ucode = &iwl5000_ucode, + .lib = &iwl5150_lib, + .hcmd = &iwl5000_hcmd, +diff --git a/drivers/net/wireless/iwlwifi/iwl-6000.c b/drivers/net/wireless/iwlwifi/iwl-6000.c +index 1473452..f07d5e1 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-6000.c ++++ b/drivers/net/wireless/iwlwifi/iwl-6000.c +@@ -146,7 +146,7 @@ static struct iwl_hcmd_utils_ops iwl6000_hcmd_utils = { + .calc_rssi = iwl5000_calc_rssi, + }; + +-static struct iwl_ops iwl6000_ops = { ++static const struct iwl_ops iwl6000_ops = { + .ucode = &iwl5000_ucode, + .lib = &iwl6000_lib, + .hcmd = &iwl5000_hcmd, +diff --git a/drivers/net/wireless/iwlwifi/iwl-agn-rs.c b/drivers/net/wireless/iwlwifi/iwl-agn-rs.c +index 1a3dfa2..b3e0a61 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-agn-rs.c ++++ b/drivers/net/wireless/iwlwifi/iwl-agn-rs.c +@@ -857,6 +857,8 @@ static void rs_tx_status(void *priv_r, struct ieee80211_supported_band *sband, + u8 active_index = 0; + s32 tpt = 0; + ++ pax_track_stack(); ++ + IWL_DEBUG_RATE_LIMIT(priv, "get frame ack response, update rate scale window\n"); + + if (!ieee80211_is_data(hdr->frame_control) || +@@ -2722,6 +2724,8 @@ static void rs_fill_link_cmd(struct iwl_priv *priv, + u8 valid_tx_ant = 0; + struct iwl_link_quality_cmd *lq_cmd = &lq_sta->lq; + ++ pax_track_stack(); ++ + /* Override starting rate (index 0) if needed for debug purposes */ + rs_dbgfs_set_mcs(lq_sta, &new_rate, index); + +diff --git a/drivers/net/wireless/iwlwifi/iwl-agn.c b/drivers/net/wireless/iwlwifi/iwl-agn.c +index 0e56d78..6a3c107 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-agn.c ++++ b/drivers/net/wireless/iwlwifi/iwl-agn.c +@@ -2911,7 +2911,9 @@ static int iwl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *ent) + if (iwl_debug_level & IWL_DL_INFO) + dev_printk(KERN_DEBUG, &(pdev->dev), + "Disabling hw_scan\n"); +- iwl_hw_ops.hw_scan = NULL; ++ pax_open_kernel(); ++ *(void **)&iwl_hw_ops.hw_scan = NULL; ++ pax_close_kernel(); + } + + hw = iwl_alloc_all(cfg, &iwl_hw_ops); +diff --git a/drivers/net/wireless/iwlwifi/iwl-debug.h b/drivers/net/wireless/iwlwifi/iwl-debug.h +index cbc6290..eb323d7 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-debug.h ++++ b/drivers/net/wireless/iwlwifi/iwl-debug.h +@@ -118,8 +118,8 @@ void iwl_dbgfs_unregister(struct iwl_priv *priv); + #endif + + #else +-#define IWL_DEBUG(__priv, level, fmt, args...) +-#define IWL_DEBUG_LIMIT(__priv, level, fmt, args...) ++#define IWL_DEBUG(__priv, level, fmt, args...) do {} while (0) ++#define IWL_DEBUG_LIMIT(__priv, level, fmt, args...) do {} while (0) + static inline void iwl_print_hex_dump(struct iwl_priv *priv, int level, + void *p, u32 len) + {} +diff --git a/drivers/net/wireless/iwlwifi/iwl-debugfs.c b/drivers/net/wireless/iwlwifi/iwl-debugfs.c +index a198bcf..8e68233 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-debugfs.c ++++ b/drivers/net/wireless/iwlwifi/iwl-debugfs.c +@@ -524,6 +524,8 @@ static ssize_t iwl_dbgfs_status_read(struct file *file, + int pos = 0; + const size_t bufsz = sizeof(buf); + ++ pax_track_stack(); ++ + pos += scnprintf(buf + pos, bufsz - pos, "STATUS_HCMD_ACTIVE:\t %d\n", + test_bit(STATUS_HCMD_ACTIVE, &priv->status)); + pos += scnprintf(buf + pos, bufsz - pos, "STATUS_HCMD_SYNC_ACTIVE: %d\n", +@@ -658,6 +660,8 @@ static ssize_t iwl_dbgfs_qos_read(struct file *file, char __user *user_buf, + const size_t bufsz = sizeof(buf); + ssize_t ret; + ++ pax_track_stack(); ++ + for (i = 0; i < AC_NUM; i++) { + pos += scnprintf(buf + pos, bufsz - pos, + "\tcw_min\tcw_max\taifsn\ttxop\n"); +diff --git a/drivers/net/wireless/iwlwifi/iwl-dev.h b/drivers/net/wireless/iwlwifi/iwl-dev.h +index 3539ea4..b174bfa 100644 +--- a/drivers/net/wireless/iwlwifi/iwl-dev.h ++++ b/drivers/net/wireless/iwlwifi/iwl-dev.h +@@ -68,7 +68,7 @@ struct iwl_tx_queue; + + /* shared structures from iwl-5000.c */ + extern struct iwl_mod_params iwl50_mod_params; +-extern struct iwl_ops iwl5000_ops; ++extern const struct iwl_ops iwl5000_ops; + extern struct iwl_ucode_ops iwl5000_ucode; + extern struct iwl_lib_ops iwl5000_lib; + extern struct iwl_hcmd_ops iwl5000_hcmd; +diff --git a/drivers/net/wireless/iwlwifi/iwl3945-base.c b/drivers/net/wireless/iwlwifi/iwl3945-base.c +index 619590d..69235ee 100644 +--- a/drivers/net/wireless/iwlwifi/iwl3945-base.c ++++ b/drivers/net/wireless/iwlwifi/iwl3945-base.c +@@ -3927,7 +3927,9 @@ static int iwl3945_pci_probe(struct pci_dev *pdev, const struct pci_device_id *e + */ + if (iwl3945_mod_params.disable_hw_scan) { + IWL_DEBUG_INFO(priv, "Disabling hw_scan\n"); +- iwl3945_hw_ops.hw_scan = NULL; ++ pax_open_kernel(); ++ *(void **)&iwl3945_hw_ops.hw_scan = NULL; ++ pax_close_kernel(); + } + + +diff --git a/drivers/net/wireless/iwmc3200wifi/debugfs.c b/drivers/net/wireless/iwmc3200wifi/debugfs.c +index 1465379..fe4d78b 100644 +--- a/drivers/net/wireless/iwmc3200wifi/debugfs.c ++++ b/drivers/net/wireless/iwmc3200wifi/debugfs.c +@@ -299,6 +299,8 @@ static ssize_t iwm_debugfs_fw_err_read(struct file *filp, + int buf_len = 512; + size_t len = 0; + ++ pax_track_stack(); ++ + if (*ppos != 0) + return 0; + if (count < sizeof(buf)) +diff --git a/drivers/net/wireless/libertas/debugfs.c b/drivers/net/wireless/libertas/debugfs.c +index 893a55c..7f66a50 100644 +--- a/drivers/net/wireless/libertas/debugfs.c ++++ b/drivers/net/wireless/libertas/debugfs.c +@@ -708,7 +708,7 @@ out_unlock: + struct lbs_debugfs_files { + const char *name; + int perm; +- struct file_operations fops; ++ const struct file_operations fops; + }; + + static const struct lbs_debugfs_files debugfs_files[] = { +diff --git a/drivers/net/wireless/rndis_wlan.c b/drivers/net/wireless/rndis_wlan.c +index 2ecbedb..42704f0 100644 +--- a/drivers/net/wireless/rndis_wlan.c ++++ b/drivers/net/wireless/rndis_wlan.c +@@ -1176,7 +1176,7 @@ static int set_rts_threshold(struct usbnet *usbdev, u32 rts_threshold) + + devdbg(usbdev, "set_rts_threshold %i", rts_threshold); + +- if (rts_threshold < 0 || rts_threshold > 2347) ++ if (rts_threshold > 2347) + rts_threshold = 2347; + + tmp = cpu_to_le32(rts_threshold); +diff --git a/drivers/oprofile/buffer_sync.c b/drivers/oprofile/buffer_sync.c +index 334ccd6..47f8944 100644 +--- a/drivers/oprofile/buffer_sync.c ++++ b/drivers/oprofile/buffer_sync.c +@@ -342,7 +342,7 @@ static void add_data(struct op_entry *entry, struct mm_struct *mm) + if (cookie == NO_COOKIE) + offset = pc; + if (cookie == INVALID_COOKIE) { +- atomic_inc(&oprofile_stats.sample_lost_no_mapping); ++ atomic_inc_unchecked(&oprofile_stats.sample_lost_no_mapping); + offset = pc; + } + if (cookie != last_cookie) { +@@ -386,14 +386,14 @@ add_sample(struct mm_struct *mm, struct op_sample *s, int in_kernel) + /* add userspace sample */ + + if (!mm) { +- atomic_inc(&oprofile_stats.sample_lost_no_mm); ++ atomic_inc_unchecked(&oprofile_stats.sample_lost_no_mm); + return 0; + } + + cookie = lookup_dcookie(mm, s->eip, &offset); + + if (cookie == INVALID_COOKIE) { +- atomic_inc(&oprofile_stats.sample_lost_no_mapping); ++ atomic_inc_unchecked(&oprofile_stats.sample_lost_no_mapping); + return 0; + } + +@@ -562,7 +562,7 @@ void sync_buffer(int cpu) + /* ignore backtraces if failed to add a sample */ + if (state == sb_bt_start) { + state = sb_bt_ignore; +- atomic_inc(&oprofile_stats.bt_lost_no_mapping); ++ atomic_inc_unchecked(&oprofile_stats.bt_lost_no_mapping); + } + } + release_mm(mm); +diff --git a/drivers/oprofile/event_buffer.c b/drivers/oprofile/event_buffer.c +index 5df60a6..72f5c1c 100644 +--- a/drivers/oprofile/event_buffer.c ++++ b/drivers/oprofile/event_buffer.c +@@ -53,7 +53,7 @@ void add_event_entry(unsigned long value) + } + + if (buffer_pos == buffer_size) { +- atomic_inc(&oprofile_stats.event_lost_overflow); ++ atomic_inc_unchecked(&oprofile_stats.event_lost_overflow); + return; + } + +diff --git a/drivers/oprofile/oprof.c b/drivers/oprofile/oprof.c +index dc8a042..fe5f315 100644 +--- a/drivers/oprofile/oprof.c ++++ b/drivers/oprofile/oprof.c +@@ -110,7 +110,7 @@ static void switch_worker(struct work_struct *work) + if (oprofile_ops.switch_events()) + return; + +- atomic_inc(&oprofile_stats.multiplex_counter); ++ atomic_inc_unchecked(&oprofile_stats.multiplex_counter); + start_switch_worker(); + } + +diff --git a/drivers/oprofile/oprofile_stats.c b/drivers/oprofile/oprofile_stats.c +index 61689e8..387f7f8 100644 +--- a/drivers/oprofile/oprofile_stats.c ++++ b/drivers/oprofile/oprofile_stats.c +@@ -30,11 +30,11 @@ void oprofile_reset_stats(void) + cpu_buf->sample_invalid_eip = 0; + } + +- atomic_set(&oprofile_stats.sample_lost_no_mm, 0); +- atomic_set(&oprofile_stats.sample_lost_no_mapping, 0); +- atomic_set(&oprofile_stats.event_lost_overflow, 0); +- atomic_set(&oprofile_stats.bt_lost_no_mapping, 0); +- atomic_set(&oprofile_stats.multiplex_counter, 0); ++ atomic_set_unchecked(&oprofile_stats.sample_lost_no_mm, 0); ++ atomic_set_unchecked(&oprofile_stats.sample_lost_no_mapping, 0); ++ atomic_set_unchecked(&oprofile_stats.event_lost_overflow, 0); ++ atomic_set_unchecked(&oprofile_stats.bt_lost_no_mapping, 0); ++ atomic_set_unchecked(&oprofile_stats.multiplex_counter, 0); + } + + +diff --git a/drivers/oprofile/oprofile_stats.h b/drivers/oprofile/oprofile_stats.h +index 0b54e46..a37c527 100644 +--- a/drivers/oprofile/oprofile_stats.h ++++ b/drivers/oprofile/oprofile_stats.h +@@ -13,11 +13,11 @@ + #include + + struct oprofile_stat_struct { +- atomic_t sample_lost_no_mm; +- atomic_t sample_lost_no_mapping; +- atomic_t bt_lost_no_mapping; +- atomic_t event_lost_overflow; +- atomic_t multiplex_counter; ++ atomic_unchecked_t sample_lost_no_mm; ++ atomic_unchecked_t sample_lost_no_mapping; ++ atomic_unchecked_t bt_lost_no_mapping; ++ atomic_unchecked_t event_lost_overflow; ++ atomic_unchecked_t multiplex_counter; + }; + + extern struct oprofile_stat_struct oprofile_stats; +diff --git a/drivers/oprofile/oprofilefs.c b/drivers/oprofile/oprofilefs.c +index 2766a6d..80c77e2 100644 +--- a/drivers/oprofile/oprofilefs.c ++++ b/drivers/oprofile/oprofilefs.c +@@ -187,7 +187,7 @@ static const struct file_operations atomic_ro_fops = { + + + int oprofilefs_create_ro_atomic(struct super_block *sb, struct dentry *root, +- char const *name, atomic_t *val) ++ char const *name, atomic_unchecked_t *val) + { + struct dentry *d = __oprofilefs_create_file(sb, root, name, + &atomic_ro_fops, 0444); +diff --git a/drivers/parisc/pdc_stable.c b/drivers/parisc/pdc_stable.c +index 13a64bc..ad62835 100644 +--- a/drivers/parisc/pdc_stable.c ++++ b/drivers/parisc/pdc_stable.c +@@ -481,7 +481,7 @@ pdcspath_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static struct sysfs_ops pdcspath_attr_ops = { ++static const struct sysfs_ops pdcspath_attr_ops = { + .show = pdcspath_attr_show, + .store = pdcspath_attr_store, + }; +diff --git a/drivers/parport/procfs.c b/drivers/parport/procfs.c +index 8eefe56..40751a7 100644 +--- a/drivers/parport/procfs.c ++++ b/drivers/parport/procfs.c +@@ -64,7 +64,7 @@ static int do_active_device(ctl_table *table, int write, + + *ppos += len; + +- return copy_to_user(result, buffer, len) ? -EFAULT : 0; ++ return (len > sizeof buffer || copy_to_user(result, buffer, len)) ? -EFAULT : 0; + } + + #ifdef CONFIG_PARPORT_1284 +@@ -106,7 +106,7 @@ static int do_autoprobe(ctl_table *table, int write, + + *ppos += len; + +- return copy_to_user (result, buffer, len) ? -EFAULT : 0; ++ return (len > sizeof buffer || copy_to_user (result, buffer, len)) ? -EFAULT : 0; + } + #endif /* IEEE1284.3 support. */ + +diff --git a/drivers/pci/hotplug/acpiphp_glue.c b/drivers/pci/hotplug/acpiphp_glue.c +index 73e7d8e..c80f3d2 100644 +--- a/drivers/pci/hotplug/acpiphp_glue.c ++++ b/drivers/pci/hotplug/acpiphp_glue.c +@@ -111,7 +111,7 @@ static int post_dock_fixups(struct notifier_block *nb, unsigned long val, + } + + +-static struct acpi_dock_ops acpiphp_dock_ops = { ++static const struct acpi_dock_ops acpiphp_dock_ops = { + .handler = handle_hotplug_event_func, + }; + +diff --git a/drivers/pci/hotplug/cpci_hotplug.h b/drivers/pci/hotplug/cpci_hotplug.h +index 9fff878..ad0ad53 100644 +--- a/drivers/pci/hotplug/cpci_hotplug.h ++++ b/drivers/pci/hotplug/cpci_hotplug.h +@@ -59,7 +59,7 @@ struct cpci_hp_controller_ops { + int (*hardware_test) (struct slot* slot, u32 value); + u8 (*get_power) (struct slot* slot); + int (*set_power) (struct slot* slot, int value); +-}; ++} __no_const; + + struct cpci_hp_controller { + unsigned int irq; +diff --git a/drivers/pci/hotplug/cpqphp_nvram.c b/drivers/pci/hotplug/cpqphp_nvram.c +index 76ba8a1..20ca857 100644 +--- a/drivers/pci/hotplug/cpqphp_nvram.c ++++ b/drivers/pci/hotplug/cpqphp_nvram.c +@@ -428,9 +428,13 @@ static u32 store_HRT (void __iomem *rom_start) + + void compaq_nvram_init (void __iomem *rom_start) + { ++ ++#ifndef CONFIG_PAX_KERNEXEC + if (rom_start) { + compaq_int15_entry_point = (rom_start + ROM_INT15_PHY_ADDR - ROM_PHY_ADDR); + } ++#endif ++ + dbg("int15 entry = %p\n", compaq_int15_entry_point); + + /* initialize our int15 lock */ +diff --git a/drivers/pci/hotplug/fakephp.c b/drivers/pci/hotplug/fakephp.c +index 6151389..0a894ef 100644 +--- a/drivers/pci/hotplug/fakephp.c ++++ b/drivers/pci/hotplug/fakephp.c +@@ -73,7 +73,7 @@ static void legacy_release(struct kobject *kobj) + } + + static struct kobj_type legacy_ktype = { +- .sysfs_ops = &(struct sysfs_ops){ ++ .sysfs_ops = &(const struct sysfs_ops){ + .store = legacy_store, .show = legacy_show + }, + .release = &legacy_release, +diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c +index 5b680df..fe05b7e 100644 +--- a/drivers/pci/intel-iommu.c ++++ b/drivers/pci/intel-iommu.c +@@ -2643,7 +2643,7 @@ error: + return 0; + } + +-static dma_addr_t intel_map_page(struct device *dev, struct page *page, ++dma_addr_t intel_map_page(struct device *dev, struct page *page, + unsigned long offset, size_t size, + enum dma_data_direction dir, + struct dma_attrs *attrs) +@@ -2719,7 +2719,7 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova) + spin_unlock_irqrestore(&async_umap_flush_lock, flags); + } + +-static void intel_unmap_page(struct device *dev, dma_addr_t dev_addr, ++void intel_unmap_page(struct device *dev, dma_addr_t dev_addr, + size_t size, enum dma_data_direction dir, + struct dma_attrs *attrs) + { +@@ -2768,7 +2768,7 @@ static void intel_unmap_page(struct device *dev, dma_addr_t dev_addr, + } + } + +-static void *intel_alloc_coherent(struct device *hwdev, size_t size, ++void *intel_alloc_coherent(struct device *hwdev, size_t size, + dma_addr_t *dma_handle, gfp_t flags) + { + void *vaddr; +@@ -2800,7 +2800,7 @@ static void *intel_alloc_coherent(struct device *hwdev, size_t size, + return NULL; + } + +-static void intel_free_coherent(struct device *hwdev, size_t size, void *vaddr, ++void intel_free_coherent(struct device *hwdev, size_t size, void *vaddr, + dma_addr_t dma_handle) + { + int order; +@@ -2812,7 +2812,7 @@ static void intel_free_coherent(struct device *hwdev, size_t size, void *vaddr, + free_pages((unsigned long)vaddr, order); + } + +-static void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist, ++void intel_unmap_sg(struct device *hwdev, struct scatterlist *sglist, + int nelems, enum dma_data_direction dir, + struct dma_attrs *attrs) + { +@@ -2872,7 +2872,7 @@ static int intel_nontranslate_map_sg(struct device *hddev, + return nelems; + } + +-static int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems, ++int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems, + enum dma_data_direction dir, struct dma_attrs *attrs) + { + int i; +@@ -2941,12 +2941,12 @@ static int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int ne + return nelems; + } + +-static int intel_mapping_error(struct device *dev, dma_addr_t dma_addr) ++int intel_mapping_error(struct device *dev, dma_addr_t dma_addr) + { + return !dma_addr; + } + +-struct dma_map_ops intel_dma_ops = { ++const struct dma_map_ops intel_dma_ops = { + .alloc_coherent = intel_alloc_coherent, + .free_coherent = intel_free_coherent, + .map_sg = intel_map_sg, +diff --git a/drivers/pci/pcie/aspm.c b/drivers/pci/pcie/aspm.c +index 5b7056c..607bc94 100644 +--- a/drivers/pci/pcie/aspm.c ++++ b/drivers/pci/pcie/aspm.c +@@ -27,9 +27,9 @@ + #define MODULE_PARAM_PREFIX "pcie_aspm." + + /* Note: those are not register definitions */ +-#define ASPM_STATE_L0S_UP (1) /* Upstream direction L0s state */ +-#define ASPM_STATE_L0S_DW (2) /* Downstream direction L0s state */ +-#define ASPM_STATE_L1 (4) /* L1 state */ ++#define ASPM_STATE_L0S_UP (1U) /* Upstream direction L0s state */ ++#define ASPM_STATE_L0S_DW (2U) /* Downstream direction L0s state */ ++#define ASPM_STATE_L1 (4U) /* L1 state */ + #define ASPM_STATE_L0S (ASPM_STATE_L0S_UP | ASPM_STATE_L0S_DW) + #define ASPM_STATE_ALL (ASPM_STATE_L0S | ASPM_STATE_L1) + +diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c +index 8105e32..ca10419 100644 +--- a/drivers/pci/probe.c ++++ b/drivers/pci/probe.c +@@ -62,14 +62,14 @@ static ssize_t pci_bus_show_cpuaffinity(struct device *dev, + return ret; + } + +-static ssize_t inline pci_bus_show_cpumaskaffinity(struct device *dev, ++static inline ssize_t pci_bus_show_cpumaskaffinity(struct device *dev, + struct device_attribute *attr, + char *buf) + { + return pci_bus_show_cpuaffinity(dev, 0, attr, buf); + } + +-static ssize_t inline pci_bus_show_cpulistaffinity(struct device *dev, ++static inline ssize_t pci_bus_show_cpulistaffinity(struct device *dev, + struct device_attribute *attr, + char *buf) + { +diff --git a/drivers/pci/proc.c b/drivers/pci/proc.c +index a03ad8c..024b0da 100644 +--- a/drivers/pci/proc.c ++++ b/drivers/pci/proc.c +@@ -480,7 +480,16 @@ static const struct file_operations proc_bus_pci_dev_operations = { + static int __init pci_proc_init(void) + { + struct pci_dev *dev = NULL; ++ ++#ifdef CONFIG_GRKERNSEC_PROC_ADD ++#ifdef CONFIG_GRKERNSEC_PROC_USER ++ proc_bus_pci_dir = proc_mkdir_mode("bus/pci", S_IRUSR | S_IXUSR, NULL); ++#elif defined(CONFIG_GRKERNSEC_PROC_USERGROUP) ++ proc_bus_pci_dir = proc_mkdir_mode("bus/pci", S_IRUSR | S_IXUSR | S_IRGRP | S_IXGRP, NULL); ++#endif ++#else + proc_bus_pci_dir = proc_mkdir("bus/pci", NULL); ++#endif + proc_create("devices", 0, proc_bus_pci_dir, + &proc_bus_pci_dev_operations); + proc_initialized = 1; +diff --git a/drivers/pci/slot.c b/drivers/pci/slot.c +index 8c02b6c..5584d8e 100644 +--- a/drivers/pci/slot.c ++++ b/drivers/pci/slot.c +@@ -29,7 +29,7 @@ static ssize_t pci_slot_attr_store(struct kobject *kobj, + return attribute->store ? attribute->store(slot, buf, len) : -EIO; + } + +-static struct sysfs_ops pci_slot_sysfs_ops = { ++static const struct sysfs_ops pci_slot_sysfs_ops = { + .show = pci_slot_attr_show, + .store = pci_slot_attr_store, + }; +diff --git a/drivers/pcmcia/pcmcia_ioctl.c b/drivers/pcmcia/pcmcia_ioctl.c +index 30cf71d2..50938f1 100644 +--- a/drivers/pcmcia/pcmcia_ioctl.c ++++ b/drivers/pcmcia/pcmcia_ioctl.c +@@ -819,7 +819,7 @@ static int ds_ioctl(struct inode * inode, struct file * file, + return -EFAULT; + } + } +- buf = kmalloc(sizeof(ds_ioctl_arg_t), GFP_KERNEL); ++ buf = kzalloc(sizeof(ds_ioctl_arg_t), GFP_KERNEL); + if (!buf) + return -ENOMEM; + +diff --git a/drivers/platform/x86/acer-wmi.c b/drivers/platform/x86/acer-wmi.c +index 52183c4..b224c69 100644 +--- a/drivers/platform/x86/acer-wmi.c ++++ b/drivers/platform/x86/acer-wmi.c +@@ -918,7 +918,7 @@ static int update_bl_status(struct backlight_device *bd) + return 0; + } + +-static struct backlight_ops acer_bl_ops = { ++static const struct backlight_ops acer_bl_ops = { + .get_brightness = read_brightness, + .update_status = update_bl_status, + }; +diff --git a/drivers/platform/x86/asus-laptop.c b/drivers/platform/x86/asus-laptop.c +index 767cb61..a87380b 100644 +--- a/drivers/platform/x86/asus-laptop.c ++++ b/drivers/platform/x86/asus-laptop.c +@@ -250,7 +250,7 @@ static struct backlight_device *asus_backlight_device; + */ + static int read_brightness(struct backlight_device *bd); + static int update_bl_status(struct backlight_device *bd); +-static struct backlight_ops asusbl_ops = { ++static const struct backlight_ops asusbl_ops = { + .get_brightness = read_brightness, + .update_status = update_bl_status, + }; +diff --git a/drivers/platform/x86/asus_acpi.c b/drivers/platform/x86/asus_acpi.c +index d66c07a..a4abaac 100644 +--- a/drivers/platform/x86/asus_acpi.c ++++ b/drivers/platform/x86/asus_acpi.c +@@ -1396,7 +1396,7 @@ static int asus_hotk_remove(struct acpi_device *device, int type) + return 0; + } + +-static struct backlight_ops asus_backlight_data = { ++static const struct backlight_ops asus_backlight_data = { + .get_brightness = read_brightness, + .update_status = set_brightness_status, + }; +diff --git a/drivers/platform/x86/compal-laptop.c b/drivers/platform/x86/compal-laptop.c +index 11003bb..550ff1b 100644 +--- a/drivers/platform/x86/compal-laptop.c ++++ b/drivers/platform/x86/compal-laptop.c +@@ -163,7 +163,7 @@ static int bl_update_status(struct backlight_device *b) + return set_lcd_level(b->props.brightness); + } + +-static struct backlight_ops compalbl_ops = { ++static const struct backlight_ops compalbl_ops = { + .get_brightness = bl_get_brightness, + .update_status = bl_update_status, + }; +diff --git a/drivers/platform/x86/dell-laptop.c b/drivers/platform/x86/dell-laptop.c +index 07a74da..9dc99fa 100644 +--- a/drivers/platform/x86/dell-laptop.c ++++ b/drivers/platform/x86/dell-laptop.c +@@ -318,7 +318,7 @@ static int dell_get_intensity(struct backlight_device *bd) + return buffer.output[1]; + } + +-static struct backlight_ops dell_ops = { ++static const struct backlight_ops dell_ops = { + .get_brightness = dell_get_intensity, + .update_status = dell_send_intensity, + }; +diff --git a/drivers/platform/x86/eeepc-laptop.c b/drivers/platform/x86/eeepc-laptop.c +index c533b1c..5c81f22 100644 +--- a/drivers/platform/x86/eeepc-laptop.c ++++ b/drivers/platform/x86/eeepc-laptop.c +@@ -245,7 +245,7 @@ static struct device *eeepc_hwmon_device; + */ + static int read_brightness(struct backlight_device *bd); + static int update_bl_status(struct backlight_device *bd); +-static struct backlight_ops eeepcbl_ops = { ++static const struct backlight_ops eeepcbl_ops = { + .get_brightness = read_brightness, + .update_status = update_bl_status, + }; +diff --git a/drivers/platform/x86/fujitsu-laptop.c b/drivers/platform/x86/fujitsu-laptop.c +index bcd4ba8..a249b35 100644 +--- a/drivers/platform/x86/fujitsu-laptop.c ++++ b/drivers/platform/x86/fujitsu-laptop.c +@@ -436,7 +436,7 @@ static int bl_update_status(struct backlight_device *b) + return ret; + } + +-static struct backlight_ops fujitsubl_ops = { ++static const struct backlight_ops fujitsubl_ops = { + .get_brightness = bl_get_brightness, + .update_status = bl_update_status, + }; +diff --git a/drivers/platform/x86/msi-laptop.c b/drivers/platform/x86/msi-laptop.c +index 759763d..1093ba2 100644 +--- a/drivers/platform/x86/msi-laptop.c ++++ b/drivers/platform/x86/msi-laptop.c +@@ -161,7 +161,7 @@ static int bl_update_status(struct backlight_device *b) + return set_lcd_level(b->props.brightness); + } + +-static struct backlight_ops msibl_ops = { ++static const struct backlight_ops msibl_ops = { + .get_brightness = bl_get_brightness, + .update_status = bl_update_status, + }; +diff --git a/drivers/platform/x86/panasonic-laptop.c b/drivers/platform/x86/panasonic-laptop.c +index fe7cf01..9012d8d 100644 +--- a/drivers/platform/x86/panasonic-laptop.c ++++ b/drivers/platform/x86/panasonic-laptop.c +@@ -352,7 +352,7 @@ static int bl_set_status(struct backlight_device *bd) + return acpi_pcc_write_sset(pcc, SINF_DC_CUR_BRIGHT, bright); + } + +-static struct backlight_ops pcc_backlight_ops = { ++static const struct backlight_ops pcc_backlight_ops = { + .get_brightness = bl_get, + .update_status = bl_set_status, + }; +diff --git a/drivers/platform/x86/sony-laptop.c b/drivers/platform/x86/sony-laptop.c +index a2a742c..b37e25e 100644 +--- a/drivers/platform/x86/sony-laptop.c ++++ b/drivers/platform/x86/sony-laptop.c +@@ -850,7 +850,7 @@ static int sony_backlight_get_brightness(struct backlight_device *bd) + } + + static struct backlight_device *sony_backlight_device; +-static struct backlight_ops sony_backlight_ops = { ++static const struct backlight_ops sony_backlight_ops = { + .update_status = sony_backlight_update_status, + .get_brightness = sony_backlight_get_brightness, + }; +diff --git a/drivers/platform/x86/thinkpad_acpi.c b/drivers/platform/x86/thinkpad_acpi.c +index 68271ae..5e8fb10 100644 +--- a/drivers/platform/x86/thinkpad_acpi.c ++++ b/drivers/platform/x86/thinkpad_acpi.c +@@ -2139,7 +2139,7 @@ static int hotkey_mask_get(void) + return 0; + } + +-void static hotkey_mask_warn_incomplete_mask(void) ++static void hotkey_mask_warn_incomplete_mask(void) + { + /* log only what the user can fix... */ + const u32 wantedmask = hotkey_driver_mask & +@@ -6125,7 +6125,7 @@ static void tpacpi_brightness_notify_change(void) + BACKLIGHT_UPDATE_HOTKEY); + } + +-static struct backlight_ops ibm_backlight_data = { ++static const struct backlight_ops ibm_backlight_data = { + .get_brightness = brightness_get, + .update_status = brightness_update_status, + }; +diff --git a/drivers/platform/x86/toshiba_acpi.c b/drivers/platform/x86/toshiba_acpi.c +index 51c0a8b..0786629 100644 +--- a/drivers/platform/x86/toshiba_acpi.c ++++ b/drivers/platform/x86/toshiba_acpi.c +@@ -671,7 +671,7 @@ static acpi_status remove_device(void) + return AE_OK; + } + +-static struct backlight_ops toshiba_backlight_data = { ++static const struct backlight_ops toshiba_backlight_data = { + .get_brightness = get_lcd, + .update_status = set_lcd_status, + }; +diff --git a/drivers/pnp/pnpbios/bioscalls.c b/drivers/pnp/pnpbios/bioscalls.c +index fc83783c..cf370d7 100644 +--- a/drivers/pnp/pnpbios/bioscalls.c ++++ b/drivers/pnp/pnpbios/bioscalls.c +@@ -60,7 +60,7 @@ do { \ + set_desc_limit(&gdt[(selname) >> 3], (size) - 1); \ + } while(0) + +-static struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4092, ++static const struct desc_struct bad_bios_desc = GDT_ENTRY_INIT(0x4093, + (unsigned long)__va(0x400UL), PAGE_SIZE - 0x400 - 1); + + /* +@@ -97,7 +97,10 @@ static inline u16 call_pnp_bios(u16 func, u16 arg1, u16 arg2, u16 arg3, + + cpu = get_cpu(); + save_desc_40 = get_cpu_gdt_table(cpu)[0x40 / 8]; ++ ++ pax_open_kernel(); + get_cpu_gdt_table(cpu)[0x40 / 8] = bad_bios_desc; ++ pax_close_kernel(); + + /* On some boxes IRQ's during PnP BIOS calls are deadly. */ + spin_lock_irqsave(&pnp_bios_lock, flags); +@@ -135,7 +138,10 @@ static inline u16 call_pnp_bios(u16 func, u16 arg1, u16 arg2, u16 arg3, + :"memory"); + spin_unlock_irqrestore(&pnp_bios_lock, flags); + ++ pax_open_kernel(); + get_cpu_gdt_table(cpu)[0x40 / 8] = save_desc_40; ++ pax_close_kernel(); ++ + put_cpu(); + + /* If we get here and this is set then the PnP BIOS faulted on us. */ +@@ -469,7 +475,7 @@ int pnp_bios_read_escd(char *data, u32 nvram_base) + return status; + } + +-void pnpbios_calls_init(union pnp_bios_install_struct *header) ++void __init pnpbios_calls_init(union pnp_bios_install_struct *header) + { + int i; + +@@ -477,6 +483,8 @@ void pnpbios_calls_init(union pnp_bios_install_struct *header) + pnp_bios_callpoint.offset = header->fields.pm16offset; + pnp_bios_callpoint.segment = PNP_CS16; + ++ pax_open_kernel(); ++ + for_each_possible_cpu(i) { + struct desc_struct *gdt = get_cpu_gdt_table(i); + if (!gdt) +@@ -488,4 +496,6 @@ void pnpbios_calls_init(union pnp_bios_install_struct *header) + set_desc_base(&gdt[GDT_ENTRY_PNPBIOS_DS], + (unsigned long)__va(header->fields.pm16dseg)); + } ++ ++ pax_close_kernel(); + } +diff --git a/drivers/pnp/resource.c b/drivers/pnp/resource.c +index ba97654..66b99d4 100644 +--- a/drivers/pnp/resource.c ++++ b/drivers/pnp/resource.c +@@ -355,7 +355,7 @@ int pnp_check_irq(struct pnp_dev *dev, struct resource *res) + return 1; + + /* check if the resource is valid */ +- if (*irq < 0 || *irq > 15) ++ if (*irq > 15) + return 0; + + /* check if the resource is reserved */ +@@ -419,7 +419,7 @@ int pnp_check_dma(struct pnp_dev *dev, struct resource *res) + return 1; + + /* check if the resource is valid */ +- if (*dma < 0 || *dma == 4 || *dma > 7) ++ if (*dma == 4 || *dma > 7) + return 0; + + /* check if the resource is reserved */ +diff --git a/drivers/power/bq27x00_battery.c b/drivers/power/bq27x00_battery.c +index 62bb981..24a2dc9 100644 +--- a/drivers/power/bq27x00_battery.c ++++ b/drivers/power/bq27x00_battery.c +@@ -44,7 +44,7 @@ struct bq27x00_device_info; + struct bq27x00_access_methods { + int (*read)(u8 reg, int *rt_value, int b_single, + struct bq27x00_device_info *di); +-}; ++} __no_const; + + struct bq27x00_device_info { + struct device *dev; +diff --git a/drivers/rtc/rtc-dev.c b/drivers/rtc/rtc-dev.c +index 62227cd..b5b538b 100644 +--- a/drivers/rtc/rtc-dev.c ++++ b/drivers/rtc/rtc-dev.c +@@ -14,6 +14,7 @@ + #include + #include + #include ++#include + #include "rtc-core.h" + + static dev_t rtc_devt; +@@ -357,6 +358,8 @@ static long rtc_dev_ioctl(struct file *file, + if (copy_from_user(&tm, uarg, sizeof(tm))) + return -EFAULT; + ++ gr_log_timechange(); ++ + return rtc_set_time(rtc, &tm); + + case RTC_PIE_ON: +diff --git a/drivers/s390/cio/qdio_perf.c b/drivers/s390/cio/qdio_perf.c +index 968e3c7..fbc637a 100644 +--- a/drivers/s390/cio/qdio_perf.c ++++ b/drivers/s390/cio/qdio_perf.c +@@ -31,51 +31,51 @@ static struct proc_dir_entry *qdio_perf_pde; + static int qdio_perf_proc_show(struct seq_file *m, void *v) + { + seq_printf(m, "Number of qdio interrupts\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.qdio_int)); ++ (long)atomic_long_read_unchecked(&perf_stats.qdio_int)); + seq_printf(m, "Number of PCI interrupts\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.pci_int)); ++ (long)atomic_long_read_unchecked(&perf_stats.pci_int)); + seq_printf(m, "Number of adapter interrupts\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.thin_int)); ++ (long)atomic_long_read_unchecked(&perf_stats.thin_int)); + seq_printf(m, "\n"); + seq_printf(m, "Inbound tasklet runs\t\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.tasklet_inbound)); ++ (long)atomic_long_read_unchecked(&perf_stats.tasklet_inbound)); + seq_printf(m, "Outbound tasklet runs\t\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.tasklet_outbound)); ++ (long)atomic_long_read_unchecked(&perf_stats.tasklet_outbound)); + seq_printf(m, "Adapter interrupt tasklet runs/loops\t\t: %li/%li\n", +- (long)atomic_long_read(&perf_stats.tasklet_thinint), +- (long)atomic_long_read(&perf_stats.tasklet_thinint_loop)); ++ (long)atomic_long_read_unchecked(&perf_stats.tasklet_thinint), ++ (long)atomic_long_read_unchecked(&perf_stats.tasklet_thinint_loop)); + seq_printf(m, "Adapter interrupt inbound tasklet runs/loops\t: %li/%li\n", +- (long)atomic_long_read(&perf_stats.thinint_inbound), +- (long)atomic_long_read(&perf_stats.thinint_inbound_loop)); ++ (long)atomic_long_read_unchecked(&perf_stats.thinint_inbound), ++ (long)atomic_long_read_unchecked(&perf_stats.thinint_inbound_loop)); + seq_printf(m, "\n"); + seq_printf(m, "Number of SIGA In issued\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.siga_in)); ++ (long)atomic_long_read_unchecked(&perf_stats.siga_in)); + seq_printf(m, "Number of SIGA Out issued\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.siga_out)); ++ (long)atomic_long_read_unchecked(&perf_stats.siga_out)); + seq_printf(m, "Number of SIGA Sync issued\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.siga_sync)); ++ (long)atomic_long_read_unchecked(&perf_stats.siga_sync)); + seq_printf(m, "\n"); + seq_printf(m, "Number of inbound transfers\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.inbound_handler)); ++ (long)atomic_long_read_unchecked(&perf_stats.inbound_handler)); + seq_printf(m, "Number of outbound transfers\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.outbound_handler)); ++ (long)atomic_long_read_unchecked(&perf_stats.outbound_handler)); + seq_printf(m, "\n"); + seq_printf(m, "Number of fast requeues (outg. SBAL w/o SIGA)\t: %li\n", +- (long)atomic_long_read(&perf_stats.fast_requeue)); ++ (long)atomic_long_read_unchecked(&perf_stats.fast_requeue)); + seq_printf(m, "Number of outbound target full condition\t: %li\n", +- (long)atomic_long_read(&perf_stats.outbound_target_full)); ++ (long)atomic_long_read_unchecked(&perf_stats.outbound_target_full)); + seq_printf(m, "Number of outbound tasklet mod_timer calls\t: %li\n", +- (long)atomic_long_read(&perf_stats.debug_tl_out_timer)); ++ (long)atomic_long_read_unchecked(&perf_stats.debug_tl_out_timer)); + seq_printf(m, "Number of stop polling calls\t\t\t: %li\n", +- (long)atomic_long_read(&perf_stats.debug_stop_polling)); ++ (long)atomic_long_read_unchecked(&perf_stats.debug_stop_polling)); + seq_printf(m, "AI inbound tasklet loops after stop polling\t: %li\n", +- (long)atomic_long_read(&perf_stats.thinint_inbound_loop2)); ++ (long)atomic_long_read_unchecked(&perf_stats.thinint_inbound_loop2)); + seq_printf(m, "QEBSM EQBS total/incomplete\t\t\t: %li/%li\n", +- (long)atomic_long_read(&perf_stats.debug_eqbs_all), +- (long)atomic_long_read(&perf_stats.debug_eqbs_incomplete)); ++ (long)atomic_long_read_unchecked(&perf_stats.debug_eqbs_all), ++ (long)atomic_long_read_unchecked(&perf_stats.debug_eqbs_incomplete)); + seq_printf(m, "QEBSM SQBS total/incomplete\t\t\t: %li/%li\n", +- (long)atomic_long_read(&perf_stats.debug_sqbs_all), +- (long)atomic_long_read(&perf_stats.debug_sqbs_incomplete)); ++ (long)atomic_long_read_unchecked(&perf_stats.debug_sqbs_all), ++ (long)atomic_long_read_unchecked(&perf_stats.debug_sqbs_incomplete)); + seq_printf(m, "\n"); + return 0; + } +diff --git a/drivers/s390/cio/qdio_perf.h b/drivers/s390/cio/qdio_perf.h +index ff4504c..b3604c3 100644 +--- a/drivers/s390/cio/qdio_perf.h ++++ b/drivers/s390/cio/qdio_perf.h +@@ -13,46 +13,46 @@ + + struct qdio_perf_stats { + /* interrupt handler calls */ +- atomic_long_t qdio_int; +- atomic_long_t pci_int; +- atomic_long_t thin_int; ++ atomic_long_unchecked_t qdio_int; ++ atomic_long_unchecked_t pci_int; ++ atomic_long_unchecked_t thin_int; + + /* tasklet runs */ +- atomic_long_t tasklet_inbound; +- atomic_long_t tasklet_outbound; +- atomic_long_t tasklet_thinint; +- atomic_long_t tasklet_thinint_loop; +- atomic_long_t thinint_inbound; +- atomic_long_t thinint_inbound_loop; +- atomic_long_t thinint_inbound_loop2; ++ atomic_long_unchecked_t tasklet_inbound; ++ atomic_long_unchecked_t tasklet_outbound; ++ atomic_long_unchecked_t tasklet_thinint; ++ atomic_long_unchecked_t tasklet_thinint_loop; ++ atomic_long_unchecked_t thinint_inbound; ++ atomic_long_unchecked_t thinint_inbound_loop; ++ atomic_long_unchecked_t thinint_inbound_loop2; + + /* signal adapter calls */ +- atomic_long_t siga_out; +- atomic_long_t siga_in; +- atomic_long_t siga_sync; ++ atomic_long_unchecked_t siga_out; ++ atomic_long_unchecked_t siga_in; ++ atomic_long_unchecked_t siga_sync; + + /* misc */ +- atomic_long_t inbound_handler; +- atomic_long_t outbound_handler; +- atomic_long_t fast_requeue; +- atomic_long_t outbound_target_full; ++ atomic_long_unchecked_t inbound_handler; ++ atomic_long_unchecked_t outbound_handler; ++ atomic_long_unchecked_t fast_requeue; ++ atomic_long_unchecked_t outbound_target_full; + + /* for debugging */ +- atomic_long_t debug_tl_out_timer; +- atomic_long_t debug_stop_polling; +- atomic_long_t debug_eqbs_all; +- atomic_long_t debug_eqbs_incomplete; +- atomic_long_t debug_sqbs_all; +- atomic_long_t debug_sqbs_incomplete; ++ atomic_long_unchecked_t debug_tl_out_timer; ++ atomic_long_unchecked_t debug_stop_polling; ++ atomic_long_unchecked_t debug_eqbs_all; ++ atomic_long_unchecked_t debug_eqbs_incomplete; ++ atomic_long_unchecked_t debug_sqbs_all; ++ atomic_long_unchecked_t debug_sqbs_incomplete; + }; + + extern struct qdio_perf_stats perf_stats; + extern int qdio_performance_stats; + +-static inline void qdio_perf_stat_inc(atomic_long_t *count) ++static inline void qdio_perf_stat_inc(atomic_long_unchecked_t *count) + { + if (qdio_performance_stats) +- atomic_long_inc(count); ++ atomic_long_inc_unchecked(count); + } + + int qdio_setup_perf_stats(void); +diff --git a/drivers/scsi/3w-sas.c b/drivers/scsi/3w-sas.c +new file mode 100644 +index 0000000..7d18a18 +--- /dev/null ++++ b/drivers/scsi/3w-sas.c +@@ -0,0 +1,1933 @@ ++/* ++ 3w-sas.c -- LSI 3ware SAS/SATA-RAID Controller device driver for Linux. ++ ++ Written By: Adam Radford ++ ++ Copyright (C) 2009 LSI Corporation. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; version 2 of the License. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ NO WARRANTY ++ THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR ++ CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT ++ LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, ++ MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is ++ solely responsible for determining the appropriateness of using and ++ distributing the Program and assumes all risks associated with its ++ exercise of rights under this Agreement, including but not limited to ++ the risks and costs of program errors, damage to or loss of data, ++ programs or equipment, and unavailability or interruption of operations. ++ ++ DISCLAIMER OF LIABILITY ++ NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY ++ DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL ++ DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND ++ ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR ++ TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE ++ USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED ++ HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ ++ Controllers supported by this driver: ++ ++ LSI 3ware 9750 6Gb/s SAS/SATA-RAID ++ ++ Bugs/Comments/Suggestions should be mailed to: ++ linuxraid@lsi.com ++ ++ For more information, goto: ++ http://www.lsi.com ++ ++ History ++ ------- ++ 3.26.00.000 - Initial driver release. ++*/ ++ ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++#include "3w-sas.h" ++ ++/* Globals */ ++#define TW_DRIVER_VERSION "3.26.00.028-2.6.32RH" ++static TW_Device_Extension *twl_device_extension_list[TW_MAX_SLOT]; ++static unsigned int twl_device_extension_count; ++static int twl_major = -1; ++extern struct timezone sys_tz; ++ ++/* Module parameters */ ++MODULE_AUTHOR ("LSI"); ++MODULE_DESCRIPTION ("LSI 3ware SAS/SATA-RAID Linux Driver"); ++MODULE_LICENSE("GPL"); ++MODULE_VERSION(TW_DRIVER_VERSION); ++ ++static int use_msi = 0; ++module_param(use_msi, int, S_IRUGO); ++MODULE_PARM_DESC(use_msi, "Use Message Signaled Interrupts. Default: 0"); ++ ++/* Function prototypes */ ++static int twl_reset_device_extension(TW_Device_Extension *tw_dev, int ioctl_reset); ++ ++/* Functions */ ++ ++/* This function returns AENs through sysfs */ ++static ssize_t twl_sysfs_aen_read(struct file *filp, struct kobject *kobj, ++ struct bin_attribute *bin_attr, ++ char *outbuf, loff_t offset, size_t count) ++{ ++ struct device *dev = container_of(kobj, struct device, kobj); ++ struct Scsi_Host *shost = class_to_shost(dev); ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)shost->hostdata; ++ unsigned long flags = 0; ++ ssize_t ret; ++ ++ if (!capable(CAP_SYS_ADMIN)) ++ return -EACCES; ++ ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ ret = memory_read_from_buffer(outbuf, count, &offset, tw_dev->event_queue[0], sizeof(TW_Event) * TW_Q_LENGTH); ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ ++ return ret; ++} /* End twl_sysfs_aen_read() */ ++ ++/* aen_read sysfs attribute initializer */ ++static struct bin_attribute twl_sysfs_aen_read_attr = { ++ .attr = { ++ .name = "3ware_aen_read", ++ .mode = S_IRUSR, ++ }, ++ .size = 0, ++ .read = twl_sysfs_aen_read ++}; ++ ++/* This function returns driver compatibility info through sysfs */ ++static ssize_t twl_sysfs_compat_info(struct file *filp, struct kobject *kobj, ++ struct bin_attribute *bin_attr, ++ char *outbuf, loff_t offset, size_t count) ++{ ++ struct device *dev = container_of(kobj, struct device, kobj); ++ struct Scsi_Host *shost = class_to_shost(dev); ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)shost->hostdata; ++ unsigned long flags = 0; ++ ssize_t ret; ++ ++ if (!capable(CAP_SYS_ADMIN)) ++ return -EACCES; ++ ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ ret = memory_read_from_buffer(outbuf, count, &offset, &tw_dev->tw_compat_info, sizeof(TW_Compatibility_Info)); ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ ++ return ret; ++} /* End twl_sysfs_compat_info() */ ++ ++/* compat_info sysfs attribute initializer */ ++static struct bin_attribute twl_sysfs_compat_info_attr = { ++ .attr = { ++ .name = "3ware_compat_info", ++ .mode = S_IRUSR, ++ }, ++ .size = 0, ++ .read = twl_sysfs_compat_info ++}; ++ ++/* Show some statistics about the card */ ++static ssize_t twl_show_stats(struct device *dev, ++ struct device_attribute *attr, char *buf) ++{ ++ struct Scsi_Host *host = class_to_shost(dev); ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)host->hostdata; ++ unsigned long flags = 0; ++ ssize_t len; ++ ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ len = snprintf(buf, PAGE_SIZE, "3w-sas Driver version: %s\n" ++ "Current commands posted: %4d\n" ++ "Max commands posted: %4d\n" ++ "Last sgl length: %4d\n" ++ "Max sgl length: %4d\n" ++ "Last sector count: %4d\n" ++ "Max sector count: %4d\n" ++ "SCSI Host Resets: %4d\n" ++ "AEN's: %4d\n", ++ TW_DRIVER_VERSION, ++ tw_dev->posted_request_count, ++ tw_dev->max_posted_request_count, ++ tw_dev->sgl_entries, ++ tw_dev->max_sgl_entries, ++ tw_dev->sector_count, ++ tw_dev->max_sector_count, ++ tw_dev->num_resets, ++ tw_dev->aen_count); ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ return len; ++} /* End twl_show_stats() */ ++ ++/* This function will set a devices queue depth */ ++static int twl_change_queue_depth(struct scsi_device *sdev, int queue_depth, ++ int reason) ++{ ++ if (reason != SCSI_QDEPTH_DEFAULT) ++ return -EOPNOTSUPP; ++ ++ if (queue_depth > TW_Q_LENGTH-2) ++ queue_depth = TW_Q_LENGTH-2; ++ scsi_adjust_queue_depth(sdev, MSG_ORDERED_TAG, queue_depth); ++ return queue_depth; ++} /* End twl_change_queue_depth() */ ++ ++/* stats sysfs attribute initializer */ ++static struct device_attribute twl_host_stats_attr = { ++ .attr = { ++ .name = "3ware_stats", ++ .mode = S_IRUGO, ++ }, ++ .show = twl_show_stats ++}; ++ ++/* Host attributes initializer */ ++static struct device_attribute *twl_host_attrs[] = { ++ &twl_host_stats_attr, ++ NULL, ++}; ++ ++/* This function will look up an AEN severity string */ ++static char *twl_aen_severity_lookup(unsigned char severity_code) ++{ ++ char *retval = NULL; ++ ++ if ((severity_code < (unsigned char) TW_AEN_SEVERITY_ERROR) || ++ (severity_code > (unsigned char) TW_AEN_SEVERITY_DEBUG)) ++ goto out; ++ ++ retval = twl_aen_severity_table[severity_code]; ++out: ++ return retval; ++} /* End twl_aen_severity_lookup() */ ++ ++/* This function will queue an event */ ++static void twl_aen_queue_event(TW_Device_Extension *tw_dev, TW_Command_Apache_Header *header) ++{ ++ u32 local_time; ++ struct timeval time; ++ TW_Event *event; ++ unsigned short aen; ++ char host[16]; ++ char *error_str; ++ ++ tw_dev->aen_count++; ++ ++ /* Fill out event info */ ++ event = tw_dev->event_queue[tw_dev->error_index]; ++ ++ host[0] = '\0'; ++ if (tw_dev->host) ++ sprintf(host, " scsi%d:", tw_dev->host->host_no); ++ ++ aen = le16_to_cpu(header->status_block.error); ++ memset(event, 0, sizeof(TW_Event)); ++ ++ event->severity = TW_SEV_OUT(header->status_block.severity__reserved); ++ do_gettimeofday(&time); ++ local_time = (u32)(time.tv_sec - (sys_tz.tz_minuteswest * 60)); ++ event->time_stamp_sec = local_time; ++ event->aen_code = aen; ++ event->retrieved = TW_AEN_NOT_RETRIEVED; ++ event->sequence_id = tw_dev->error_sequence_id; ++ tw_dev->error_sequence_id++; ++ ++ /* Check for embedded error string */ ++ error_str = &(header->err_specific_desc[strlen(header->err_specific_desc)+1]); ++ ++ header->err_specific_desc[sizeof(header->err_specific_desc) - 1] = '\0'; ++ event->parameter_len = strlen(header->err_specific_desc); ++ memcpy(event->parameter_data, header->err_specific_desc, event->parameter_len + 1 + strlen(error_str)); ++ if (event->severity != TW_AEN_SEVERITY_DEBUG) ++ printk(KERN_WARNING "3w-sas:%s AEN: %s (0x%02X:0x%04X): %s:%s.\n", ++ host, ++ twl_aen_severity_lookup(TW_SEV_OUT(header->status_block.severity__reserved)), ++ TW_MESSAGE_SOURCE_CONTROLLER_EVENT, aen, error_str, ++ header->err_specific_desc); ++ else ++ tw_dev->aen_count--; ++ ++ tw_dev->error_index = (tw_dev->error_index + 1 ) % TW_Q_LENGTH; ++} /* End twl_aen_queue_event() */ ++ ++/* This function will attempt to post a command packet to the board */ ++static int twl_post_command_packet(TW_Device_Extension *tw_dev, int request_id) ++{ ++ dma_addr_t command_que_value; ++ ++ command_que_value = tw_dev->command_packet_phys[request_id]; ++ command_que_value += TW_COMMAND_OFFSET; ++ ++ /* First write upper 4 bytes */ ++ writel((u32)((u64)command_que_value >> 32), TWL_HIBQPH_REG_ADDR(tw_dev)); ++ /* Then the lower 4 bytes */ ++ writel((u32)(command_que_value | TWL_PULL_MODE), TWL_HIBQPL_REG_ADDR(tw_dev)); ++ ++ tw_dev->state[request_id] = TW_S_POSTED; ++ tw_dev->posted_request_count++; ++ if (tw_dev->posted_request_count > tw_dev->max_posted_request_count) ++ tw_dev->max_posted_request_count = tw_dev->posted_request_count; ++ ++ return 0; ++} /* End twl_post_command_packet() */ ++ ++/* This function will perform a pci-dma mapping for a scatter gather list */ ++static int twl_map_scsi_sg_data(TW_Device_Extension *tw_dev, int request_id) ++{ ++ int use_sg; ++ struct scsi_cmnd *cmd = tw_dev->srb[request_id]; ++ ++ use_sg = scsi_dma_map(cmd); ++ if (!use_sg) ++ return 0; ++ else if (use_sg < 0) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1, "Failed to map scatter gather list"); ++ return 0; ++ } ++ ++ cmd->SCp.phase = TW_PHASE_SGLIST; ++ cmd->SCp.have_data_in = use_sg; ++ ++ return use_sg; ++} /* End twl_map_scsi_sg_data() */ ++ ++/* This function hands scsi cdb's to the firmware */ ++static int twl_scsiop_execute_scsi(TW_Device_Extension *tw_dev, int request_id, char *cdb, int use_sg, TW_SG_Entry_ISO *sglistarg) ++{ ++ TW_Command_Full *full_command_packet; ++ TW_Command_Apache *command_packet; ++ int i, sg_count; ++ struct scsi_cmnd *srb = NULL; ++ struct scatterlist *sglist = NULL, *sg; ++ int retval = 1; ++ ++ if (tw_dev->srb[request_id]) { ++ srb = tw_dev->srb[request_id]; ++ if (scsi_sglist(srb)) ++ sglist = scsi_sglist(srb); ++ } ++ ++ /* Initialize command packet */ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ full_command_packet->header.header_desc.size_header = 128; ++ full_command_packet->header.status_block.error = 0; ++ full_command_packet->header.status_block.severity__reserved = 0; ++ ++ command_packet = &full_command_packet->command.newcommand; ++ command_packet->status = 0; ++ command_packet->opcode__reserved = TW_OPRES_IN(0, TW_OP_EXECUTE_SCSI); ++ ++ /* We forced 16 byte cdb use earlier */ ++ if (!cdb) ++ memcpy(command_packet->cdb, srb->cmnd, TW_MAX_CDB_LEN); ++ else ++ memcpy(command_packet->cdb, cdb, TW_MAX_CDB_LEN); ++ ++ if (srb) { ++ command_packet->unit = srb->device->id; ++ command_packet->request_id__lunl = ++ cpu_to_le16(TW_REQ_LUN_IN(srb->device->lun, request_id)); ++ } else { ++ command_packet->request_id__lunl = ++ cpu_to_le16(TW_REQ_LUN_IN(0, request_id)); ++ command_packet->unit = 0; ++ } ++ ++ command_packet->sgl_offset = 16; ++ ++ if (!sglistarg) { ++ /* Map sglist from scsi layer to cmd packet */ ++ if (scsi_sg_count(srb)) { ++ sg_count = twl_map_scsi_sg_data(tw_dev, request_id); ++ if (sg_count == 0) ++ goto out; ++ ++ scsi_for_each_sg(srb, sg, sg_count, i) { ++ command_packet->sg_list[i].address = TW_CPU_TO_SGL(sg_dma_address(sg)); ++ command_packet->sg_list[i].length = TW_CPU_TO_SGL(sg_dma_len(sg)); ++ } ++ command_packet->sgl_entries__lunh = cpu_to_le16(TW_REQ_LUN_IN((srb->device->lun >> 4), scsi_sg_count(tw_dev->srb[request_id]))); ++ } ++ } else { ++ /* Internal cdb post */ ++ for (i = 0; i < use_sg; i++) { ++ command_packet->sg_list[i].address = TW_CPU_TO_SGL(sglistarg[i].address); ++ command_packet->sg_list[i].length = TW_CPU_TO_SGL(sglistarg[i].length); ++ } ++ command_packet->sgl_entries__lunh = cpu_to_le16(TW_REQ_LUN_IN(0, use_sg)); ++ } ++ ++ /* Update some stats */ ++ if (srb) { ++ tw_dev->sector_count = scsi_bufflen(srb) / 512; ++ if (tw_dev->sector_count > tw_dev->max_sector_count) ++ tw_dev->max_sector_count = tw_dev->sector_count; ++ tw_dev->sgl_entries = scsi_sg_count(srb); ++ if (tw_dev->sgl_entries > tw_dev->max_sgl_entries) ++ tw_dev->max_sgl_entries = tw_dev->sgl_entries; ++ } ++ ++ /* Now post the command to the board */ ++ retval = twl_post_command_packet(tw_dev, request_id); ++ ++out: ++ return retval; ++} /* End twl_scsiop_execute_scsi() */ ++ ++/* This function will read the aen queue from the isr */ ++static int twl_aen_read_queue(TW_Device_Extension *tw_dev, int request_id) ++{ ++ char cdb[TW_MAX_CDB_LEN]; ++ TW_SG_Entry_ISO sglist[1]; ++ TW_Command_Full *full_command_packet; ++ int retval = 1; ++ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ memset(full_command_packet, 0, sizeof(TW_Command_Full)); ++ ++ /* Initialize cdb */ ++ memset(&cdb, 0, TW_MAX_CDB_LEN); ++ cdb[0] = REQUEST_SENSE; /* opcode */ ++ cdb[4] = TW_ALLOCATION_LENGTH; /* allocation length */ ++ ++ /* Initialize sglist */ ++ memset(&sglist, 0, sizeof(TW_SG_Entry_ISO)); ++ sglist[0].length = TW_SECTOR_SIZE; ++ sglist[0].address = tw_dev->generic_buffer_phys[request_id]; ++ ++ /* Mark internal command */ ++ tw_dev->srb[request_id] = NULL; ++ ++ /* Now post the command packet */ ++ if (twl_scsiop_execute_scsi(tw_dev, request_id, cdb, 1, sglist)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x2, "Post failed while reading AEN queue"); ++ goto out; ++ } ++ retval = 0; ++out: ++ return retval; ++} /* End twl_aen_read_queue() */ ++ ++/* This function will sync firmware time with the host time */ ++static void twl_aen_sync_time(TW_Device_Extension *tw_dev, int request_id) ++{ ++ u32 schedulertime; ++ struct timeval utc; ++ TW_Command_Full *full_command_packet; ++ TW_Command *command_packet; ++ TW_Param_Apache *param; ++ u32 local_time; ++ ++ /* Fill out the command packet */ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ memset(full_command_packet, 0, sizeof(TW_Command_Full)); ++ command_packet = &full_command_packet->command.oldcommand; ++ command_packet->opcode__sgloffset = TW_OPSGL_IN(2, TW_OP_SET_PARAM); ++ command_packet->request_id = request_id; ++ command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); ++ command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE); ++ command_packet->size = TW_COMMAND_SIZE; ++ command_packet->byte6_offset.parameter_count = cpu_to_le16(1); ++ ++ /* Setup the param */ ++ param = (TW_Param_Apache *)tw_dev->generic_buffer_virt[request_id]; ++ memset(param, 0, TW_SECTOR_SIZE); ++ param->table_id = cpu_to_le16(TW_TIMEKEEP_TABLE | 0x8000); /* Controller time keep table */ ++ param->parameter_id = cpu_to_le16(0x3); /* SchedulerTime */ ++ param->parameter_size_bytes = cpu_to_le16(4); ++ ++ /* Convert system time in UTC to local time seconds since last ++ Sunday 12:00AM */ ++ do_gettimeofday(&utc); ++ local_time = (u32)(utc.tv_sec - (sys_tz.tz_minuteswest * 60)); ++ schedulertime = local_time - (3 * 86400); ++ schedulertime = cpu_to_le32(schedulertime % 604800); ++ ++ memcpy(param->data, &schedulertime, sizeof(u32)); ++ ++ /* Mark internal command */ ++ tw_dev->srb[request_id] = NULL; ++ ++ /* Now post the command */ ++ twl_post_command_packet(tw_dev, request_id); ++} /* End twl_aen_sync_time() */ ++ ++/* This function will assign an available request id */ ++static void twl_get_request_id(TW_Device_Extension *tw_dev, int *request_id) ++{ ++ *request_id = tw_dev->free_queue[tw_dev->free_head]; ++ tw_dev->free_head = (tw_dev->free_head + 1) % TW_Q_LENGTH; ++ tw_dev->state[*request_id] = TW_S_STARTED; ++} /* End twl_get_request_id() */ ++ ++/* This function will free a request id */ ++static void twl_free_request_id(TW_Device_Extension *tw_dev, int request_id) ++{ ++ tw_dev->free_queue[tw_dev->free_tail] = request_id; ++ tw_dev->state[request_id] = TW_S_FINISHED; ++ tw_dev->free_tail = (tw_dev->free_tail + 1) % TW_Q_LENGTH; ++} /* End twl_free_request_id() */ ++ ++/* This function will complete an aen request from the isr */ ++static int twl_aen_complete(TW_Device_Extension *tw_dev, int request_id) ++{ ++ TW_Command_Full *full_command_packet; ++ TW_Command *command_packet; ++ TW_Command_Apache_Header *header; ++ unsigned short aen; ++ int retval = 1; ++ ++ header = (TW_Command_Apache_Header *)tw_dev->generic_buffer_virt[request_id]; ++ tw_dev->posted_request_count--; ++ aen = le16_to_cpu(header->status_block.error); ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ command_packet = &full_command_packet->command.oldcommand; ++ ++ /* First check for internal completion of set param for time sync */ ++ if (TW_OP_OUT(command_packet->opcode__sgloffset) == TW_OP_SET_PARAM) { ++ /* Keep reading the queue in case there are more aen's */ ++ if (twl_aen_read_queue(tw_dev, request_id)) ++ goto out2; ++ else { ++ retval = 0; ++ goto out; ++ } ++ } ++ ++ switch (aen) { ++ case TW_AEN_QUEUE_EMPTY: ++ /* Quit reading the queue if this is the last one */ ++ break; ++ case TW_AEN_SYNC_TIME_WITH_HOST: ++ twl_aen_sync_time(tw_dev, request_id); ++ retval = 0; ++ goto out; ++ default: ++ twl_aen_queue_event(tw_dev, header); ++ ++ /* If there are more aen's, keep reading the queue */ ++ if (twl_aen_read_queue(tw_dev, request_id)) ++ goto out2; ++ else { ++ retval = 0; ++ goto out; ++ } ++ } ++ retval = 0; ++out2: ++ tw_dev->state[request_id] = TW_S_COMPLETED; ++ twl_free_request_id(tw_dev, request_id); ++ clear_bit(TW_IN_ATTENTION_LOOP, &tw_dev->flags); ++out: ++ return retval; ++} /* End twl_aen_complete() */ ++ ++/* This function will poll for a response */ ++static int twl_poll_response(TW_Device_Extension *tw_dev, int request_id, int seconds) ++{ ++ unsigned long before; ++ dma_addr_t mfa; ++ u32 regh, regl; ++ u32 response; ++ int retval = 1; ++ int found = 0; ++ ++ before = jiffies; ++ ++ while (!found) { ++ if (sizeof(dma_addr_t) > 4) { ++ regh = readl(TWL_HOBQPH_REG_ADDR(tw_dev)); ++ regl = readl(TWL_HOBQPL_REG_ADDR(tw_dev)); ++ mfa = ((u64)regh << 32) | regl; ++ } else ++ mfa = readl(TWL_HOBQPL_REG_ADDR(tw_dev)); ++ ++ response = (u32)mfa; ++ ++ if (TW_RESID_OUT(response) == request_id) ++ found = 1; ++ ++ if (time_after(jiffies, before + HZ * seconds)) ++ goto out; ++ ++ msleep(50); ++ } ++ retval = 0; ++out: ++ return retval; ++} /* End twl_poll_response() */ ++ ++/* This function will drain the aen queue */ ++static int twl_aen_drain_queue(TW_Device_Extension *tw_dev, int no_check_reset) ++{ ++ int request_id = 0; ++ char cdb[TW_MAX_CDB_LEN]; ++ TW_SG_Entry_ISO sglist[1]; ++ int finished = 0, count = 0; ++ TW_Command_Full *full_command_packet; ++ TW_Command_Apache_Header *header; ++ unsigned short aen; ++ int first_reset = 0, queue = 0, retval = 1; ++ ++ if (no_check_reset) ++ first_reset = 0; ++ else ++ first_reset = 1; ++ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ memset(full_command_packet, 0, sizeof(TW_Command_Full)); ++ ++ /* Initialize cdb */ ++ memset(&cdb, 0, TW_MAX_CDB_LEN); ++ cdb[0] = REQUEST_SENSE; /* opcode */ ++ cdb[4] = TW_ALLOCATION_LENGTH; /* allocation length */ ++ ++ /* Initialize sglist */ ++ memset(&sglist, 0, sizeof(TW_SG_Entry_ISO)); ++ sglist[0].length = TW_SECTOR_SIZE; ++ sglist[0].address = tw_dev->generic_buffer_phys[request_id]; ++ ++ /* Mark internal command */ ++ tw_dev->srb[request_id] = NULL; ++ ++ do { ++ /* Send command to the board */ ++ if (twl_scsiop_execute_scsi(tw_dev, request_id, cdb, 1, sglist)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x3, "Error posting request sense"); ++ goto out; ++ } ++ ++ /* Now poll for completion */ ++ if (twl_poll_response(tw_dev, request_id, 30)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x4, "No valid response while draining AEN queue"); ++ tw_dev->posted_request_count--; ++ goto out; ++ } ++ ++ tw_dev->posted_request_count--; ++ header = (TW_Command_Apache_Header *)tw_dev->generic_buffer_virt[request_id]; ++ aen = le16_to_cpu(header->status_block.error); ++ queue = 0; ++ count++; ++ ++ switch (aen) { ++ case TW_AEN_QUEUE_EMPTY: ++ if (first_reset != 1) ++ goto out; ++ else ++ finished = 1; ++ break; ++ case TW_AEN_SOFT_RESET: ++ if (first_reset == 0) ++ first_reset = 1; ++ else ++ queue = 1; ++ break; ++ case TW_AEN_SYNC_TIME_WITH_HOST: ++ break; ++ default: ++ queue = 1; ++ } ++ ++ /* Now queue an event info */ ++ if (queue) ++ twl_aen_queue_event(tw_dev, header); ++ } while ((finished == 0) && (count < TW_MAX_AEN_DRAIN)); ++ ++ if (count == TW_MAX_AEN_DRAIN) ++ goto out; ++ ++ retval = 0; ++out: ++ tw_dev->state[request_id] = TW_S_INITIAL; ++ return retval; ++} /* End twl_aen_drain_queue() */ ++ ++/* This function will allocate memory and check if it is correctly aligned */ ++static int twl_allocate_memory(TW_Device_Extension *tw_dev, int size, int which) ++{ ++ int i; ++ dma_addr_t dma_handle; ++ unsigned long *cpu_addr; ++ int retval = 1; ++ ++ cpu_addr = pci_alloc_consistent(tw_dev->tw_pci_dev, size*TW_Q_LENGTH, &dma_handle); ++ if (!cpu_addr) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x5, "Memory allocation failed"); ++ goto out; ++ } ++ ++ memset(cpu_addr, 0, size*TW_Q_LENGTH); ++ ++ for (i = 0; i < TW_Q_LENGTH; i++) { ++ switch(which) { ++ case 0: ++ tw_dev->command_packet_phys[i] = dma_handle+(i*size); ++ tw_dev->command_packet_virt[i] = (TW_Command_Full *)((unsigned char *)cpu_addr + (i*size)); ++ break; ++ case 1: ++ tw_dev->generic_buffer_phys[i] = dma_handle+(i*size); ++ tw_dev->generic_buffer_virt[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size)); ++ break; ++ case 2: ++ tw_dev->sense_buffer_phys[i] = dma_handle+(i*size); ++ tw_dev->sense_buffer_virt[i] = (TW_Command_Apache_Header *)((unsigned char *)cpu_addr + (i*size)); ++ break; ++ } ++ } ++ retval = 0; ++out: ++ return retval; ++} /* End twl_allocate_memory() */ ++ ++/* This function will load the request id and various sgls for ioctls */ ++static void twl_load_sgl(TW_Device_Extension *tw_dev, TW_Command_Full *full_command_packet, int request_id, dma_addr_t dma_handle, int length) ++{ ++ TW_Command *oldcommand; ++ TW_Command_Apache *newcommand; ++ TW_SG_Entry_ISO *sgl; ++ unsigned int pae = 0; ++ ++ if ((sizeof(long) < 8) && (sizeof(dma_addr_t) > 4)) ++ pae = 1; ++ ++ if (TW_OP_OUT(full_command_packet->command.newcommand.opcode__reserved) == TW_OP_EXECUTE_SCSI) { ++ newcommand = &full_command_packet->command.newcommand; ++ newcommand->request_id__lunl = ++ cpu_to_le16(TW_REQ_LUN_IN(TW_LUN_OUT(newcommand->request_id__lunl), request_id)); ++ if (length) { ++ newcommand->sg_list[0].address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); ++ newcommand->sg_list[0].length = TW_CPU_TO_SGL(length); ++ } ++ newcommand->sgl_entries__lunh = ++ cpu_to_le16(TW_REQ_LUN_IN(TW_LUN_OUT(newcommand->sgl_entries__lunh), length ? 1 : 0)); ++ } else { ++ oldcommand = &full_command_packet->command.oldcommand; ++ oldcommand->request_id = request_id; ++ ++ if (TW_SGL_OUT(oldcommand->opcode__sgloffset)) { ++ /* Load the sg list */ ++ sgl = (TW_SG_Entry_ISO *)((u32 *)oldcommand+oldcommand->size - (sizeof(TW_SG_Entry_ISO)/4) + pae + (sizeof(dma_addr_t) > 4 ? 1 : 0)); ++ sgl->address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); ++ sgl->length = TW_CPU_TO_SGL(length); ++ oldcommand->size += pae; ++ oldcommand->size += sizeof(dma_addr_t) > 4 ? 1 : 0; ++ } ++ } ++} /* End twl_load_sgl() */ ++ ++/* This function handles ioctl for the character device ++ This interface is used by smartmontools open source software */ ++static int twl_chrdev_ioctl(struct inode *inode, struct file *file, unsigned int cmd, unsigned long arg) ++{ ++ long timeout; ++ unsigned long *cpu_addr, data_buffer_length_adjusted = 0, flags = 0; ++ dma_addr_t dma_handle; ++ int request_id = 0; ++ TW_Ioctl_Driver_Command driver_command; ++ TW_Ioctl_Buf_Apache *tw_ioctl; ++ TW_Command_Full *full_command_packet; ++ TW_Device_Extension *tw_dev = twl_device_extension_list[iminor(inode)]; ++ int retval = -EFAULT; ++ void __user *argp = (void __user *)arg; ++ ++ /* Only let one of these through at a time */ ++ if (mutex_lock_interruptible(&tw_dev->ioctl_lock)) { ++ retval = -EINTR; ++ goto out; ++ } ++ ++ /* First copy down the driver command */ ++ if (copy_from_user(&driver_command, argp, sizeof(TW_Ioctl_Driver_Command))) ++ goto out2; ++ ++ /* Check data buffer size */ ++ if (driver_command.buffer_length > TW_MAX_SECTORS * 2048) { ++ retval = -EINVAL; ++ goto out2; ++ } ++ ++ /* Hardware can only do multiple of 512 byte transfers */ ++ data_buffer_length_adjusted = (driver_command.buffer_length + 511) & ~511; ++ ++ /* Now allocate ioctl buf memory */ ++ cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, &dma_handle, GFP_KERNEL); ++ if (!cpu_addr) { ++ retval = -ENOMEM; ++ goto out2; ++ } ++ ++ tw_ioctl = (TW_Ioctl_Buf_Apache *)cpu_addr; ++ ++ /* Now copy down the entire ioctl */ ++ if (copy_from_user(tw_ioctl, argp, driver_command.buffer_length + sizeof(TW_Ioctl_Buf_Apache) - 1)) ++ goto out3; ++ ++ /* See which ioctl we are doing */ ++ switch (cmd) { ++ case TW_IOCTL_FIRMWARE_PASS_THROUGH: ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ twl_get_request_id(tw_dev, &request_id); ++ ++ /* Flag internal command */ ++ tw_dev->srb[request_id] = NULL; ++ ++ /* Flag chrdev ioctl */ ++ tw_dev->chrdev_request_id = request_id; ++ ++ full_command_packet = (TW_Command_Full *)&tw_ioctl->firmware_command; ++ ++ /* Load request id and sglist for both command types */ ++ twl_load_sgl(tw_dev, full_command_packet, request_id, dma_handle, data_buffer_length_adjusted); ++ ++ memcpy(tw_dev->command_packet_virt[request_id], &(tw_ioctl->firmware_command), sizeof(TW_Command_Full)); ++ ++ /* Now post the command packet to the controller */ ++ twl_post_command_packet(tw_dev, request_id); ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ ++ timeout = TW_IOCTL_CHRDEV_TIMEOUT*HZ; ++ ++ /* Now wait for command to complete */ ++ timeout = wait_event_timeout(tw_dev->ioctl_wqueue, tw_dev->chrdev_request_id == TW_IOCTL_CHRDEV_FREE, timeout); ++ ++ /* We timed out, and didn't get an interrupt */ ++ if (tw_dev->chrdev_request_id != TW_IOCTL_CHRDEV_FREE) { ++ /* Now we need to reset the board */ ++ printk(KERN_WARNING "3w-sas: scsi%d: WARNING: (0x%02X:0x%04X): Character ioctl (0x%x) timed out, resetting card.\n", ++ tw_dev->host->host_no, TW_DRIVER, 0x6, ++ cmd); ++ retval = -EIO; ++ twl_reset_device_extension(tw_dev, 1); ++ goto out3; ++ } ++ ++ /* Now copy in the command packet response */ ++ memcpy(&(tw_ioctl->firmware_command), tw_dev->command_packet_virt[request_id], sizeof(TW_Command_Full)); ++ ++ /* Now complete the io */ ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ tw_dev->posted_request_count--; ++ tw_dev->state[request_id] = TW_S_COMPLETED; ++ twl_free_request_id(tw_dev, request_id); ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ break; ++ default: ++ retval = -ENOTTY; ++ goto out3; ++ } ++ ++ /* Now copy the entire response to userspace */ ++ if (copy_to_user(argp, tw_ioctl, sizeof(TW_Ioctl_Buf_Apache) + driver_command.buffer_length - 1) == 0) ++ retval = 0; ++out3: ++ /* Now free ioctl buf memory */ ++ dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, cpu_addr, dma_handle); ++out2: ++ mutex_unlock(&tw_dev->ioctl_lock); ++out: ++ return retval; ++} /* End twl_chrdev_ioctl() */ ++ ++/* This function handles open for the character device */ ++static int twl_chrdev_open(struct inode *inode, struct file *file) ++{ ++ unsigned int minor_number; ++ int retval = -ENODEV; ++ ++ if (!capable(CAP_SYS_ADMIN)) { ++ retval = -EACCES; ++ goto out; ++ } ++ ++ cycle_kernel_lock(); ++ minor_number = iminor(inode); ++ if (minor_number >= twl_device_extension_count) ++ goto out; ++ retval = 0; ++out: ++ return retval; ++} /* End twl_chrdev_open() */ ++ ++/* File operations struct for character device */ ++static const struct file_operations twl_fops = { ++ .owner = THIS_MODULE, ++ .ioctl = twl_chrdev_ioctl, ++ .open = twl_chrdev_open, ++ .release = NULL ++}; ++ ++/* This function passes sense data from firmware to scsi layer */ ++static int twl_fill_sense(TW_Device_Extension *tw_dev, int i, int request_id, int copy_sense, int print_host) ++{ ++ TW_Command_Apache_Header *header; ++ TW_Command_Full *full_command_packet; ++ unsigned short error; ++ char *error_str; ++ int retval = 1; ++ ++ header = tw_dev->sense_buffer_virt[i]; ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ ++ /* Get embedded firmware error string */ ++ error_str = &(header->err_specific_desc[strlen(header->err_specific_desc) + 1]); ++ ++ /* Don't print error for Logical unit not supported during rollcall */ ++ error = le16_to_cpu(header->status_block.error); ++ if ((error != TW_ERROR_LOGICAL_UNIT_NOT_SUPPORTED) && (error != TW_ERROR_UNIT_OFFLINE) && (error != TW_ERROR_INVALID_FIELD_IN_CDB)) { ++ if (print_host) ++ printk(KERN_WARNING "3w-sas: scsi%d: ERROR: (0x%02X:0x%04X): %s:%s.\n", ++ tw_dev->host->host_no, ++ TW_MESSAGE_SOURCE_CONTROLLER_ERROR, ++ header->status_block.error, ++ error_str, ++ header->err_specific_desc); ++ else ++ printk(KERN_WARNING "3w-sas: ERROR: (0x%02X:0x%04X): %s:%s.\n", ++ TW_MESSAGE_SOURCE_CONTROLLER_ERROR, ++ header->status_block.error, ++ error_str, ++ header->err_specific_desc); ++ } ++ ++ if (copy_sense) { ++ memcpy(tw_dev->srb[request_id]->sense_buffer, header->sense_data, TW_SENSE_DATA_LENGTH); ++ tw_dev->srb[request_id]->result = (full_command_packet->command.newcommand.status << 1); ++ goto out; ++ } ++out: ++ return retval; ++} /* End twl_fill_sense() */ ++ ++/* This function will free up device extension resources */ ++static void twl_free_device_extension(TW_Device_Extension *tw_dev) ++{ ++ if (tw_dev->command_packet_virt[0]) ++ pci_free_consistent(tw_dev->tw_pci_dev, ++ sizeof(TW_Command_Full)*TW_Q_LENGTH, ++ tw_dev->command_packet_virt[0], ++ tw_dev->command_packet_phys[0]); ++ ++ if (tw_dev->generic_buffer_virt[0]) ++ pci_free_consistent(tw_dev->tw_pci_dev, ++ TW_SECTOR_SIZE*TW_Q_LENGTH, ++ tw_dev->generic_buffer_virt[0], ++ tw_dev->generic_buffer_phys[0]); ++ ++ if (tw_dev->sense_buffer_virt[0]) ++ pci_free_consistent(tw_dev->tw_pci_dev, ++ sizeof(TW_Command_Apache_Header)* ++ TW_Q_LENGTH, ++ tw_dev->sense_buffer_virt[0], ++ tw_dev->sense_buffer_phys[0]); ++ ++ kfree(tw_dev->event_queue[0]); ++} /* End twl_free_device_extension() */ ++ ++/* This function will get parameter table entries from the firmware */ ++static void *twl_get_param(TW_Device_Extension *tw_dev, int request_id, int table_id, int parameter_id, int parameter_size_bytes) ++{ ++ TW_Command_Full *full_command_packet; ++ TW_Command *command_packet; ++ TW_Param_Apache *param; ++ void *retval = NULL; ++ ++ /* Setup the command packet */ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ memset(full_command_packet, 0, sizeof(TW_Command_Full)); ++ command_packet = &full_command_packet->command.oldcommand; ++ ++ command_packet->opcode__sgloffset = TW_OPSGL_IN(2, TW_OP_GET_PARAM); ++ command_packet->size = TW_COMMAND_SIZE; ++ command_packet->request_id = request_id; ++ command_packet->byte6_offset.block_count = cpu_to_le16(1); ++ ++ /* Now setup the param */ ++ param = (TW_Param_Apache *)tw_dev->generic_buffer_virt[request_id]; ++ memset(param, 0, TW_SECTOR_SIZE); ++ param->table_id = cpu_to_le16(table_id | 0x8000); ++ param->parameter_id = cpu_to_le16(parameter_id); ++ param->parameter_size_bytes = cpu_to_le16(parameter_size_bytes); ++ ++ command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); ++ command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE); ++ ++ /* Post the command packet to the board */ ++ twl_post_command_packet(tw_dev, request_id); ++ ++ /* Poll for completion */ ++ if (twl_poll_response(tw_dev, request_id, 30)) ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x7, "No valid response during get param") ++ else ++ retval = (void *)&(param->data[0]); ++ ++ tw_dev->posted_request_count--; ++ tw_dev->state[request_id] = TW_S_INITIAL; ++ ++ return retval; ++} /* End twl_get_param() */ ++ ++/* This function will send an initconnection command to controller */ ++static int twl_initconnection(TW_Device_Extension *tw_dev, int message_credits, ++ u32 set_features, unsigned short current_fw_srl, ++ unsigned short current_fw_arch_id, ++ unsigned short current_fw_branch, ++ unsigned short current_fw_build, ++ unsigned short *fw_on_ctlr_srl, ++ unsigned short *fw_on_ctlr_arch_id, ++ unsigned short *fw_on_ctlr_branch, ++ unsigned short *fw_on_ctlr_build, ++ u32 *init_connect_result) ++{ ++ TW_Command_Full *full_command_packet; ++ TW_Initconnect *tw_initconnect; ++ int request_id = 0, retval = 1; ++ ++ /* Initialize InitConnection command packet */ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ memset(full_command_packet, 0, sizeof(TW_Command_Full)); ++ full_command_packet->header.header_desc.size_header = 128; ++ ++ tw_initconnect = (TW_Initconnect *)&full_command_packet->command.oldcommand; ++ tw_initconnect->opcode__reserved = TW_OPRES_IN(0, TW_OP_INIT_CONNECTION); ++ tw_initconnect->request_id = request_id; ++ tw_initconnect->message_credits = cpu_to_le16(message_credits); ++ tw_initconnect->features = set_features; ++ ++ /* Turn on 64-bit sgl support if we need to */ ++ tw_initconnect->features |= sizeof(dma_addr_t) > 4 ? 1 : 0; ++ ++ tw_initconnect->features = cpu_to_le32(tw_initconnect->features); ++ ++ if (set_features & TW_EXTENDED_INIT_CONNECT) { ++ tw_initconnect->size = TW_INIT_COMMAND_PACKET_SIZE_EXTENDED; ++ tw_initconnect->fw_srl = cpu_to_le16(current_fw_srl); ++ tw_initconnect->fw_arch_id = cpu_to_le16(current_fw_arch_id); ++ tw_initconnect->fw_branch = cpu_to_le16(current_fw_branch); ++ tw_initconnect->fw_build = cpu_to_le16(current_fw_build); ++ } else ++ tw_initconnect->size = TW_INIT_COMMAND_PACKET_SIZE; ++ ++ /* Send command packet to the board */ ++ twl_post_command_packet(tw_dev, request_id); ++ ++ /* Poll for completion */ ++ if (twl_poll_response(tw_dev, request_id, 30)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x8, "No valid response during init connection"); ++ } else { ++ if (set_features & TW_EXTENDED_INIT_CONNECT) { ++ *fw_on_ctlr_srl = le16_to_cpu(tw_initconnect->fw_srl); ++ *fw_on_ctlr_arch_id = le16_to_cpu(tw_initconnect->fw_arch_id); ++ *fw_on_ctlr_branch = le16_to_cpu(tw_initconnect->fw_branch); ++ *fw_on_ctlr_build = le16_to_cpu(tw_initconnect->fw_build); ++ *init_connect_result = le32_to_cpu(tw_initconnect->result); ++ } ++ retval = 0; ++ } ++ ++ tw_dev->posted_request_count--; ++ tw_dev->state[request_id] = TW_S_INITIAL; ++ ++ return retval; ++} /* End twl_initconnection() */ ++ ++/* This function will initialize the fields of a device extension */ ++static int twl_initialize_device_extension(TW_Device_Extension *tw_dev) ++{ ++ int i, retval = 1; ++ ++ /* Initialize command packet buffers */ ++ if (twl_allocate_memory(tw_dev, sizeof(TW_Command_Full), 0)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x9, "Command packet memory allocation failed"); ++ goto out; ++ } ++ ++ /* Initialize generic buffer */ ++ if (twl_allocate_memory(tw_dev, TW_SECTOR_SIZE, 1)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xa, "Generic memory allocation failed"); ++ goto out; ++ } ++ ++ /* Allocate sense buffers */ ++ if (twl_allocate_memory(tw_dev, sizeof(TW_Command_Apache_Header), 2)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xb, "Sense buffer allocation failed"); ++ goto out; ++ } ++ ++ /* Allocate event info space */ ++ tw_dev->event_queue[0] = kcalloc(TW_Q_LENGTH, sizeof(TW_Event), GFP_KERNEL); ++ if (!tw_dev->event_queue[0]) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xc, "Event info memory allocation failed"); ++ goto out; ++ } ++ ++ for (i = 0; i < TW_Q_LENGTH; i++) { ++ tw_dev->event_queue[i] = (TW_Event *)((unsigned char *)tw_dev->event_queue[0] + (i * sizeof(TW_Event))); ++ tw_dev->free_queue[i] = i; ++ tw_dev->state[i] = TW_S_INITIAL; ++ } ++ ++ tw_dev->free_head = TW_Q_START; ++ tw_dev->free_tail = TW_Q_START; ++ tw_dev->error_sequence_id = 1; ++ tw_dev->chrdev_request_id = TW_IOCTL_CHRDEV_FREE; ++ ++ mutex_init(&tw_dev->ioctl_lock); ++ init_waitqueue_head(&tw_dev->ioctl_wqueue); ++ ++ retval = 0; ++out: ++ return retval; ++} /* End twl_initialize_device_extension() */ ++ ++/* This function will perform a pci-dma unmap */ ++static void twl_unmap_scsi_data(TW_Device_Extension *tw_dev, int request_id) ++{ ++ struct scsi_cmnd *cmd = tw_dev->srb[request_id]; ++ ++ if (cmd->SCp.phase == TW_PHASE_SGLIST) ++ scsi_dma_unmap(cmd); ++} /* End twl_unmap_scsi_data() */ ++ ++/* This function will handle attention interrupts */ ++static int twl_handle_attention_interrupt(TW_Device_Extension *tw_dev) ++{ ++ int retval = 1; ++ u32 request_id, doorbell; ++ ++ /* Read doorbell status */ ++ doorbell = readl(TWL_HOBDB_REG_ADDR(tw_dev)); ++ ++ /* Check for controller errors */ ++ if (doorbell & TWL_DOORBELL_CONTROLLER_ERROR) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xd, "Microcontroller Error: clearing"); ++ goto out; ++ } ++ ++ /* Check if we need to perform an AEN drain */ ++ if (doorbell & TWL_DOORBELL_ATTENTION_INTERRUPT) { ++ if (!(test_and_set_bit(TW_IN_ATTENTION_LOOP, &tw_dev->flags))) { ++ twl_get_request_id(tw_dev, &request_id); ++ if (twl_aen_read_queue(tw_dev, request_id)) { ++ tw_dev->state[request_id] = TW_S_COMPLETED; ++ twl_free_request_id(tw_dev, request_id); ++ clear_bit(TW_IN_ATTENTION_LOOP, &tw_dev->flags); ++ } ++ } ++ } ++ ++ retval = 0; ++out: ++ /* Clear doorbell interrupt */ ++ TWL_CLEAR_DB_INTERRUPT(tw_dev); ++ ++ /* Make sure the clear was flushed by reading it back */ ++ readl(TWL_HOBDBC_REG_ADDR(tw_dev)); ++ ++ return retval; ++} /* End twl_handle_attention_interrupt() */ ++ ++/* Interrupt service routine */ ++static irqreturn_t twl_interrupt(int irq, void *dev_instance) ++{ ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)dev_instance; ++ int i, handled = 0, error = 0; ++ dma_addr_t mfa = 0; ++ u32 reg, regl, regh, response, request_id = 0; ++ struct scsi_cmnd *cmd; ++ TW_Command_Full *full_command_packet; ++ ++ spin_lock(tw_dev->host->host_lock); ++ ++ /* Read host interrupt status */ ++ reg = readl(TWL_HISTAT_REG_ADDR(tw_dev)); ++ ++ /* Check if this is our interrupt, otherwise bail */ ++ if (!(reg & TWL_HISTATUS_VALID_INTERRUPT)) ++ goto twl_interrupt_bail; ++ ++ handled = 1; ++ ++ /* If we are resetting, bail */ ++ if (test_bit(TW_IN_RESET, &tw_dev->flags)) ++ goto twl_interrupt_bail; ++ ++ /* Attention interrupt */ ++ if (reg & TWL_HISTATUS_ATTENTION_INTERRUPT) { ++ if (twl_handle_attention_interrupt(tw_dev)) { ++ TWL_MASK_INTERRUPTS(tw_dev); ++ goto twl_interrupt_bail; ++ } ++ } ++ ++ /* Response interrupt */ ++ while (reg & TWL_HISTATUS_RESPONSE_INTERRUPT) { ++ if (sizeof(dma_addr_t) > 4) { ++ regh = readl(TWL_HOBQPH_REG_ADDR(tw_dev)); ++ regl = readl(TWL_HOBQPL_REG_ADDR(tw_dev)); ++ mfa = ((u64)regh << 32) | regl; ++ } else ++ mfa = readl(TWL_HOBQPL_REG_ADDR(tw_dev)); ++ ++ error = 0; ++ response = (u32)mfa; ++ ++ /* Check for command packet error */ ++ if (!TW_NOTMFA_OUT(response)) { ++ for (i=0;isense_buffer_phys[i] == mfa) { ++ request_id = le16_to_cpu(tw_dev->sense_buffer_virt[i]->header_desc.request_id); ++ if (tw_dev->srb[request_id] != NULL) ++ error = twl_fill_sense(tw_dev, i, request_id, 1, 1); ++ else { ++ /* Skip ioctl error prints */ ++ if (request_id != tw_dev->chrdev_request_id) ++ error = twl_fill_sense(tw_dev, i, request_id, 0, 1); ++ else ++ memcpy(tw_dev->command_packet_virt[request_id], tw_dev->sense_buffer_virt[i], sizeof(TW_Command_Apache_Header)); ++ } ++ ++ /* Now re-post the sense buffer */ ++ writel((u32)((u64)tw_dev->sense_buffer_phys[i] >> 32), TWL_HOBQPH_REG_ADDR(tw_dev)); ++ writel((u32)tw_dev->sense_buffer_phys[i], TWL_HOBQPL_REG_ADDR(tw_dev)); ++ break; ++ } ++ } ++ } else ++ request_id = TW_RESID_OUT(response); ++ ++ full_command_packet = tw_dev->command_packet_virt[request_id]; ++ ++ /* Check for correct state */ ++ if (tw_dev->state[request_id] != TW_S_POSTED) { ++ if (tw_dev->srb[request_id] != NULL) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xe, "Received a request id that wasn't posted"); ++ TWL_MASK_INTERRUPTS(tw_dev); ++ goto twl_interrupt_bail; ++ } ++ } ++ ++ /* Check for internal command completion */ ++ if (tw_dev->srb[request_id] == NULL) { ++ if (request_id != tw_dev->chrdev_request_id) { ++ if (twl_aen_complete(tw_dev, request_id)) ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0xf, "Error completing AEN during attention interrupt"); ++ } else { ++ tw_dev->chrdev_request_id = TW_IOCTL_CHRDEV_FREE; ++ wake_up(&tw_dev->ioctl_wqueue); ++ } ++ } else { ++ cmd = tw_dev->srb[request_id]; ++ ++ if (!error) ++ cmd->result = (DID_OK << 16); ++ ++ /* Report residual bytes for single sgl */ ++ if ((scsi_sg_count(cmd) <= 1) && (full_command_packet->command.newcommand.status == 0)) { ++ if (full_command_packet->command.newcommand.sg_list[0].length < scsi_bufflen(tw_dev->srb[request_id])) ++ scsi_set_resid(cmd, scsi_bufflen(cmd) - full_command_packet->command.newcommand.sg_list[0].length); ++ } ++ ++ /* Now complete the io */ ++ tw_dev->state[request_id] = TW_S_COMPLETED; ++ twl_free_request_id(tw_dev, request_id); ++ tw_dev->posted_request_count--; ++ tw_dev->srb[request_id]->scsi_done(tw_dev->srb[request_id]); ++ twl_unmap_scsi_data(tw_dev, request_id); ++ } ++ ++ /* Check for another response interrupt */ ++ reg = readl(TWL_HISTAT_REG_ADDR(tw_dev)); ++ } ++ ++twl_interrupt_bail: ++ spin_unlock(tw_dev->host->host_lock); ++ return IRQ_RETVAL(handled); ++} /* End twl_interrupt() */ ++ ++/* This function will poll for a register change */ ++static int twl_poll_register(TW_Device_Extension *tw_dev, void *reg, u32 value, u32 result, int seconds) ++{ ++ unsigned long before; ++ int retval = 1; ++ u32 reg_value; ++ ++ reg_value = readl(reg); ++ before = jiffies; ++ ++ while ((reg_value & value) != result) { ++ reg_value = readl(reg); ++ if (time_after(jiffies, before + HZ * seconds)) ++ goto out; ++ msleep(50); ++ } ++ retval = 0; ++out: ++ return retval; ++} /* End twl_poll_register() */ ++ ++/* This function will reset a controller */ ++static int twl_reset_sequence(TW_Device_Extension *tw_dev, int soft_reset) ++{ ++ int retval = 1; ++ int i = 0; ++ u32 status = 0; ++ unsigned short fw_on_ctlr_srl = 0, fw_on_ctlr_arch_id = 0; ++ unsigned short fw_on_ctlr_branch = 0, fw_on_ctlr_build = 0; ++ u32 init_connect_result = 0; ++ int tries = 0; ++ int do_soft_reset = soft_reset; ++ ++ while (tries < TW_MAX_RESET_TRIES) { ++ /* Do a soft reset if one is needed */ ++ if (do_soft_reset) { ++ TWL_SOFT_RESET(tw_dev); ++ ++ /* Make sure controller is in a good state */ ++ if (twl_poll_register(tw_dev, TWL_SCRPD3_REG_ADDR(tw_dev), TWL_CONTROLLER_READY, 0x0, 30)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x10, "Controller never went non-ready during reset sequence"); ++ tries++; ++ continue; ++ } ++ if (twl_poll_register(tw_dev, TWL_SCRPD3_REG_ADDR(tw_dev), TWL_CONTROLLER_READY, TWL_CONTROLLER_READY, 60)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x11, "Controller not ready during reset sequence"); ++ tries++; ++ continue; ++ } ++ } ++ ++ /* Initconnect */ ++ if (twl_initconnection(tw_dev, TW_INIT_MESSAGE_CREDITS, ++ TW_EXTENDED_INIT_CONNECT, TW_CURRENT_DRIVER_SRL, ++ TW_9750_ARCH_ID, TW_CURRENT_DRIVER_BRANCH, ++ TW_CURRENT_DRIVER_BUILD, &fw_on_ctlr_srl, ++ &fw_on_ctlr_arch_id, &fw_on_ctlr_branch, ++ &fw_on_ctlr_build, &init_connect_result)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x12, "Initconnection failed while checking SRL"); ++ do_soft_reset = 1; ++ tries++; ++ continue; ++ } ++ ++ /* Load sense buffers */ ++ while (i < TW_Q_LENGTH) { ++ writel((u32)((u64)tw_dev->sense_buffer_phys[i] >> 32), TWL_HOBQPH_REG_ADDR(tw_dev)); ++ writel((u32)tw_dev->sense_buffer_phys[i], TWL_HOBQPL_REG_ADDR(tw_dev)); ++ ++ /* Check status for over-run after each write */ ++ status = readl(TWL_STATUS_REG_ADDR(tw_dev)); ++ if (!(status & TWL_STATUS_OVERRUN_SUBMIT)) ++ i++; ++ } ++ ++ /* Now check status */ ++ status = readl(TWL_STATUS_REG_ADDR(tw_dev)); ++ if (status) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x13, "Bad controller status after loading sense buffers"); ++ do_soft_reset = 1; ++ tries++; ++ continue; ++ } ++ ++ /* Drain the AEN queue */ ++ if (twl_aen_drain_queue(tw_dev, soft_reset)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x14, "AEN drain failed during reset sequence"); ++ do_soft_reset = 1; ++ tries++; ++ continue; ++ } ++ ++ /* Load rest of compatibility struct */ ++ strncpy(tw_dev->tw_compat_info.driver_version, TW_DRIVER_VERSION, strlen(TW_DRIVER_VERSION)); ++ tw_dev->tw_compat_info.driver_srl_high = TW_CURRENT_DRIVER_SRL; ++ tw_dev->tw_compat_info.driver_branch_high = TW_CURRENT_DRIVER_BRANCH; ++ tw_dev->tw_compat_info.driver_build_high = TW_CURRENT_DRIVER_BUILD; ++ tw_dev->tw_compat_info.driver_srl_low = TW_BASE_FW_SRL; ++ tw_dev->tw_compat_info.driver_branch_low = TW_BASE_FW_BRANCH; ++ tw_dev->tw_compat_info.driver_build_low = TW_BASE_FW_BUILD; ++ tw_dev->tw_compat_info.fw_on_ctlr_srl = fw_on_ctlr_srl; ++ tw_dev->tw_compat_info.fw_on_ctlr_branch = fw_on_ctlr_branch; ++ tw_dev->tw_compat_info.fw_on_ctlr_build = fw_on_ctlr_build; ++ ++ /* If we got here, controller is in a good state */ ++ retval = 0; ++ goto out; ++ } ++out: ++ return retval; ++} /* End twl_reset_sequence() */ ++ ++/* This function will reset a device extension */ ++static int twl_reset_device_extension(TW_Device_Extension *tw_dev, int ioctl_reset) ++{ ++ int i = 0, retval = 1; ++ unsigned long flags = 0; ++ ++ /* Block SCSI requests while we are resetting */ ++ if (ioctl_reset) ++ scsi_block_requests(tw_dev->host); ++ ++ set_bit(TW_IN_RESET, &tw_dev->flags); ++ TWL_MASK_INTERRUPTS(tw_dev); ++ TWL_CLEAR_DB_INTERRUPT(tw_dev); ++ ++ spin_lock_irqsave(tw_dev->host->host_lock, flags); ++ ++ /* Abort all requests that are in progress */ ++ for (i = 0; i < TW_Q_LENGTH; i++) { ++ if ((tw_dev->state[i] != TW_S_FINISHED) && ++ (tw_dev->state[i] != TW_S_INITIAL) && ++ (tw_dev->state[i] != TW_S_COMPLETED)) { ++ if (tw_dev->srb[i]) { ++ tw_dev->srb[i]->result = (DID_RESET << 16); ++ tw_dev->srb[i]->scsi_done(tw_dev->srb[i]); ++ twl_unmap_scsi_data(tw_dev, i); ++ } ++ } ++ } ++ ++ /* Reset queues and counts */ ++ for (i = 0; i < TW_Q_LENGTH; i++) { ++ tw_dev->free_queue[i] = i; ++ tw_dev->state[i] = TW_S_INITIAL; ++ } ++ tw_dev->free_head = TW_Q_START; ++ tw_dev->free_tail = TW_Q_START; ++ tw_dev->posted_request_count = 0; ++ ++ spin_unlock_irqrestore(tw_dev->host->host_lock, flags); ++ ++ if (twl_reset_sequence(tw_dev, 1)) ++ goto out; ++ ++ TWL_UNMASK_INTERRUPTS(tw_dev); ++ ++ clear_bit(TW_IN_RESET, &tw_dev->flags); ++ tw_dev->chrdev_request_id = TW_IOCTL_CHRDEV_FREE; ++ ++ retval = 0; ++out: ++ if (ioctl_reset) ++ scsi_unblock_requests(tw_dev->host); ++ return retval; ++} /* End twl_reset_device_extension() */ ++ ++/* This funciton returns unit geometry in cylinders/heads/sectors */ ++static int twl_scsi_biosparam(struct scsi_device *sdev, struct block_device *bdev, sector_t capacity, int geom[]) ++{ ++ int heads, sectors, cylinders; ++ TW_Device_Extension *tw_dev; ++ ++ tw_dev = (TW_Device_Extension *)sdev->host->hostdata; ++ ++ if (capacity >= 0x200000) { ++ heads = 255; ++ sectors = 63; ++ cylinders = sector_div(capacity, heads * sectors); ++ } else { ++ heads = 64; ++ sectors = 32; ++ cylinders = sector_div(capacity, heads * sectors); ++ } ++ ++ geom[0] = heads; ++ geom[1] = sectors; ++ geom[2] = cylinders; ++ ++ return 0; ++} /* End twl_scsi_biosparam() */ ++ ++/* This is the new scsi eh reset function */ ++static int twl_scsi_eh_reset(struct scsi_cmnd *SCpnt) ++{ ++ TW_Device_Extension *tw_dev = NULL; ++ int retval = FAILED; ++ ++ tw_dev = (TW_Device_Extension *)SCpnt->device->host->hostdata; ++ ++ tw_dev->num_resets++; ++ ++ sdev_printk(KERN_WARNING, SCpnt->device, ++ "WARNING: (0x%02X:0x%04X): Command (0x%x) timed out, resetting card.\n", ++ TW_DRIVER, 0x2c, SCpnt->cmnd[0]); ++ ++ /* Make sure we are not issuing an ioctl or resetting from ioctl */ ++ mutex_lock(&tw_dev->ioctl_lock); ++ ++ /* Now reset the card and some of the device extension data */ ++ if (twl_reset_device_extension(tw_dev, 0)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x15, "Controller reset failed during scsi host reset"); ++ goto out; ++ } ++ ++ retval = SUCCESS; ++out: ++ mutex_unlock(&tw_dev->ioctl_lock); ++ return retval; ++} /* End twl_scsi_eh_reset() */ ++ ++/* This is the main scsi queue function to handle scsi opcodes */ ++static int twl_scsi_queue(struct scsi_cmnd *SCpnt, void (*done)(struct scsi_cmnd *)) ++{ ++ int request_id, retval; ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)SCpnt->device->host->hostdata; ++ ++ /* If we are resetting due to timed out ioctl, report as busy */ ++ if (test_bit(TW_IN_RESET, &tw_dev->flags)) { ++ retval = SCSI_MLQUEUE_HOST_BUSY; ++ goto out; ++ } ++ ++ /* Save done function into scsi_cmnd struct */ ++ SCpnt->scsi_done = done; ++ ++ /* Get a free request id */ ++ twl_get_request_id(tw_dev, &request_id); ++ ++ /* Save the scsi command for use by the ISR */ ++ tw_dev->srb[request_id] = SCpnt; ++ ++ /* Initialize phase to zero */ ++ SCpnt->SCp.phase = TW_PHASE_INITIAL; ++ ++ retval = twl_scsiop_execute_scsi(tw_dev, request_id, NULL, 0, NULL); ++ if (retval) { ++ tw_dev->state[request_id] = TW_S_COMPLETED; ++ twl_free_request_id(tw_dev, request_id); ++ SCpnt->result = (DID_ERROR << 16); ++ done(SCpnt); ++ retval = 0; ++ } ++out: ++ return retval; ++} /* End twl_scsi_queue() */ ++ ++/* This function tells the controller to shut down */ ++static void __twl_shutdown(TW_Device_Extension *tw_dev) ++{ ++ /* Disable interrupts */ ++ TWL_MASK_INTERRUPTS(tw_dev); ++ ++ /* Free up the IRQ */ ++ free_irq(tw_dev->tw_pci_dev->irq, tw_dev); ++ ++ printk(KERN_WARNING "3w-sas: Shutting down host %d.\n", tw_dev->host->host_no); ++ ++ /* Tell the card we are shutting down */ ++ if (twl_initconnection(tw_dev, 1, 0, 0, 0, 0, 0, NULL, NULL, NULL, NULL, NULL)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x16, "Connection shutdown failed"); ++ } else { ++ printk(KERN_WARNING "3w-sas: Shutdown complete.\n"); ++ } ++ ++ /* Clear doorbell interrupt just before exit */ ++ TWL_CLEAR_DB_INTERRUPT(tw_dev); ++} /* End __twl_shutdown() */ ++ ++/* Wrapper for __twl_shutdown */ ++static void twl_shutdown(struct pci_dev *pdev) ++{ ++ struct Scsi_Host *host = pci_get_drvdata(pdev); ++ TW_Device_Extension *tw_dev; ++ ++ if (!host) ++ return; ++ ++ tw_dev = (TW_Device_Extension *)host->hostdata; ++ ++ if (tw_dev->online) ++ __twl_shutdown(tw_dev); ++} /* End twl_shutdown() */ ++ ++/* This function configures unit settings when a unit is coming on-line */ ++static int twl_slave_configure(struct scsi_device *sdev) ++{ ++ /* Force 60 second timeout */ ++ blk_queue_rq_timeout(sdev->request_queue, 60 * HZ); ++ ++ return 0; ++} /* End twl_slave_configure() */ ++ ++/* scsi_host_template initializer */ ++static struct scsi_host_template driver_template = { ++ .module = THIS_MODULE, ++ .name = "3w-sas", ++ .queuecommand = twl_scsi_queue, ++ .eh_host_reset_handler = twl_scsi_eh_reset, ++ .bios_param = twl_scsi_biosparam, ++ .change_queue_depth = twl_change_queue_depth, ++ .can_queue = TW_Q_LENGTH-2, ++ .slave_configure = twl_slave_configure, ++ .this_id = -1, ++ .sg_tablesize = TW_LIBERATOR_MAX_SGL_LENGTH, ++ .max_sectors = TW_MAX_SECTORS, ++ .cmd_per_lun = TW_MAX_CMDS_PER_LUN, ++ .use_clustering = ENABLE_CLUSTERING, ++ .shost_attrs = twl_host_attrs, ++ .emulated = 1 ++}; ++ ++/* This function will probe and initialize a card */ ++static int __devinit twl_probe(struct pci_dev *pdev, const struct pci_device_id *dev_id) ++{ ++ struct Scsi_Host *host = NULL; ++ TW_Device_Extension *tw_dev; ++ resource_size_t mem_addr, mem_len; ++ int retval = -ENODEV; ++ int *ptr_phycount, phycount=0; ++ ++ retval = pci_enable_device(pdev); ++ if (retval) { ++ TW_PRINTK(host, TW_DRIVER, 0x17, "Failed to enable pci device"); ++ goto out_disable_device; ++ } ++ ++ pci_set_master(pdev); ++ pci_try_set_mwi(pdev); ++ ++ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) ++ || pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64))) ++ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32)) ++ || pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32))) { ++ TW_PRINTK(host, TW_DRIVER, 0x18, "Failed to set dma mask"); ++ retval = -ENODEV; ++ goto out_disable_device; ++ } ++ ++ host = scsi_host_alloc(&driver_template, sizeof(TW_Device_Extension)); ++ if (!host) { ++ TW_PRINTK(host, TW_DRIVER, 0x19, "Failed to allocate memory for device extension"); ++ retval = -ENOMEM; ++ goto out_disable_device; ++ } ++ tw_dev = (TW_Device_Extension *)host->hostdata; ++ ++ /* Save values to device extension */ ++ tw_dev->host = host; ++ tw_dev->tw_pci_dev = pdev; ++ ++ if (twl_initialize_device_extension(tw_dev)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1a, "Failed to initialize device extension"); ++ goto out_free_device_extension; ++ } ++ ++ /* Request IO regions */ ++ retval = pci_request_regions(pdev, "3w-sas"); ++ if (retval) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1b, "Failed to get mem region"); ++ goto out_free_device_extension; ++ } ++ ++ /* Use region 1 */ ++ mem_addr = pci_resource_start(pdev, 1); ++ mem_len = pci_resource_len(pdev, 1); ++ ++ /* Save base address */ ++ tw_dev->base_addr = ioremap(mem_addr, mem_len); ++ ++ if (!tw_dev->base_addr) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1c, "Failed to ioremap"); ++ goto out_release_mem_region; ++ } ++ ++ /* Disable interrupts on the card */ ++ TWL_MASK_INTERRUPTS(tw_dev); ++ ++ /* Initialize the card */ ++ if (twl_reset_sequence(tw_dev, 0)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1d, "Controller reset failed during probe"); ++ goto out_iounmap; ++ } ++ ++ /* Set host specific parameters */ ++ host->max_id = TW_MAX_UNITS; ++ host->max_cmd_len = TW_MAX_CDB_LEN; ++ host->max_lun = TW_MAX_LUNS; ++ host->max_channel = 0; ++ ++ /* Register the card with the kernel SCSI layer */ ++ retval = scsi_add_host(host, &pdev->dev); ++ if (retval) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1e, "scsi add host failed"); ++ goto out_iounmap; ++ } ++ ++ pci_set_drvdata(pdev, host); ++ ++ printk(KERN_WARNING "3w-sas: scsi%d: Found an LSI 3ware %s Controller at 0x%llx, IRQ: %d.\n", ++ host->host_no, ++ (char *)twl_get_param(tw_dev, 1, TW_VERSION_TABLE, ++ TW_PARAM_MODEL, TW_PARAM_MODEL_LENGTH), ++ (u64)mem_addr, pdev->irq); ++ ++ ptr_phycount = twl_get_param(tw_dev, 2, TW_PARAM_PHY_SUMMARY_TABLE, ++ TW_PARAM_PHYCOUNT, TW_PARAM_PHYCOUNT_LENGTH); ++ if (ptr_phycount) ++ phycount = le32_to_cpu(*(int *)ptr_phycount); ++ ++ printk(KERN_WARNING "3w-sas: scsi%d: Firmware %s, BIOS %s, Phys: %d.\n", ++ host->host_no, ++ (char *)twl_get_param(tw_dev, 1, TW_VERSION_TABLE, ++ TW_PARAM_FWVER, TW_PARAM_FWVER_LENGTH), ++ (char *)twl_get_param(tw_dev, 2, TW_VERSION_TABLE, ++ TW_PARAM_BIOSVER, TW_PARAM_BIOSVER_LENGTH), ++ phycount); ++ ++ /* Try to enable MSI */ ++ if (use_msi && !pci_enable_msi(pdev)) ++ set_bit(TW_USING_MSI, &tw_dev->flags); ++ ++ /* Now setup the interrupt handler */ ++ retval = request_irq(pdev->irq, twl_interrupt, IRQF_SHARED, "3w-sas", tw_dev); ++ if (retval) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x1f, "Error requesting IRQ"); ++ goto out_remove_host; ++ } ++ ++ twl_device_extension_list[twl_device_extension_count] = tw_dev; ++ twl_device_extension_count++; ++ ++ /* Re-enable interrupts on the card */ ++ TWL_UNMASK_INTERRUPTS(tw_dev); ++ ++ /* Finally, scan the host */ ++ scsi_scan_host(host); ++ ++ /* Add sysfs binary files */ ++ if (sysfs_create_bin_file(&host->shost_dev.kobj, &twl_sysfs_aen_read_attr)) ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x20, "Failed to create sysfs binary file: 3ware_aen_read"); ++ if (sysfs_create_bin_file(&host->shost_dev.kobj, &twl_sysfs_compat_info_attr)) ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x21, "Failed to create sysfs binary file: 3ware_compat_info"); ++ ++ if (twl_major == -1) { ++ if ((twl_major = register_chrdev (0, "twl", &twl_fops)) < 0) ++ TW_PRINTK(host, TW_DRIVER, 0x22, "Failed to register character device"); ++ } ++ tw_dev->online = 1; ++ return 0; ++ ++out_remove_host: ++ if (test_bit(TW_USING_MSI, &tw_dev->flags)) ++ pci_disable_msi(pdev); ++ scsi_remove_host(host); ++out_iounmap: ++ iounmap(tw_dev->base_addr); ++out_release_mem_region: ++ pci_release_regions(pdev); ++out_free_device_extension: ++ twl_free_device_extension(tw_dev); ++ scsi_host_put(host); ++out_disable_device: ++ pci_disable_device(pdev); ++ ++ return retval; ++} /* End twl_probe() */ ++ ++/* This function is called to remove a device */ ++static void twl_remove(struct pci_dev *pdev) ++{ ++ struct Scsi_Host *host = pci_get_drvdata(pdev); ++ TW_Device_Extension *tw_dev; ++ ++ if (!host) ++ return; ++ ++ tw_dev = (TW_Device_Extension *)host->hostdata; ++ ++ if (!tw_dev->online) ++ return; ++ ++ /* Remove sysfs binary files */ ++ sysfs_remove_bin_file(&host->shost_dev.kobj, &twl_sysfs_aen_read_attr); ++ sysfs_remove_bin_file(&host->shost_dev.kobj, &twl_sysfs_compat_info_attr); ++ ++ scsi_remove_host(tw_dev->host); ++ ++ /* Unregister character device */ ++ if (twl_major >= 0) { ++ unregister_chrdev(twl_major, "twl"); ++ twl_major = -1; ++ } ++ ++ /* Shutdown the card */ ++ __twl_shutdown(tw_dev); ++ ++ /* Disable MSI if enabled */ ++ if (test_bit(TW_USING_MSI, &tw_dev->flags)) ++ pci_disable_msi(pdev); ++ ++ /* Free IO remapping */ ++ iounmap(tw_dev->base_addr); ++ ++ /* Free up the mem region */ ++ pci_release_regions(pdev); ++ ++ /* Free up device extension resources */ ++ twl_free_device_extension(tw_dev); ++ ++ scsi_host_put(tw_dev->host); ++ pci_disable_device(pdev); ++ twl_device_extension_count--; ++} /* End twl_remove() */ ++ ++#ifdef CONFIG_PM ++/* This function is called on PCI suspend */ ++static int twl_suspend(struct pci_dev *pdev, pm_message_t state) ++{ ++ struct Scsi_Host *host = pci_get_drvdata(pdev); ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)host->hostdata; ++ ++ printk(KERN_WARNING "3w-sas: Suspending host %d.\n", tw_dev->host->host_no); ++ /* Disable interrupts */ ++ TWL_MASK_INTERRUPTS(tw_dev); ++ ++ free_irq(tw_dev->tw_pci_dev->irq, tw_dev); ++ ++ /* Tell the card we are shutting down */ ++ if (twl_initconnection(tw_dev, 1, 0, 0, 0, 0, 0, NULL, NULL, NULL, NULL, NULL)) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x23, "Connection shutdown failed during suspend"); ++ } else { ++ printk(KERN_WARNING "3w-sas: Suspend complete.\n"); ++ } ++ ++ /* Clear doorbell interrupt */ ++ TWL_CLEAR_DB_INTERRUPT(tw_dev); ++ ++ pci_save_state(pdev); ++ pci_disable_device(pdev); ++ pci_set_power_state(pdev, pci_choose_state(pdev, state)); ++ ++ return 0; ++} /* End twl_suspend() */ ++ ++/* This function is called on PCI resume */ ++static int twl_resume(struct pci_dev *pdev) ++{ ++ int retval = 0; ++ struct Scsi_Host *host = pci_get_drvdata(pdev); ++ TW_Device_Extension *tw_dev = (TW_Device_Extension *)host->hostdata; ++ ++ printk(KERN_WARNING "3w-sas: Resuming host %d.\n", tw_dev->host->host_no); ++ pci_set_power_state(pdev, PCI_D0); ++ pci_enable_wake(pdev, PCI_D0, 0); ++ pci_restore_state(pdev); ++ ++ retval = pci_enable_device(pdev); ++ if (retval) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x24, "Enable device failed during resume"); ++ return retval; ++ } ++ ++ pci_set_master(pdev); ++ pci_try_set_mwi(pdev); ++ ++ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) ++ || pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64))) ++ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32)) ++ || pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32))) { ++ TW_PRINTK(host, TW_DRIVER, 0x25, "Failed to set dma mask during resume"); ++ retval = -ENODEV; ++ goto out_disable_device; ++ } ++ ++ /* Initialize the card */ ++ if (twl_reset_sequence(tw_dev, 0)) { ++ retval = -ENODEV; ++ goto out_disable_device; ++ } ++ ++ /* Now setup the interrupt handler */ ++ retval = request_irq(pdev->irq, twl_interrupt, IRQF_SHARED, "3w-sas", tw_dev); ++ if (retval) { ++ TW_PRINTK(tw_dev->host, TW_DRIVER, 0x26, "Error requesting IRQ during resume"); ++ retval = -ENODEV; ++ goto out_disable_device; ++ } ++ ++ /* Now enable MSI if enabled */ ++ if (test_bit(TW_USING_MSI, &tw_dev->flags)) ++ pci_enable_msi(pdev); ++ ++ /* Re-enable interrupts on the card */ ++ TWL_UNMASK_INTERRUPTS(tw_dev); ++ ++ printk(KERN_WARNING "3w-sas: Resume complete.\n"); ++ return 0; ++ ++out_disable_device: ++ scsi_remove_host(host); ++ pci_disable_device(pdev); ++ ++ return retval; ++} /* End twl_resume() */ ++#endif ++ ++/* PCI Devices supported by this driver */ ++static struct pci_device_id twl_pci_tbl[] __devinitdata = { ++ { PCI_VENDOR_ID_3WARE, PCI_DEVICE_ID_3WARE_9750, ++ PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0}, ++ { } ++}; ++MODULE_DEVICE_TABLE(pci, twl_pci_tbl); ++ ++/* pci_driver initializer */ ++static struct pci_driver twl_driver = { ++ .name = "3w-sas", ++ .id_table = twl_pci_tbl, ++ .probe = twl_probe, ++ .remove = twl_remove, ++#ifdef CONFIG_PM ++ .suspend = twl_suspend, ++ .resume = twl_resume, ++#endif ++ .shutdown = twl_shutdown ++}; ++ ++/* This function is called on driver initialization */ ++static int __init twl_init(void) ++{ ++ printk(KERN_WARNING "LSI 3ware SAS/SATA-RAID Controller device driver for Linux v%s.\n", TW_DRIVER_VERSION); ++ ++ return pci_register_driver(&twl_driver); ++} /* End twl_init() */ ++ ++/* This function is called on driver exit */ ++static void __exit twl_exit(void) ++{ ++ pci_unregister_driver(&twl_driver); ++} /* End twl_exit() */ ++ ++module_init(twl_init); ++module_exit(twl_exit); ++ +diff --git a/drivers/scsi/3w-sas.h b/drivers/scsi/3w-sas.h +new file mode 100644 +index 0000000..e620505 +--- /dev/null ++++ b/drivers/scsi/3w-sas.h +@@ -0,0 +1,396 @@ ++/* ++ 3w-sas.h -- LSI 3ware SAS/SATA-RAID Controller device driver for Linux. ++ ++ Written By: Adam Radford ++ ++ Copyright (C) 2009 LSI Corporation. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; version 2 of the License. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ NO WARRANTY ++ THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR ++ CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT ++ LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, ++ MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is ++ solely responsible for determining the appropriateness of using and ++ distributing the Program and assumes all risks associated with its ++ exercise of rights under this Agreement, including but not limited to ++ the risks and costs of program errors, damage to or loss of data, ++ programs or equipment, and unavailability or interruption of operations. ++ ++ DISCLAIMER OF LIABILITY ++ NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY ++ DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL ++ DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND ++ ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR ++ TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE ++ USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED ++ HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ ++ Bugs/Comments/Suggestions should be mailed to: ++ linuxraid@lsi.com ++ ++ For more information, goto: ++ http://www.lsi.com ++*/ ++ ++#ifndef _3W_SAS_H ++#define _3W_SAS_H ++ ++/* AEN severity table */ ++static char *twl_aen_severity_table[] = ++{ ++ "None", "ERROR", "WARNING", "INFO", "DEBUG", (char*) 0 ++}; ++ ++/* Liberator register offsets */ ++#define TWL_STATUS 0x0 /* Status */ ++#define TWL_HIBDB 0x20 /* Inbound doorbell */ ++#define TWL_HISTAT 0x30 /* Host interrupt status */ ++#define TWL_HIMASK 0x34 /* Host interrupt mask */ ++#define TWL_HOBDB 0x9C /* Outbound doorbell */ ++#define TWL_HOBDBC 0xA0 /* Outbound doorbell clear */ ++#define TWL_SCRPD3 0xBC /* Scratchpad */ ++#define TWL_HIBQPL 0xC0 /* Host inbound Q low */ ++#define TWL_HIBQPH 0xC4 /* Host inbound Q high */ ++#define TWL_HOBQPL 0xC8 /* Host outbound Q low */ ++#define TWL_HOBQPH 0xCC /* Host outbound Q high */ ++#define TWL_HISTATUS_VALID_INTERRUPT 0xC ++#define TWL_HISTATUS_ATTENTION_INTERRUPT 0x4 ++#define TWL_HISTATUS_RESPONSE_INTERRUPT 0x8 ++#define TWL_STATUS_OVERRUN_SUBMIT 0x2000 ++#define TWL_ISSUE_SOFT_RESET 0x100 ++#define TWL_CONTROLLER_READY 0x2000 ++#define TWL_DOORBELL_CONTROLLER_ERROR 0x200000 ++#define TWL_DOORBELL_ATTENTION_INTERRUPT 0x40000 ++#define TWL_PULL_MODE 0x1 ++ ++/* Command packet opcodes used by the driver */ ++#define TW_OP_INIT_CONNECTION 0x1 ++#define TW_OP_GET_PARAM 0x12 ++#define TW_OP_SET_PARAM 0x13 ++#define TW_OP_EXECUTE_SCSI 0x10 ++ ++/* Asynchronous Event Notification (AEN) codes used by the driver */ ++#define TW_AEN_QUEUE_EMPTY 0x0000 ++#define TW_AEN_SOFT_RESET 0x0001 ++#define TW_AEN_SYNC_TIME_WITH_HOST 0x031 ++#define TW_AEN_SEVERITY_ERROR 0x1 ++#define TW_AEN_SEVERITY_DEBUG 0x4 ++#define TW_AEN_NOT_RETRIEVED 0x1 ++ ++/* Command state defines */ ++#define TW_S_INITIAL 0x1 /* Initial state */ ++#define TW_S_STARTED 0x2 /* Id in use */ ++#define TW_S_POSTED 0x4 /* Posted to the controller */ ++#define TW_S_COMPLETED 0x8 /* Completed by isr */ ++#define TW_S_FINISHED 0x10 /* I/O completely done */ ++ ++/* Compatibility defines */ ++#define TW_9750_ARCH_ID 10 ++#define TW_CURRENT_DRIVER_SRL 40 ++#define TW_CURRENT_DRIVER_BUILD 0 ++#define TW_CURRENT_DRIVER_BRANCH 0 ++ ++/* Phase defines */ ++#define TW_PHASE_INITIAL 0 ++#define TW_PHASE_SGLIST 2 ++ ++/* Misc defines */ ++#define TW_SECTOR_SIZE 512 ++#define TW_MAX_UNITS 32 ++#define TW_INIT_MESSAGE_CREDITS 0x100 ++#define TW_INIT_COMMAND_PACKET_SIZE 0x3 ++#define TW_INIT_COMMAND_PACKET_SIZE_EXTENDED 0x6 ++#define TW_EXTENDED_INIT_CONNECT 0x2 ++#define TW_BASE_FW_SRL 24 ++#define TW_BASE_FW_BRANCH 0 ++#define TW_BASE_FW_BUILD 1 ++#define TW_Q_LENGTH 256 ++#define TW_Q_START 0 ++#define TW_MAX_SLOT 32 ++#define TW_MAX_RESET_TRIES 2 ++#define TW_MAX_CMDS_PER_LUN 254 ++#define TW_MAX_AEN_DRAIN 255 ++#define TW_IN_RESET 2 ++#define TW_USING_MSI 3 ++#define TW_IN_ATTENTION_LOOP 4 ++#define TW_MAX_SECTORS 256 ++#define TW_MAX_CDB_LEN 16 ++#define TW_IOCTL_CHRDEV_TIMEOUT 60 /* 60 seconds */ ++#define TW_IOCTL_CHRDEV_FREE -1 ++#define TW_COMMAND_OFFSET 128 /* 128 bytes */ ++#define TW_VERSION_TABLE 0x0402 ++#define TW_TIMEKEEP_TABLE 0x040A ++#define TW_INFORMATION_TABLE 0x0403 ++#define TW_PARAM_FWVER 3 ++#define TW_PARAM_FWVER_LENGTH 16 ++#define TW_PARAM_BIOSVER 4 ++#define TW_PARAM_BIOSVER_LENGTH 16 ++#define TW_PARAM_MODEL 8 ++#define TW_PARAM_MODEL_LENGTH 16 ++#define TW_PARAM_PHY_SUMMARY_TABLE 1 ++#define TW_PARAM_PHYCOUNT 2 ++#define TW_PARAM_PHYCOUNT_LENGTH 1 ++#define TW_IOCTL_FIRMWARE_PASS_THROUGH 0x108 // Used by smartmontools ++#define TW_ALLOCATION_LENGTH 128 ++#define TW_SENSE_DATA_LENGTH 18 ++#define TW_ERROR_LOGICAL_UNIT_NOT_SUPPORTED 0x10a ++#define TW_ERROR_INVALID_FIELD_IN_CDB 0x10d ++#define TW_ERROR_UNIT_OFFLINE 0x128 ++#define TW_MESSAGE_SOURCE_CONTROLLER_ERROR 3 ++#define TW_MESSAGE_SOURCE_CONTROLLER_EVENT 4 ++#define TW_DRIVER 6 ++#ifndef PCI_DEVICE_ID_3WARE_9750 ++#define PCI_DEVICE_ID_3WARE_9750 0x1010 ++#endif ++ ++/* Bitmask macros to eliminate bitfields */ ++ ++/* opcode: 5, reserved: 3 */ ++#define TW_OPRES_IN(x,y) ((x << 5) | (y & 0x1f)) ++#define TW_OP_OUT(x) (x & 0x1f) ++ ++/* opcode: 5, sgloffset: 3 */ ++#define TW_OPSGL_IN(x,y) ((x << 5) | (y & 0x1f)) ++#define TW_SGL_OUT(x) ((x >> 5) & 0x7) ++ ++/* severity: 3, reserved: 5 */ ++#define TW_SEV_OUT(x) (x & 0x7) ++ ++/* not_mfa: 1, reserved: 7, status: 8, request_id: 16 */ ++#define TW_RESID_OUT(x) ((x >> 16) & 0xffff) ++#define TW_NOTMFA_OUT(x) (x & 0x1) ++ ++/* request_id: 12, lun: 4 */ ++#define TW_REQ_LUN_IN(lun, request_id) (((lun << 12) & 0xf000) | (request_id & 0xfff)) ++#define TW_LUN_OUT(lun) ((lun >> 12) & 0xf) ++ ++/* Register access macros */ ++#define TWL_STATUS_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_STATUS) ++#define TWL_HOBQPL_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HOBQPL) ++#define TWL_HOBQPH_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HOBQPH) ++#define TWL_HOBDB_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HOBDB) ++#define TWL_HOBDBC_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HOBDBC) ++#define TWL_HIMASK_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HIMASK) ++#define TWL_HISTAT_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HISTAT) ++#define TWL_HIBQPH_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HIBQPH) ++#define TWL_HIBQPL_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HIBQPL) ++#define TWL_HIBDB_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_HIBDB) ++#define TWL_SCRPD3_REG_ADDR(x) ((unsigned char __iomem *)x->base_addr + TWL_SCRPD3) ++#define TWL_MASK_INTERRUPTS(x) (writel(~0, TWL_HIMASK_REG_ADDR(tw_dev))) ++#define TWL_UNMASK_INTERRUPTS(x) (writel(~TWL_HISTATUS_VALID_INTERRUPT, TWL_HIMASK_REG_ADDR(tw_dev))) ++#define TWL_CLEAR_DB_INTERRUPT(x) (writel(~0, TWL_HOBDBC_REG_ADDR(tw_dev))) ++#define TWL_SOFT_RESET(x) (writel(TWL_ISSUE_SOFT_RESET, TWL_HIBDB_REG_ADDR(tw_dev))) ++ ++/* Macros */ ++#define TW_PRINTK(h,a,b,c) { \ ++if (h) \ ++printk(KERN_WARNING "3w-sas: scsi%d: ERROR: (0x%02X:0x%04X): %s.\n",h->host_no,a,b,c); \ ++else \ ++printk(KERN_WARNING "3w-sas: ERROR: (0x%02X:0x%04X): %s.\n",a,b,c); \ ++} ++#define TW_MAX_LUNS 16 ++#define TW_COMMAND_SIZE (sizeof(dma_addr_t) > 4 ? 6 : 4) ++#define TW_LIBERATOR_MAX_SGL_LENGTH (sizeof(dma_addr_t) > 4 ? 46 : 92) ++#define TW_LIBERATOR_MAX_SGL_LENGTH_OLD (sizeof(dma_addr_t) > 4 ? 47 : 94) ++#define TW_PADDING_LENGTH_LIBERATOR 136 ++#define TW_PADDING_LENGTH_LIBERATOR_OLD 132 ++#define TW_CPU_TO_SGL(x) (sizeof(dma_addr_t) > 4 ? cpu_to_le64(x) : cpu_to_le32(x)) ++ ++#pragma pack(1) ++ ++/* SGL entry */ ++typedef struct TAG_TW_SG_Entry_ISO { ++ dma_addr_t address; ++ dma_addr_t length; ++} TW_SG_Entry_ISO; ++ ++/* Old Command Packet with ISO SGL */ ++typedef struct TW_Command { ++ unsigned char opcode__sgloffset; ++ unsigned char size; ++ unsigned char request_id; ++ unsigned char unit__hostid; ++ /* Second DWORD */ ++ unsigned char status; ++ unsigned char flags; ++ union { ++ unsigned short block_count; ++ unsigned short parameter_count; ++ } byte6_offset; ++ union { ++ struct { ++ u32 lba; ++ TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; ++ unsigned char padding[TW_PADDING_LENGTH_LIBERATOR_OLD]; ++ } io; ++ struct { ++ TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; ++ u32 padding; ++ unsigned char padding2[TW_PADDING_LENGTH_LIBERATOR_OLD]; ++ } param; ++ } byte8_offset; ++} TW_Command; ++ ++/* New Command Packet with ISO SGL */ ++typedef struct TAG_TW_Command_Apache { ++ unsigned char opcode__reserved; ++ unsigned char unit; ++ unsigned short request_id__lunl; ++ unsigned char status; ++ unsigned char sgl_offset; ++ unsigned short sgl_entries__lunh; ++ unsigned char cdb[16]; ++ TW_SG_Entry_ISO sg_list[TW_LIBERATOR_MAX_SGL_LENGTH]; ++ unsigned char padding[TW_PADDING_LENGTH_LIBERATOR]; ++} TW_Command_Apache; ++ ++/* New command packet header */ ++typedef struct TAG_TW_Command_Apache_Header { ++ unsigned char sense_data[TW_SENSE_DATA_LENGTH]; ++ struct { ++ char reserved[4]; ++ unsigned short error; ++ unsigned char padding; ++ unsigned char severity__reserved; ++ } status_block; ++ unsigned char err_specific_desc[98]; ++ struct { ++ unsigned char size_header; ++ unsigned short request_id; ++ unsigned char size_sense; ++ } header_desc; ++} TW_Command_Apache_Header; ++ ++/* This struct is a union of the 2 command packets */ ++typedef struct TAG_TW_Command_Full { ++ TW_Command_Apache_Header header; ++ union { ++ TW_Command oldcommand; ++ TW_Command_Apache newcommand; ++ } command; ++} TW_Command_Full; ++ ++/* Initconnection structure */ ++typedef struct TAG_TW_Initconnect { ++ unsigned char opcode__reserved; ++ unsigned char size; ++ unsigned char request_id; ++ unsigned char res2; ++ unsigned char status; ++ unsigned char flags; ++ unsigned short message_credits; ++ u32 features; ++ unsigned short fw_srl; ++ unsigned short fw_arch_id; ++ unsigned short fw_branch; ++ unsigned short fw_build; ++ u32 result; ++} TW_Initconnect; ++ ++/* Event info structure */ ++typedef struct TAG_TW_Event ++{ ++ unsigned int sequence_id; ++ unsigned int time_stamp_sec; ++ unsigned short aen_code; ++ unsigned char severity; ++ unsigned char retrieved; ++ unsigned char repeat_count; ++ unsigned char parameter_len; ++ unsigned char parameter_data[98]; ++} TW_Event; ++ ++typedef struct TAG_TW_Ioctl_Driver_Command { ++ unsigned int control_code; ++ unsigned int status; ++ unsigned int unique_id; ++ unsigned int sequence_id; ++ unsigned int os_specific; ++ unsigned int buffer_length; ++} TW_Ioctl_Driver_Command; ++ ++typedef struct TAG_TW_Ioctl_Apache { ++ TW_Ioctl_Driver_Command driver_command; ++ char padding[488]; ++ TW_Command_Full firmware_command; ++ char data_buffer[1]; ++} TW_Ioctl_Buf_Apache; ++ ++/* GetParam descriptor */ ++typedef struct { ++ unsigned short table_id; ++ unsigned short parameter_id; ++ unsigned short parameter_size_bytes; ++ unsigned short actual_parameter_size_bytes; ++ unsigned char data[1]; ++} TW_Param_Apache; ++ ++/* Compatibility information structure */ ++typedef struct TAG_TW_Compatibility_Info ++{ ++ char driver_version[32]; ++ unsigned short working_srl; ++ unsigned short working_branch; ++ unsigned short working_build; ++ unsigned short driver_srl_high; ++ unsigned short driver_branch_high; ++ unsigned short driver_build_high; ++ unsigned short driver_srl_low; ++ unsigned short driver_branch_low; ++ unsigned short driver_build_low; ++ unsigned short fw_on_ctlr_srl; ++ unsigned short fw_on_ctlr_branch; ++ unsigned short fw_on_ctlr_build; ++} TW_Compatibility_Info; ++ ++#pragma pack() ++ ++typedef struct TAG_TW_Device_Extension { ++ void __iomem *base_addr; ++ unsigned long *generic_buffer_virt[TW_Q_LENGTH]; ++ dma_addr_t generic_buffer_phys[TW_Q_LENGTH]; ++ TW_Command_Full *command_packet_virt[TW_Q_LENGTH]; ++ dma_addr_t command_packet_phys[TW_Q_LENGTH]; ++ TW_Command_Apache_Header *sense_buffer_virt[TW_Q_LENGTH]; ++ dma_addr_t sense_buffer_phys[TW_Q_LENGTH]; ++ struct pci_dev *tw_pci_dev; ++ struct scsi_cmnd *srb[TW_Q_LENGTH]; ++ unsigned char free_queue[TW_Q_LENGTH]; ++ unsigned char free_head; ++ unsigned char free_tail; ++ int state[TW_Q_LENGTH]; ++ unsigned int posted_request_count; ++ unsigned int max_posted_request_count; ++ unsigned int max_sgl_entries; ++ unsigned int sgl_entries; ++ unsigned int num_resets; ++ unsigned int sector_count; ++ unsigned int max_sector_count; ++ unsigned int aen_count; ++ struct Scsi_Host *host; ++ long flags; ++ TW_Event *event_queue[TW_Q_LENGTH]; ++ unsigned char error_index; ++ unsigned int error_sequence_id; ++ int chrdev_request_id; ++ wait_queue_head_t ioctl_wqueue; ++ struct mutex ioctl_lock; ++ TW_Compatibility_Info tw_compat_info; ++ char online; ++} TW_Device_Extension; ++ ++#endif /* _3W_SAS_H */ ++ +diff --git a/drivers/scsi/BusLogic.c b/drivers/scsi/BusLogic.c +index 1ddcf40..a85f062 100644 +--- a/drivers/scsi/BusLogic.c ++++ b/drivers/scsi/BusLogic.c +@@ -961,6 +961,8 @@ static int __init BusLogic_InitializeFlashPointProbeInfo(struct BusLogic_HostAda + static void __init BusLogic_InitializeProbeInfoList(struct BusLogic_HostAdapter + *PrototypeHostAdapter) + { ++ pax_track_stack(); ++ + /* + If a PCI BIOS is present, interrogate it for MultiMaster and FlashPoint + Host Adapters; otherwise, default to the standard ISA MultiMaster probe. +diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig +index e11cca4..4295679 100644 +--- a/drivers/scsi/Kconfig ++++ b/drivers/scsi/Kconfig +@@ -399,6 +399,17 @@ config SCSI_3W_9XXX + Please read the comments at the top of + . + ++config SCSI_3W_SAS ++ tristate "3ware 97xx SAS/SATA-RAID support" ++ depends on PCI && SCSI ++ help ++ This driver supports the LSI 3ware 9750 6Gb/s SAS/SATA-RAID cards. ++ ++ ++ ++ Please read the comments at the top of ++ . ++ + config SCSI_7000FASST + tristate "7000FASST SCSI support" + depends on ISA && SCSI && ISA_DMA_API +@@ -621,6 +632,14 @@ config SCSI_FLASHPOINT + substantial, so users of MultiMaster Host Adapters may not + wish to include it. + ++config VMWARE_PVSCSI ++ tristate "VMware PVSCSI driver support" ++ depends on PCI && SCSI && X86 ++ help ++ This driver supports VMware's para virtualized SCSI HBA. ++ To compile this driver as a module, choose M here: the ++ module will be called vmw_pvscsi. ++ + config LIBFC + tristate "LibFC module" + select SCSI_FC_ATTRS +diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile +index 3ad61db..c938975 100644 +--- a/drivers/scsi/Makefile ++++ b/drivers/scsi/Makefile +@@ -113,6 +113,7 @@ obj-$(CONFIG_SCSI_MESH) += mesh.o + obj-$(CONFIG_SCSI_MAC53C94) += mac53c94.o + obj-$(CONFIG_BLK_DEV_3W_XXXX_RAID) += 3w-xxxx.o + obj-$(CONFIG_SCSI_3W_9XXX) += 3w-9xxx.o ++obj-$(CONFIG_SCSI_3W_SAS) += 3w-sas.o + obj-$(CONFIG_SCSI_PPA) += ppa.o + obj-$(CONFIG_SCSI_IMM) += imm.o + obj-$(CONFIG_JAZZ_ESP) += esp_scsi.o jazz_esp.o +@@ -133,6 +134,7 @@ obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgb3i/ + obj-$(CONFIG_SCSI_BNX2_ISCSI) += libiscsi.o bnx2i/ + obj-$(CONFIG_BE2ISCSI) += libiscsi.o be2iscsi/ + obj-$(CONFIG_SCSI_PMCRAID) += pmcraid.o ++obj-$(CONFIG_VMWARE_PVSCSI) += vmw_pvscsi.o + + obj-$(CONFIG_ARM) += arm/ + +diff --git a/drivers/scsi/aacraid/aacraid.h b/drivers/scsi/aacraid/aacraid.h +index cdbdec9..b7d560b 100644 +--- a/drivers/scsi/aacraid/aacraid.h ++++ b/drivers/scsi/aacraid/aacraid.h +@@ -471,7 +471,7 @@ struct adapter_ops + int (*adapter_scsi)(struct fib * fib, struct scsi_cmnd * cmd); + /* Administrative operations */ + int (*adapter_comm)(struct aac_dev * dev, int comm); +-}; ++} __no_const; + + /* + * Define which interrupt handler needs to be installed +diff --git a/drivers/scsi/aacraid/commctrl.c b/drivers/scsi/aacraid/commctrl.c +index a5b8e7b..a6a0e43 100644 +--- a/drivers/scsi/aacraid/commctrl.c ++++ b/drivers/scsi/aacraid/commctrl.c +@@ -481,6 +481,7 @@ static int aac_send_raw_srb(struct aac_dev* dev, void __user * arg) + u32 actual_fibsize64, actual_fibsize = 0; + int i; + ++ pax_track_stack(); + + if (dev->in_reset) { + dprintk((KERN_DEBUG"aacraid: send raw srb -EBUSY\n")); +diff --git a/drivers/scsi/aacraid/linit.c b/drivers/scsi/aacraid/linit.c +index 9b97c3e..f099725 100644 +--- a/drivers/scsi/aacraid/linit.c ++++ b/drivers/scsi/aacraid/linit.c +@@ -91,7 +91,7 @@ static DECLARE_PCI_DEVICE_TABLE(aac_pci_tbl) = { + #elif defined(__devinitconst) + static const struct pci_device_id aac_pci_tbl[] __devinitconst = { + #else +-static const struct pci_device_id aac_pci_tbl[] __devinitdata = { ++static const struct pci_device_id aac_pci_tbl[] __devinitconst = { + #endif + { 0x1028, 0x0001, 0x1028, 0x0001, 0, 0, 0 }, /* PERC 2/Si (Iguana/PERC2Si) */ + { 0x1028, 0x0002, 0x1028, 0x0002, 0, 0, 1 }, /* PERC 3/Di (Opal/PERC3Di) */ +diff --git a/drivers/scsi/aic94xx/aic94xx_init.c b/drivers/scsi/aic94xx/aic94xx_init.c +index 996f722..9127845 100644 +--- a/drivers/scsi/aic94xx/aic94xx_init.c ++++ b/drivers/scsi/aic94xx/aic94xx_init.c +@@ -485,7 +485,7 @@ static ssize_t asd_show_update_bios(struct device *dev, + flash_error_table[i].reason); + } + +-static DEVICE_ATTR(update_bios, S_IRUGO|S_IWUGO, ++static DEVICE_ATTR(update_bios, S_IRUGO|S_IWUSR, + asd_show_update_bios, asd_store_update_bios); + + static int asd_create_dev_attrs(struct asd_ha_struct *asd_ha) +@@ -1011,7 +1011,7 @@ static struct sas_domain_function_template aic94xx_transport_functions = { + .lldd_control_phy = asd_control_phy, + }; + +-static const struct pci_device_id aic94xx_pci_table[] __devinitdata = { ++static const struct pci_device_id aic94xx_pci_table[] __devinitconst = { + {PCI_DEVICE(PCI_VENDOR_ID_ADAPTEC2, 0x410),0, 0, 1}, + {PCI_DEVICE(PCI_VENDOR_ID_ADAPTEC2, 0x412),0, 0, 1}, + {PCI_DEVICE(PCI_VENDOR_ID_ADAPTEC2, 0x416),0, 0, 1}, +diff --git a/drivers/scsi/bfa/bfa_ioc.h b/drivers/scsi/bfa/bfa_ioc.h +index 58efd4b..cb48dc7 100644 +--- a/drivers/scsi/bfa/bfa_ioc.h ++++ b/drivers/scsi/bfa/bfa_ioc.h +@@ -127,7 +127,7 @@ struct bfa_ioc_cbfn_s { + bfa_ioc_disable_cbfn_t disable_cbfn; + bfa_ioc_hbfail_cbfn_t hbfail_cbfn; + bfa_ioc_reset_cbfn_t reset_cbfn; +-}; ++} __no_const; + + /** + * Heartbeat failure notification queue element. +diff --git a/drivers/scsi/bfa/bfa_iocfc.h b/drivers/scsi/bfa/bfa_iocfc.h +index 7ad177e..5503586 100644 +--- a/drivers/scsi/bfa/bfa_iocfc.h ++++ b/drivers/scsi/bfa/bfa_iocfc.h +@@ -61,7 +61,7 @@ struct bfa_hwif_s { + void (*hw_isr_mode_set)(struct bfa_s *bfa, bfa_boolean_t msix); + void (*hw_msix_getvecs)(struct bfa_s *bfa, u32 *vecmap, + u32 *nvecs, u32 *maxvec); +-}; ++} __no_const; + typedef void (*bfa_cb_iocfc_t) (void *cbarg, enum bfa_status status); + + struct bfa_iocfc_s { +diff --git a/drivers/scsi/dpt_i2o.c b/drivers/scsi/dpt_i2o.c +index 4967643..cbec06b 100644 +--- a/drivers/scsi/dpt_i2o.c ++++ b/drivers/scsi/dpt_i2o.c +@@ -1804,6 +1804,8 @@ static int adpt_i2o_passthru(adpt_hba* pHba, u32 __user *arg) + dma_addr_t addr; + ulong flags = 0; + ++ pax_track_stack(); ++ + memset(&msg, 0, MAX_MESSAGE_SIZE*4); + // get user msg size in u32s + if(get_user(size, &user_msg[0])){ +@@ -2297,6 +2299,8 @@ static s32 adpt_scsi_to_i2o(adpt_hba* pHba, struct scsi_cmnd* cmd, struct adpt_d + s32 rcode; + dma_addr_t addr; + ++ pax_track_stack(); ++ + memset(msg, 0 , sizeof(msg)); + len = scsi_bufflen(cmd); + direction = 0x00000000; +diff --git a/drivers/scsi/eata.c b/drivers/scsi/eata.c +index c7076ce..e20c67c 100644 +--- a/drivers/scsi/eata.c ++++ b/drivers/scsi/eata.c +@@ -1087,6 +1087,8 @@ static int port_detect(unsigned long port_base, unsigned int j, + struct hostdata *ha; + char name[16]; + ++ pax_track_stack(); ++ + sprintf(name, "%s%d", driver_name, j); + + if (!request_region(port_base, REGION_SIZE, driver_name)) { +diff --git a/drivers/scsi/fcoe/libfcoe.c b/drivers/scsi/fcoe/libfcoe.c +index 11ae5c9..891daec 100644 +--- a/drivers/scsi/fcoe/libfcoe.c ++++ b/drivers/scsi/fcoe/libfcoe.c +@@ -809,6 +809,8 @@ static void fcoe_ctlr_recv_els(struct fcoe_ctlr *fip, struct sk_buff *skb) + size_t rlen; + size_t dlen; + ++ pax_track_stack(); ++ + fiph = (struct fip_header *)skb->data; + sub = fiph->fip_subcode; + if (sub != FIP_SC_REQ && sub != FIP_SC_REP) +diff --git a/drivers/scsi/fnic/fnic_main.c b/drivers/scsi/fnic/fnic_main.c +index 71c7bbe..e93088a 100644 +--- a/drivers/scsi/fnic/fnic_main.c ++++ b/drivers/scsi/fnic/fnic_main.c +@@ -669,7 +669,7 @@ static int __devinit fnic_probe(struct pci_dev *pdev, + /* Start local port initiatialization */ + + lp->link_up = 0; +- lp->tt = fnic_transport_template; ++ memcpy((void *)&lp->tt, &fnic_transport_template, sizeof(fnic_transport_template)); + + lp->max_retry_count = fnic->config.flogi_retries; + lp->max_rport_retry_count = fnic->config.plogi_retries; +diff --git a/drivers/scsi/gdth.c b/drivers/scsi/gdth.c +index bb96d74..9ec3ce4 100644 +--- a/drivers/scsi/gdth.c ++++ b/drivers/scsi/gdth.c +@@ -4102,6 +4102,8 @@ static int ioc_lockdrv(void __user *arg) + ulong flags; + gdth_ha_str *ha; + ++ pax_track_stack(); ++ + if (copy_from_user(&ldrv, arg, sizeof(gdth_ioctl_lockdrv))) + return -EFAULT; + ha = gdth_find_ha(ldrv.ionode); +@@ -4134,6 +4136,8 @@ static int ioc_resetdrv(void __user *arg, char *cmnd) + gdth_ha_str *ha; + int rval; + ++ pax_track_stack(); ++ + if (copy_from_user(&res, arg, sizeof(gdth_ioctl_reset)) || + res.number >= MAX_HDRIVES) + return -EFAULT; +@@ -4169,6 +4173,8 @@ static int ioc_general(void __user *arg, char *cmnd) + gdth_ha_str *ha; + int rval; + ++ pax_track_stack(); ++ + if (copy_from_user(&gen, arg, sizeof(gdth_ioctl_general))) + return -EFAULT; + ha = gdth_find_ha(gen.ionode); +@@ -4625,6 +4631,9 @@ static void gdth_flush(gdth_ha_str *ha) + int i; + gdth_cmd_str gdtcmd; + char cmnd[MAX_COMMAND_SIZE]; ++ ++ pax_track_stack(); ++ + memset(cmnd, 0xff, MAX_COMMAND_SIZE); + + TRACE2(("gdth_flush() hanum %d\n", ha->hanum)); +diff --git a/drivers/scsi/gdth_proc.c b/drivers/scsi/gdth_proc.c +index 1258da3..20d8ae6 100644 +--- a/drivers/scsi/gdth_proc.c ++++ b/drivers/scsi/gdth_proc.c +@@ -46,6 +46,9 @@ static int gdth_set_asc_info(struct Scsi_Host *host, char *buffer, + ulong64 paddr; + + char cmnd[MAX_COMMAND_SIZE]; ++ ++ pax_track_stack(); ++ + memset(cmnd, 0xff, 12); + memset(&gdtcmd, 0, sizeof(gdth_cmd_str)); + +@@ -174,6 +177,8 @@ static int gdth_get_info(char *buffer,char **start,off_t offset,int length, + gdth_hget_str *phg; + char cmnd[MAX_COMMAND_SIZE]; + ++ pax_track_stack(); ++ + gdtcmd = kmalloc(sizeof(*gdtcmd), GFP_KERNEL); + estr = kmalloc(sizeof(*estr), GFP_KERNEL); + if (!gdtcmd || !estr) +diff --git a/drivers/scsi/hosts.c b/drivers/scsi/hosts.c +index d03a926..f324286 100644 +--- a/drivers/scsi/hosts.c ++++ b/drivers/scsi/hosts.c +@@ -40,7 +40,7 @@ + #include "scsi_logging.h" + + +-static atomic_t scsi_host_next_hn; /* host_no for next new host */ ++static atomic_unchecked_t scsi_host_next_hn; /* host_no for next new host */ + + + static void scsi_host_cls_release(struct device *dev) +@@ -347,7 +347,7 @@ struct Scsi_Host *scsi_host_alloc(struct scsi_host_template *sht, int privsize) + * subtract one because we increment first then return, but we need to + * know what the next host number was before increment + */ +- shost->host_no = atomic_inc_return(&scsi_host_next_hn) - 1; ++ shost->host_no = atomic_inc_return_unchecked(&scsi_host_next_hn) - 1; + shost->dma_channel = 0xff; + + /* These three are default values which can be overridden */ +diff --git a/drivers/scsi/ipr.c b/drivers/scsi/ipr.c +index a601159..55e19d2 100644 +--- a/drivers/scsi/ipr.c ++++ b/drivers/scsi/ipr.c +@@ -5286,7 +5286,7 @@ static bool ipr_qc_fill_rtf(struct ata_queued_cmd *qc) + return true; + } + +-static struct ata_port_operations ipr_sata_ops = { ++static const struct ata_port_operations ipr_sata_ops = { + .phy_reset = ipr_ata_phy_reset, + .hardreset = ipr_sata_reset, + .post_internal_cmd = ipr_ata_post_internal, +diff --git a/drivers/scsi/ips.h b/drivers/scsi/ips.h +index 4e49fbc..97907ff 100644 +--- a/drivers/scsi/ips.h ++++ b/drivers/scsi/ips.h +@@ -1027,7 +1027,7 @@ typedef struct { + int (*intr)(struct ips_ha *); + void (*enableint)(struct ips_ha *); + uint32_t (*statupd)(struct ips_ha *); +-} ips_hw_func_t; ++} __no_const ips_hw_func_t; + + typedef struct ips_ha { + uint8_t ha_id[IPS_MAX_CHANNELS+1]; +diff --git a/drivers/scsi/libfc/fc_exch.c b/drivers/scsi/libfc/fc_exch.c +index c1c1574..a9c9348 100644 +--- a/drivers/scsi/libfc/fc_exch.c ++++ b/drivers/scsi/libfc/fc_exch.c +@@ -86,12 +86,12 @@ struct fc_exch_mgr { + * all together if not used XXX + */ + struct { +- atomic_t no_free_exch; +- atomic_t no_free_exch_xid; +- atomic_t xid_not_found; +- atomic_t xid_busy; +- atomic_t seq_not_found; +- atomic_t non_bls_resp; ++ atomic_unchecked_t no_free_exch; ++ atomic_unchecked_t no_free_exch_xid; ++ atomic_unchecked_t xid_not_found; ++ atomic_unchecked_t xid_busy; ++ atomic_unchecked_t seq_not_found; ++ atomic_unchecked_t non_bls_resp; + } stats; + }; + #define fc_seq_exch(sp) container_of(sp, struct fc_exch, seq) +@@ -510,7 +510,7 @@ static struct fc_exch *fc_exch_em_alloc(struct fc_lport *lport, + /* allocate memory for exchange */ + ep = mempool_alloc(mp->ep_pool, GFP_ATOMIC); + if (!ep) { +- atomic_inc(&mp->stats.no_free_exch); ++ atomic_inc_unchecked(&mp->stats.no_free_exch); + goto out; + } + memset(ep, 0, sizeof(*ep)); +@@ -557,7 +557,7 @@ out: + return ep; + err: + spin_unlock_bh(&pool->lock); +- atomic_inc(&mp->stats.no_free_exch_xid); ++ atomic_inc_unchecked(&mp->stats.no_free_exch_xid); + mempool_free(ep, mp->ep_pool); + return NULL; + } +@@ -690,7 +690,7 @@ static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_lport *lport, + xid = ntohs(fh->fh_ox_id); /* we originated exch */ + ep = fc_exch_find(mp, xid); + if (!ep) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + reject = FC_RJT_OX_ID; + goto out; + } +@@ -720,7 +720,7 @@ static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_lport *lport, + ep = fc_exch_find(mp, xid); + if ((f_ctl & FC_FC_FIRST_SEQ) && fc_sof_is_init(fr_sof(fp))) { + if (ep) { +- atomic_inc(&mp->stats.xid_busy); ++ atomic_inc_unchecked(&mp->stats.xid_busy); + reject = FC_RJT_RX_ID; + goto rel; + } +@@ -731,7 +731,7 @@ static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_lport *lport, + } + xid = ep->xid; /* get our XID */ + } else if (!ep) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + reject = FC_RJT_RX_ID; /* XID not found */ + goto out; + } +@@ -752,7 +752,7 @@ static enum fc_pf_rjt_reason fc_seq_lookup_recip(struct fc_lport *lport, + } else { + sp = &ep->seq; + if (sp->id != fh->fh_seq_id) { +- atomic_inc(&mp->stats.seq_not_found); ++ atomic_inc_unchecked(&mp->stats.seq_not_found); + reject = FC_RJT_SEQ_ID; /* sequence/exch should exist */ + goto rel; + } +@@ -1163,22 +1163,22 @@ static void fc_exch_recv_seq_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) + + ep = fc_exch_find(mp, ntohs(fh->fh_ox_id)); + if (!ep) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + goto out; + } + if (ep->esb_stat & ESB_ST_COMPLETE) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + goto out; + } + if (ep->rxid == FC_XID_UNKNOWN) + ep->rxid = ntohs(fh->fh_rx_id); + if (ep->sid != 0 && ep->sid != ntoh24(fh->fh_d_id)) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + goto rel; + } + if (ep->did != ntoh24(fh->fh_s_id) && + ep->did != FC_FID_FLOGI) { +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + goto rel; + } + sof = fr_sof(fp); +@@ -1189,7 +1189,7 @@ static void fc_exch_recv_seq_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) + } else { + sp = &ep->seq; + if (sp->id != fh->fh_seq_id) { +- atomic_inc(&mp->stats.seq_not_found); ++ atomic_inc_unchecked(&mp->stats.seq_not_found); + goto rel; + } + } +@@ -1249,9 +1249,9 @@ static void fc_exch_recv_resp(struct fc_exch_mgr *mp, struct fc_frame *fp) + sp = fc_seq_lookup_orig(mp, fp); /* doesn't hold sequence */ + + if (!sp) +- atomic_inc(&mp->stats.xid_not_found); ++ atomic_inc_unchecked(&mp->stats.xid_not_found); + else +- atomic_inc(&mp->stats.non_bls_resp); ++ atomic_inc_unchecked(&mp->stats.non_bls_resp); + + fc_frame_free(fp); + } +diff --git a/drivers/scsi/libsas/sas_ata.c b/drivers/scsi/libsas/sas_ata.c +index 0ee989f..a582241 100644 +--- a/drivers/scsi/libsas/sas_ata.c ++++ b/drivers/scsi/libsas/sas_ata.c +@@ -343,7 +343,7 @@ static int sas_ata_scr_read(struct ata_link *link, unsigned int sc_reg_in, + } + } + +-static struct ata_port_operations sas_sata_ops = { ++static const struct ata_port_operations sas_sata_ops = { + .phy_reset = sas_ata_phy_reset, + .post_internal_cmd = sas_ata_post_internal, + .qc_defer = ata_std_qc_defer, +diff --git a/drivers/scsi/lpfc/lpfc.h b/drivers/scsi/lpfc/lpfc.h +index aa10f79..5cc79e4 100644 +--- a/drivers/scsi/lpfc/lpfc.h ++++ b/drivers/scsi/lpfc/lpfc.h +@@ -400,7 +400,7 @@ struct lpfc_vport { + struct dentry *debug_nodelist; + struct dentry *vport_debugfs_root; + struct lpfc_debugfs_trc *disc_trc; +- atomic_t disc_trc_cnt; ++ atomic_unchecked_t disc_trc_cnt; + #endif + uint8_t stat_data_enabled; + uint8_t stat_data_blocked; +@@ -725,8 +725,8 @@ struct lpfc_hba { + struct timer_list fabric_block_timer; + unsigned long bit_flags; + #define FABRIC_COMANDS_BLOCKED 0 +- atomic_t num_rsrc_err; +- atomic_t num_cmd_success; ++ atomic_unchecked_t num_rsrc_err; ++ atomic_unchecked_t num_cmd_success; + unsigned long last_rsrc_error_time; + unsigned long last_ramp_down_time; + unsigned long last_ramp_up_time; +@@ -740,7 +740,7 @@ struct lpfc_hba { + struct dentry *debug_dumpDif; /* BlockGuard BPL*/ + struct dentry *debug_slow_ring_trc; + struct lpfc_debugfs_trc *slow_ring_trc; +- atomic_t slow_ring_trc_cnt; ++ atomic_unchecked_t slow_ring_trc_cnt; + #endif + + /* Used for deferred freeing of ELS data buffers */ +diff --git a/drivers/scsi/lpfc/lpfc_debugfs.c b/drivers/scsi/lpfc/lpfc_debugfs.c +index 8d0f0de..7c77a62 100644 +--- a/drivers/scsi/lpfc/lpfc_debugfs.c ++++ b/drivers/scsi/lpfc/lpfc_debugfs.c +@@ -124,7 +124,7 @@ struct lpfc_debug { + int len; + }; + +-static atomic_t lpfc_debugfs_seq_trc_cnt = ATOMIC_INIT(0); ++static atomic_unchecked_t lpfc_debugfs_seq_trc_cnt = ATOMIC_INIT(0); + static unsigned long lpfc_debugfs_start_time = 0L; + + /** +@@ -158,7 +158,7 @@ lpfc_debugfs_disc_trc_data(struct lpfc_vport *vport, char *buf, int size) + lpfc_debugfs_enable = 0; + + len = 0; +- index = (atomic_read(&vport->disc_trc_cnt) + 1) & ++ index = (atomic_read_unchecked(&vport->disc_trc_cnt) + 1) & + (lpfc_debugfs_max_disc_trc - 1); + for (i = index; i < lpfc_debugfs_max_disc_trc; i++) { + dtp = vport->disc_trc + i; +@@ -219,7 +219,7 @@ lpfc_debugfs_slow_ring_trc_data(struct lpfc_hba *phba, char *buf, int size) + lpfc_debugfs_enable = 0; + + len = 0; +- index = (atomic_read(&phba->slow_ring_trc_cnt) + 1) & ++ index = (atomic_read_unchecked(&phba->slow_ring_trc_cnt) + 1) & + (lpfc_debugfs_max_slow_ring_trc - 1); + for (i = index; i < lpfc_debugfs_max_slow_ring_trc; i++) { + dtp = phba->slow_ring_trc + i; +@@ -397,6 +397,8 @@ lpfc_debugfs_dumpHBASlim_data(struct lpfc_hba *phba, char *buf, int size) + uint32_t *ptr; + char buffer[1024]; + ++ pax_track_stack(); ++ + off = 0; + spin_lock_irq(&phba->hbalock); + +@@ -634,14 +636,14 @@ lpfc_debugfs_disc_trc(struct lpfc_vport *vport, int mask, char *fmt, + !vport || !vport->disc_trc) + return; + +- index = atomic_inc_return(&vport->disc_trc_cnt) & ++ index = atomic_inc_return_unchecked(&vport->disc_trc_cnt) & + (lpfc_debugfs_max_disc_trc - 1); + dtp = vport->disc_trc + index; + dtp->fmt = fmt; + dtp->data1 = data1; + dtp->data2 = data2; + dtp->data3 = data3; +- dtp->seq_cnt = atomic_inc_return(&lpfc_debugfs_seq_trc_cnt); ++ dtp->seq_cnt = atomic_inc_return_unchecked(&lpfc_debugfs_seq_trc_cnt); + dtp->jif = jiffies; + #endif + return; +@@ -672,14 +674,14 @@ lpfc_debugfs_slow_ring_trc(struct lpfc_hba *phba, char *fmt, + !phba || !phba->slow_ring_trc) + return; + +- index = atomic_inc_return(&phba->slow_ring_trc_cnt) & ++ index = atomic_inc_return_unchecked(&phba->slow_ring_trc_cnt) & + (lpfc_debugfs_max_slow_ring_trc - 1); + dtp = phba->slow_ring_trc + index; + dtp->fmt = fmt; + dtp->data1 = data1; + dtp->data2 = data2; + dtp->data3 = data3; +- dtp->seq_cnt = atomic_inc_return(&lpfc_debugfs_seq_trc_cnt); ++ dtp->seq_cnt = atomic_inc_return_unchecked(&lpfc_debugfs_seq_trc_cnt); + dtp->jif = jiffies; + #endif + return; +@@ -1364,7 +1366,7 @@ lpfc_debugfs_initialize(struct lpfc_vport *vport) + "slow_ring buffer\n"); + goto debug_failed; + } +- atomic_set(&phba->slow_ring_trc_cnt, 0); ++ atomic_set_unchecked(&phba->slow_ring_trc_cnt, 0); + memset(phba->slow_ring_trc, 0, + (sizeof(struct lpfc_debugfs_trc) * + lpfc_debugfs_max_slow_ring_trc)); +@@ -1410,7 +1412,7 @@ lpfc_debugfs_initialize(struct lpfc_vport *vport) + "buffer\n"); + goto debug_failed; + } +- atomic_set(&vport->disc_trc_cnt, 0); ++ atomic_set_unchecked(&vport->disc_trc_cnt, 0); + + snprintf(name, sizeof(name), "discovery_trace"); + vport->debug_disc_trc = +diff --git a/drivers/scsi/lpfc/lpfc_init.c b/drivers/scsi/lpfc/lpfc_init.c +index 549bc7d..8189dbb 100644 +--- a/drivers/scsi/lpfc/lpfc_init.c ++++ b/drivers/scsi/lpfc/lpfc_init.c +@@ -8021,8 +8021,10 @@ lpfc_init(void) + printk(LPFC_COPYRIGHT "\n"); + + if (lpfc_enable_npiv) { +- lpfc_transport_functions.vport_create = lpfc_vport_create; +- lpfc_transport_functions.vport_delete = lpfc_vport_delete; ++ pax_open_kernel(); ++ *(void **)&lpfc_transport_functions.vport_create = lpfc_vport_create; ++ *(void **)&lpfc_transport_functions.vport_delete = lpfc_vport_delete; ++ pax_close_kernel(); + } + lpfc_transport_template = + fc_attach_transport(&lpfc_transport_functions); +diff --git a/drivers/scsi/lpfc/lpfc_scsi.c b/drivers/scsi/lpfc/lpfc_scsi.c +index c88f59f..ff2a42f 100644 +--- a/drivers/scsi/lpfc/lpfc_scsi.c ++++ b/drivers/scsi/lpfc/lpfc_scsi.c +@@ -259,7 +259,7 @@ lpfc_rampdown_queue_depth(struct lpfc_hba *phba) + uint32_t evt_posted; + + spin_lock_irqsave(&phba->hbalock, flags); +- atomic_inc(&phba->num_rsrc_err); ++ atomic_inc_unchecked(&phba->num_rsrc_err); + phba->last_rsrc_error_time = jiffies; + + if ((phba->last_ramp_down_time + QUEUE_RAMP_DOWN_INTERVAL) > jiffies) { +@@ -300,7 +300,7 @@ lpfc_rampup_queue_depth(struct lpfc_vport *vport, + unsigned long flags; + struct lpfc_hba *phba = vport->phba; + uint32_t evt_posted; +- atomic_inc(&phba->num_cmd_success); ++ atomic_inc_unchecked(&phba->num_cmd_success); + + if (vport->cfg_lun_queue_depth <= queue_depth) + return; +@@ -343,8 +343,8 @@ lpfc_ramp_down_queue_handler(struct lpfc_hba *phba) + int i; + struct lpfc_rport_data *rdata; + +- num_rsrc_err = atomic_read(&phba->num_rsrc_err); +- num_cmd_success = atomic_read(&phba->num_cmd_success); ++ num_rsrc_err = atomic_read_unchecked(&phba->num_rsrc_err); ++ num_cmd_success = atomic_read_unchecked(&phba->num_cmd_success); + + vports = lpfc_create_vport_work_array(phba); + if (vports != NULL) +@@ -378,8 +378,8 @@ lpfc_ramp_down_queue_handler(struct lpfc_hba *phba) + } + } + lpfc_destroy_vport_work_array(phba, vports); +- atomic_set(&phba->num_rsrc_err, 0); +- atomic_set(&phba->num_cmd_success, 0); ++ atomic_set_unchecked(&phba->num_rsrc_err, 0); ++ atomic_set_unchecked(&phba->num_cmd_success, 0); + } + + /** +@@ -427,8 +427,8 @@ lpfc_ramp_up_queue_handler(struct lpfc_hba *phba) + } + } + lpfc_destroy_vport_work_array(phba, vports); +- atomic_set(&phba->num_rsrc_err, 0); +- atomic_set(&phba->num_cmd_success, 0); ++ atomic_set_unchecked(&phba->num_rsrc_err, 0); ++ atomic_set_unchecked(&phba->num_cmd_success, 0); + } + + /** +diff --git a/drivers/scsi/megaraid/megaraid_mbox.c b/drivers/scsi/megaraid/megaraid_mbox.c +index 234f0b7..3020aea 100644 +--- a/drivers/scsi/megaraid/megaraid_mbox.c ++++ b/drivers/scsi/megaraid/megaraid_mbox.c +@@ -3503,6 +3503,8 @@ megaraid_cmm_register(adapter_t *adapter) + int rval; + int i; + ++ pax_track_stack(); ++ + // Allocate memory for the base list of scb for management module. + adapter->uscb_list = kcalloc(MBOX_MAX_USER_CMDS, sizeof(scb_t), GFP_KERNEL); + +diff --git a/drivers/scsi/osd/osd_initiator.c b/drivers/scsi/osd/osd_initiator.c +index 7a117c1..ee01e9e 100644 +--- a/drivers/scsi/osd/osd_initiator.c ++++ b/drivers/scsi/osd/osd_initiator.c +@@ -94,6 +94,8 @@ static int _osd_print_system_info(struct osd_dev *od, void *caps) + int nelem = ARRAY_SIZE(get_attrs), a = 0; + int ret; + ++ pax_track_stack(); ++ + or = osd_start_request(od, GFP_KERNEL); + if (!or) + return -ENOMEM; +diff --git a/drivers/scsi/pmcraid.c b/drivers/scsi/pmcraid.c +index 9ab8c86..9425ad3 100644 +--- a/drivers/scsi/pmcraid.c ++++ b/drivers/scsi/pmcraid.c +@@ -189,8 +189,8 @@ static int pmcraid_slave_alloc(struct scsi_device *scsi_dev) + res->scsi_dev = scsi_dev; + scsi_dev->hostdata = res; + res->change_detected = 0; +- atomic_set(&res->read_failures, 0); +- atomic_set(&res->write_failures, 0); ++ atomic_set_unchecked(&res->read_failures, 0); ++ atomic_set_unchecked(&res->write_failures, 0); + rc = 0; + } + spin_unlock_irqrestore(&pinstance->resource_lock, lock_flags); +@@ -2396,9 +2396,9 @@ static int pmcraid_error_handler(struct pmcraid_cmd *cmd) + + /* If this was a SCSI read/write command keep count of errors */ + if (SCSI_CMD_TYPE(scsi_cmd->cmnd[0]) == SCSI_READ_CMD) +- atomic_inc(&res->read_failures); ++ atomic_inc_unchecked(&res->read_failures); + else if (SCSI_CMD_TYPE(scsi_cmd->cmnd[0]) == SCSI_WRITE_CMD) +- atomic_inc(&res->write_failures); ++ atomic_inc_unchecked(&res->write_failures); + + if (!RES_IS_GSCSI(res->cfg_entry) && + masked_ioasc != PMCRAID_IOASC_HW_DEVICE_BUS_STATUS_ERROR) { +@@ -4116,7 +4116,7 @@ static void pmcraid_worker_function(struct work_struct *workp) + + pinstance = container_of(workp, struct pmcraid_instance, worker_q); + /* add resources only after host is added into system */ +- if (!atomic_read(&pinstance->expose_resources)) ++ if (!atomic_read_unchecked(&pinstance->expose_resources)) + return; + + spin_lock_irqsave(&pinstance->resource_lock, lock_flags); +@@ -4850,7 +4850,7 @@ static int __devinit pmcraid_init_instance( + init_waitqueue_head(&pinstance->reset_wait_q); + + atomic_set(&pinstance->outstanding_cmds, 0); +- atomic_set(&pinstance->expose_resources, 0); ++ atomic_set_unchecked(&pinstance->expose_resources, 0); + + INIT_LIST_HEAD(&pinstance->free_res_q); + INIT_LIST_HEAD(&pinstance->used_res_q); +@@ -5502,7 +5502,7 @@ static int __devinit pmcraid_probe( + /* Schedule worker thread to handle CCN and take care of adding and + * removing devices to OS + */ +- atomic_set(&pinstance->expose_resources, 1); ++ atomic_set_unchecked(&pinstance->expose_resources, 1); + schedule_work(&pinstance->worker_q); + return rc; + +diff --git a/drivers/scsi/pmcraid.h b/drivers/scsi/pmcraid.h +index 3441b3f..6cbe8f7 100644 +--- a/drivers/scsi/pmcraid.h ++++ b/drivers/scsi/pmcraid.h +@@ -690,7 +690,7 @@ struct pmcraid_instance { + atomic_t outstanding_cmds; + + /* should add/delete resources to mid-layer now ?*/ +- atomic_t expose_resources; ++ atomic_unchecked_t expose_resources; + + /* Tasklet to handle deferred processing */ + struct tasklet_struct isr_tasklet[PMCRAID_NUM_MSIX_VECTORS]; +@@ -727,8 +727,8 @@ struct pmcraid_resource_entry { + struct list_head queue; /* link to "to be exposed" resources */ + struct pmcraid_config_table_entry cfg_entry; + struct scsi_device *scsi_dev; /* Link scsi_device structure */ +- atomic_t read_failures; /* count of failed READ commands */ +- atomic_t write_failures; /* count of failed WRITE commands */ ++ atomic_unchecked_t read_failures; /* count of failed READ commands */ ++ atomic_unchecked_t write_failures; /* count of failed WRITE commands */ + + /* To indicate add/delete/modify during CCN */ + u8 change_detected; +diff --git a/drivers/scsi/qla2xxx/qla_def.h b/drivers/scsi/qla2xxx/qla_def.h +index 2150618..7034215 100644 +--- a/drivers/scsi/qla2xxx/qla_def.h ++++ b/drivers/scsi/qla2xxx/qla_def.h +@@ -2089,7 +2089,7 @@ struct isp_operations { + + int (*get_flash_version) (struct scsi_qla_host *, void *); + int (*start_scsi) (srb_t *); +-}; ++} __no_const; + + /* MSI-X Support *************************************************************/ + +diff --git a/drivers/scsi/qla4xxx/ql4_def.h b/drivers/scsi/qla4xxx/ql4_def.h +index 81b5f29..2ae1fad 100644 +--- a/drivers/scsi/qla4xxx/ql4_def.h ++++ b/drivers/scsi/qla4xxx/ql4_def.h +@@ -240,7 +240,7 @@ struct ddb_entry { + atomic_t retry_relogin_timer; /* Min Time between relogins + * (4000 only) */ + atomic_t relogin_timer; /* Max Time to wait for relogin to complete */ +- atomic_t relogin_retry_count; /* Num of times relogin has been ++ atomic_unchecked_t relogin_retry_count; /* Num of times relogin has been + * retried */ + + uint16_t port; +diff --git a/drivers/scsi/qla4xxx/ql4_init.c b/drivers/scsi/qla4xxx/ql4_init.c +index af8c323..515dd51 100644 +--- a/drivers/scsi/qla4xxx/ql4_init.c ++++ b/drivers/scsi/qla4xxx/ql4_init.c +@@ -482,7 +482,7 @@ static struct ddb_entry * qla4xxx_alloc_ddb(struct scsi_qla_host *ha, + atomic_set(&ddb_entry->port_down_timer, ha->port_down_retry_count); + atomic_set(&ddb_entry->retry_relogin_timer, INVALID_ENTRY); + atomic_set(&ddb_entry->relogin_timer, 0); +- atomic_set(&ddb_entry->relogin_retry_count, 0); ++ atomic_set_unchecked(&ddb_entry->relogin_retry_count, 0); + atomic_set(&ddb_entry->state, DDB_STATE_ONLINE); + list_add_tail(&ddb_entry->list, &ha->ddb_list); + ha->fw_ddb_index_map[fw_ddb_index] = ddb_entry; +@@ -1308,7 +1308,7 @@ int qla4xxx_process_ddb_changed(struct scsi_qla_host *ha, + atomic_set(&ddb_entry->state, DDB_STATE_ONLINE); + atomic_set(&ddb_entry->port_down_timer, + ha->port_down_retry_count); +- atomic_set(&ddb_entry->relogin_retry_count, 0); ++ atomic_set_unchecked(&ddb_entry->relogin_retry_count, 0); + atomic_set(&ddb_entry->relogin_timer, 0); + clear_bit(DF_RELOGIN, &ddb_entry->flags); + clear_bit(DF_NO_RELOGIN, &ddb_entry->flags); +diff --git a/drivers/scsi/qla4xxx/ql4_os.c b/drivers/scsi/qla4xxx/ql4_os.c +index 83c8b5e..a82b348 100644 +--- a/drivers/scsi/qla4xxx/ql4_os.c ++++ b/drivers/scsi/qla4xxx/ql4_os.c +@@ -641,13 +641,13 @@ static void qla4xxx_timer(struct scsi_qla_host *ha) + ddb_entry->fw_ddb_device_state == + DDB_DS_SESSION_FAILED) { + /* Reset retry relogin timer */ +- atomic_inc(&ddb_entry->relogin_retry_count); ++ atomic_inc_unchecked(&ddb_entry->relogin_retry_count); + DEBUG2(printk("scsi%ld: index[%d] relogin" + " timed out-retrying" + " relogin (%d)\n", + ha->host_no, + ddb_entry->fw_ddb_index, +- atomic_read(&ddb_entry-> ++ atomic_read_unchecked(&ddb_entry-> + relogin_retry_count)) + ); + start_dpc++; +diff --git a/drivers/scsi/scsi.c b/drivers/scsi/scsi.c +index dd098ca..686ce01 100644 +--- a/drivers/scsi/scsi.c ++++ b/drivers/scsi/scsi.c +@@ -652,7 +652,7 @@ int scsi_dispatch_cmd(struct scsi_cmnd *cmd) + unsigned long timeout; + int rtn = 0; + +- atomic_inc(&cmd->device->iorequest_cnt); ++ atomic_inc_unchecked(&cmd->device->iorequest_cnt); + + /* check if the device is still usable */ + if (unlikely(cmd->device->sdev_state == SDEV_DEL)) { +diff --git a/drivers/scsi/scsi_debug.c b/drivers/scsi/scsi_debug.c +index bc3e363..e1a8e50 100644 +--- a/drivers/scsi/scsi_debug.c ++++ b/drivers/scsi/scsi_debug.c +@@ -1395,6 +1395,8 @@ static int resp_mode_select(struct scsi_cmnd * scp, int mselect6, + unsigned char arr[SDEBUG_MAX_MSELECT_SZ]; + unsigned char *cmd = (unsigned char *)scp->cmnd; + ++ pax_track_stack(); ++ + if ((errsts = check_readiness(scp, 1, devip))) + return errsts; + memset(arr, 0, sizeof(arr)); +@@ -1492,6 +1494,8 @@ static int resp_log_sense(struct scsi_cmnd * scp, + unsigned char arr[SDEBUG_MAX_LSENSE_SZ]; + unsigned char *cmd = (unsigned char *)scp->cmnd; + ++ pax_track_stack(); ++ + if ((errsts = check_readiness(scp, 1, devip))) + return errsts; + memset(arr, 0, sizeof(arr)); +diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c +index 8df12522..c4c1472 100644 +--- a/drivers/scsi/scsi_lib.c ++++ b/drivers/scsi/scsi_lib.c +@@ -1389,7 +1389,7 @@ static void scsi_kill_request(struct request *req, struct request_queue *q) + shost = sdev->host; + scsi_init_cmd_errh(cmd); + cmd->result = DID_NO_CONNECT << 16; +- atomic_inc(&cmd->device->iorequest_cnt); ++ atomic_inc_unchecked(&cmd->device->iorequest_cnt); + + /* + * SCSI request completion path will do scsi_device_unbusy(), +@@ -1420,9 +1420,9 @@ static void scsi_softirq_done(struct request *rq) + */ + cmd->serial_number = 0; + +- atomic_inc(&cmd->device->iodone_cnt); ++ atomic_inc_unchecked(&cmd->device->iodone_cnt); + if (cmd->result) +- atomic_inc(&cmd->device->ioerr_cnt); ++ atomic_inc_unchecked(&cmd->device->ioerr_cnt); + + disposition = scsi_decide_disposition(cmd); + if (disposition != SUCCESS && +diff --git a/drivers/scsi/scsi_sysfs.c b/drivers/scsi/scsi_sysfs.c +index 91a93e0..eae0fe3 100644 +--- a/drivers/scsi/scsi_sysfs.c ++++ b/drivers/scsi/scsi_sysfs.c +@@ -662,7 +662,7 @@ show_iostat_##field(struct device *dev, struct device_attribute *attr, \ + char *buf) \ + { \ + struct scsi_device *sdev = to_scsi_device(dev); \ +- unsigned long long count = atomic_read(&sdev->field); \ ++ unsigned long long count = atomic_read_unchecked(&sdev->field); \ + return snprintf(buf, 20, "0x%llx\n", count); \ + } \ + static DEVICE_ATTR(field, S_IRUGO, show_iostat_##field, NULL) +diff --git a/drivers/scsi/scsi_tgt_lib.c b/drivers/scsi/scsi_tgt_lib.c +index 1030327..f91fd30 100644 +--- a/drivers/scsi/scsi_tgt_lib.c ++++ b/drivers/scsi/scsi_tgt_lib.c +@@ -362,7 +362,7 @@ static int scsi_map_user_pages(struct scsi_tgt_cmd *tcmd, struct scsi_cmnd *cmd, + int err; + + dprintk("%lx %u\n", uaddr, len); +- err = blk_rq_map_user(q, rq, NULL, (void *)uaddr, len, GFP_KERNEL); ++ err = blk_rq_map_user(q, rq, NULL, (void __user *)uaddr, len, GFP_KERNEL); + if (err) { + /* + * TODO: need to fixup sg_tablesize, max_segment_size, +diff --git a/drivers/scsi/scsi_transport_fc.c b/drivers/scsi/scsi_transport_fc.c +index db02e31..1b42ea9 100644 +--- a/drivers/scsi/scsi_transport_fc.c ++++ b/drivers/scsi/scsi_transport_fc.c +@@ -480,7 +480,7 @@ MODULE_PARM_DESC(dev_loss_tmo, + * Netlink Infrastructure + */ + +-static atomic_t fc_event_seq; ++static atomic_unchecked_t fc_event_seq; + + /** + * fc_get_event_number - Obtain the next sequential FC event number +@@ -493,7 +493,7 @@ static atomic_t fc_event_seq; + u32 + fc_get_event_number(void) + { +- return atomic_add_return(1, &fc_event_seq); ++ return atomic_add_return_unchecked(1, &fc_event_seq); + } + EXPORT_SYMBOL(fc_get_event_number); + +@@ -641,7 +641,7 @@ static __init int fc_transport_init(void) + { + int error; + +- atomic_set(&fc_event_seq, 0); ++ atomic_set_unchecked(&fc_event_seq, 0); + + error = transport_class_register(&fc_host_class); + if (error) +diff --git a/drivers/scsi/scsi_transport_iscsi.c b/drivers/scsi/scsi_transport_iscsi.c +index de2f8c4..63c5278 100644 +--- a/drivers/scsi/scsi_transport_iscsi.c ++++ b/drivers/scsi/scsi_transport_iscsi.c +@@ -81,7 +81,7 @@ struct iscsi_internal { + struct device_attribute *session_attrs[ISCSI_SESSION_ATTRS + 1]; + }; + +-static atomic_t iscsi_session_nr; /* sysfs session id for next new session */ ++static atomic_unchecked_t iscsi_session_nr; /* sysfs session id for next new session */ + static struct workqueue_struct *iscsi_eh_timer_workq; + + /* +@@ -728,7 +728,7 @@ int iscsi_add_session(struct iscsi_cls_session *session, unsigned int target_id) + int err; + + ihost = shost->shost_data; +- session->sid = atomic_add_return(1, &iscsi_session_nr); ++ session->sid = atomic_add_return_unchecked(1, &iscsi_session_nr); + + if (id == ISCSI_MAX_TARGET) { + for (id = 0; id < ISCSI_MAX_TARGET; id++) { +@@ -2060,7 +2060,7 @@ static __init int iscsi_transport_init(void) + printk(KERN_INFO "Loading iSCSI transport class v%s.\n", + ISCSI_TRANSPORT_VERSION); + +- atomic_set(&iscsi_session_nr, 0); ++ atomic_set_unchecked(&iscsi_session_nr, 0); + + err = class_register(&iscsi_transport_class); + if (err) +diff --git a/drivers/scsi/scsi_transport_srp.c b/drivers/scsi/scsi_transport_srp.c +index 21a045e..ec89e03 100644 +--- a/drivers/scsi/scsi_transport_srp.c ++++ b/drivers/scsi/scsi_transport_srp.c +@@ -33,7 +33,7 @@ + #include "scsi_transport_srp_internal.h" + + struct srp_host_attrs { +- atomic_t next_port_id; ++ atomic_unchecked_t next_port_id; + }; + #define to_srp_host_attrs(host) ((struct srp_host_attrs *)(host)->shost_data) + +@@ -62,7 +62,7 @@ static int srp_host_setup(struct transport_container *tc, struct device *dev, + struct Scsi_Host *shost = dev_to_shost(dev); + struct srp_host_attrs *srp_host = to_srp_host_attrs(shost); + +- atomic_set(&srp_host->next_port_id, 0); ++ atomic_set_unchecked(&srp_host->next_port_id, 0); + return 0; + } + +@@ -211,7 +211,7 @@ struct srp_rport *srp_rport_add(struct Scsi_Host *shost, + memcpy(rport->port_id, ids->port_id, sizeof(rport->port_id)); + rport->roles = ids->roles; + +- id = atomic_inc_return(&to_srp_host_attrs(shost)->next_port_id); ++ id = atomic_inc_return_unchecked(&to_srp_host_attrs(shost)->next_port_id); + dev_set_name(&rport->dev, "port-%d:%d", shost->host_no, id); + + transport_setup_device(&rport->dev); +diff --git a/drivers/scsi/sg.c b/drivers/scsi/sg.c +index 040f751..98a5ed2 100644 +--- a/drivers/scsi/sg.c ++++ b/drivers/scsi/sg.c +@@ -1064,7 +1064,7 @@ sg_ioctl(struct inode *inode, struct file *filp, + sdp->disk->disk_name, + MKDEV(SCSI_GENERIC_MAJOR, sdp->index), + NULL, +- (char *)arg); ++ (char __user *)arg); + case BLKTRACESTART: + return blk_trace_startstop(sdp->device->request_queue, 1); + case BLKTRACESTOP: +@@ -2292,7 +2292,7 @@ struct sg_proc_leaf { + const struct file_operations * fops; + }; + +-static struct sg_proc_leaf sg_proc_leaf_arr[] = { ++static const struct sg_proc_leaf sg_proc_leaf_arr[] = { + {"allow_dio", &adio_fops}, + {"debug", &debug_fops}, + {"def_reserved_size", &dressz_fops}, +@@ -2307,7 +2307,7 @@ sg_proc_init(void) + { + int k, mask; + int num_leaves = ARRAY_SIZE(sg_proc_leaf_arr); +- struct sg_proc_leaf * leaf; ++ const struct sg_proc_leaf * leaf; + + sg_proc_sgp = proc_mkdir(sg_proc_sg_dirname, NULL); + if (!sg_proc_sgp) +diff --git a/drivers/scsi/sym53c8xx_2/sym_glue.c b/drivers/scsi/sym53c8xx_2/sym_glue.c +index c19ca5e..3eb5959 100644 +--- a/drivers/scsi/sym53c8xx_2/sym_glue.c ++++ b/drivers/scsi/sym53c8xx_2/sym_glue.c +@@ -1758,6 +1758,8 @@ static int __devinit sym2_probe(struct pci_dev *pdev, + int do_iounmap = 0; + int do_disable_device = 1; + ++ pax_track_stack(); ++ + memset(&sym_dev, 0, sizeof(sym_dev)); + memset(&nvram, 0, sizeof(nvram)); + sym_dev.pdev = pdev; +diff --git a/drivers/scsi/vmw_pvscsi.c b/drivers/scsi/vmw_pvscsi.c +new file mode 100644 +index 0000000..eabb432 +--- /dev/null ++++ b/drivers/scsi/vmw_pvscsi.c +@@ -0,0 +1,1401 @@ ++/* ++ * Linux driver for VMware's para-virtualized SCSI HBA. ++ * ++ * Copyright (C) 2008-2009, VMware, Inc. All Rights Reserved. ++ * ++ * This program is free software; you can redistribute it and/or modify it ++ * under the terms of the GNU General Public License as published by the ++ * Free Software Foundation; version 2 of the License and no later version. ++ * ++ * This program is distributed in the hope that it will be useful, but ++ * WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or ++ * NON INFRINGEMENT. See the GNU General Public License for more ++ * details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. ++ * ++ * Maintained by: Alok N Kataria ++ * ++ */ ++ ++#include ++#include ++#include ++#include ++#include ++#include ++#include ++ ++#include ++#include ++#include ++#include ++ ++#include "vmw_pvscsi.h" ++ ++#define PVSCSI_LINUX_DRIVER_DESC "VMware PVSCSI driver" ++ ++MODULE_DESCRIPTION(PVSCSI_LINUX_DRIVER_DESC); ++MODULE_AUTHOR("VMware, Inc."); ++MODULE_LICENSE("GPL"); ++MODULE_VERSION(PVSCSI_DRIVER_VERSION_STRING); ++ ++#define PVSCSI_DEFAULT_NUM_PAGES_PER_RING 8 ++#define PVSCSI_DEFAULT_NUM_PAGES_MSG_RING 1 ++#define PVSCSI_DEFAULT_QUEUE_DEPTH 64 ++#define SGL_SIZE PAGE_SIZE ++ ++#define pvscsi_dev(adapter) (&(adapter->dev->dev)) ++ ++struct pvscsi_sg_list { ++ struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT]; ++}; ++ ++struct pvscsi_ctx { ++ /* ++ * The index of the context in cmd_map serves as the context ID for a ++ * 1-to-1 mapping completions back to requests. ++ */ ++ struct scsi_cmnd *cmd; ++ struct pvscsi_sg_list *sgl; ++ struct list_head list; ++ dma_addr_t dataPA; ++ dma_addr_t sensePA; ++ dma_addr_t sglPA; ++}; ++ ++struct pvscsi_adapter { ++ char *mmioBase; ++ unsigned int irq; ++ u8 rev; ++ bool use_msi; ++ bool use_msix; ++ bool use_msg; ++ ++ spinlock_t hw_lock; ++ ++ struct workqueue_struct *workqueue; ++ struct work_struct work; ++ ++ struct PVSCSIRingReqDesc *req_ring; ++ unsigned req_pages; ++ unsigned req_depth; ++ dma_addr_t reqRingPA; ++ ++ struct PVSCSIRingCmpDesc *cmp_ring; ++ unsigned cmp_pages; ++ dma_addr_t cmpRingPA; ++ ++ struct PVSCSIRingMsgDesc *msg_ring; ++ unsigned msg_pages; ++ dma_addr_t msgRingPA; ++ ++ struct PVSCSIRingsState *rings_state; ++ dma_addr_t ringStatePA; ++ ++ struct pci_dev *dev; ++ struct Scsi_Host *host; ++ ++ struct list_head cmd_pool; ++ struct pvscsi_ctx *cmd_map; ++}; ++ ++ ++/* Command line parameters */ ++static int pvscsi_ring_pages = PVSCSI_DEFAULT_NUM_PAGES_PER_RING; ++static int pvscsi_msg_ring_pages = PVSCSI_DEFAULT_NUM_PAGES_MSG_RING; ++static int pvscsi_cmd_per_lun = PVSCSI_DEFAULT_QUEUE_DEPTH; ++static bool pvscsi_disable_msi; ++static bool pvscsi_disable_msix; ++static bool pvscsi_use_msg = true; ++ ++#define PVSCSI_RW (S_IRUSR | S_IWUSR) ++ ++module_param_named(ring_pages, pvscsi_ring_pages, int, PVSCSI_RW); ++MODULE_PARM_DESC(ring_pages, "Number of pages per req/cmp ring - (default=" ++ __stringify(PVSCSI_DEFAULT_NUM_PAGES_PER_RING) ")"); ++ ++module_param_named(msg_ring_pages, pvscsi_msg_ring_pages, int, PVSCSI_RW); ++MODULE_PARM_DESC(msg_ring_pages, "Number of pages for the msg ring - (default=" ++ __stringify(PVSCSI_DEFAULT_NUM_PAGES_MSG_RING) ")"); ++ ++module_param_named(cmd_per_lun, pvscsi_cmd_per_lun, int, PVSCSI_RW); ++MODULE_PARM_DESC(cmd_per_lun, "Maximum commands per lun - (default=" ++ __stringify(PVSCSI_MAX_REQ_QUEUE_DEPTH) ")"); ++ ++module_param_named(disable_msi, pvscsi_disable_msi, bool, PVSCSI_RW); ++MODULE_PARM_DESC(disable_msi, "Disable MSI use in driver - (default=0)"); ++ ++module_param_named(disable_msix, pvscsi_disable_msix, bool, PVSCSI_RW); ++MODULE_PARM_DESC(disable_msix, "Disable MSI-X use in driver - (default=0)"); ++ ++module_param_named(use_msg, pvscsi_use_msg, bool, PVSCSI_RW); ++MODULE_PARM_DESC(use_msg, "Use msg ring when available - (default=1)"); ++ ++static const struct pci_device_id pvscsi_pci_tbl[] = { ++ { PCI_VDEVICE(VMWARE, PCI_DEVICE_ID_VMWARE_PVSCSI) }, ++ { 0 } ++}; ++ ++MODULE_DEVICE_TABLE(pci, pvscsi_pci_tbl); ++ ++static struct pvscsi_ctx * ++pvscsi_find_context(const struct pvscsi_adapter *adapter, struct scsi_cmnd *cmd) ++{ ++ struct pvscsi_ctx *ctx, *end; ++ ++ end = &adapter->cmd_map[adapter->req_depth]; ++ for (ctx = adapter->cmd_map; ctx < end; ctx++) ++ if (ctx->cmd == cmd) ++ return ctx; ++ ++ return NULL; ++} ++ ++static struct pvscsi_ctx * ++pvscsi_acquire_context(struct pvscsi_adapter *adapter, struct scsi_cmnd *cmd) ++{ ++ struct pvscsi_ctx *ctx; ++ ++ if (list_empty(&adapter->cmd_pool)) ++ return NULL; ++ ++ ctx = list_first_entry(&adapter->cmd_pool, struct pvscsi_ctx, list); ++ ctx->cmd = cmd; ++ list_del(&ctx->list); ++ ++ return ctx; ++} ++ ++static void pvscsi_release_context(struct pvscsi_adapter *adapter, ++ struct pvscsi_ctx *ctx) ++{ ++ ctx->cmd = NULL; ++ list_add(&ctx->list, &adapter->cmd_pool); ++} ++ ++/* ++ * Map a pvscsi_ctx struct to a context ID field value; we map to a simple ++ * non-zero integer. ctx always points to an entry in cmd_map array, hence ++ * the return value is always >=1. ++ */ ++static u64 pvscsi_map_context(const struct pvscsi_adapter *adapter, ++ const struct pvscsi_ctx *ctx) ++{ ++ return ctx - adapter->cmd_map + 1; ++} ++ ++static struct pvscsi_ctx * ++pvscsi_get_context(const struct pvscsi_adapter *adapter, u64 context) ++{ ++ return &adapter->cmd_map[context - 1]; ++} ++ ++static void pvscsi_reg_write(const struct pvscsi_adapter *adapter, ++ u32 offset, u32 val) ++{ ++ writel(val, adapter->mmioBase + offset); ++} ++ ++static u32 pvscsi_reg_read(const struct pvscsi_adapter *adapter, u32 offset) ++{ ++ return readl(adapter->mmioBase + offset); ++} ++ ++static u32 pvscsi_read_intr_status(const struct pvscsi_adapter *adapter) ++{ ++ return pvscsi_reg_read(adapter, PVSCSI_REG_OFFSET_INTR_STATUS); ++} ++ ++static void pvscsi_write_intr_status(const struct pvscsi_adapter *adapter, ++ u32 val) ++{ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_INTR_STATUS, val); ++} ++ ++static void pvscsi_unmask_intr(const struct pvscsi_adapter *adapter) ++{ ++ u32 intr_bits; ++ ++ intr_bits = PVSCSI_INTR_CMPL_MASK; ++ if (adapter->use_msg) ++ intr_bits |= PVSCSI_INTR_MSG_MASK; ++ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_INTR_MASK, intr_bits); ++} ++ ++static void pvscsi_mask_intr(const struct pvscsi_adapter *adapter) ++{ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_INTR_MASK, 0); ++} ++ ++static void pvscsi_write_cmd_desc(const struct pvscsi_adapter *adapter, ++ u32 cmd, const void *desc, size_t len) ++{ ++ const u32 *ptr = desc; ++ size_t i; ++ ++ len /= sizeof(*ptr); ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_COMMAND, cmd); ++ for (i = 0; i < len; i++) ++ pvscsi_reg_write(adapter, ++ PVSCSI_REG_OFFSET_COMMAND_DATA, ptr[i]); ++} ++ ++static void pvscsi_abort_cmd(const struct pvscsi_adapter *adapter, ++ const struct pvscsi_ctx *ctx) ++{ ++ struct PVSCSICmdDescAbortCmd cmd = { 0 }; ++ ++ cmd.target = ctx->cmd->device->id; ++ cmd.context = pvscsi_map_context(adapter, ctx); ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_ABORT_CMD, &cmd, sizeof(cmd)); ++} ++ ++static void pvscsi_kick_rw_io(const struct pvscsi_adapter *adapter) ++{ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_KICK_RW_IO, 0); ++} ++ ++static void pvscsi_process_request_ring(const struct pvscsi_adapter *adapter) ++{ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_KICK_NON_RW_IO, 0); ++} ++ ++static int scsi_is_rw(unsigned char op) ++{ ++ return op == READ_6 || op == WRITE_6 || ++ op == READ_10 || op == WRITE_10 || ++ op == READ_12 || op == WRITE_12 || ++ op == READ_16 || op == WRITE_16; ++} ++ ++static void pvscsi_kick_io(const struct pvscsi_adapter *adapter, ++ unsigned char op) ++{ ++ if (scsi_is_rw(op)) ++ pvscsi_kick_rw_io(adapter); ++ else ++ pvscsi_process_request_ring(adapter); ++} ++ ++static void ll_adapter_reset(const struct pvscsi_adapter *adapter) ++{ ++ dev_dbg(pvscsi_dev(adapter), "Adapter Reset on %p\n", adapter); ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_ADAPTER_RESET, NULL, 0); ++} ++ ++static void ll_bus_reset(const struct pvscsi_adapter *adapter) ++{ ++ dev_dbg(pvscsi_dev(adapter), "Reseting bus on %p\n", adapter); ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_RESET_BUS, NULL, 0); ++} ++ ++static void ll_device_reset(const struct pvscsi_adapter *adapter, u32 target) ++{ ++ struct PVSCSICmdDescResetDevice cmd = { 0 }; ++ ++ dev_dbg(pvscsi_dev(adapter), "Reseting device: target=%u\n", target); ++ ++ cmd.target = target; ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_RESET_DEVICE, ++ &cmd, sizeof(cmd)); ++} ++ ++static void pvscsi_create_sg(struct pvscsi_ctx *ctx, ++ struct scatterlist *sg, unsigned count) ++{ ++ unsigned i; ++ struct PVSCSISGElement *sge; ++ ++ BUG_ON(count > PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT); ++ ++ sge = &ctx->sgl->sge[0]; ++ for (i = 0; i < count; i++, sg++) { ++ sge[i].addr = sg_dma_address(sg); ++ sge[i].length = sg_dma_len(sg); ++ sge[i].flags = 0; ++ } ++} ++ ++/* ++ * Map all data buffers for a command into PCI space and ++ * setup the scatter/gather list if needed. ++ */ ++static void pvscsi_map_buffers(struct pvscsi_adapter *adapter, ++ struct pvscsi_ctx *ctx, struct scsi_cmnd *cmd, ++ struct PVSCSIRingReqDesc *e) ++{ ++ unsigned count; ++ unsigned bufflen = scsi_bufflen(cmd); ++ struct scatterlist *sg; ++ ++ e->dataLen = bufflen; ++ e->dataAddr = 0; ++ if (bufflen == 0) ++ return; ++ ++ sg = scsi_sglist(cmd); ++ count = scsi_sg_count(cmd); ++ if (count != 0) { ++ int segs = scsi_dma_map(cmd); ++ if (segs > 1) { ++ pvscsi_create_sg(ctx, sg, segs); ++ ++ e->flags |= PVSCSI_FLAG_CMD_WITH_SG_LIST; ++ ctx->sglPA = pci_map_single(adapter->dev, ctx->sgl, ++ SGL_SIZE, PCI_DMA_TODEVICE); ++ e->dataAddr = ctx->sglPA; ++ } else ++ e->dataAddr = sg_dma_address(sg); ++ } else { ++ /* ++ * In case there is no S/G list, scsi_sglist points ++ * directly to the buffer. ++ */ ++ ctx->dataPA = pci_map_single(adapter->dev, sg, bufflen, ++ cmd->sc_data_direction); ++ e->dataAddr = ctx->dataPA; ++ } ++} ++ ++static void pvscsi_unmap_buffers(const struct pvscsi_adapter *adapter, ++ struct pvscsi_ctx *ctx) ++{ ++ struct scsi_cmnd *cmd; ++ unsigned bufflen; ++ ++ cmd = ctx->cmd; ++ bufflen = scsi_bufflen(cmd); ++ ++ if (bufflen != 0) { ++ unsigned count = scsi_sg_count(cmd); ++ ++ if (count != 0) { ++ scsi_dma_unmap(cmd); ++ if (ctx->sglPA) { ++ pci_unmap_single(adapter->dev, ctx->sglPA, ++ SGL_SIZE, PCI_DMA_TODEVICE); ++ ctx->sglPA = 0; ++ } ++ } else ++ pci_unmap_single(adapter->dev, ctx->dataPA, bufflen, ++ cmd->sc_data_direction); ++ } ++ if (cmd->sense_buffer) ++ pci_unmap_single(adapter->dev, ctx->sensePA, ++ SCSI_SENSE_BUFFERSIZE, PCI_DMA_FROMDEVICE); ++} ++ ++static int __devinit pvscsi_allocate_rings(struct pvscsi_adapter *adapter) ++{ ++ adapter->rings_state = pci_alloc_consistent(adapter->dev, PAGE_SIZE, ++ &adapter->ringStatePA); ++ if (!adapter->rings_state) ++ return -ENOMEM; ++ ++ adapter->req_pages = min(PVSCSI_MAX_NUM_PAGES_REQ_RING, ++ pvscsi_ring_pages); ++ adapter->req_depth = adapter->req_pages ++ * PVSCSI_MAX_NUM_REQ_ENTRIES_PER_PAGE; ++ adapter->req_ring = pci_alloc_consistent(adapter->dev, ++ adapter->req_pages * PAGE_SIZE, ++ &adapter->reqRingPA); ++ if (!adapter->req_ring) ++ return -ENOMEM; ++ ++ adapter->cmp_pages = min(PVSCSI_MAX_NUM_PAGES_CMP_RING, ++ pvscsi_ring_pages); ++ adapter->cmp_ring = pci_alloc_consistent(adapter->dev, ++ adapter->cmp_pages * PAGE_SIZE, ++ &adapter->cmpRingPA); ++ if (!adapter->cmp_ring) ++ return -ENOMEM; ++ ++ BUG_ON(!IS_ALIGNED(adapter->ringStatePA, PAGE_SIZE)); ++ BUG_ON(!IS_ALIGNED(adapter->reqRingPA, PAGE_SIZE)); ++ BUG_ON(!IS_ALIGNED(adapter->cmpRingPA, PAGE_SIZE)); ++ ++ if (!adapter->use_msg) ++ return 0; ++ ++ adapter->msg_pages = min(PVSCSI_MAX_NUM_PAGES_MSG_RING, ++ pvscsi_msg_ring_pages); ++ adapter->msg_ring = pci_alloc_consistent(adapter->dev, ++ adapter->msg_pages * PAGE_SIZE, ++ &adapter->msgRingPA); ++ if (!adapter->msg_ring) ++ return -ENOMEM; ++ BUG_ON(!IS_ALIGNED(adapter->msgRingPA, PAGE_SIZE)); ++ ++ return 0; ++} ++ ++static void pvscsi_setup_all_rings(const struct pvscsi_adapter *adapter) ++{ ++ struct PVSCSICmdDescSetupRings cmd = { 0 }; ++ dma_addr_t base; ++ unsigned i; ++ ++ cmd.ringsStatePPN = adapter->ringStatePA >> PAGE_SHIFT; ++ cmd.reqRingNumPages = adapter->req_pages; ++ cmd.cmpRingNumPages = adapter->cmp_pages; ++ ++ base = adapter->reqRingPA; ++ for (i = 0; i < adapter->req_pages; i++) { ++ cmd.reqRingPPNs[i] = base >> PAGE_SHIFT; ++ base += PAGE_SIZE; ++ } ++ ++ base = adapter->cmpRingPA; ++ for (i = 0; i < adapter->cmp_pages; i++) { ++ cmd.cmpRingPPNs[i] = base >> PAGE_SHIFT; ++ base += PAGE_SIZE; ++ } ++ ++ memset(adapter->rings_state, 0, PAGE_SIZE); ++ memset(adapter->req_ring, 0, adapter->req_pages * PAGE_SIZE); ++ memset(adapter->cmp_ring, 0, adapter->cmp_pages * PAGE_SIZE); ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_SETUP_RINGS, ++ &cmd, sizeof(cmd)); ++ ++ if (adapter->use_msg) { ++ struct PVSCSICmdDescSetupMsgRing cmd_msg = { 0 }; ++ ++ cmd_msg.numPages = adapter->msg_pages; ++ ++ base = adapter->msgRingPA; ++ for (i = 0; i < adapter->msg_pages; i++) { ++ cmd_msg.ringPPNs[i] = base >> PAGE_SHIFT; ++ base += PAGE_SIZE; ++ } ++ memset(adapter->msg_ring, 0, adapter->msg_pages * PAGE_SIZE); ++ ++ pvscsi_write_cmd_desc(adapter, PVSCSI_CMD_SETUP_MSG_RING, ++ &cmd_msg, sizeof(cmd_msg)); ++ } ++} ++ ++/* ++ * Pull a completion descriptor off and pass the completion back ++ * to the SCSI mid layer. ++ */ ++static void pvscsi_complete_request(struct pvscsi_adapter *adapter, ++ const struct PVSCSIRingCmpDesc *e) ++{ ++ struct pvscsi_ctx *ctx; ++ struct scsi_cmnd *cmd; ++ u32 btstat = e->hostStatus; ++ u32 sdstat = e->scsiStatus; ++ ++ ctx = pvscsi_get_context(adapter, e->context); ++ cmd = ctx->cmd; ++ pvscsi_unmap_buffers(adapter, ctx); ++ pvscsi_release_context(adapter, ctx); ++ cmd->result = 0; ++ ++ if (sdstat != SAM_STAT_GOOD && ++ (btstat == BTSTAT_SUCCESS || ++ btstat == BTSTAT_LINKED_COMMAND_COMPLETED || ++ btstat == BTSTAT_LINKED_COMMAND_COMPLETED_WITH_FLAG)) { ++ cmd->result = (DID_OK << 16) | sdstat; ++ if (sdstat == SAM_STAT_CHECK_CONDITION && cmd->sense_buffer) ++ cmd->result |= (DRIVER_SENSE << 24); ++ } else ++ switch (btstat) { ++ case BTSTAT_SUCCESS: ++ case BTSTAT_LINKED_COMMAND_COMPLETED: ++ case BTSTAT_LINKED_COMMAND_COMPLETED_WITH_FLAG: ++ /* If everything went fine, let's move on.. */ ++ cmd->result = (DID_OK << 16); ++ break; ++ ++ case BTSTAT_DATARUN: ++ case BTSTAT_DATA_UNDERRUN: ++ /* Report residual data in underruns */ ++ scsi_set_resid(cmd, scsi_bufflen(cmd) - e->dataLen); ++ cmd->result = (DID_ERROR << 16); ++ break; ++ ++ case BTSTAT_SELTIMEO: ++ /* Our emulation returns this for non-connected devs */ ++ cmd->result = (DID_BAD_TARGET << 16); ++ break; ++ ++ case BTSTAT_LUNMISMATCH: ++ case BTSTAT_TAGREJECT: ++ case BTSTAT_BADMSG: ++ cmd->result = (DRIVER_INVALID << 24); ++ /* fall through */ ++ ++ case BTSTAT_HAHARDWARE: ++ case BTSTAT_INVPHASE: ++ case BTSTAT_HATIMEOUT: ++ case BTSTAT_NORESPONSE: ++ case BTSTAT_DISCONNECT: ++ case BTSTAT_HASOFTWARE: ++ case BTSTAT_BUSFREE: ++ case BTSTAT_SENSFAILED: ++ cmd->result |= (DID_ERROR << 16); ++ break; ++ ++ case BTSTAT_SENTRST: ++ case BTSTAT_RECVRST: ++ case BTSTAT_BUSRESET: ++ cmd->result = (DID_RESET << 16); ++ break; ++ ++ case BTSTAT_ABORTQUEUE: ++ cmd->result = (DID_ABORT << 16); ++ break; ++ ++ case BTSTAT_SCSIPARITY: ++ cmd->result = (DID_PARITY << 16); ++ break; ++ ++ default: ++ cmd->result = (DID_ERROR << 16); ++ scmd_printk(KERN_DEBUG, cmd, ++ "Unknown completion status: 0x%x\n", ++ btstat); ++ } ++ ++ dev_dbg(&cmd->device->sdev_gendev, ++ "cmd=%p %x ctx=%p result=0x%x status=0x%x,%x\n", ++ cmd, cmd->cmnd[0], ctx, cmd->result, btstat, sdstat); ++ ++ cmd->scsi_done(cmd); ++} ++ ++/* ++ * barrier usage : Since the PVSCSI device is emulated, there could be cases ++ * where we may want to serialize some accesses between the driver and the ++ * emulation layer. We use compiler barriers instead of the more expensive ++ * memory barriers because PVSCSI is only supported on X86 which has strong ++ * memory access ordering. ++ */ ++static void pvscsi_process_completion_ring(struct pvscsi_adapter *adapter) ++{ ++ struct PVSCSIRingsState *s = adapter->rings_state; ++ struct PVSCSIRingCmpDesc *ring = adapter->cmp_ring; ++ u32 cmp_entries = s->cmpNumEntriesLog2; ++ ++ while (s->cmpConsIdx != s->cmpProdIdx) { ++ struct PVSCSIRingCmpDesc *e = ring + (s->cmpConsIdx & ++ MASK(cmp_entries)); ++ /* ++ * This barrier() ensures that *e is not dereferenced while ++ * the device emulation still writes data into the slot. ++ * Since the device emulation advances s->cmpProdIdx only after ++ * updating the slot we want to check it first. ++ */ ++ barrier(); ++ pvscsi_complete_request(adapter, e); ++ /* ++ * This barrier() ensures that compiler doesn't reorder write ++ * to s->cmpConsIdx before the read of (*e) inside ++ * pvscsi_complete_request. Otherwise, device emulation may ++ * overwrite *e before we had a chance to read it. ++ */ ++ barrier(); ++ s->cmpConsIdx++; ++ } ++} ++ ++/* ++ * Translate a Linux SCSI request into a request ring entry. ++ */ ++static int pvscsi_queue_ring(struct pvscsi_adapter *adapter, ++ struct pvscsi_ctx *ctx, struct scsi_cmnd *cmd) ++{ ++ struct PVSCSIRingsState *s; ++ struct PVSCSIRingReqDesc *e; ++ struct scsi_device *sdev; ++ u32 req_entries; ++ ++ s = adapter->rings_state; ++ sdev = cmd->device; ++ req_entries = s->reqNumEntriesLog2; ++ ++ /* ++ * If this condition holds, we might have room on the request ring, but ++ * we might not have room on the completion ring for the response. ++ * However, we have already ruled out this possibility - we would not ++ * have successfully allocated a context if it were true, since we only ++ * have one context per request entry. Check for it anyway, since it ++ * would be a serious bug. ++ */ ++ if (s->reqProdIdx - s->cmpConsIdx >= 1 << req_entries) { ++ scmd_printk(KERN_ERR, cmd, "vmw_pvscsi: " ++ "ring full: reqProdIdx=%d cmpConsIdx=%d\n", ++ s->reqProdIdx, s->cmpConsIdx); ++ return -1; ++ } ++ ++ e = adapter->req_ring + (s->reqProdIdx & MASK(req_entries)); ++ ++ e->bus = sdev->channel; ++ e->target = sdev->id; ++ memset(e->lun, 0, sizeof(e->lun)); ++ e->lun[1] = sdev->lun; ++ ++ if (cmd->sense_buffer) { ++ ctx->sensePA = pci_map_single(adapter->dev, cmd->sense_buffer, ++ SCSI_SENSE_BUFFERSIZE, ++ PCI_DMA_FROMDEVICE); ++ e->senseAddr = ctx->sensePA; ++ e->senseLen = SCSI_SENSE_BUFFERSIZE; ++ } else { ++ e->senseLen = 0; ++ e->senseAddr = 0; ++ } ++ e->cdbLen = cmd->cmd_len; ++ e->vcpuHint = smp_processor_id(); ++ memcpy(e->cdb, cmd->cmnd, e->cdbLen); ++ ++ e->tag = SIMPLE_QUEUE_TAG; ++ if (sdev->tagged_supported && ++ (cmd->tag == HEAD_OF_QUEUE_TAG || ++ cmd->tag == ORDERED_QUEUE_TAG)) ++ e->tag = cmd->tag; ++ ++ if (cmd->sc_data_direction == DMA_FROM_DEVICE) ++ e->flags = PVSCSI_FLAG_CMD_DIR_TOHOST; ++ else if (cmd->sc_data_direction == DMA_TO_DEVICE) ++ e->flags = PVSCSI_FLAG_CMD_DIR_TODEVICE; ++ else if (cmd->sc_data_direction == DMA_NONE) ++ e->flags = PVSCSI_FLAG_CMD_DIR_NONE; ++ else ++ e->flags = 0; ++ ++ pvscsi_map_buffers(adapter, ctx, cmd, e); ++ ++ e->context = pvscsi_map_context(adapter, ctx); ++ ++ barrier(); ++ ++ s->reqProdIdx++; ++ ++ return 0; ++} ++ ++static int pvscsi_queue(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *)) ++{ ++ struct Scsi_Host *host = cmd->device->host; ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ struct pvscsi_ctx *ctx; ++ unsigned long flags; ++ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ ctx = pvscsi_acquire_context(adapter, cmd); ++ if (!ctx || pvscsi_queue_ring(adapter, ctx, cmd) != 0) { ++ if (ctx) ++ pvscsi_release_context(adapter, ctx); ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ return SCSI_MLQUEUE_HOST_BUSY; ++ } ++ ++ cmd->scsi_done = done; ++ ++ dev_dbg(&cmd->device->sdev_gendev, ++ "queued cmd %p, ctx %p, op=%x\n", cmd, ctx, cmd->cmnd[0]); ++ ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ ++ pvscsi_kick_io(adapter, cmd->cmnd[0]); ++ ++ return 0; ++} ++ ++static int pvscsi_abort(struct scsi_cmnd *cmd) ++{ ++ struct pvscsi_adapter *adapter = shost_priv(cmd->device->host); ++ struct pvscsi_ctx *ctx; ++ unsigned long flags; ++ ++ scmd_printk(KERN_DEBUG, cmd, "task abort on host %u, %p\n", ++ adapter->host->host_no, cmd); ++ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ /* ++ * Poll the completion ring first - we might be trying to abort ++ * a command that is waiting to be dispatched in the completion ring. ++ */ ++ pvscsi_process_completion_ring(adapter); ++ ++ /* ++ * If there is no context for the command, it either already succeeded ++ * or else was never properly issued. Not our problem. ++ */ ++ ctx = pvscsi_find_context(adapter, cmd); ++ if (!ctx) { ++ scmd_printk(KERN_DEBUG, cmd, "Failed to abort cmd %p\n", cmd); ++ goto out; ++ } ++ ++ pvscsi_abort_cmd(adapter, ctx); ++ ++ pvscsi_process_completion_ring(adapter); ++ ++out: ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ return SUCCESS; ++} ++ ++/* ++ * Abort all outstanding requests. This is only safe to use if the completion ++ * ring will never be walked again or the device has been reset, because it ++ * destroys the 1-1 mapping between context field passed to emulation and our ++ * request structure. ++ */ ++static void pvscsi_reset_all(struct pvscsi_adapter *adapter) ++{ ++ unsigned i; ++ ++ for (i = 0; i < adapter->req_depth; i++) { ++ struct pvscsi_ctx *ctx = &adapter->cmd_map[i]; ++ struct scsi_cmnd *cmd = ctx->cmd; ++ if (cmd) { ++ scmd_printk(KERN_ERR, cmd, ++ "Forced reset on cmd %p\n", cmd); ++ pvscsi_unmap_buffers(adapter, ctx); ++ pvscsi_release_context(adapter, ctx); ++ cmd->result = (DID_RESET << 16); ++ cmd->scsi_done(cmd); ++ } ++ } ++} ++ ++static int pvscsi_host_reset(struct scsi_cmnd *cmd) ++{ ++ struct Scsi_Host *host = cmd->device->host; ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ unsigned long flags; ++ bool use_msg; ++ ++ scmd_printk(KERN_INFO, cmd, "SCSI Host reset\n"); ++ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ use_msg = adapter->use_msg; ++ ++ if (use_msg) { ++ adapter->use_msg = 0; ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ ++ /* ++ * Now that we know that the ISR won't add more work on the ++ * workqueue we can safely flush any outstanding work. ++ */ ++ flush_workqueue(adapter->workqueue); ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ } ++ ++ /* ++ * We're going to tear down the entire ring structure and set it back ++ * up, so stalling new requests until all completions are flushed and ++ * the rings are back in place. ++ */ ++ ++ pvscsi_process_request_ring(adapter); ++ ++ ll_adapter_reset(adapter); ++ ++ /* ++ * Now process any completions. Note we do this AFTER adapter reset, ++ * which is strange, but stops races where completions get posted ++ * between processing the ring and issuing the reset. The backend will ++ * not touch the ring memory after reset, so the immediately pre-reset ++ * completion ring state is still valid. ++ */ ++ pvscsi_process_completion_ring(adapter); ++ ++ pvscsi_reset_all(adapter); ++ adapter->use_msg = use_msg; ++ pvscsi_setup_all_rings(adapter); ++ pvscsi_unmask_intr(adapter); ++ ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ ++ return SUCCESS; ++} ++ ++static int pvscsi_bus_reset(struct scsi_cmnd *cmd) ++{ ++ struct Scsi_Host *host = cmd->device->host; ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ unsigned long flags; ++ ++ scmd_printk(KERN_INFO, cmd, "SCSI Bus reset\n"); ++ ++ /* ++ * We don't want to queue new requests for this bus after ++ * flushing all pending requests to emulation, since new ++ * requests could then sneak in during this bus reset phase, ++ * so take the lock now. ++ */ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ pvscsi_process_request_ring(adapter); ++ ll_bus_reset(adapter); ++ pvscsi_process_completion_ring(adapter); ++ ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ ++ return SUCCESS; ++} ++ ++static int pvscsi_device_reset(struct scsi_cmnd *cmd) ++{ ++ struct Scsi_Host *host = cmd->device->host; ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ unsigned long flags; ++ ++ scmd_printk(KERN_INFO, cmd, "SCSI device reset on scsi%u:%u\n", ++ host->host_no, cmd->device->id); ++ ++ /* ++ * We don't want to queue new requests for this device after flushing ++ * all pending requests to emulation, since new requests could then ++ * sneak in during this device reset phase, so take the lock now. ++ */ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ pvscsi_process_request_ring(adapter); ++ ll_device_reset(adapter, cmd->device->id); ++ pvscsi_process_completion_ring(adapter); ++ ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ ++ return SUCCESS; ++} ++ ++static struct scsi_host_template pvscsi_template; ++ ++static const char *pvscsi_info(struct Scsi_Host *host) ++{ ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ static char buf[256]; ++ ++ sprintf(buf, "VMware PVSCSI storage adapter rev %d, req/cmp/msg rings: " ++ "%u/%u/%u pages, cmd_per_lun=%u", adapter->rev, ++ adapter->req_pages, adapter->cmp_pages, adapter->msg_pages, ++ pvscsi_template.cmd_per_lun); ++ ++ return buf; ++} ++ ++static struct scsi_host_template pvscsi_template = { ++ .module = THIS_MODULE, ++ .name = "VMware PVSCSI Host Adapter", ++ .proc_name = "vmw_pvscsi", ++ .info = pvscsi_info, ++ .queuecommand = pvscsi_queue, ++ .this_id = -1, ++ .sg_tablesize = PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT, ++ .dma_boundary = UINT_MAX, ++ .max_sectors = 0xffff, ++ .use_clustering = ENABLE_CLUSTERING, ++ .eh_abort_handler = pvscsi_abort, ++ .eh_device_reset_handler = pvscsi_device_reset, ++ .eh_bus_reset_handler = pvscsi_bus_reset, ++ .eh_host_reset_handler = pvscsi_host_reset, ++}; ++ ++static void pvscsi_process_msg(const struct pvscsi_adapter *adapter, ++ const struct PVSCSIRingMsgDesc *e) ++{ ++ struct PVSCSIRingsState *s = adapter->rings_state; ++ struct Scsi_Host *host = adapter->host; ++ struct scsi_device *sdev; ++ ++ printk(KERN_INFO "vmw_pvscsi: msg type: 0x%x - MSG RING: %u/%u (%u) \n", ++ e->type, s->msgProdIdx, s->msgConsIdx, s->msgNumEntriesLog2); ++ ++ BUILD_BUG_ON(PVSCSI_MSG_LAST != 2); ++ ++ if (e->type == PVSCSI_MSG_DEV_ADDED) { ++ struct PVSCSIMsgDescDevStatusChanged *desc; ++ desc = (struct PVSCSIMsgDescDevStatusChanged *)e; ++ ++ printk(KERN_INFO ++ "vmw_pvscsi: msg: device added at scsi%u:%u:%u\n", ++ desc->bus, desc->target, desc->lun[1]); ++ ++ if (!scsi_host_get(host)) ++ return; ++ ++ sdev = scsi_device_lookup(host, desc->bus, desc->target, ++ desc->lun[1]); ++ if (sdev) { ++ printk(KERN_INFO "vmw_pvscsi: device already exists\n"); ++ scsi_device_put(sdev); ++ } else ++ scsi_add_device(adapter->host, desc->bus, ++ desc->target, desc->lun[1]); ++ ++ scsi_host_put(host); ++ } else if (e->type == PVSCSI_MSG_DEV_REMOVED) { ++ struct PVSCSIMsgDescDevStatusChanged *desc; ++ desc = (struct PVSCSIMsgDescDevStatusChanged *)e; ++ ++ printk(KERN_INFO ++ "vmw_pvscsi: msg: device removed at scsi%u:%u:%u\n", ++ desc->bus, desc->target, desc->lun[1]); ++ ++ if (!scsi_host_get(host)) ++ return; ++ ++ sdev = scsi_device_lookup(host, desc->bus, desc->target, ++ desc->lun[1]); ++ if (sdev) { ++ scsi_remove_device(sdev); ++ scsi_device_put(sdev); ++ } else ++ printk(KERN_INFO ++ "vmw_pvscsi: failed to lookup scsi%u:%u:%u\n", ++ desc->bus, desc->target, desc->lun[1]); ++ ++ scsi_host_put(host); ++ } ++} ++ ++static int pvscsi_msg_pending(const struct pvscsi_adapter *adapter) ++{ ++ struct PVSCSIRingsState *s = adapter->rings_state; ++ ++ return s->msgProdIdx != s->msgConsIdx; ++} ++ ++static void pvscsi_process_msg_ring(const struct pvscsi_adapter *adapter) ++{ ++ struct PVSCSIRingsState *s = adapter->rings_state; ++ struct PVSCSIRingMsgDesc *ring = adapter->msg_ring; ++ u32 msg_entries = s->msgNumEntriesLog2; ++ ++ while (pvscsi_msg_pending(adapter)) { ++ struct PVSCSIRingMsgDesc *e = ring + (s->msgConsIdx & ++ MASK(msg_entries)); ++ ++ barrier(); ++ pvscsi_process_msg(adapter, e); ++ barrier(); ++ s->msgConsIdx++; ++ } ++} ++ ++static void pvscsi_msg_workqueue_handler(struct work_struct *data) ++{ ++ struct pvscsi_adapter *adapter; ++ ++ adapter = container_of(data, struct pvscsi_adapter, work); ++ ++ pvscsi_process_msg_ring(adapter); ++} ++ ++static int pvscsi_setup_msg_workqueue(struct pvscsi_adapter *adapter) ++{ ++ char name[32]; ++ ++ if (!pvscsi_use_msg) ++ return 0; ++ ++ pvscsi_reg_write(adapter, PVSCSI_REG_OFFSET_COMMAND, ++ PVSCSI_CMD_SETUP_MSG_RING); ++ ++ if (pvscsi_reg_read(adapter, PVSCSI_REG_OFFSET_COMMAND_STATUS) == -1) ++ return 0; ++ ++ snprintf(name, sizeof(name), ++ "vmw_pvscsi_wq_%u", adapter->host->host_no); ++ ++ adapter->workqueue = create_singlethread_workqueue(name); ++ if (!adapter->workqueue) { ++ printk(KERN_ERR "vmw_pvscsi: failed to create work queue\n"); ++ return 0; ++ } ++ INIT_WORK(&adapter->work, pvscsi_msg_workqueue_handler); ++ ++ return 1; ++} ++ ++static irqreturn_t pvscsi_isr(int irq, void *devp) ++{ ++ struct pvscsi_adapter *adapter = devp; ++ int handled; ++ ++ if (adapter->use_msi || adapter->use_msix) ++ handled = true; ++ else { ++ u32 val = pvscsi_read_intr_status(adapter); ++ handled = (val & PVSCSI_INTR_ALL_SUPPORTED) != 0; ++ if (handled) ++ pvscsi_write_intr_status(devp, val); ++ } ++ ++ if (handled) { ++ unsigned long flags; ++ ++ spin_lock_irqsave(&adapter->hw_lock, flags); ++ ++ pvscsi_process_completion_ring(adapter); ++ if (adapter->use_msg && pvscsi_msg_pending(adapter)) ++ queue_work(adapter->workqueue, &adapter->work); ++ ++ spin_unlock_irqrestore(&adapter->hw_lock, flags); ++ } ++ ++ return IRQ_RETVAL(handled); ++} ++ ++static void pvscsi_free_sgls(const struct pvscsi_adapter *adapter) ++{ ++ struct pvscsi_ctx *ctx = adapter->cmd_map; ++ unsigned i; ++ ++ for (i = 0; i < adapter->req_depth; ++i, ++ctx) ++ kfree(ctx->sgl); ++} ++ ++static int pvscsi_setup_msix(const struct pvscsi_adapter *adapter, int *irq) ++{ ++ struct msix_entry entry = { 0, PVSCSI_VECTOR_COMPLETION }; ++ int ret; ++ ++ ret = pci_enable_msix(adapter->dev, &entry, 1); ++ if (ret) ++ return ret; ++ ++ *irq = entry.vector; ++ ++ return 0; ++} ++ ++static void pvscsi_shutdown_intr(struct pvscsi_adapter *adapter) ++{ ++ if (adapter->irq) { ++ free_irq(adapter->irq, adapter); ++ adapter->irq = 0; ++ } ++ if (adapter->use_msi) { ++ pci_disable_msi(adapter->dev); ++ adapter->use_msi = 0; ++ } else if (adapter->use_msix) { ++ pci_disable_msix(adapter->dev); ++ adapter->use_msix = 0; ++ } ++} ++ ++static void pvscsi_release_resources(struct pvscsi_adapter *adapter) ++{ ++ pvscsi_shutdown_intr(adapter); ++ ++ if (adapter->workqueue) ++ destroy_workqueue(adapter->workqueue); ++ ++ if (adapter->mmioBase) ++ pci_iounmap(adapter->dev, adapter->mmioBase); ++ ++ pci_release_regions(adapter->dev); ++ ++ if (adapter->cmd_map) { ++ pvscsi_free_sgls(adapter); ++ kfree(adapter->cmd_map); ++ } ++ ++ if (adapter->rings_state) ++ pci_free_consistent(adapter->dev, PAGE_SIZE, ++ adapter->rings_state, adapter->ringStatePA); ++ ++ if (adapter->req_ring) ++ pci_free_consistent(adapter->dev, ++ adapter->req_pages * PAGE_SIZE, ++ adapter->req_ring, adapter->reqRingPA); ++ ++ if (adapter->cmp_ring) ++ pci_free_consistent(adapter->dev, ++ adapter->cmp_pages * PAGE_SIZE, ++ adapter->cmp_ring, adapter->cmpRingPA); ++ ++ if (adapter->msg_ring) ++ pci_free_consistent(adapter->dev, ++ adapter->msg_pages * PAGE_SIZE, ++ adapter->msg_ring, adapter->msgRingPA); ++} ++ ++/* ++ * Allocate scatter gather lists. ++ * ++ * These are statically allocated. Trying to be clever was not worth it. ++ * ++ * Dynamic allocation can fail, and we can't go deeep into the memory ++ * allocator, since we're a SCSI driver, and trying too hard to allocate ++ * memory might generate disk I/O. We also don't want to fail disk I/O ++ * in that case because we can't get an allocation - the I/O could be ++ * trying to swap out data to free memory. Since that is pathological, ++ * just use a statically allocated scatter list. ++ * ++ */ ++static int __devinit pvscsi_allocate_sg(struct pvscsi_adapter *adapter) ++{ ++ struct pvscsi_ctx *ctx; ++ int i; ++ ++ ctx = adapter->cmd_map; ++ BUILD_BUG_ON(sizeof(struct pvscsi_sg_list) > SGL_SIZE); ++ ++ for (i = 0; i < adapter->req_depth; ++i, ++ctx) { ++ ctx->sgl = kmalloc(SGL_SIZE, GFP_KERNEL); ++ ctx->sglPA = 0; ++ BUG_ON(!IS_ALIGNED(((unsigned long)ctx->sgl), PAGE_SIZE)); ++ if (!ctx->sgl) { ++ for (; i >= 0; --i, --ctx) { ++ kfree(ctx->sgl); ++ ctx->sgl = NULL; ++ } ++ return -ENOMEM; ++ } ++ } ++ ++ return 0; ++} ++ ++static int __devinit pvscsi_probe(struct pci_dev *pdev, ++ const struct pci_device_id *id) ++{ ++ struct pvscsi_adapter *adapter; ++ struct Scsi_Host *host; ++ unsigned int i; ++ int error; ++ ++ error = -ENODEV; ++ ++ if (pci_enable_device(pdev)) ++ return error; ++ ++ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) == 0 && ++ pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)) == 0) { ++ printk(KERN_INFO "vmw_pvscsi: using 64bit dma\n"); ++ } else if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32)) == 0 && ++ pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) == 0) { ++ printk(KERN_INFO "vmw_pvscsi: using 32bit dma\n"); ++ } else { ++ printk(KERN_ERR "vmw_pvscsi: failed to set DMA mask\n"); ++ goto out_disable_device; ++ } ++ ++ pvscsi_template.can_queue = ++ min(PVSCSI_MAX_NUM_PAGES_REQ_RING, pvscsi_ring_pages) * ++ PVSCSI_MAX_NUM_REQ_ENTRIES_PER_PAGE; ++ pvscsi_template.cmd_per_lun = ++ min(pvscsi_template.can_queue, pvscsi_cmd_per_lun); ++ host = scsi_host_alloc(&pvscsi_template, sizeof(struct pvscsi_adapter)); ++ if (!host) { ++ printk(KERN_ERR "vmw_pvscsi: failed to allocate host\n"); ++ goto out_disable_device; ++ } ++ ++ adapter = shost_priv(host); ++ memset(adapter, 0, sizeof(*adapter)); ++ adapter->dev = pdev; ++ adapter->host = host; ++ ++ spin_lock_init(&adapter->hw_lock); ++ ++ host->max_channel = 0; ++ host->max_id = 16; ++ host->max_lun = 1; ++ host->max_cmd_len = 16; ++ ++ adapter->rev = pdev->revision; ++ ++ if (pci_request_regions(pdev, "vmw_pvscsi")) { ++ printk(KERN_ERR "vmw_pvscsi: pci memory selection failed\n"); ++ goto out_free_host; ++ } ++ ++ for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) { ++ if ((pci_resource_flags(pdev, i) & PCI_BASE_ADDRESS_SPACE_IO)) ++ continue; ++ ++ if (pci_resource_len(pdev, i) < PVSCSI_MEM_SPACE_SIZE) ++ continue; ++ ++ break; ++ } ++ ++ if (i == DEVICE_COUNT_RESOURCE) { ++ printk(KERN_ERR ++ "vmw_pvscsi: adapter has no suitable MMIO region\n"); ++ goto out_release_resources; ++ } ++ ++ adapter->mmioBase = pci_iomap(pdev, i, PVSCSI_MEM_SPACE_SIZE); ++ ++ if (!adapter->mmioBase) { ++ printk(KERN_ERR ++ "vmw_pvscsi: can't iomap for BAR %d memsize %lu\n", ++ i, PVSCSI_MEM_SPACE_SIZE); ++ goto out_release_resources; ++ } ++ ++ pci_set_master(pdev); ++ pci_set_drvdata(pdev, host); ++ ++ ll_adapter_reset(adapter); ++ ++ adapter->use_msg = pvscsi_setup_msg_workqueue(adapter); ++ ++ error = pvscsi_allocate_rings(adapter); ++ if (error) { ++ printk(KERN_ERR "vmw_pvscsi: unable to allocate ring memory\n"); ++ goto out_release_resources; ++ } ++ ++ /* ++ * From this point on we should reset the adapter if anything goes ++ * wrong. ++ */ ++ pvscsi_setup_all_rings(adapter); ++ ++ adapter->cmd_map = kcalloc(adapter->req_depth, ++ sizeof(struct pvscsi_ctx), GFP_KERNEL); ++ if (!adapter->cmd_map) { ++ printk(KERN_ERR "vmw_pvscsi: failed to allocate memory.\n"); ++ error = -ENOMEM; ++ goto out_reset_adapter; ++ } ++ ++ INIT_LIST_HEAD(&adapter->cmd_pool); ++ for (i = 0; i < adapter->req_depth; i++) { ++ struct pvscsi_ctx *ctx = adapter->cmd_map + i; ++ list_add(&ctx->list, &adapter->cmd_pool); ++ } ++ ++ error = pvscsi_allocate_sg(adapter); ++ if (error) { ++ printk(KERN_ERR "vmw_pvscsi: unable to allocate s/g table\n"); ++ goto out_reset_adapter; ++ } ++ ++ if (!pvscsi_disable_msix && ++ pvscsi_setup_msix(adapter, &adapter->irq) == 0) { ++ printk(KERN_INFO "vmw_pvscsi: using MSI-X\n"); ++ adapter->use_msix = 1; ++ } else if (!pvscsi_disable_msi && pci_enable_msi(pdev) == 0) { ++ printk(KERN_INFO "vmw_pvscsi: using MSI\n"); ++ adapter->use_msi = 1; ++ adapter->irq = pdev->irq; ++ } else { ++ printk(KERN_INFO "vmw_pvscsi: using INTx\n"); ++ adapter->irq = pdev->irq; ++ } ++ ++ error = request_irq(adapter->irq, pvscsi_isr, IRQF_SHARED, ++ "vmw_pvscsi", adapter); ++ if (error) { ++ printk(KERN_ERR ++ "vmw_pvscsi: unable to request IRQ: %d\n", error); ++ adapter->irq = 0; ++ goto out_reset_adapter; ++ } ++ ++ error = scsi_add_host(host, &pdev->dev); ++ if (error) { ++ printk(KERN_ERR ++ "vmw_pvscsi: scsi_add_host failed: %d\n", error); ++ goto out_reset_adapter; ++ } ++ ++ dev_info(&pdev->dev, "VMware PVSCSI rev %d host #%u\n", ++ adapter->rev, host->host_no); ++ ++ pvscsi_unmask_intr(adapter); ++ ++ scsi_scan_host(host); ++ ++ return 0; ++ ++out_reset_adapter: ++ ll_adapter_reset(adapter); ++out_release_resources: ++ pvscsi_release_resources(adapter); ++out_free_host: ++ scsi_host_put(host); ++out_disable_device: ++ pci_set_drvdata(pdev, NULL); ++ pci_disable_device(pdev); ++ ++ return error; ++} ++ ++static void __pvscsi_shutdown(struct pvscsi_adapter *adapter) ++{ ++ pvscsi_mask_intr(adapter); ++ ++ if (adapter->workqueue) ++ flush_workqueue(adapter->workqueue); ++ ++ pvscsi_shutdown_intr(adapter); ++ ++ pvscsi_process_request_ring(adapter); ++ pvscsi_process_completion_ring(adapter); ++ ll_adapter_reset(adapter); ++} ++ ++static void pvscsi_shutdown(struct pci_dev *dev) ++{ ++ struct Scsi_Host *host = pci_get_drvdata(dev); ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ ++ __pvscsi_shutdown(adapter); ++} ++ ++static void pvscsi_remove(struct pci_dev *pdev) ++{ ++ struct Scsi_Host *host = pci_get_drvdata(pdev); ++ struct pvscsi_adapter *adapter = shost_priv(host); ++ ++ scsi_remove_host(host); ++ ++ __pvscsi_shutdown(adapter); ++ pvscsi_release_resources(adapter); ++ ++ scsi_host_put(host); ++ ++ pci_set_drvdata(pdev, NULL); ++ pci_disable_device(pdev); ++} ++ ++static struct pci_driver pvscsi_pci_driver = { ++ .name = "vmw_pvscsi", ++ .id_table = pvscsi_pci_tbl, ++ .probe = pvscsi_probe, ++ .remove = __devexit_p(pvscsi_remove), ++ .shutdown = pvscsi_shutdown, ++}; ++ ++static int __init pvscsi_init(void) ++{ ++ pr_info("%s - version %s\n", ++ PVSCSI_LINUX_DRIVER_DESC, PVSCSI_DRIVER_VERSION_STRING); ++ return pci_register_driver(&pvscsi_pci_driver); ++} ++ ++static void __exit pvscsi_exit(void) ++{ ++ pci_unregister_driver(&pvscsi_pci_driver); ++} ++ ++module_init(pvscsi_init); ++module_exit(pvscsi_exit); +diff --git a/drivers/scsi/vmw_pvscsi.h b/drivers/scsi/vmw_pvscsi.h +new file mode 100644 +index 0000000..62e36e7 +--- /dev/null ++++ b/drivers/scsi/vmw_pvscsi.h +@@ -0,0 +1,397 @@ ++/* ++ * VMware PVSCSI header file ++ * ++ * Copyright (C) 2008-2009, VMware, Inc. All Rights Reserved. ++ * ++ * This program is free software; you can redistribute it and/or modify it ++ * under the terms of the GNU General Public License as published by the ++ * Free Software Foundation; version 2 of the License and no later version. ++ * ++ * This program is distributed in the hope that it will be useful, but ++ * WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or ++ * NON INFRINGEMENT. See the GNU General Public License for more ++ * details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. ++ * ++ * Maintained by: Alok N Kataria ++ * ++ */ ++ ++#ifndef _VMW_PVSCSI_H_ ++#define _VMW_PVSCSI_H_ ++ ++#include ++ ++#define PVSCSI_DRIVER_VERSION_STRING "1.0.1.0-k" ++ ++#define PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT 128 ++ ++#define MASK(n) ((1 << (n)) - 1) /* make an n-bit mask */ ++ ++#define PCI_VENDOR_ID_VMWARE 0x15AD ++#define PCI_DEVICE_ID_VMWARE_PVSCSI 0x07C0 ++ ++/* ++ * host adapter status/error codes ++ */ ++enum HostBusAdapterStatus { ++ BTSTAT_SUCCESS = 0x00, /* CCB complete normally with no errors */ ++ BTSTAT_LINKED_COMMAND_COMPLETED = 0x0a, ++ BTSTAT_LINKED_COMMAND_COMPLETED_WITH_FLAG = 0x0b, ++ BTSTAT_DATA_UNDERRUN = 0x0c, ++ BTSTAT_SELTIMEO = 0x11, /* SCSI selection timeout */ ++ BTSTAT_DATARUN = 0x12, /* data overrun/underrun */ ++ BTSTAT_BUSFREE = 0x13, /* unexpected bus free */ ++ BTSTAT_INVPHASE = 0x14, /* invalid bus phase or sequence requested by target */ ++ BTSTAT_LUNMISMATCH = 0x17, /* linked CCB has different LUN from first CCB */ ++ BTSTAT_SENSFAILED = 0x1b, /* auto request sense failed */ ++ BTSTAT_TAGREJECT = 0x1c, /* SCSI II tagged queueing message rejected by target */ ++ BTSTAT_BADMSG = 0x1d, /* unsupported message received by the host adapter */ ++ BTSTAT_HAHARDWARE = 0x20, /* host adapter hardware failed */ ++ BTSTAT_NORESPONSE = 0x21, /* target did not respond to SCSI ATN, sent a SCSI RST */ ++ BTSTAT_SENTRST = 0x22, /* host adapter asserted a SCSI RST */ ++ BTSTAT_RECVRST = 0x23, /* other SCSI devices asserted a SCSI RST */ ++ BTSTAT_DISCONNECT = 0x24, /* target device reconnected improperly (w/o tag) */ ++ BTSTAT_BUSRESET = 0x25, /* host adapter issued BUS device reset */ ++ BTSTAT_ABORTQUEUE = 0x26, /* abort queue generated */ ++ BTSTAT_HASOFTWARE = 0x27, /* host adapter software error */ ++ BTSTAT_HATIMEOUT = 0x30, /* host adapter hardware timeout error */ ++ BTSTAT_SCSIPARITY = 0x34, /* SCSI parity error detected */ ++}; ++ ++/* ++ * Register offsets. ++ * ++ * These registers are accessible both via i/o space and mm i/o. ++ */ ++ ++enum PVSCSIRegOffset { ++ PVSCSI_REG_OFFSET_COMMAND = 0x0, ++ PVSCSI_REG_OFFSET_COMMAND_DATA = 0x4, ++ PVSCSI_REG_OFFSET_COMMAND_STATUS = 0x8, ++ PVSCSI_REG_OFFSET_LAST_STS_0 = 0x100, ++ PVSCSI_REG_OFFSET_LAST_STS_1 = 0x104, ++ PVSCSI_REG_OFFSET_LAST_STS_2 = 0x108, ++ PVSCSI_REG_OFFSET_LAST_STS_3 = 0x10c, ++ PVSCSI_REG_OFFSET_INTR_STATUS = 0x100c, ++ PVSCSI_REG_OFFSET_INTR_MASK = 0x2010, ++ PVSCSI_REG_OFFSET_KICK_NON_RW_IO = 0x3014, ++ PVSCSI_REG_OFFSET_DEBUG = 0x3018, ++ PVSCSI_REG_OFFSET_KICK_RW_IO = 0x4018, ++}; ++ ++/* ++ * Virtual h/w commands. ++ */ ++ ++enum PVSCSICommands { ++ PVSCSI_CMD_FIRST = 0, /* has to be first */ ++ ++ PVSCSI_CMD_ADAPTER_RESET = 1, ++ PVSCSI_CMD_ISSUE_SCSI = 2, ++ PVSCSI_CMD_SETUP_RINGS = 3, ++ PVSCSI_CMD_RESET_BUS = 4, ++ PVSCSI_CMD_RESET_DEVICE = 5, ++ PVSCSI_CMD_ABORT_CMD = 6, ++ PVSCSI_CMD_CONFIG = 7, ++ PVSCSI_CMD_SETUP_MSG_RING = 8, ++ PVSCSI_CMD_DEVICE_UNPLUG = 9, ++ ++ PVSCSI_CMD_LAST = 10 /* has to be last */ ++}; ++ ++/* ++ * Command descriptor for PVSCSI_CMD_RESET_DEVICE -- ++ */ ++ ++struct PVSCSICmdDescResetDevice { ++ u32 target; ++ u8 lun[8]; ++} __packed; ++ ++/* ++ * Command descriptor for PVSCSI_CMD_ABORT_CMD -- ++ * ++ * - currently does not support specifying the LUN. ++ * - _pad should be 0. ++ */ ++ ++struct PVSCSICmdDescAbortCmd { ++ u64 context; ++ u32 target; ++ u32 _pad; ++} __packed; ++ ++/* ++ * Command descriptor for PVSCSI_CMD_SETUP_RINGS -- ++ * ++ * Notes: ++ * - reqRingNumPages and cmpRingNumPages need to be power of two. ++ * - reqRingNumPages and cmpRingNumPages need to be different from 0, ++ * - reqRingNumPages and cmpRingNumPages need to be inferior to ++ * PVSCSI_SETUP_RINGS_MAX_NUM_PAGES. ++ */ ++ ++#define PVSCSI_SETUP_RINGS_MAX_NUM_PAGES 32 ++struct PVSCSICmdDescSetupRings { ++ u32 reqRingNumPages; ++ u32 cmpRingNumPages; ++ u64 ringsStatePPN; ++ u64 reqRingPPNs[PVSCSI_SETUP_RINGS_MAX_NUM_PAGES]; ++ u64 cmpRingPPNs[PVSCSI_SETUP_RINGS_MAX_NUM_PAGES]; ++} __packed; ++ ++/* ++ * Command descriptor for PVSCSI_CMD_SETUP_MSG_RING -- ++ * ++ * Notes: ++ * - this command was not supported in the initial revision of the h/w ++ * interface. Before using it, you need to check that it is supported by ++ * writing PVSCSI_CMD_SETUP_MSG_RING to the 'command' register, then ++ * immediately after read the 'command status' register: ++ * * a value of -1 means that the cmd is NOT supported, ++ * * a value != -1 means that the cmd IS supported. ++ * If it's supported the 'command status' register should return: ++ * sizeof(PVSCSICmdDescSetupMsgRing) / sizeof(u32). ++ * - this command should be issued _after_ the usual SETUP_RINGS so that the ++ * RingsState page is already setup. If not, the command is a nop. ++ * - numPages needs to be a power of two, ++ * - numPages needs to be different from 0, ++ * - _pad should be zero. ++ */ ++ ++#define PVSCSI_SETUP_MSG_RING_MAX_NUM_PAGES 16 ++ ++struct PVSCSICmdDescSetupMsgRing { ++ u32 numPages; ++ u32 _pad; ++ u64 ringPPNs[PVSCSI_SETUP_MSG_RING_MAX_NUM_PAGES]; ++} __packed; ++ ++enum PVSCSIMsgType { ++ PVSCSI_MSG_DEV_ADDED = 0, ++ PVSCSI_MSG_DEV_REMOVED = 1, ++ PVSCSI_MSG_LAST = 2, ++}; ++ ++/* ++ * Msg descriptor. ++ * ++ * sizeof(struct PVSCSIRingMsgDesc) == 128. ++ * ++ * - type is of type enum PVSCSIMsgType. ++ * - the content of args depend on the type of event being delivered. ++ */ ++ ++struct PVSCSIRingMsgDesc { ++ u32 type; ++ u32 args[31]; ++} __packed; ++ ++struct PVSCSIMsgDescDevStatusChanged { ++ u32 type; /* PVSCSI_MSG_DEV _ADDED / _REMOVED */ ++ u32 bus; ++ u32 target; ++ u8 lun[8]; ++ u32 pad[27]; ++} __packed; ++ ++/* ++ * Rings state. ++ * ++ * - the fields: ++ * . msgProdIdx, ++ * . msgConsIdx, ++ * . msgNumEntriesLog2, ++ * .. are only used once the SETUP_MSG_RING cmd has been issued. ++ * - '_pad' helps to ensure that the msg related fields are on their own ++ * cache-line. ++ */ ++ ++struct PVSCSIRingsState { ++ u32 reqProdIdx; ++ u32 reqConsIdx; ++ u32 reqNumEntriesLog2; ++ ++ u32 cmpProdIdx; ++ u32 cmpConsIdx; ++ u32 cmpNumEntriesLog2; ++ ++ u8 _pad[104]; ++ ++ u32 msgProdIdx; ++ u32 msgConsIdx; ++ u32 msgNumEntriesLog2; ++} __packed; ++ ++/* ++ * Request descriptor. ++ * ++ * sizeof(RingReqDesc) = 128 ++ * ++ * - context: is a unique identifier of a command. It could normally be any ++ * 64bit value, however we currently store it in the serialNumber variable ++ * of struct SCSI_Command, so we have the following restrictions due to the ++ * way this field is handled in the vmkernel storage stack: ++ * * this value can't be 0, ++ * * the upper 32bit need to be 0 since serialNumber is as a u32. ++ * Currently tracked as PR 292060. ++ * - dataLen: contains the total number of bytes that need to be transferred. ++ * - dataAddr: ++ * * if PVSCSI_FLAG_CMD_WITH_SG_LIST is set: dataAddr is the PA of the first ++ * s/g table segment, each s/g segment is entirely contained on a single ++ * page of physical memory, ++ * * if PVSCSI_FLAG_CMD_WITH_SG_LIST is NOT set, then dataAddr is the PA of ++ * the buffer used for the DMA transfer, ++ * - flags: ++ * * PVSCSI_FLAG_CMD_WITH_SG_LIST: see dataAddr above, ++ * * PVSCSI_FLAG_CMD_DIR_NONE: no DMA involved, ++ * * PVSCSI_FLAG_CMD_DIR_TOHOST: transfer from device to main memory, ++ * * PVSCSI_FLAG_CMD_DIR_TODEVICE: transfer from main memory to device, ++ * * PVSCSI_FLAG_CMD_OUT_OF_BAND_CDB: reserved to handle CDBs larger than ++ * 16bytes. To be specified. ++ * - vcpuHint: vcpuId of the processor that will be most likely waiting for the ++ * completion of the i/o. For guest OSes that use lowest priority message ++ * delivery mode (such as windows), we use this "hint" to deliver the ++ * completion action to the proper vcpu. For now, we can use the vcpuId of ++ * the processor that initiated the i/o as a likely candidate for the vcpu ++ * that will be waiting for the completion.. ++ * - bus should be 0: we currently only support bus 0 for now. ++ * - unused should be zero'd. ++ */ ++ ++#define PVSCSI_FLAG_CMD_WITH_SG_LIST (1 << 0) ++#define PVSCSI_FLAG_CMD_OUT_OF_BAND_CDB (1 << 1) ++#define PVSCSI_FLAG_CMD_DIR_NONE (1 << 2) ++#define PVSCSI_FLAG_CMD_DIR_TOHOST (1 << 3) ++#define PVSCSI_FLAG_CMD_DIR_TODEVICE (1 << 4) ++ ++struct PVSCSIRingReqDesc { ++ u64 context; ++ u64 dataAddr; ++ u64 dataLen; ++ u64 senseAddr; ++ u32 senseLen; ++ u32 flags; ++ u8 cdb[16]; ++ u8 cdbLen; ++ u8 lun[8]; ++ u8 tag; ++ u8 bus; ++ u8 target; ++ u8 vcpuHint; ++ u8 unused[59]; ++} __packed; ++ ++/* ++ * Scatter-gather list management. ++ * ++ * As described above, when PVSCSI_FLAG_CMD_WITH_SG_LIST is set in the ++ * RingReqDesc.flags, then RingReqDesc.dataAddr is the PA of the first s/g ++ * table segment. ++ * ++ * - each segment of the s/g table contain a succession of struct ++ * PVSCSISGElement. ++ * - each segment is entirely contained on a single physical page of memory. ++ * - a "chain" s/g element has the flag PVSCSI_SGE_FLAG_CHAIN_ELEMENT set in ++ * PVSCSISGElement.flags and in this case: ++ * * addr is the PA of the next s/g segment, ++ * * length is undefined, assumed to be 0. ++ */ ++ ++struct PVSCSISGElement { ++ u64 addr; ++ u32 length; ++ u32 flags; ++} __packed; ++ ++/* ++ * Completion descriptor. ++ * ++ * sizeof(RingCmpDesc) = 32 ++ * ++ * - context: identifier of the command. The same thing that was specified ++ * under "context" as part of struct RingReqDesc at initiation time, ++ * - dataLen: number of bytes transferred for the actual i/o operation, ++ * - senseLen: number of bytes written into the sense buffer, ++ * - hostStatus: adapter status, ++ * - scsiStatus: device status, ++ * - _pad should be zero. ++ */ ++ ++struct PVSCSIRingCmpDesc { ++ u64 context; ++ u64 dataLen; ++ u32 senseLen; ++ u16 hostStatus; ++ u16 scsiStatus; ++ u32 _pad[2]; ++} __packed; ++ ++/* ++ * Interrupt status / IRQ bits. ++ */ ++ ++#define PVSCSI_INTR_CMPL_0 (1 << 0) ++#define PVSCSI_INTR_CMPL_1 (1 << 1) ++#define PVSCSI_INTR_CMPL_MASK MASK(2) ++ ++#define PVSCSI_INTR_MSG_0 (1 << 2) ++#define PVSCSI_INTR_MSG_1 (1 << 3) ++#define PVSCSI_INTR_MSG_MASK (MASK(2) << 2) ++ ++#define PVSCSI_INTR_ALL_SUPPORTED MASK(4) ++ ++/* ++ * Number of MSI-X vectors supported. ++ */ ++#define PVSCSI_MAX_INTRS 24 ++ ++/* ++ * Enumeration of supported MSI-X vectors ++ */ ++#define PVSCSI_VECTOR_COMPLETION 0 ++ ++/* ++ * Misc constants for the rings. ++ */ ++ ++#define PVSCSI_MAX_NUM_PAGES_REQ_RING PVSCSI_SETUP_RINGS_MAX_NUM_PAGES ++#define PVSCSI_MAX_NUM_PAGES_CMP_RING PVSCSI_SETUP_RINGS_MAX_NUM_PAGES ++#define PVSCSI_MAX_NUM_PAGES_MSG_RING PVSCSI_SETUP_MSG_RING_MAX_NUM_PAGES ++ ++#define PVSCSI_MAX_NUM_REQ_ENTRIES_PER_PAGE \ ++ (PAGE_SIZE / sizeof(struct PVSCSIRingReqDesc)) ++ ++#define PVSCSI_MAX_REQ_QUEUE_DEPTH \ ++ (PVSCSI_MAX_NUM_PAGES_REQ_RING * PVSCSI_MAX_NUM_REQ_ENTRIES_PER_PAGE) ++ ++#define PVSCSI_MEM_SPACE_COMMAND_NUM_PAGES 1 ++#define PVSCSI_MEM_SPACE_INTR_STATUS_NUM_PAGES 1 ++#define PVSCSI_MEM_SPACE_MISC_NUM_PAGES 2 ++#define PVSCSI_MEM_SPACE_KICK_IO_NUM_PAGES 2 ++#define PVSCSI_MEM_SPACE_MSIX_NUM_PAGES 2 ++ ++enum PVSCSIMemSpace { ++ PVSCSI_MEM_SPACE_COMMAND_PAGE = 0, ++ PVSCSI_MEM_SPACE_INTR_STATUS_PAGE = 1, ++ PVSCSI_MEM_SPACE_MISC_PAGE = 2, ++ PVSCSI_MEM_SPACE_KICK_IO_PAGE = 4, ++ PVSCSI_MEM_SPACE_MSIX_TABLE_PAGE = 6, ++ PVSCSI_MEM_SPACE_MSIX_PBA_PAGE = 7, ++}; ++ ++#define PVSCSI_MEM_SPACE_NUM_PAGES \ ++ (PVSCSI_MEM_SPACE_COMMAND_NUM_PAGES + \ ++ PVSCSI_MEM_SPACE_INTR_STATUS_NUM_PAGES + \ ++ PVSCSI_MEM_SPACE_MISC_NUM_PAGES + \ ++ PVSCSI_MEM_SPACE_KICK_IO_NUM_PAGES + \ ++ PVSCSI_MEM_SPACE_MSIX_NUM_PAGES) ++ ++#define PVSCSI_MEM_SPACE_SIZE (PVSCSI_MEM_SPACE_NUM_PAGES * PAGE_SIZE) ++ ++#endif /* _VMW_PVSCSI_H_ */ +diff --git a/drivers/serial/kgdboc.c b/drivers/serial/kgdboc.c +index eadc1ab..2d81457 100644 +--- a/drivers/serial/kgdboc.c ++++ b/drivers/serial/kgdboc.c +@@ -18,7 +18,7 @@ + + #define MAX_CONFIG_LEN 40 + +-static struct kgdb_io kgdboc_io_ops; ++static const struct kgdb_io kgdboc_io_ops; + + /* -1 = init not run yet, 0 = unconfigured, 1 = configured. */ + static int configured = -1; +@@ -154,7 +154,7 @@ static void kgdboc_post_exp_handler(void) + module_put(THIS_MODULE); + } + +-static struct kgdb_io kgdboc_io_ops = { ++static const struct kgdb_io kgdboc_io_ops = { + .name = "kgdboc", + .read_char = kgdboc_get_char, + .write_char = kgdboc_put_char, +diff --git a/drivers/spi/spi.c b/drivers/spi/spi.c +index b76f246..7f41af7 100644 +--- a/drivers/spi/spi.c ++++ b/drivers/spi/spi.c +@@ -774,7 +774,7 @@ int spi_sync(struct spi_device *spi, struct spi_message *message) + EXPORT_SYMBOL_GPL(spi_sync); + + /* portable code must never pass more than 32 bytes */ +-#define SPI_BUFSIZ max(32,SMP_CACHE_BYTES) ++#define SPI_BUFSIZ max(32U,SMP_CACHE_BYTES) + + static u8 *buf; + +diff --git a/drivers/staging/android/binder.c b/drivers/staging/android/binder.c +index b9b37ff..19dfa23 100644 +--- a/drivers/staging/android/binder.c ++++ b/drivers/staging/android/binder.c +@@ -2761,7 +2761,7 @@ static void binder_vma_close(struct vm_area_struct *vma) + binder_defer_work(proc, BINDER_DEFERRED_PUT_FILES); + } + +-static struct vm_operations_struct binder_vm_ops = { ++static const struct vm_operations_struct binder_vm_ops = { + .open = binder_vma_open, + .close = binder_vma_close, + }; +diff --git a/drivers/staging/b3dfg/b3dfg.c b/drivers/staging/b3dfg/b3dfg.c +index cda26bb..39fed3f 100644 +--- a/drivers/staging/b3dfg/b3dfg.c ++++ b/drivers/staging/b3dfg/b3dfg.c +@@ -455,7 +455,7 @@ static int b3dfg_vma_fault(struct vm_area_struct *vma, + return VM_FAULT_NOPAGE; + } + +-static struct vm_operations_struct b3dfg_vm_ops = { ++static const struct vm_operations_struct b3dfg_vm_ops = { + .fault = b3dfg_vma_fault, + }; + +@@ -848,7 +848,7 @@ static int b3dfg_mmap(struct file *filp, struct vm_area_struct *vma) + return r; + } + +-static struct file_operations b3dfg_fops = { ++static const struct file_operations b3dfg_fops = { + .owner = THIS_MODULE, + .open = b3dfg_open, + .release = b3dfg_release, +diff --git a/drivers/staging/comedi/comedi_fops.c b/drivers/staging/comedi/comedi_fops.c +index 908f25a..c9a579b 100644 +--- a/drivers/staging/comedi/comedi_fops.c ++++ b/drivers/staging/comedi/comedi_fops.c +@@ -1389,7 +1389,7 @@ void comedi_unmap(struct vm_area_struct *area) + mutex_unlock(&dev->mutex); + } + +-static struct vm_operations_struct comedi_vm_ops = { ++static const struct vm_operations_struct comedi_vm_ops = { + .close = comedi_unmap, + }; + +diff --git a/drivers/staging/dream/qdsp5/adsp_driver.c b/drivers/staging/dream/qdsp5/adsp_driver.c +index e55a0db..577b776 100644 +--- a/drivers/staging/dream/qdsp5/adsp_driver.c ++++ b/drivers/staging/dream/qdsp5/adsp_driver.c +@@ -576,7 +576,7 @@ static struct adsp_device *inode_to_device(struct inode *inode) + static dev_t adsp_devno; + static struct class *adsp_class; + +-static struct file_operations adsp_fops = { ++static const struct file_operations adsp_fops = { + .owner = THIS_MODULE, + .open = adsp_open, + .unlocked_ioctl = adsp_ioctl, +diff --git a/drivers/staging/dream/qdsp5/audio_aac.c b/drivers/staging/dream/qdsp5/audio_aac.c +index ad2390f..4116ee8 100644 +--- a/drivers/staging/dream/qdsp5/audio_aac.c ++++ b/drivers/staging/dream/qdsp5/audio_aac.c +@@ -1022,7 +1022,7 @@ done: + return rc; + } + +-static struct file_operations audio_aac_fops = { ++static const struct file_operations audio_aac_fops = { + .owner = THIS_MODULE, + .open = audio_open, + .release = audio_release, +diff --git a/drivers/staging/dream/qdsp5/audio_amrnb.c b/drivers/staging/dream/qdsp5/audio_amrnb.c +index cd818a5..870b37b 100644 +--- a/drivers/staging/dream/qdsp5/audio_amrnb.c ++++ b/drivers/staging/dream/qdsp5/audio_amrnb.c +@@ -833,7 +833,7 @@ done: + return rc; + } + +-static struct file_operations audio_amrnb_fops = { ++static const struct file_operations audio_amrnb_fops = { + .owner = THIS_MODULE, + .open = audamrnb_open, + .release = audamrnb_release, +diff --git a/drivers/staging/dream/qdsp5/audio_evrc.c b/drivers/staging/dream/qdsp5/audio_evrc.c +index 4b43e18..cedafda 100644 +--- a/drivers/staging/dream/qdsp5/audio_evrc.c ++++ b/drivers/staging/dream/qdsp5/audio_evrc.c +@@ -805,7 +805,7 @@ dma_fail: + return rc; + } + +-static struct file_operations audio_evrc_fops = { ++static const struct file_operations audio_evrc_fops = { + .owner = THIS_MODULE, + .open = audevrc_open, + .release = audevrc_release, +diff --git a/drivers/staging/dream/qdsp5/audio_in.c b/drivers/staging/dream/qdsp5/audio_in.c +index 3d950a2..9431118 100644 +--- a/drivers/staging/dream/qdsp5/audio_in.c ++++ b/drivers/staging/dream/qdsp5/audio_in.c +@@ -913,7 +913,7 @@ static int audpre_open(struct inode *inode, struct file *file) + return 0; + } + +-static struct file_operations audio_fops = { ++static const struct file_operations audio_fops = { + .owner = THIS_MODULE, + .open = audio_in_open, + .release = audio_in_release, +@@ -922,7 +922,7 @@ static struct file_operations audio_fops = { + .unlocked_ioctl = audio_in_ioctl, + }; + +-static struct file_operations audpre_fops = { ++static const struct file_operations audpre_fops = { + .owner = THIS_MODULE, + .open = audpre_open, + .unlocked_ioctl = audpre_ioctl, +diff --git a/drivers/staging/dream/qdsp5/audio_mp3.c b/drivers/staging/dream/qdsp5/audio_mp3.c +index b95574f..286c2f4 100644 +--- a/drivers/staging/dream/qdsp5/audio_mp3.c ++++ b/drivers/staging/dream/qdsp5/audio_mp3.c +@@ -941,7 +941,7 @@ done: + return rc; + } + +-static struct file_operations audio_mp3_fops = { ++static const struct file_operations audio_mp3_fops = { + .owner = THIS_MODULE, + .open = audio_open, + .release = audio_release, +diff --git a/drivers/staging/dream/qdsp5/audio_out.c b/drivers/staging/dream/qdsp5/audio_out.c +index d1adcf6..f8f9833 100644 +--- a/drivers/staging/dream/qdsp5/audio_out.c ++++ b/drivers/staging/dream/qdsp5/audio_out.c +@@ -810,7 +810,7 @@ static int audpp_open(struct inode *inode, struct file *file) + return 0; + } + +-static struct file_operations audio_fops = { ++static const struct file_operations audio_fops = { + .owner = THIS_MODULE, + .open = audio_open, + .release = audio_release, +@@ -819,7 +819,7 @@ static struct file_operations audio_fops = { + .unlocked_ioctl = audio_ioctl, + }; + +-static struct file_operations audpp_fops = { ++static const struct file_operations audpp_fops = { + .owner = THIS_MODULE, + .open = audpp_open, + .unlocked_ioctl = audpp_ioctl, +diff --git a/drivers/staging/dream/qdsp5/audio_qcelp.c b/drivers/staging/dream/qdsp5/audio_qcelp.c +index f0f50e3..f6b9dbc 100644 +--- a/drivers/staging/dream/qdsp5/audio_qcelp.c ++++ b/drivers/staging/dream/qdsp5/audio_qcelp.c +@@ -816,7 +816,7 @@ err: + return rc; + } + +-static struct file_operations audio_qcelp_fops = { ++static const struct file_operations audio_qcelp_fops = { + .owner = THIS_MODULE, + .open = audqcelp_open, + .release = audqcelp_release, +diff --git a/drivers/staging/dream/qdsp5/snd.c b/drivers/staging/dream/qdsp5/snd.c +index 037d7ff..5469ec3 100644 +--- a/drivers/staging/dream/qdsp5/snd.c ++++ b/drivers/staging/dream/qdsp5/snd.c +@@ -242,7 +242,7 @@ err: + return rc; + } + +-static struct file_operations snd_fops = { ++static const struct file_operations snd_fops = { + .owner = THIS_MODULE, + .open = snd_open, + .release = snd_release, +diff --git a/drivers/staging/dream/smd/smd_qmi.c b/drivers/staging/dream/smd/smd_qmi.c +index d4e7d88..0ea632a 100644 +--- a/drivers/staging/dream/smd/smd_qmi.c ++++ b/drivers/staging/dream/smd/smd_qmi.c +@@ -793,7 +793,7 @@ static int qmi_release(struct inode *ip, struct file *fp) + return 0; + } + +-static struct file_operations qmi_fops = { ++static const struct file_operations qmi_fops = { + .owner = THIS_MODULE, + .read = qmi_read, + .write = qmi_write, +diff --git a/drivers/staging/dream/smd/smd_rpcrouter_device.c b/drivers/staging/dream/smd/smd_rpcrouter_device.c +index cd3910b..ff053d3 100644 +--- a/drivers/staging/dream/smd/smd_rpcrouter_device.c ++++ b/drivers/staging/dream/smd/smd_rpcrouter_device.c +@@ -214,7 +214,7 @@ static long rpcrouter_ioctl(struct file *filp, unsigned int cmd, + return rc; + } + +-static struct file_operations rpcrouter_server_fops = { ++static const struct file_operations rpcrouter_server_fops = { + .owner = THIS_MODULE, + .open = rpcrouter_open, + .release = rpcrouter_release, +@@ -224,7 +224,7 @@ static struct file_operations rpcrouter_server_fops = { + .unlocked_ioctl = rpcrouter_ioctl, + }; + +-static struct file_operations rpcrouter_router_fops = { ++static const struct file_operations rpcrouter_router_fops = { + .owner = THIS_MODULE, + .open = rpcrouter_open, + .release = rpcrouter_release, +diff --git a/drivers/staging/dst/dcore.c b/drivers/staging/dst/dcore.c +index c24e4e0..07665be 100644 +--- a/drivers/staging/dst/dcore.c ++++ b/drivers/staging/dst/dcore.c +@@ -149,7 +149,7 @@ static int dst_bdev_release(struct gendisk *disk, fmode_t mode) + return 0; + } + +-static struct block_device_operations dst_blk_ops = { ++static const struct block_device_operations dst_blk_ops = { + .open = dst_bdev_open, + .release = dst_bdev_release, + .owner = THIS_MODULE, +@@ -588,7 +588,7 @@ static struct dst_node *dst_alloc_node(struct dst_ctl *ctl, + n->size = ctl->size; + + atomic_set(&n->refcnt, 1); +- atomic_long_set(&n->gen, 0); ++ atomic_long_set_unchecked(&n->gen, 0); + snprintf(n->name, sizeof(n->name), "%s", ctl->name); + + err = dst_node_sysfs_init(n); +diff --git a/drivers/staging/dst/trans.c b/drivers/staging/dst/trans.c +index 557d372..8d84422 100644 +--- a/drivers/staging/dst/trans.c ++++ b/drivers/staging/dst/trans.c +@@ -169,7 +169,7 @@ int dst_process_bio(struct dst_node *n, struct bio *bio) + t->error = 0; + t->retries = 0; + atomic_set(&t->refcnt, 1); +- t->gen = atomic_long_inc_return(&n->gen); ++ t->gen = atomic_long_inc_return_unchecked(&n->gen); + + t->enc = bio_data_dir(bio); + dst_bio_to_cmd(bio, &t->cmd, DST_IO, t->gen); +diff --git a/drivers/staging/et131x/et1310_tx.c b/drivers/staging/et131x/et1310_tx.c +index 94f7752..d051514 100644 +--- a/drivers/staging/et131x/et1310_tx.c ++++ b/drivers/staging/et131x/et1310_tx.c +@@ -710,11 +710,11 @@ inline void et131x_free_send_packet(struct et131x_adapter *etdev, + struct net_device_stats *stats = &etdev->net_stats; + + if (pMpTcb->Flags & fMP_DEST_BROAD) +- atomic_inc(&etdev->Stats.brdcstxmt); ++ atomic_inc_unchecked(&etdev->Stats.brdcstxmt); + else if (pMpTcb->Flags & fMP_DEST_MULTI) +- atomic_inc(&etdev->Stats.multixmt); ++ atomic_inc_unchecked(&etdev->Stats.multixmt); + else +- atomic_inc(&etdev->Stats.unixmt); ++ atomic_inc_unchecked(&etdev->Stats.unixmt); + + if (pMpTcb->Packet) { + stats->tx_bytes += pMpTcb->Packet->len; +diff --git a/drivers/staging/et131x/et131x_adapter.h b/drivers/staging/et131x/et131x_adapter.h +index 1dfe06f..f469b4d 100644 +--- a/drivers/staging/et131x/et131x_adapter.h ++++ b/drivers/staging/et131x/et131x_adapter.h +@@ -145,11 +145,11 @@ typedef struct _ce_stats_t { + * operations + */ + u32 unircv; /* # multicast packets received */ +- atomic_t unixmt; /* # multicast packets for Tx */ ++ atomic_unchecked_t unixmt; /* # multicast packets for Tx */ + u32 multircv; /* # multicast packets received */ +- atomic_t multixmt; /* # multicast packets for Tx */ ++ atomic_unchecked_t multixmt; /* # multicast packets for Tx */ + u32 brdcstrcv; /* # broadcast packets received */ +- atomic_t brdcstxmt; /* # broadcast packets for Tx */ ++ atomic_unchecked_t brdcstxmt; /* # broadcast packets for Tx */ + u32 norcvbuf; /* # Rx packets discarded */ + u32 noxmtbuf; /* # Tx packets discarded */ + +diff --git a/drivers/staging/go7007/go7007-v4l2.c b/drivers/staging/go7007/go7007-v4l2.c +index 4bd353a..e28f455 100644 +--- a/drivers/staging/go7007/go7007-v4l2.c ++++ b/drivers/staging/go7007/go7007-v4l2.c +@@ -1700,7 +1700,7 @@ static int go7007_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) + return 0; + } + +-static struct vm_operations_struct go7007_vm_ops = { ++static const struct vm_operations_struct go7007_vm_ops = { + .open = go7007_vm_open, + .close = go7007_vm_close, + .fault = go7007_vm_fault, +diff --git a/drivers/staging/hv/Channel.c b/drivers/staging/hv/Channel.c +index 366dc95..b974d87 100644 +--- a/drivers/staging/hv/Channel.c ++++ b/drivers/staging/hv/Channel.c +@@ -464,8 +464,8 @@ int VmbusChannelEstablishGpadl(struct vmbus_channel *Channel, void *Kbuffer, + + DPRINT_ENTER(VMBUS); + +- nextGpadlHandle = atomic_read(&gVmbusConnection.NextGpadlHandle); +- atomic_inc(&gVmbusConnection.NextGpadlHandle); ++ nextGpadlHandle = atomic_read_unchecked(&gVmbusConnection.NextGpadlHandle); ++ atomic_inc_unchecked(&gVmbusConnection.NextGpadlHandle); + + VmbusChannelCreateGpadlHeader(Kbuffer, Size, &msgInfo, &msgCount); + ASSERT(msgInfo != NULL); +diff --git a/drivers/staging/hv/Hv.c b/drivers/staging/hv/Hv.c +index b12237f..01ae28a 100644 +--- a/drivers/staging/hv/Hv.c ++++ b/drivers/staging/hv/Hv.c +@@ -161,7 +161,7 @@ static u64 HvDoHypercall(u64 Control, void *Input, void *Output) + u64 outputAddress = (Output) ? virt_to_phys(Output) : 0; + u32 outputAddressHi = outputAddress >> 32; + u32 outputAddressLo = outputAddress & 0xFFFFFFFF; +- volatile void *hypercallPage = gHvContext.HypercallPage; ++ volatile void *hypercallPage = ktva_ktla(gHvContext.HypercallPage); + + DPRINT_DBG(VMBUS, "Hypercall ", + Control, Input, Output); +diff --git a/drivers/staging/hv/VmbusApi.h b/drivers/staging/hv/VmbusApi.h +index d089bb1..2ebc158 100644 +--- a/drivers/staging/hv/VmbusApi.h ++++ b/drivers/staging/hv/VmbusApi.h +@@ -109,7 +109,7 @@ struct vmbus_channel_interface { + u32 *GpadlHandle); + int (*TeardownGpadl)(struct hv_device *device, u32 GpadlHandle); + void (*GetInfo)(struct hv_device *dev, struct hv_device_info *devinfo); +-}; ++} __no_const; + + /* Base driver object */ + struct hv_driver { +diff --git a/drivers/staging/hv/VmbusPrivate.h b/drivers/staging/hv/VmbusPrivate.h +index 5a37cce..6ecc88c 100644 +--- a/drivers/staging/hv/VmbusPrivate.h ++++ b/drivers/staging/hv/VmbusPrivate.h +@@ -59,7 +59,7 @@ enum VMBUS_CONNECT_STATE { + struct VMBUS_CONNECTION { + enum VMBUS_CONNECT_STATE ConnectState; + +- atomic_t NextGpadlHandle; ++ atomic_unchecked_t NextGpadlHandle; + + /* + * Represents channel interrupts. Each bit position represents a +diff --git a/drivers/staging/hv/blkvsc_drv.c b/drivers/staging/hv/blkvsc_drv.c +index 871a202..ca50ddf 100644 +--- a/drivers/staging/hv/blkvsc_drv.c ++++ b/drivers/staging/hv/blkvsc_drv.c +@@ -153,7 +153,7 @@ static int blkvsc_ringbuffer_size = BLKVSC_RING_BUFFER_SIZE; + /* The one and only one */ + static struct blkvsc_driver_context g_blkvsc_drv; + +-static struct block_device_operations block_ops = { ++static const struct block_device_operations block_ops = { + .owner = THIS_MODULE, + .open = blkvsc_open, + .release = blkvsc_release, +diff --git a/drivers/staging/hv/vmbus_drv.c b/drivers/staging/hv/vmbus_drv.c +index 6acc49a..fbc8d46 100644 +--- a/drivers/staging/hv/vmbus_drv.c ++++ b/drivers/staging/hv/vmbus_drv.c +@@ -532,7 +532,7 @@ static int vmbus_child_device_register(struct hv_device *root_device_obj, + to_device_context(root_device_obj); + struct device_context *child_device_ctx = + to_device_context(child_device_obj); +- static atomic_t device_num = ATOMIC_INIT(0); ++ static atomic_unchecked_t device_num = ATOMIC_INIT(0); + + DPRINT_ENTER(VMBUS_DRV); + +@@ -541,7 +541,7 @@ static int vmbus_child_device_register(struct hv_device *root_device_obj, + + /* Set the device name. Otherwise, device_register() will fail. */ + dev_set_name(&child_device_ctx->device, "vmbus_0_%d", +- atomic_inc_return(&device_num)); ++ atomic_inc_return_unchecked(&device_num)); + + /* The new device belongs to this bus */ + child_device_ctx->device.bus = &g_vmbus_drv.bus; /* device->dev.bus; */ +diff --git a/drivers/staging/iio/ring_generic.h b/drivers/staging/iio/ring_generic.h +index d926189..17b19fd 100644 +--- a/drivers/staging/iio/ring_generic.h ++++ b/drivers/staging/iio/ring_generic.h +@@ -87,7 +87,7 @@ struct iio_ring_access_funcs { + + int (*is_enabled)(struct iio_ring_buffer *ring); + int (*enable)(struct iio_ring_buffer *ring); +-}; ++} __no_const; + + /** + * struct iio_ring_buffer - general ring buffer structure +diff --git a/drivers/staging/octeon/ethernet-rx.c b/drivers/staging/octeon/ethernet-rx.c +index 1b237b7..88c624e 100644 +--- a/drivers/staging/octeon/ethernet-rx.c ++++ b/drivers/staging/octeon/ethernet-rx.c +@@ -406,11 +406,11 @@ void cvm_oct_tasklet_rx(unsigned long unused) + /* Increment RX stats for virtual ports */ + if (work->ipprt >= CVMX_PIP_NUM_INPUT_PORTS) { + #ifdef CONFIG_64BIT +- atomic64_add(1, (atomic64_t *)&priv->stats.rx_packets); +- atomic64_add(skb->len, (atomic64_t *)&priv->stats.rx_bytes); ++ atomic64_add_unchecked(1, (atomic64_unchecked_t *)&priv->stats.rx_packets); ++ atomic64_add_unchecked(skb->len, (atomic64_unchecked_t *)&priv->stats.rx_bytes); + #else +- atomic_add(1, (atomic_t *)&priv->stats.rx_packets); +- atomic_add(skb->len, (atomic_t *)&priv->stats.rx_bytes); ++ atomic_add_unchecked(1, (atomic_unchecked_t *)&priv->stats.rx_packets); ++ atomic_add_unchecked(skb->len, (atomic_unchecked_t *)&priv->stats.rx_bytes); + #endif + } + netif_receive_skb(skb); +@@ -424,9 +424,9 @@ void cvm_oct_tasklet_rx(unsigned long unused) + dev->name); + */ + #ifdef CONFIG_64BIT +- atomic64_add(1, (atomic64_t *)&priv->stats.rx_dropped); ++ atomic64_add_unchecked(1, (atomic64_t *)&priv->stats.rx_dropped); + #else +- atomic_add(1, (atomic_t *)&priv->stats.rx_dropped); ++ atomic_add_unchecked(1, (atomic_t *)&priv->stats.rx_dropped); + #endif + dev_kfree_skb_irq(skb); + } +diff --git a/drivers/staging/octeon/ethernet.c b/drivers/staging/octeon/ethernet.c +index 492c502..d9909f1 100644 +--- a/drivers/staging/octeon/ethernet.c ++++ b/drivers/staging/octeon/ethernet.c +@@ -294,11 +294,11 @@ static struct net_device_stats *cvm_oct_common_get_stats(struct net_device *dev) + * since the RX tasklet also increments it. + */ + #ifdef CONFIG_64BIT +- atomic64_add(rx_status.dropped_packets, +- (atomic64_t *)&priv->stats.rx_dropped); ++ atomic64_add_unchecked(rx_status.dropped_packets, ++ (atomic64_unchecked_t *)&priv->stats.rx_dropped); + #else +- atomic_add(rx_status.dropped_packets, +- (atomic_t *)&priv->stats.rx_dropped); ++ atomic_add_unchecked(rx_status.dropped_packets, ++ (atomic_unchecked_t *)&priv->stats.rx_dropped); + #endif + } + +diff --git a/drivers/staging/otus/80211core/pub_zfi.h b/drivers/staging/otus/80211core/pub_zfi.h +index a35bd5d..28fff45 100644 +--- a/drivers/staging/otus/80211core/pub_zfi.h ++++ b/drivers/staging/otus/80211core/pub_zfi.h +@@ -531,7 +531,7 @@ struct zsCbFuncTbl + u8_t (*zfcbClassifyTxPacket)(zdev_t* dev, zbuf_t* buf); + + void (*zfcbHwWatchDogNotify)(zdev_t* dev); +-}; ++} __no_const; + + extern void zfZeroMemory(u8_t* va, u16_t length); + #define ZM_INIT_CB_FUNC_TABLE(p) zfZeroMemory((u8_t *)p, sizeof(struct zsCbFuncTbl)); +diff --git a/drivers/staging/panel/panel.c b/drivers/staging/panel/panel.c +index c39a25f..696f5aa 100644 +--- a/drivers/staging/panel/panel.c ++++ b/drivers/staging/panel/panel.c +@@ -1305,7 +1305,7 @@ static int lcd_release(struct inode *inode, struct file *file) + return 0; + } + +-static struct file_operations lcd_fops = { ++static const struct file_operations lcd_fops = { + .write = lcd_write, + .open = lcd_open, + .release = lcd_release, +@@ -1565,7 +1565,7 @@ static int keypad_release(struct inode *inode, struct file *file) + return 0; + } + +-static struct file_operations keypad_fops = { ++static const struct file_operations keypad_fops = { + .read = keypad_read, /* read */ + .open = keypad_open, /* open */ + .release = keypad_release, /* close */ +diff --git a/drivers/staging/phison/phison.c b/drivers/staging/phison/phison.c +index 270ebcb..37e46af 100644 +--- a/drivers/staging/phison/phison.c ++++ b/drivers/staging/phison/phison.c +@@ -43,7 +43,7 @@ static struct scsi_host_template phison_sht = { + ATA_BMDMA_SHT(DRV_NAME), + }; + +-static struct ata_port_operations phison_ops = { ++static const struct ata_port_operations phison_ops = { + .inherits = &ata_bmdma_port_ops, + .prereset = phison_pre_reset, + }; +diff --git a/drivers/staging/poch/poch.c b/drivers/staging/poch/poch.c +index 2eb8e3d..57616a7 100644 +--- a/drivers/staging/poch/poch.c ++++ b/drivers/staging/poch/poch.c +@@ -1057,7 +1057,7 @@ static int poch_ioctl(struct inode *inode, struct file *filp, + return 0; + } + +-static struct file_operations poch_fops = { ++static const struct file_operations poch_fops = { + .owner = THIS_MODULE, + .open = poch_open, + .release = poch_release, +diff --git a/drivers/staging/pohmelfs/inode.c b/drivers/staging/pohmelfs/inode.c +index c94de31..19402bc 100644 +--- a/drivers/staging/pohmelfs/inode.c ++++ b/drivers/staging/pohmelfs/inode.c +@@ -1850,7 +1850,7 @@ static int pohmelfs_fill_super(struct super_block *sb, void *data, int silent) + mutex_init(&psb->mcache_lock); + psb->mcache_root = RB_ROOT; + psb->mcache_timeout = msecs_to_jiffies(5000); +- atomic_long_set(&psb->mcache_gen, 0); ++ atomic_long_set_unchecked(&psb->mcache_gen, 0); + + psb->trans_max_pages = 100; + +@@ -1865,7 +1865,7 @@ static int pohmelfs_fill_super(struct super_block *sb, void *data, int silent) + INIT_LIST_HEAD(&psb->crypto_ready_list); + INIT_LIST_HEAD(&psb->crypto_active_list); + +- atomic_set(&psb->trans_gen, 1); ++ atomic_set_unchecked(&psb->trans_gen, 1); + atomic_long_set(&psb->total_inodes, 0); + + mutex_init(&psb->state_lock); +diff --git a/drivers/staging/pohmelfs/mcache.c b/drivers/staging/pohmelfs/mcache.c +index e22665c..a2a9390 100644 +--- a/drivers/staging/pohmelfs/mcache.c ++++ b/drivers/staging/pohmelfs/mcache.c +@@ -121,7 +121,7 @@ struct pohmelfs_mcache *pohmelfs_mcache_alloc(struct pohmelfs_sb *psb, u64 start + m->data = data; + m->start = start; + m->size = size; +- m->gen = atomic_long_inc_return(&psb->mcache_gen); ++ m->gen = atomic_long_inc_return_unchecked(&psb->mcache_gen); + + mutex_lock(&psb->mcache_lock); + err = pohmelfs_mcache_insert(psb, m); +diff --git a/drivers/staging/pohmelfs/netfs.h b/drivers/staging/pohmelfs/netfs.h +index 623a07d..4035c19 100644 +--- a/drivers/staging/pohmelfs/netfs.h ++++ b/drivers/staging/pohmelfs/netfs.h +@@ -570,14 +570,14 @@ struct pohmelfs_config; + struct pohmelfs_sb { + struct rb_root mcache_root; + struct mutex mcache_lock; +- atomic_long_t mcache_gen; ++ atomic_long_unchecked_t mcache_gen; + unsigned long mcache_timeout; + + unsigned int idx; + + unsigned int trans_retries; + +- atomic_t trans_gen; ++ atomic_unchecked_t trans_gen; + + unsigned int crypto_attached_size; + unsigned int crypto_align_size; +diff --git a/drivers/staging/pohmelfs/trans.c b/drivers/staging/pohmelfs/trans.c +index 36a2535..0591bf4 100644 +--- a/drivers/staging/pohmelfs/trans.c ++++ b/drivers/staging/pohmelfs/trans.c +@@ -492,7 +492,7 @@ int netfs_trans_finish(struct netfs_trans *t, struct pohmelfs_sb *psb) + int err; + struct netfs_cmd *cmd = t->iovec.iov_base; + +- t->gen = atomic_inc_return(&psb->trans_gen); ++ t->gen = atomic_inc_return_unchecked(&psb->trans_gen); + + cmd->size = t->iovec.iov_len - sizeof(struct netfs_cmd) + + t->attached_size + t->attached_pages * sizeof(struct netfs_cmd); +diff --git a/drivers/staging/sep/sep_driver.c b/drivers/staging/sep/sep_driver.c +index f890a16..509ece8 100644 +--- a/drivers/staging/sep/sep_driver.c ++++ b/drivers/staging/sep/sep_driver.c +@@ -2603,7 +2603,7 @@ static struct pci_driver sep_pci_driver = { + static dev_t sep_devno; + + /* the files operations structure of the driver */ +-static struct file_operations sep_file_operations = { ++static const struct file_operations sep_file_operations = { + .owner = THIS_MODULE, + .ioctl = sep_ioctl, + .poll = sep_poll, +diff --git a/drivers/staging/usbip/usbip_common.h b/drivers/staging/usbip/usbip_common.h +index 5e16bc3..7655b10 100644 +--- a/drivers/staging/usbip/usbip_common.h ++++ b/drivers/staging/usbip/usbip_common.h +@@ -374,7 +374,7 @@ struct usbip_device { + void (*shutdown)(struct usbip_device *); + void (*reset)(struct usbip_device *); + void (*unusable)(struct usbip_device *); +- } eh_ops; ++ } __no_const eh_ops; + }; + + +diff --git a/drivers/staging/usbip/vhci.h b/drivers/staging/usbip/vhci.h +index 57f7946..d9df23d 100644 +--- a/drivers/staging/usbip/vhci.h ++++ b/drivers/staging/usbip/vhci.h +@@ -92,7 +92,7 @@ struct vhci_hcd { + unsigned resuming:1; + unsigned long re_timeout; + +- atomic_t seqnum; ++ atomic_unchecked_t seqnum; + + /* + * NOTE: +diff --git a/drivers/staging/usbip/vhci_hcd.c b/drivers/staging/usbip/vhci_hcd.c +index 20cd7db..c2693ff 100644 +--- a/drivers/staging/usbip/vhci_hcd.c ++++ b/drivers/staging/usbip/vhci_hcd.c +@@ -534,7 +534,7 @@ static void vhci_tx_urb(struct urb *urb) + return; + } + +- priv->seqnum = atomic_inc_return(&the_controller->seqnum); ++ priv->seqnum = atomic_inc_return_unchecked(&the_controller->seqnum); + if (priv->seqnum == 0xffff) + usbip_uinfo("seqnum max\n"); + +@@ -793,7 +793,7 @@ static int vhci_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status) + return -ENOMEM; + } + +- unlink->seqnum = atomic_inc_return(&the_controller->seqnum); ++ unlink->seqnum = atomic_inc_return_unchecked(&the_controller->seqnum); + if (unlink->seqnum == 0xffff) + usbip_uinfo("seqnum max\n"); + +@@ -988,7 +988,7 @@ static int vhci_start(struct usb_hcd *hcd) + vdev->rhport = rhport; + } + +- atomic_set(&vhci->seqnum, 0); ++ atomic_set_unchecked(&vhci->seqnum, 0); + spin_lock_init(&vhci->lock); + + +diff --git a/drivers/staging/usbip/vhci_rx.c b/drivers/staging/usbip/vhci_rx.c +index 7fd76fe..673695a 100644 +--- a/drivers/staging/usbip/vhci_rx.c ++++ b/drivers/staging/usbip/vhci_rx.c +@@ -79,7 +79,7 @@ static void vhci_recv_ret_submit(struct vhci_device *vdev, + usbip_uerr("cannot find a urb of seqnum %u\n", + pdu->base.seqnum); + usbip_uinfo("max seqnum %d\n", +- atomic_read(&the_controller->seqnum)); ++ atomic_read_unchecked(&the_controller->seqnum)); + usbip_event_add(ud, VDEV_EVENT_ERROR_TCP); + return; + } +diff --git a/drivers/staging/vme/devices/vme_user.c b/drivers/staging/vme/devices/vme_user.c +index 7891288..8e31300 100644 +--- a/drivers/staging/vme/devices/vme_user.c ++++ b/drivers/staging/vme/devices/vme_user.c +@@ -136,7 +136,7 @@ static int vme_user_ioctl(struct inode *, struct file *, unsigned int, + static int __init vme_user_probe(struct device *, int, int); + static int __exit vme_user_remove(struct device *, int, int); + +-static struct file_operations vme_user_fops = { ++static const struct file_operations vme_user_fops = { + .open = vme_user_open, + .release = vme_user_release, + .read = vme_user_read, +diff --git a/drivers/staging/vt6655/hostap.c b/drivers/staging/vt6655/hostap.c +index 58abf44..00c1fc8 100644 +--- a/drivers/staging/vt6655/hostap.c ++++ b/drivers/staging/vt6655/hostap.c +@@ -84,7 +84,7 @@ static int hostap_enable_hostapd(PSDevice pDevice, int rtnl_locked) + PSDevice apdev_priv; + struct net_device *dev = pDevice->dev; + int ret; +- const struct net_device_ops apdev_netdev_ops = { ++ net_device_ops_no_const apdev_netdev_ops = { + .ndo_start_xmit = pDevice->tx_80211, + }; + +diff --git a/drivers/staging/vt6656/hostap.c b/drivers/staging/vt6656/hostap.c +index 0c8267a..db1f363 100644 +--- a/drivers/staging/vt6656/hostap.c ++++ b/drivers/staging/vt6656/hostap.c +@@ -86,7 +86,7 @@ static int hostap_enable_hostapd(PSDevice pDevice, int rtnl_locked) + PSDevice apdev_priv; + struct net_device *dev = pDevice->dev; + int ret; +- const struct net_device_ops apdev_netdev_ops = { ++ net_device_ops_no_const apdev_netdev_ops = { + .ndo_start_xmit = pDevice->tx_80211, + }; + +diff --git a/drivers/staging/wlan-ng/hfa384x_usb.c b/drivers/staging/wlan-ng/hfa384x_usb.c +index 925678b..da7f5ed 100644 +--- a/drivers/staging/wlan-ng/hfa384x_usb.c ++++ b/drivers/staging/wlan-ng/hfa384x_usb.c +@@ -205,7 +205,7 @@ static void unlocked_usbctlx_complete(hfa384x_t *hw, hfa384x_usbctlx_t *ctlx); + + struct usbctlx_completor { + int (*complete) (struct usbctlx_completor *); +-}; ++} __no_const; + typedef struct usbctlx_completor usbctlx_completor_t; + + static int +diff --git a/drivers/telephony/ixj.c b/drivers/telephony/ixj.c +index 40de151..924f268 100644 +--- a/drivers/telephony/ixj.c ++++ b/drivers/telephony/ixj.c +@@ -4976,6 +4976,8 @@ static int ixj_daa_cid_read(IXJ *j) + bool mContinue; + char *pIn, *pOut; + ++ pax_track_stack(); ++ + if (!SCI_Prepare(j)) + return 0; + +diff --git a/drivers/uio/uio.c b/drivers/uio/uio.c +index e941367..b631f5a 100644 +--- a/drivers/uio/uio.c ++++ b/drivers/uio/uio.c +@@ -23,6 +23,7 @@ + #include + #include + #include ++#include + + #define UIO_MAX_DEVICES 255 + +@@ -30,10 +31,10 @@ struct uio_device { + struct module *owner; + struct device *dev; + int minor; +- atomic_t event; ++ atomic_unchecked_t event; + struct fasync_struct *async_queue; + wait_queue_head_t wait; +- int vma_count; ++ local_t vma_count; + struct uio_info *info; + struct kobject *map_dir; + struct kobject *portio_dir; +@@ -129,7 +130,7 @@ static ssize_t map_type_show(struct kobject *kobj, struct attribute *attr, + return entry->show(mem, buf); + } + +-static struct sysfs_ops map_sysfs_ops = { ++static const struct sysfs_ops map_sysfs_ops = { + .show = map_type_show, + }; + +@@ -217,7 +218,7 @@ static ssize_t portio_type_show(struct kobject *kobj, struct attribute *attr, + return entry->show(port, buf); + } + +-static struct sysfs_ops portio_sysfs_ops = { ++static const struct sysfs_ops portio_sysfs_ops = { + .show = portio_type_show, + }; + +@@ -255,7 +256,7 @@ static ssize_t show_event(struct device *dev, + struct uio_device *idev = dev_get_drvdata(dev); + if (idev) + return sprintf(buf, "%u\n", +- (unsigned int)atomic_read(&idev->event)); ++ (unsigned int)atomic_read_unchecked(&idev->event)); + else + return -ENODEV; + } +@@ -424,7 +425,7 @@ void uio_event_notify(struct uio_info *info) + { + struct uio_device *idev = info->uio_dev; + +- atomic_inc(&idev->event); ++ atomic_inc_unchecked(&idev->event); + wake_up_interruptible(&idev->wait); + kill_fasync(&idev->async_queue, SIGIO, POLL_IN); + } +@@ -477,7 +478,7 @@ static int uio_open(struct inode *inode, struct file *filep) + } + + listener->dev = idev; +- listener->event_count = atomic_read(&idev->event); ++ listener->event_count = atomic_read_unchecked(&idev->event); + filep->private_data = listener; + + if (idev->info->open) { +@@ -528,7 +529,7 @@ static unsigned int uio_poll(struct file *filep, poll_table *wait) + return -EIO; + + poll_wait(filep, &idev->wait, wait); +- if (listener->event_count != atomic_read(&idev->event)) ++ if (listener->event_count != atomic_read_unchecked(&idev->event)) + return POLLIN | POLLRDNORM; + return 0; + } +@@ -553,7 +554,7 @@ static ssize_t uio_read(struct file *filep, char __user *buf, + do { + set_current_state(TASK_INTERRUPTIBLE); + +- event_count = atomic_read(&idev->event); ++ event_count = atomic_read_unchecked(&idev->event); + if (event_count != listener->event_count) { + if (copy_to_user(buf, &event_count, count)) + retval = -EFAULT; +@@ -624,13 +625,13 @@ static int uio_find_mem_index(struct vm_area_struct *vma) + static void uio_vma_open(struct vm_area_struct *vma) + { + struct uio_device *idev = vma->vm_private_data; +- idev->vma_count++; ++ local_inc(&idev->vma_count); + } + + static void uio_vma_close(struct vm_area_struct *vma) + { + struct uio_device *idev = vma->vm_private_data; +- idev->vma_count--; ++ local_dec(&idev->vma_count); + } + + static int uio_vma_fault(struct vm_area_struct *vma, struct vm_fault *vmf) +@@ -840,7 +841,7 @@ int __uio_register_device(struct module *owner, + idev->owner = owner; + idev->info = info; + init_waitqueue_head(&idev->wait); +- atomic_set(&idev->event, 0); ++ atomic_set_unchecked(&idev->event, 0); + + ret = uio_get_minor(idev); + if (ret) +diff --git a/drivers/usb/atm/usbatm.c b/drivers/usb/atm/usbatm.c +index fbea856..06efea6 100644 +--- a/drivers/usb/atm/usbatm.c ++++ b/drivers/usb/atm/usbatm.c +@@ -333,7 +333,7 @@ static void usbatm_extract_one_cell(struct usbatm_data *instance, unsigned char + if (printk_ratelimit()) + atm_warn(instance, "%s: OAM not supported (vpi %d, vci %d)!\n", + __func__, vpi, vci); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + return; + } + +@@ -361,7 +361,7 @@ static void usbatm_extract_one_cell(struct usbatm_data *instance, unsigned char + if (length > ATM_MAX_AAL5_PDU) { + atm_rldbg(instance, "%s: bogus length %u (vcc: 0x%p)!\n", + __func__, length, vcc); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto out; + } + +@@ -370,14 +370,14 @@ static void usbatm_extract_one_cell(struct usbatm_data *instance, unsigned char + if (sarb->len < pdu_length) { + atm_rldbg(instance, "%s: bogus pdu_length %u (sarb->len: %u, vcc: 0x%p)!\n", + __func__, pdu_length, sarb->len, vcc); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto out; + } + + if (crc32_be(~0, skb_tail_pointer(sarb) - pdu_length, pdu_length) != 0xc704dd7b) { + atm_rldbg(instance, "%s: packet failed crc check (vcc: 0x%p)!\n", + __func__, vcc); +- atomic_inc(&vcc->stats->rx_err); ++ atomic_inc_unchecked(&vcc->stats->rx_err); + goto out; + } + +@@ -387,7 +387,7 @@ static void usbatm_extract_one_cell(struct usbatm_data *instance, unsigned char + if (printk_ratelimit()) + atm_err(instance, "%s: no memory for skb (length: %u)!\n", + __func__, length); +- atomic_inc(&vcc->stats->rx_drop); ++ atomic_inc_unchecked(&vcc->stats->rx_drop); + goto out; + } + +@@ -412,7 +412,7 @@ static void usbatm_extract_one_cell(struct usbatm_data *instance, unsigned char + + vcc->push(vcc, skb); + +- atomic_inc(&vcc->stats->rx); ++ atomic_inc_unchecked(&vcc->stats->rx); + out: + skb_trim(sarb, 0); + } +@@ -616,7 +616,7 @@ static void usbatm_tx_process(unsigned long data) + struct atm_vcc *vcc = UDSL_SKB(skb)->atm.vcc; + + usbatm_pop(vcc, skb); +- atomic_inc(&vcc->stats->tx); ++ atomic_inc_unchecked(&vcc->stats->tx); + + skb = skb_dequeue(&instance->sndqueue); + } +@@ -775,11 +775,11 @@ static int usbatm_atm_proc_read(struct atm_dev *atm_dev, loff_t * pos, char *pag + if (!left--) + return sprintf(page, + "AAL5: tx %d ( %d err ), rx %d ( %d err, %d drop )\n", +- atomic_read(&atm_dev->stats.aal5.tx), +- atomic_read(&atm_dev->stats.aal5.tx_err), +- atomic_read(&atm_dev->stats.aal5.rx), +- atomic_read(&atm_dev->stats.aal5.rx_err), +- atomic_read(&atm_dev->stats.aal5.rx_drop)); ++ atomic_read_unchecked(&atm_dev->stats.aal5.tx), ++ atomic_read_unchecked(&atm_dev->stats.aal5.tx_err), ++ atomic_read_unchecked(&atm_dev->stats.aal5.rx), ++ atomic_read_unchecked(&atm_dev->stats.aal5.rx_err), ++ atomic_read_unchecked(&atm_dev->stats.aal5.rx_drop)); + + if (!left--) { + if (instance->disconnected) +diff --git a/drivers/usb/core/hcd.c b/drivers/usb/core/hcd.c +index 24e6205..fe5a5d4 100644 +--- a/drivers/usb/core/hcd.c ++++ b/drivers/usb/core/hcd.c +@@ -2216,7 +2216,7 @@ EXPORT_SYMBOL_GPL(usb_hcd_platform_shutdown); + + #if defined(CONFIG_USB_MON) || defined(CONFIG_USB_MON_MODULE) + +-struct usb_mon_operations *mon_ops; ++const struct usb_mon_operations *mon_ops; + + /* + * The registration is unlocked. +@@ -2226,7 +2226,7 @@ struct usb_mon_operations *mon_ops; + * symbols from usbcore, usbcore gets referenced and cannot be unloaded first. + */ + +-int usb_mon_register (struct usb_mon_operations *ops) ++int usb_mon_register (const struct usb_mon_operations *ops) + { + + if (mon_ops) +diff --git a/drivers/usb/core/hcd.h b/drivers/usb/core/hcd.h +index bcbe104..9cfd1c6 100644 +--- a/drivers/usb/core/hcd.h ++++ b/drivers/usb/core/hcd.h +@@ -486,13 +486,13 @@ static inline void usbfs_cleanup(void) { } + #if defined(CONFIG_USB_MON) || defined(CONFIG_USB_MON_MODULE) + + struct usb_mon_operations { +- void (*urb_submit)(struct usb_bus *bus, struct urb *urb); +- void (*urb_submit_error)(struct usb_bus *bus, struct urb *urb, int err); +- void (*urb_complete)(struct usb_bus *bus, struct urb *urb, int status); ++ void (* const urb_submit)(struct usb_bus *bus, struct urb *urb); ++ void (* const urb_submit_error)(struct usb_bus *bus, struct urb *urb, int err); ++ void (* const urb_complete)(struct usb_bus *bus, struct urb *urb, int status); + /* void (*urb_unlink)(struct usb_bus *bus, struct urb *urb); */ + }; + +-extern struct usb_mon_operations *mon_ops; ++extern const struct usb_mon_operations *mon_ops; + + static inline void usbmon_urb_submit(struct usb_bus *bus, struct urb *urb) + { +@@ -514,7 +514,7 @@ static inline void usbmon_urb_complete(struct usb_bus *bus, struct urb *urb, + (*mon_ops->urb_complete)(bus, urb, status); + } + +-int usb_mon_register(struct usb_mon_operations *ops); ++int usb_mon_register(const struct usb_mon_operations *ops); + void usb_mon_deregister(void); + + #else +diff --git a/drivers/usb/core/message.c b/drivers/usb/core/message.c +index 409cc94..a673bad 100644 +--- a/drivers/usb/core/message.c ++++ b/drivers/usb/core/message.c +@@ -914,8 +914,8 @@ char *usb_cache_string(struct usb_device *udev, int index) + buf = kmalloc(MAX_USB_STRING_SIZE, GFP_NOIO); + if (buf) { + len = usb_string(udev, index, buf, MAX_USB_STRING_SIZE); +- if (len > 0) { +- smallbuf = kmalloc(++len, GFP_NOIO); ++ if (len++ > 0) { ++ smallbuf = kmalloc(len, GFP_NOIO); + if (!smallbuf) + return buf; + memcpy(smallbuf, buf, len); +diff --git a/drivers/usb/misc/appledisplay.c b/drivers/usb/misc/appledisplay.c +index 62ff5e7..530b74e 100644 +--- a/drivers/usb/misc/appledisplay.c ++++ b/drivers/usb/misc/appledisplay.c +@@ -178,7 +178,7 @@ static int appledisplay_bl_get_brightness(struct backlight_device *bd) + return pdata->msgdata[1]; + } + +-static struct backlight_ops appledisplay_bl_data = { ++static const struct backlight_ops appledisplay_bl_data = { + .get_brightness = appledisplay_bl_get_brightness, + .update_status = appledisplay_bl_update_status, + }; +diff --git a/drivers/usb/mon/mon_main.c b/drivers/usb/mon/mon_main.c +index e0c2db3..bd8cb66 100644 +--- a/drivers/usb/mon/mon_main.c ++++ b/drivers/usb/mon/mon_main.c +@@ -238,7 +238,7 @@ static struct notifier_block mon_nb = { + /* + * Ops + */ +-static struct usb_mon_operations mon_ops_0 = { ++static const struct usb_mon_operations mon_ops_0 = { + .urb_submit = mon_submit, + .urb_submit_error = mon_submit_error, + .urb_complete = mon_complete, +diff --git a/drivers/usb/wusbcore/wa-hc.h b/drivers/usb/wusbcore/wa-hc.h +index d6bea3e..60b250e 100644 +--- a/drivers/usb/wusbcore/wa-hc.h ++++ b/drivers/usb/wusbcore/wa-hc.h +@@ -192,7 +192,7 @@ struct wahc { + struct list_head xfer_delayed_list; + spinlock_t xfer_list_lock; + struct work_struct xfer_work; +- atomic_t xfer_id_count; ++ atomic_unchecked_t xfer_id_count; + }; + + +@@ -246,7 +246,7 @@ static inline void wa_init(struct wahc *wa) + INIT_LIST_HEAD(&wa->xfer_delayed_list); + spin_lock_init(&wa->xfer_list_lock); + INIT_WORK(&wa->xfer_work, wa_urb_enqueue_run); +- atomic_set(&wa->xfer_id_count, 1); ++ atomic_set_unchecked(&wa->xfer_id_count, 1); + } + + /** +diff --git a/drivers/usb/wusbcore/wa-xfer.c b/drivers/usb/wusbcore/wa-xfer.c +index 613a5fc..3174865 100644 +--- a/drivers/usb/wusbcore/wa-xfer.c ++++ b/drivers/usb/wusbcore/wa-xfer.c +@@ -293,7 +293,7 @@ out: + */ + static void wa_xfer_id_init(struct wa_xfer *xfer) + { +- xfer->id = atomic_add_return(1, &xfer->wa->xfer_id_count); ++ xfer->id = atomic_add_return_unchecked(1, &xfer->wa->xfer_id_count); + } + + /* +diff --git a/drivers/uwb/wlp/messages.c b/drivers/uwb/wlp/messages.c +index aa42fce..f8a828c 100644 +--- a/drivers/uwb/wlp/messages.c ++++ b/drivers/uwb/wlp/messages.c +@@ -903,7 +903,7 @@ int wlp_parse_f0(struct wlp *wlp, struct sk_buff *skb) + size_t len = skb->len; + size_t used; + ssize_t result; +- struct wlp_nonce enonce, rnonce; ++ struct wlp_nonce enonce = {{0}}, rnonce = {{0}}; + enum wlp_assc_error assc_err; + char enonce_buf[WLP_WSS_NONCE_STRSIZE]; + char rnonce_buf[WLP_WSS_NONCE_STRSIZE]; +diff --git a/drivers/uwb/wlp/sysfs.c b/drivers/uwb/wlp/sysfs.c +index 0370399..6627c94 100644 +--- a/drivers/uwb/wlp/sysfs.c ++++ b/drivers/uwb/wlp/sysfs.c +@@ -615,8 +615,7 @@ ssize_t wlp_wss_attr_store(struct kobject *kobj, struct attribute *attr, + return ret; + } + +-static +-struct sysfs_ops wss_sysfs_ops = { ++static const struct sysfs_ops wss_sysfs_ops = { + .show = wlp_wss_attr_show, + .store = wlp_wss_attr_store, + }; +diff --git a/drivers/video/atmel_lcdfb.c b/drivers/video/atmel_lcdfb.c +index 8c5e432..5ee90ea 100644 +--- a/drivers/video/atmel_lcdfb.c ++++ b/drivers/video/atmel_lcdfb.c +@@ -110,7 +110,7 @@ static int atmel_bl_get_brightness(struct backlight_device *bl) + return lcdc_readl(sinfo, ATMEL_LCDC_CONTRAST_VAL); + } + +-static struct backlight_ops atmel_lcdc_bl_ops = { ++static const struct backlight_ops atmel_lcdc_bl_ops = { + .update_status = atmel_bl_update_status, + .get_brightness = atmel_bl_get_brightness, + }; +diff --git a/drivers/video/aty/aty128fb.c b/drivers/video/aty/aty128fb.c +index e4e4d43..66bcbcc 100644 +--- a/drivers/video/aty/aty128fb.c ++++ b/drivers/video/aty/aty128fb.c +@@ -149,7 +149,7 @@ enum { + }; + + /* Must match above enum */ +-static const char *r128_family[] __devinitdata = { ++static const char *r128_family[] __devinitconst = { + "AGP", + "PCI", + "PRO AGP", +@@ -1787,7 +1787,7 @@ static int aty128_bl_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops aty128_bl_data = { ++static const struct backlight_ops aty128_bl_data = { + .get_brightness = aty128_bl_get_brightness, + .update_status = aty128_bl_update_status, + }; +diff --git a/drivers/video/aty/atyfb_base.c b/drivers/video/aty/atyfb_base.c +index 913b4a4..9295a38 100644 +--- a/drivers/video/aty/atyfb_base.c ++++ b/drivers/video/aty/atyfb_base.c +@@ -2225,7 +2225,7 @@ static int aty_bl_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops aty_bl_data = { ++static const struct backlight_ops aty_bl_data = { + .get_brightness = aty_bl_get_brightness, + .update_status = aty_bl_update_status, + }; +diff --git a/drivers/video/aty/radeon_backlight.c b/drivers/video/aty/radeon_backlight.c +index 1a056ad..221bd6a 100644 +--- a/drivers/video/aty/radeon_backlight.c ++++ b/drivers/video/aty/radeon_backlight.c +@@ -127,7 +127,7 @@ static int radeon_bl_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops radeon_bl_data = { ++static const struct backlight_ops radeon_bl_data = { + .get_brightness = radeon_bl_get_brightness, + .update_status = radeon_bl_update_status, + }; +diff --git a/drivers/video/backlight/adp5520_bl.c b/drivers/video/backlight/adp5520_bl.c +index ad05da5..3cb2cb9 100644 +--- a/drivers/video/backlight/adp5520_bl.c ++++ b/drivers/video/backlight/adp5520_bl.c +@@ -84,7 +84,7 @@ static int adp5520_bl_get_brightness(struct backlight_device *bl) + return error ? data->current_brightness : reg_val; + } + +-static struct backlight_ops adp5520_bl_ops = { ++static const struct backlight_ops adp5520_bl_ops = { + .update_status = adp5520_bl_update_status, + .get_brightness = adp5520_bl_get_brightness, + }; +diff --git a/drivers/video/backlight/adx_bl.c b/drivers/video/backlight/adx_bl.c +index 2c3bdfc..d769b0b 100644 +--- a/drivers/video/backlight/adx_bl.c ++++ b/drivers/video/backlight/adx_bl.c +@@ -61,7 +61,7 @@ static int adx_backlight_check_fb(struct fb_info *fb) + return 1; + } + +-static struct backlight_ops adx_backlight_ops = { ++static const struct backlight_ops adx_backlight_ops = { + .options = 0, + .update_status = adx_backlight_update_status, + .get_brightness = adx_backlight_get_brightness, +diff --git a/drivers/video/backlight/atmel-pwm-bl.c b/drivers/video/backlight/atmel-pwm-bl.c +index 505c082..6b6b3cc 100644 +--- a/drivers/video/backlight/atmel-pwm-bl.c ++++ b/drivers/video/backlight/atmel-pwm-bl.c +@@ -113,7 +113,7 @@ static int atmel_pwm_bl_init_pwm(struct atmel_pwm_bl *pwmbl) + return pwm_channel_enable(&pwmbl->pwmc); + } + +-static struct backlight_ops atmel_pwm_bl_ops = { ++static const struct backlight_ops atmel_pwm_bl_ops = { + .get_brightness = atmel_pwm_bl_get_intensity, + .update_status = atmel_pwm_bl_set_intensity, + }; +diff --git a/drivers/video/backlight/backlight.c b/drivers/video/backlight/backlight.c +index 5e20e6e..89025e6 100644 +--- a/drivers/video/backlight/backlight.c ++++ b/drivers/video/backlight/backlight.c +@@ -269,7 +269,7 @@ EXPORT_SYMBOL(backlight_force_update); + * ERR_PTR() or a pointer to the newly allocated device. + */ + struct backlight_device *backlight_device_register(const char *name, +- struct device *parent, void *devdata, struct backlight_ops *ops) ++ struct device *parent, void *devdata, const struct backlight_ops *ops) + { + struct backlight_device *new_bd; + int rc; +diff --git a/drivers/video/backlight/corgi_lcd.c b/drivers/video/backlight/corgi_lcd.c +index 9677494..b4bcf80 100644 +--- a/drivers/video/backlight/corgi_lcd.c ++++ b/drivers/video/backlight/corgi_lcd.c +@@ -451,7 +451,7 @@ void corgi_lcd_limit_intensity(int limit) + } + EXPORT_SYMBOL(corgi_lcd_limit_intensity); + +-static struct backlight_ops corgi_bl_ops = { ++static const struct backlight_ops corgi_bl_ops = { + .get_brightness = corgi_bl_get_intensity, + .update_status = corgi_bl_update_status, + }; +diff --git a/drivers/video/backlight/cr_bllcd.c b/drivers/video/backlight/cr_bllcd.c +index b9fe62b..2914bf1 100644 +--- a/drivers/video/backlight/cr_bllcd.c ++++ b/drivers/video/backlight/cr_bllcd.c +@@ -108,7 +108,7 @@ static int cr_backlight_get_intensity(struct backlight_device *bd) + return intensity; + } + +-static struct backlight_ops cr_backlight_ops = { ++static const struct backlight_ops cr_backlight_ops = { + .get_brightness = cr_backlight_get_intensity, + .update_status = cr_backlight_set_intensity, + }; +diff --git a/drivers/video/backlight/da903x_bl.c b/drivers/video/backlight/da903x_bl.c +index 701a108..feacfd5 100644 +--- a/drivers/video/backlight/da903x_bl.c ++++ b/drivers/video/backlight/da903x_bl.c +@@ -94,7 +94,7 @@ static int da903x_backlight_get_brightness(struct backlight_device *bl) + return data->current_brightness; + } + +-static struct backlight_ops da903x_backlight_ops = { ++static const struct backlight_ops da903x_backlight_ops = { + .update_status = da903x_backlight_update_status, + .get_brightness = da903x_backlight_get_brightness, + }; +diff --git a/drivers/video/backlight/generic_bl.c b/drivers/video/backlight/generic_bl.c +index 6d27f62..e6d348e 100644 +--- a/drivers/video/backlight/generic_bl.c ++++ b/drivers/video/backlight/generic_bl.c +@@ -70,7 +70,7 @@ void corgibl_limit_intensity(int limit) + } + EXPORT_SYMBOL(corgibl_limit_intensity); + +-static struct backlight_ops genericbl_ops = { ++static const struct backlight_ops genericbl_ops = { + .options = BL_CORE_SUSPENDRESUME, + .get_brightness = genericbl_get_intensity, + .update_status = genericbl_send_intensity, +diff --git a/drivers/video/backlight/hp680_bl.c b/drivers/video/backlight/hp680_bl.c +index 7fb4eef..f7cc528 100644 +--- a/drivers/video/backlight/hp680_bl.c ++++ b/drivers/video/backlight/hp680_bl.c +@@ -98,7 +98,7 @@ static int hp680bl_get_intensity(struct backlight_device *bd) + return current_intensity; + } + +-static struct backlight_ops hp680bl_ops = { ++static const struct backlight_ops hp680bl_ops = { + .get_brightness = hp680bl_get_intensity, + .update_status = hp680bl_set_intensity, + }; +diff --git a/drivers/video/backlight/jornada720_bl.c b/drivers/video/backlight/jornada720_bl.c +index 7aed256..db9071f 100644 +--- a/drivers/video/backlight/jornada720_bl.c ++++ b/drivers/video/backlight/jornada720_bl.c +@@ -93,7 +93,7 @@ out: + return ret; + } + +-static struct backlight_ops jornada_bl_ops = { ++static const struct backlight_ops jornada_bl_ops = { + .get_brightness = jornada_bl_get_brightness, + .update_status = jornada_bl_update_status, + .options = BL_CORE_SUSPENDRESUME, +diff --git a/drivers/video/backlight/kb3886_bl.c b/drivers/video/backlight/kb3886_bl.c +index a38fda1..939e7b8 100644 +--- a/drivers/video/backlight/kb3886_bl.c ++++ b/drivers/video/backlight/kb3886_bl.c +@@ -134,7 +134,7 @@ static int kb3886bl_get_intensity(struct backlight_device *bd) + return kb3886bl_intensity; + } + +-static struct backlight_ops kb3886bl_ops = { ++static const struct backlight_ops kb3886bl_ops = { + .get_brightness = kb3886bl_get_intensity, + .update_status = kb3886bl_send_intensity, + }; +diff --git a/drivers/video/backlight/locomolcd.c b/drivers/video/backlight/locomolcd.c +index 6b488b8..00a9591 100644 +--- a/drivers/video/backlight/locomolcd.c ++++ b/drivers/video/backlight/locomolcd.c +@@ -141,7 +141,7 @@ static int locomolcd_get_intensity(struct backlight_device *bd) + return current_intensity; + } + +-static struct backlight_ops locomobl_data = { ++static const struct backlight_ops locomobl_data = { + .get_brightness = locomolcd_get_intensity, + .update_status = locomolcd_set_intensity, + }; +diff --git a/drivers/video/backlight/mbp_nvidia_bl.c b/drivers/video/backlight/mbp_nvidia_bl.c +index 99bdfa8..3dac448 100644 +--- a/drivers/video/backlight/mbp_nvidia_bl.c ++++ b/drivers/video/backlight/mbp_nvidia_bl.c +@@ -33,7 +33,7 @@ struct dmi_match_data { + unsigned long iostart; + unsigned long iolen; + /* Backlight operations structure. */ +- struct backlight_ops backlight_ops; ++ const struct backlight_ops backlight_ops; + }; + + /* Module parameters. */ +diff --git a/drivers/video/backlight/omap1_bl.c b/drivers/video/backlight/omap1_bl.c +index cbad67e..3cf900e 100644 +--- a/drivers/video/backlight/omap1_bl.c ++++ b/drivers/video/backlight/omap1_bl.c +@@ -125,7 +125,7 @@ static int omapbl_get_intensity(struct backlight_device *dev) + return bl->current_intensity; + } + +-static struct backlight_ops omapbl_ops = { ++static const struct backlight_ops omapbl_ops = { + .get_brightness = omapbl_get_intensity, + .update_status = omapbl_update_status, + }; +diff --git a/drivers/video/backlight/progear_bl.c b/drivers/video/backlight/progear_bl.c +index 9edaf24..075786e 100644 +--- a/drivers/video/backlight/progear_bl.c ++++ b/drivers/video/backlight/progear_bl.c +@@ -54,7 +54,7 @@ static int progearbl_get_intensity(struct backlight_device *bd) + return intensity - HW_LEVEL_MIN; + } + +-static struct backlight_ops progearbl_ops = { ++static const struct backlight_ops progearbl_ops = { + .get_brightness = progearbl_get_intensity, + .update_status = progearbl_set_intensity, + }; +diff --git a/drivers/video/backlight/pwm_bl.c b/drivers/video/backlight/pwm_bl.c +index 8871662..df9e0b3 100644 +--- a/drivers/video/backlight/pwm_bl.c ++++ b/drivers/video/backlight/pwm_bl.c +@@ -56,7 +56,7 @@ static int pwm_backlight_get_brightness(struct backlight_device *bl) + return bl->props.brightness; + } + +-static struct backlight_ops pwm_backlight_ops = { ++static const struct backlight_ops pwm_backlight_ops = { + .update_status = pwm_backlight_update_status, + .get_brightness = pwm_backlight_get_brightness, + }; +diff --git a/drivers/video/backlight/tosa_bl.c b/drivers/video/backlight/tosa_bl.c +index 43edbad..e14ce4d 100644 +--- a/drivers/video/backlight/tosa_bl.c ++++ b/drivers/video/backlight/tosa_bl.c +@@ -72,7 +72,7 @@ static int tosa_bl_get_brightness(struct backlight_device *dev) + return props->brightness; + } + +-static struct backlight_ops bl_ops = { ++static const struct backlight_ops bl_ops = { + .get_brightness = tosa_bl_get_brightness, + .update_status = tosa_bl_update_status, + }; +diff --git a/drivers/video/backlight/wm831x_bl.c b/drivers/video/backlight/wm831x_bl.c +index 467bdb7..e32add3 100644 +--- a/drivers/video/backlight/wm831x_bl.c ++++ b/drivers/video/backlight/wm831x_bl.c +@@ -112,7 +112,7 @@ static int wm831x_backlight_get_brightness(struct backlight_device *bl) + return data->current_brightness; + } + +-static struct backlight_ops wm831x_backlight_ops = { ++static const struct backlight_ops wm831x_backlight_ops = { + .options = BL_CORE_SUSPENDRESUME, + .update_status = wm831x_backlight_update_status, + .get_brightness = wm831x_backlight_get_brightness, +diff --git a/drivers/video/bf54x-lq043fb.c b/drivers/video/bf54x-lq043fb.c +index e49ae5e..db4e6f7 100644 +--- a/drivers/video/bf54x-lq043fb.c ++++ b/drivers/video/bf54x-lq043fb.c +@@ -463,7 +463,7 @@ static int bl_get_brightness(struct backlight_device *bd) + return 0; + } + +-static struct backlight_ops bfin_lq043fb_bl_ops = { ++static const struct backlight_ops bfin_lq043fb_bl_ops = { + .get_brightness = bl_get_brightness, + }; + +diff --git a/drivers/video/bfin-t350mcqb-fb.c b/drivers/video/bfin-t350mcqb-fb.c +index 2c72a7c..d523e52 100644 +--- a/drivers/video/bfin-t350mcqb-fb.c ++++ b/drivers/video/bfin-t350mcqb-fb.c +@@ -381,7 +381,7 @@ static int bl_get_brightness(struct backlight_device *bd) + return 0; + } + +-static struct backlight_ops bfin_lq043fb_bl_ops = { ++static const struct backlight_ops bfin_lq043fb_bl_ops = { + .get_brightness = bl_get_brightness, + }; + +diff --git a/drivers/video/fbcmap.c b/drivers/video/fbcmap.c +index f53b9f1..958bf4e 100644 +--- a/drivers/video/fbcmap.c ++++ b/drivers/video/fbcmap.c +@@ -266,8 +266,7 @@ int fb_set_user_cmap(struct fb_cmap_user *cmap, struct fb_info *info) + rc = -ENODEV; + goto out; + } +- if (cmap->start < 0 || (!info->fbops->fb_setcolreg && +- !info->fbops->fb_setcmap)) { ++ if (!info->fbops->fb_setcolreg && !info->fbops->fb_setcmap) { + rc = -EINVAL; + goto out1; + } +diff --git a/drivers/video/fbmem.c b/drivers/video/fbmem.c +index 99bbd28..ad3829e 100644 +--- a/drivers/video/fbmem.c ++++ b/drivers/video/fbmem.c +@@ -403,7 +403,7 @@ static void fb_do_show_logo(struct fb_info *info, struct fb_image *image, + image->dx += image->width + 8; + } + } else if (rotate == FB_ROTATE_UD) { +- for (x = 0; x < num && image->dx >= 0; x++) { ++ for (x = 0; x < num && (__s32)image->dx >= 0; x++) { + info->fbops->fb_imageblit(info, image); + image->dx -= image->width + 8; + } +@@ -415,7 +415,7 @@ static void fb_do_show_logo(struct fb_info *info, struct fb_image *image, + image->dy += image->height + 8; + } + } else if (rotate == FB_ROTATE_CCW) { +- for (x = 0; x < num && image->dy >= 0; x++) { ++ for (x = 0; x < num && (__s32)image->dy >= 0; x++) { + info->fbops->fb_imageblit(info, image); + image->dy -= image->height + 8; + } +@@ -915,6 +915,8 @@ fb_set_var(struct fb_info *info, struct fb_var_screeninfo *var) + int flags = info->flags; + int ret = 0; + ++ pax_track_stack(); ++ + if (var->activate & FB_ACTIVATE_INV_MODE) { + struct fb_videomode mode1, mode2; + +@@ -1040,6 +1042,8 @@ static long do_fb_ioctl(struct fb_info *info, unsigned int cmd, + void __user *argp = (void __user *)arg; + long ret = 0; + ++ pax_track_stack(); ++ + switch (cmd) { + case FBIOGET_VSCREENINFO: + if (!lock_fb_info(info)) +@@ -1119,7 +1123,7 @@ static long do_fb_ioctl(struct fb_info *info, unsigned int cmd, + return -EFAULT; + if (con2fb.console < 1 || con2fb.console > MAX_NR_CONSOLES) + return -EINVAL; +- if (con2fb.framebuffer < 0 || con2fb.framebuffer >= FB_MAX) ++ if (con2fb.framebuffer >= FB_MAX) + return -EINVAL; + if (!registered_fb[con2fb.framebuffer]) + request_module("fb%d", con2fb.framebuffer); +diff --git a/drivers/video/geode/gx1fb_core.c b/drivers/video/geode/gx1fb_core.c +index f20eff8..3e4f622 100644 +--- a/drivers/video/geode/gx1fb_core.c ++++ b/drivers/video/geode/gx1fb_core.c +@@ -30,7 +30,7 @@ static int crt_option = 1; + static char panel_option[32] = ""; + + /* Modes relevant to the GX1 (taken from modedb.c) */ +-static const struct fb_videomode __initdata gx1_modedb[] = { ++static const struct fb_videomode __initconst gx1_modedb[] = { + /* 640x480-60 VESA */ + { NULL, 60, 640, 480, 39682, 48, 16, 33, 10, 96, 2, + 0, FB_VMODE_NONINTERLACED, FB_MODE_IS_VESA }, +diff --git a/drivers/video/gxt4500.c b/drivers/video/gxt4500.c +index 896e53d..4d87d0b 100644 +--- a/drivers/video/gxt4500.c ++++ b/drivers/video/gxt4500.c +@@ -156,7 +156,7 @@ struct gxt4500_par { + static char *mode_option; + + /* default mode: 1280x1024 @ 60 Hz, 8 bpp */ +-static const struct fb_videomode defaultmode __devinitdata = { ++static const struct fb_videomode defaultmode __devinitconst = { + .refresh = 60, + .xres = 1280, + .yres = 1024, +@@ -581,7 +581,7 @@ static int gxt4500_blank(int blank, struct fb_info *info) + return 0; + } + +-static const struct fb_fix_screeninfo gxt4500_fix __devinitdata = { ++static const struct fb_fix_screeninfo gxt4500_fix __devinitconst = { + .id = "IBM GXT4500P", + .type = FB_TYPE_PACKED_PIXELS, + .visual = FB_VISUAL_PSEUDOCOLOR, +diff --git a/drivers/video/i810/i810_accel.c b/drivers/video/i810/i810_accel.c +index f5bedee..28c6028 100644 +--- a/drivers/video/i810/i810_accel.c ++++ b/drivers/video/i810/i810_accel.c +@@ -73,6 +73,7 @@ static inline int wait_for_space(struct fb_info *info, u32 space) + } + } + printk("ringbuffer lockup!!!\n"); ++ printk("head:%u tail:%u iring.size:%u space:%u\n", head, tail, par->iring.size, space); + i810_report_error(mmio); + par->dev_flags |= LOCKUP; + info->pixmap.scan_align = 1; +diff --git a/drivers/video/i810/i810_main.c b/drivers/video/i810/i810_main.c +index 5743ea2..457f82c 100644 +--- a/drivers/video/i810/i810_main.c ++++ b/drivers/video/i810/i810_main.c +@@ -97,7 +97,7 @@ static int i810fb_blank (int blank_mode, struct fb_info *info); + static void i810fb_release_resource (struct fb_info *info, struct i810fb_par *par); + + /* PCI */ +-static const char *i810_pci_list[] __devinitdata = { ++static const char *i810_pci_list[] __devinitconst = { + "Intel(R) 810 Framebuffer Device" , + "Intel(R) 810-DC100 Framebuffer Device" , + "Intel(R) 810E Framebuffer Device" , +diff --git a/drivers/video/logo/logo_linux_clut224.ppm b/drivers/video/logo/logo_linux_clut224.ppm +index 3c14e43..eafa544 100644 +--- a/drivers/video/logo/logo_linux_clut224.ppm ++++ b/drivers/video/logo/logo_linux_clut224.ppm +@@ -1,1604 +1,1123 @@ + P3 +-# Standard 224-color Linux logo + 80 80 + 255 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 6 6 6 10 10 10 10 10 10 +- 10 10 10 6 6 6 6 6 6 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 10 10 10 14 14 14 +- 22 22 22 26 26 26 30 30 30 34 34 34 +- 30 30 30 30 30 30 26 26 26 18 18 18 +- 14 14 14 10 10 10 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 14 14 14 26 26 26 42 42 42 +- 54 54 54 66 66 66 78 78 78 78 78 78 +- 78 78 78 74 74 74 66 66 66 54 54 54 +- 42 42 42 26 26 26 18 18 18 10 10 10 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 22 22 22 42 42 42 66 66 66 86 86 86 +- 66 66 66 38 38 38 38 38 38 22 22 22 +- 26 26 26 34 34 34 54 54 54 66 66 66 +- 86 86 86 70 70 70 46 46 46 26 26 26 +- 14 14 14 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 10 10 10 26 26 26 +- 50 50 50 82 82 82 58 58 58 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 6 6 6 54 54 54 86 86 86 66 66 66 +- 38 38 38 18 18 18 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 22 22 22 50 50 50 +- 78 78 78 34 34 34 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 6 6 6 70 70 70 +- 78 78 78 46 46 46 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 18 18 18 42 42 42 82 82 82 +- 26 26 26 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 14 14 14 +- 46 46 46 34 34 34 6 6 6 2 2 6 +- 42 42 42 78 78 78 42 42 42 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 0 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 10 10 10 30 30 30 66 66 66 58 58 58 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 26 26 26 +- 86 86 86 101 101 101 46 46 46 10 10 10 +- 2 2 6 58 58 58 70 70 70 34 34 34 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 14 14 14 42 42 42 86 86 86 10 10 10 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 30 30 30 +- 94 94 94 94 94 94 58 58 58 26 26 26 +- 2 2 6 6 6 6 78 78 78 54 54 54 +- 22 22 22 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 22 22 22 62 62 62 62 62 62 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 26 26 26 +- 54 54 54 38 38 38 18 18 18 10 10 10 +- 2 2 6 2 2 6 34 34 34 82 82 82 +- 38 38 38 14 14 14 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 30 30 30 78 78 78 30 30 30 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 10 10 10 +- 10 10 10 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 78 78 78 +- 50 50 50 18 18 18 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 38 38 38 86 86 86 14 14 14 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 54 54 54 +- 66 66 66 26 26 26 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 42 42 42 82 82 82 2 2 6 2 2 6 +- 2 2 6 6 6 6 10 10 10 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 6 6 6 +- 14 14 14 10 10 10 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 18 18 18 +- 82 82 82 34 34 34 10 10 10 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 46 46 46 86 86 86 2 2 6 2 2 6 +- 6 6 6 6 6 6 22 22 22 34 34 34 +- 6 6 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 18 18 18 34 34 34 +- 10 10 10 50 50 50 22 22 22 2 2 6 +- 2 2 6 2 2 6 2 2 6 10 10 10 +- 86 86 86 42 42 42 14 14 14 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 46 46 46 86 86 86 2 2 6 2 2 6 +- 38 38 38 116 116 116 94 94 94 22 22 22 +- 22 22 22 2 2 6 2 2 6 2 2 6 +- 14 14 14 86 86 86 138 138 138 162 162 162 +-154 154 154 38 38 38 26 26 26 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 86 86 86 46 46 46 14 14 14 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 46 46 46 86 86 86 2 2 6 14 14 14 +-134 134 134 198 198 198 195 195 195 116 116 116 +- 10 10 10 2 2 6 2 2 6 6 6 6 +-101 98 89 187 187 187 210 210 210 218 218 218 +-214 214 214 134 134 134 14 14 14 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 86 86 86 50 50 50 18 18 18 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 1 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 46 46 46 86 86 86 2 2 6 54 54 54 +-218 218 218 195 195 195 226 226 226 246 246 246 +- 58 58 58 2 2 6 2 2 6 30 30 30 +-210 210 210 253 253 253 174 174 174 123 123 123 +-221 221 221 234 234 234 74 74 74 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 70 70 70 58 58 58 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 46 46 46 82 82 82 2 2 6 106 106 106 +-170 170 170 26 26 26 86 86 86 226 226 226 +-123 123 123 10 10 10 14 14 14 46 46 46 +-231 231 231 190 190 190 6 6 6 70 70 70 +- 90 90 90 238 238 238 158 158 158 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 70 70 70 58 58 58 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 1 0 0 0 +- 0 0 1 0 0 1 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 42 42 42 86 86 86 6 6 6 116 116 116 +-106 106 106 6 6 6 70 70 70 149 149 149 +-128 128 128 18 18 18 38 38 38 54 54 54 +-221 221 221 106 106 106 2 2 6 14 14 14 +- 46 46 46 190 190 190 198 198 198 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 74 74 74 62 62 62 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 1 0 0 0 +- 0 0 1 0 0 0 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 42 42 42 94 94 94 14 14 14 101 101 101 +-128 128 128 2 2 6 18 18 18 116 116 116 +-118 98 46 121 92 8 121 92 8 98 78 10 +-162 162 162 106 106 106 2 2 6 2 2 6 +- 2 2 6 195 195 195 195 195 195 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 74 74 74 62 62 62 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 1 0 0 1 +- 0 0 1 0 0 0 0 0 1 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 38 38 38 90 90 90 14 14 14 58 58 58 +-210 210 210 26 26 26 54 38 6 154 114 10 +-226 170 11 236 186 11 225 175 15 184 144 12 +-215 174 15 175 146 61 37 26 9 2 2 6 +- 70 70 70 246 246 246 138 138 138 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 70 70 70 66 66 66 26 26 26 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 38 38 38 86 86 86 14 14 14 10 10 10 +-195 195 195 188 164 115 192 133 9 225 175 15 +-239 182 13 234 190 10 232 195 16 232 200 30 +-245 207 45 241 208 19 232 195 16 184 144 12 +-218 194 134 211 206 186 42 42 42 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 50 50 50 74 74 74 30 30 30 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 34 34 34 86 86 86 14 14 14 2 2 6 +-121 87 25 192 133 9 219 162 10 239 182 13 +-236 186 11 232 195 16 241 208 19 244 214 54 +-246 218 60 246 218 38 246 215 20 241 208 19 +-241 208 19 226 184 13 121 87 25 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 50 50 50 82 82 82 34 34 34 10 10 10 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 34 34 34 82 82 82 30 30 30 61 42 6 +-180 123 7 206 145 10 230 174 11 239 182 13 +-234 190 10 238 202 15 241 208 19 246 218 74 +-246 218 38 246 215 20 246 215 20 246 215 20 +-226 184 13 215 174 15 184 144 12 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 26 26 26 94 94 94 42 42 42 14 14 14 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 78 78 78 50 50 50 104 69 6 +-192 133 9 216 158 10 236 178 12 236 186 11 +-232 195 16 241 208 19 244 214 54 245 215 43 +-246 215 20 246 215 20 241 208 19 198 155 10 +-200 144 11 216 158 10 156 118 10 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 6 6 6 90 90 90 54 54 54 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 78 78 78 46 46 46 22 22 22 +-137 92 6 210 162 10 239 182 13 238 190 10 +-238 202 15 241 208 19 246 215 20 246 215 20 +-241 208 19 203 166 17 185 133 11 210 150 10 +-216 158 10 210 150 10 102 78 10 2 2 6 +- 6 6 6 54 54 54 14 14 14 2 2 6 +- 2 2 6 62 62 62 74 74 74 30 30 30 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 34 34 34 78 78 78 50 50 50 6 6 6 +- 94 70 30 139 102 15 190 146 13 226 184 13 +-232 200 30 232 195 16 215 174 15 190 146 13 +-168 122 10 192 133 9 210 150 10 213 154 11 +-202 150 34 182 157 106 101 98 89 2 2 6 +- 2 2 6 78 78 78 116 116 116 58 58 58 +- 2 2 6 22 22 22 90 90 90 46 46 46 +- 18 18 18 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 38 38 38 86 86 86 50 50 50 6 6 6 +-128 128 128 174 154 114 156 107 11 168 122 10 +-198 155 10 184 144 12 197 138 11 200 144 11 +-206 145 10 206 145 10 197 138 11 188 164 115 +-195 195 195 198 198 198 174 174 174 14 14 14 +- 2 2 6 22 22 22 116 116 116 116 116 116 +- 22 22 22 2 2 6 74 74 74 70 70 70 +- 30 30 30 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 50 50 50 101 101 101 26 26 26 10 10 10 +-138 138 138 190 190 190 174 154 114 156 107 11 +-197 138 11 200 144 11 197 138 11 192 133 9 +-180 123 7 190 142 34 190 178 144 187 187 187 +-202 202 202 221 221 221 214 214 214 66 66 66 +- 2 2 6 2 2 6 50 50 50 62 62 62 +- 6 6 6 2 2 6 10 10 10 90 90 90 +- 50 50 50 18 18 18 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 10 10 10 34 34 34 +- 74 74 74 74 74 74 2 2 6 6 6 6 +-144 144 144 198 198 198 190 190 190 178 166 146 +-154 121 60 156 107 11 156 107 11 168 124 44 +-174 154 114 187 187 187 190 190 190 210 210 210 +-246 246 246 253 253 253 253 253 253 182 182 182 +- 6 6 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 62 62 62 +- 74 74 74 34 34 34 14 14 14 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 10 10 10 22 22 22 54 54 54 +- 94 94 94 18 18 18 2 2 6 46 46 46 +-234 234 234 221 221 221 190 190 190 190 190 190 +-190 190 190 187 187 187 187 187 187 190 190 190 +-190 190 190 195 195 195 214 214 214 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +- 82 82 82 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 14 14 14 +- 86 86 86 54 54 54 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 18 18 18 46 46 46 90 90 90 +- 46 46 46 18 18 18 6 6 6 182 182 182 +-253 253 253 246 246 246 206 206 206 190 190 190 +-190 190 190 190 190 190 190 190 190 190 190 190 +-206 206 206 231 231 231 250 250 250 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-202 202 202 14 14 14 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 42 42 42 86 86 86 42 42 42 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 14 14 14 38 38 38 74 74 74 66 66 66 +- 2 2 6 6 6 6 90 90 90 250 250 250 +-253 253 253 253 253 253 238 238 238 198 198 198 +-190 190 190 190 190 190 195 195 195 221 221 221 +-246 246 246 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 82 82 82 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 78 78 78 70 70 70 34 34 34 +- 14 14 14 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 34 34 34 66 66 66 78 78 78 6 6 6 +- 2 2 6 18 18 18 218 218 218 253 253 253 +-253 253 253 253 253 253 253 253 253 246 246 246 +-226 226 226 231 231 231 246 246 246 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 178 178 178 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 18 18 18 90 90 90 62 62 62 +- 30 30 30 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 10 10 10 26 26 26 +- 58 58 58 90 90 90 18 18 18 2 2 6 +- 2 2 6 110 110 110 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-250 250 250 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 231 231 231 18 18 18 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 18 18 18 94 94 94 +- 54 54 54 26 26 26 10 10 10 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 22 22 22 50 50 50 +- 90 90 90 26 26 26 2 2 6 2 2 6 +- 14 14 14 195 195 195 250 250 250 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-250 250 250 242 242 242 54 54 54 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 38 38 38 +- 86 86 86 50 50 50 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 14 14 14 38 38 38 82 82 82 +- 34 34 34 2 2 6 2 2 6 2 2 6 +- 42 42 42 195 195 195 246 246 246 253 253 253 +-253 253 253 253 253 253 253 253 253 250 250 250 +-242 242 242 242 242 242 250 250 250 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 250 250 250 246 246 246 238 238 238 +-226 226 226 231 231 231 101 101 101 6 6 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 38 38 38 82 82 82 42 42 42 14 14 14 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 10 10 10 26 26 26 62 62 62 66 66 66 +- 2 2 6 2 2 6 2 2 6 6 6 6 +- 70 70 70 170 170 170 206 206 206 234 234 234 +-246 246 246 250 250 250 250 250 250 238 238 238 +-226 226 226 231 231 231 238 238 238 250 250 250 +-250 250 250 250 250 250 246 246 246 231 231 231 +-214 214 214 206 206 206 202 202 202 202 202 202 +-198 198 198 202 202 202 182 182 182 18 18 18 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 62 62 62 66 66 66 30 30 30 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 14 14 14 42 42 42 82 82 82 18 18 18 +- 2 2 6 2 2 6 2 2 6 10 10 10 +- 94 94 94 182 182 182 218 218 218 242 242 242 +-250 250 250 253 253 253 253 253 253 250 250 250 +-234 234 234 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 246 246 246 +-238 238 238 226 226 226 210 210 210 202 202 202 +-195 195 195 195 195 195 210 210 210 158 158 158 +- 6 6 6 14 14 14 50 50 50 14 14 14 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 6 6 6 86 86 86 46 46 46 +- 18 18 18 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 22 22 22 54 54 54 70 70 70 2 2 6 +- 2 2 6 10 10 10 2 2 6 22 22 22 +-166 166 166 231 231 231 250 250 250 253 253 253 +-253 253 253 253 253 253 253 253 253 250 250 250 +-242 242 242 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 246 246 246 +-231 231 231 206 206 206 198 198 198 226 226 226 +- 94 94 94 2 2 6 6 6 6 38 38 38 +- 30 30 30 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 62 62 62 66 66 66 +- 26 26 26 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 74 74 74 50 50 50 2 2 6 +- 26 26 26 26 26 26 2 2 6 106 106 106 +-238 238 238 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 246 246 246 218 218 218 202 202 202 +-210 210 210 14 14 14 2 2 6 2 2 6 +- 30 30 30 22 22 22 2 2 6 2 2 6 +- 2 2 6 2 2 6 18 18 18 86 86 86 +- 42 42 42 14 14 14 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 42 42 42 90 90 90 22 22 22 2 2 6 +- 42 42 42 2 2 6 18 18 18 218 218 218 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 250 250 250 221 221 221 +-218 218 218 101 101 101 2 2 6 14 14 14 +- 18 18 18 38 38 38 10 10 10 2 2 6 +- 2 2 6 2 2 6 2 2 6 78 78 78 +- 58 58 58 22 22 22 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 54 54 54 82 82 82 2 2 6 26 26 26 +- 22 22 22 2 2 6 123 123 123 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 250 250 250 +-238 238 238 198 198 198 6 6 6 38 38 38 +- 58 58 58 26 26 26 38 38 38 2 2 6 +- 2 2 6 2 2 6 2 2 6 46 46 46 +- 78 78 78 30 30 30 10 10 10 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 10 10 10 30 30 30 +- 74 74 74 58 58 58 2 2 6 42 42 42 +- 2 2 6 22 22 22 231 231 231 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 250 250 250 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 246 246 246 46 46 46 38 38 38 +- 42 42 42 14 14 14 38 38 38 14 14 14 +- 2 2 6 2 2 6 2 2 6 6 6 6 +- 86 86 86 46 46 46 14 14 14 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 14 14 14 42 42 42 +- 90 90 90 18 18 18 18 18 18 26 26 26 +- 2 2 6 116 116 116 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 250 250 250 238 238 238 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 94 94 94 6 6 6 +- 2 2 6 2 2 6 10 10 10 34 34 34 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 74 74 74 58 58 58 22 22 22 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 10 10 10 26 26 26 66 66 66 +- 82 82 82 2 2 6 38 38 38 6 6 6 +- 14 14 14 210 210 210 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 246 246 246 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 144 144 144 2 2 6 +- 2 2 6 2 2 6 2 2 6 46 46 46 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 42 42 42 74 74 74 30 30 30 10 10 10 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 14 14 14 42 42 42 90 90 90 +- 26 26 26 6 6 6 42 42 42 2 2 6 +- 74 74 74 250 250 250 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 242 242 242 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 182 182 182 2 2 6 +- 2 2 6 2 2 6 2 2 6 46 46 46 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 10 10 10 86 86 86 38 38 38 10 10 10 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 10 10 10 26 26 26 66 66 66 82 82 82 +- 2 2 6 22 22 22 18 18 18 2 2 6 +-149 149 149 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 234 234 234 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 206 206 206 2 2 6 +- 2 2 6 2 2 6 2 2 6 38 38 38 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 6 6 6 86 86 86 46 46 46 14 14 14 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 18 18 18 46 46 46 86 86 86 18 18 18 +- 2 2 6 34 34 34 10 10 10 6 6 6 +-210 210 210 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 234 234 234 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 221 221 221 6 6 6 +- 2 2 6 2 2 6 6 6 6 30 30 30 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 82 82 82 54 54 54 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 26 26 26 66 66 66 62 62 62 2 2 6 +- 2 2 6 38 38 38 10 10 10 26 26 26 +-238 238 238 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 238 238 238 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 6 6 6 +- 2 2 6 2 2 6 10 10 10 30 30 30 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 66 66 66 58 58 58 22 22 22 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 38 38 38 78 78 78 6 6 6 2 2 6 +- 2 2 6 46 46 46 14 14 14 42 42 42 +-246 246 246 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 234 234 234 10 10 10 +- 2 2 6 2 2 6 22 22 22 14 14 14 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 66 66 66 62 62 62 22 22 22 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 50 50 50 74 74 74 2 2 6 2 2 6 +- 14 14 14 70 70 70 34 34 34 62 62 62 +-250 250 250 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 246 246 246 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 234 234 234 14 14 14 +- 2 2 6 2 2 6 30 30 30 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 66 66 66 62 62 62 22 22 22 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 54 54 54 62 62 62 2 2 6 2 2 6 +- 2 2 6 30 30 30 46 46 46 70 70 70 +-250 250 250 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 246 246 246 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 226 226 226 10 10 10 +- 2 2 6 6 6 6 30 30 30 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 66 66 66 58 58 58 22 22 22 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 22 22 22 +- 58 58 58 62 62 62 2 2 6 2 2 6 +- 2 2 6 2 2 6 30 30 30 78 78 78 +-250 250 250 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 246 246 246 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 206 206 206 2 2 6 +- 22 22 22 34 34 34 18 14 6 22 22 22 +- 26 26 26 18 18 18 6 6 6 2 2 6 +- 2 2 6 82 82 82 54 54 54 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 26 26 26 +- 62 62 62 106 106 106 74 54 14 185 133 11 +-210 162 10 121 92 8 6 6 6 62 62 62 +-238 238 238 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 246 246 246 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 158 158 158 18 18 18 +- 14 14 14 2 2 6 2 2 6 2 2 6 +- 6 6 6 18 18 18 66 66 66 38 38 38 +- 6 6 6 94 94 94 50 50 50 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 10 10 10 10 10 10 18 18 18 38 38 38 +- 78 78 78 142 134 106 216 158 10 242 186 14 +-246 190 14 246 190 14 156 118 10 10 10 10 +- 90 90 90 238 238 238 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 250 250 250 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 246 230 190 +-238 204 91 238 204 91 181 142 44 37 26 9 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 38 38 38 46 46 46 +- 26 26 26 106 106 106 54 54 54 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 14 14 14 22 22 22 +- 30 30 30 38 38 38 50 50 50 70 70 70 +-106 106 106 190 142 34 226 170 11 242 186 14 +-246 190 14 246 190 14 246 190 14 154 114 10 +- 6 6 6 74 74 74 226 226 226 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 231 231 231 250 250 250 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 228 184 62 +-241 196 14 241 208 19 232 195 16 38 30 10 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 6 6 6 30 30 30 26 26 26 +-203 166 17 154 142 90 66 66 66 26 26 26 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 18 18 18 38 38 38 58 58 58 +- 78 78 78 86 86 86 101 101 101 123 123 123 +-175 146 61 210 150 10 234 174 13 246 186 14 +-246 190 14 246 190 14 246 190 14 238 190 10 +-102 78 10 2 2 6 46 46 46 198 198 198 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 234 234 234 242 242 242 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 224 178 62 +-242 186 14 241 196 14 210 166 10 22 18 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 6 6 6 121 92 8 +-238 202 15 232 195 16 82 82 82 34 34 34 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 14 14 14 38 38 38 70 70 70 154 122 46 +-190 142 34 200 144 11 197 138 11 197 138 11 +-213 154 11 226 170 11 242 186 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-225 175 15 46 32 6 2 2 6 22 22 22 +-158 158 158 250 250 250 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 250 250 250 242 242 242 224 178 62 +-239 182 13 236 186 11 213 154 11 46 32 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 61 42 6 225 175 15 +-238 190 10 236 186 11 112 100 78 42 42 42 +- 14 14 14 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 22 22 22 54 54 54 154 122 46 213 154 11 +-226 170 11 230 174 11 226 170 11 226 170 11 +-236 178 12 242 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-241 196 14 184 144 12 10 10 10 2 2 6 +- 6 6 6 116 116 116 242 242 242 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 231 231 231 198 198 198 214 170 54 +-236 178 12 236 178 12 210 150 10 137 92 6 +- 18 14 6 2 2 6 2 2 6 2 2 6 +- 6 6 6 70 47 6 200 144 11 236 178 12 +-239 182 13 239 182 13 124 112 88 58 58 58 +- 22 22 22 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 70 70 70 180 133 36 226 170 11 +-239 182 13 242 186 14 242 186 14 246 186 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 232 195 16 98 70 6 2 2 6 +- 2 2 6 2 2 6 66 66 66 221 221 221 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 206 206 206 198 198 198 214 166 58 +-230 174 11 230 174 11 216 158 10 192 133 9 +-163 110 8 116 81 8 102 78 10 116 81 8 +-167 114 7 197 138 11 226 170 11 239 182 13 +-242 186 14 242 186 14 162 146 94 78 78 78 +- 34 34 34 14 14 14 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 30 30 30 78 78 78 190 142 34 226 170 11 +-239 182 13 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 241 196 14 203 166 17 22 18 6 +- 2 2 6 2 2 6 2 2 6 38 38 38 +-218 218 218 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-250 250 250 206 206 206 198 198 198 202 162 69 +-226 170 11 236 178 12 224 166 10 210 150 10 +-200 144 11 197 138 11 192 133 9 197 138 11 +-210 150 10 226 170 11 242 186 14 246 190 14 +-246 190 14 246 186 14 225 175 15 124 112 88 +- 62 62 62 30 30 30 14 14 14 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 78 78 78 174 135 50 224 166 10 +-239 182 13 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 241 196 14 139 102 15 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 78 78 78 250 250 250 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-250 250 250 214 214 214 198 198 198 190 150 46 +-219 162 10 236 178 12 234 174 13 224 166 10 +-216 158 10 213 154 11 213 154 11 216 158 10 +-226 170 11 239 182 13 246 190 14 246 190 14 +-246 190 14 246 190 14 242 186 14 206 162 42 +-101 101 101 58 58 58 30 30 30 14 14 14 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 74 74 74 174 135 50 216 158 10 +-236 178 12 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 241 196 14 226 184 13 +- 61 42 6 2 2 6 2 2 6 2 2 6 +- 22 22 22 238 238 238 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 226 226 226 187 187 187 180 133 36 +-216 158 10 236 178 12 239 182 13 236 178 12 +-230 174 11 226 170 11 226 170 11 230 174 11 +-236 178 12 242 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 186 14 239 182 13 +-206 162 42 106 106 106 66 66 66 34 34 34 +- 14 14 14 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 26 26 26 70 70 70 163 133 67 213 154 11 +-236 178 12 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 241 196 14 +-190 146 13 18 14 6 2 2 6 2 2 6 +- 46 46 46 246 246 246 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 221 221 221 86 86 86 156 107 11 +-216 158 10 236 178 12 242 186 14 246 186 14 +-242 186 14 239 182 13 239 182 13 242 186 14 +-242 186 14 246 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-242 186 14 225 175 15 142 122 72 66 66 66 +- 30 30 30 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 26 26 26 70 70 70 163 133 67 210 150 10 +-236 178 12 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-232 195 16 121 92 8 34 34 34 106 106 106 +-221 221 221 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-242 242 242 82 82 82 18 14 6 163 110 8 +-216 158 10 236 178 12 242 186 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 242 186 14 163 133 67 +- 46 46 46 18 18 18 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 10 10 10 +- 30 30 30 78 78 78 163 133 67 210 150 10 +-236 178 12 246 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-241 196 14 215 174 15 190 178 144 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 218 218 218 +- 58 58 58 2 2 6 22 18 6 167 114 7 +-216 158 10 236 178 12 246 186 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 186 14 242 186 14 190 150 46 +- 54 54 54 22 22 22 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 38 38 38 86 86 86 180 133 36 213 154 11 +-236 178 12 246 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 232 195 16 190 146 13 214 214 214 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 250 250 250 170 170 170 26 26 26 +- 2 2 6 2 2 6 37 26 9 163 110 8 +-219 162 10 239 182 13 246 186 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 186 14 236 178 12 224 166 10 142 122 72 +- 46 46 46 18 18 18 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 50 50 50 109 106 95 192 133 9 224 166 10 +-242 186 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-242 186 14 226 184 13 210 162 10 142 110 46 +-226 226 226 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-253 253 253 253 253 253 253 253 253 253 253 253 +-198 198 198 66 66 66 2 2 6 2 2 6 +- 2 2 6 2 2 6 50 34 6 156 107 11 +-219 162 10 239 182 13 246 186 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 242 186 14 +-234 174 13 213 154 11 154 122 46 66 66 66 +- 30 30 30 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 22 22 22 +- 58 58 58 154 121 60 206 145 10 234 174 13 +-242 186 14 246 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 186 14 236 178 12 210 162 10 163 110 8 +- 61 42 6 138 138 138 218 218 218 250 250 250 +-253 253 253 253 253 253 253 253 253 250 250 250 +-242 242 242 210 210 210 144 144 144 66 66 66 +- 6 6 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 61 42 6 163 110 8 +-216 158 10 236 178 12 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 239 182 13 230 174 11 216 158 10 +-190 142 34 124 112 88 70 70 70 38 38 38 +- 18 18 18 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 22 22 22 +- 62 62 62 168 124 44 206 145 10 224 166 10 +-236 178 12 239 182 13 242 186 14 242 186 14 +-246 186 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 236 178 12 216 158 10 175 118 6 +- 80 54 7 2 2 6 6 6 6 30 30 30 +- 54 54 54 62 62 62 50 50 50 38 38 38 +- 14 14 14 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 6 6 6 80 54 7 167 114 7 +-213 154 11 236 178 12 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 190 14 242 186 14 239 182 13 239 182 13 +-230 174 11 210 150 10 174 135 50 124 112 88 +- 82 82 82 54 54 54 34 34 34 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 18 18 18 +- 50 50 50 158 118 36 192 133 9 200 144 11 +-216 158 10 219 162 10 224 166 10 226 170 11 +-230 174 11 236 178 12 239 182 13 239 182 13 +-242 186 14 246 186 14 246 190 14 246 190 14 +-246 190 14 246 190 14 246 190 14 246 190 14 +-246 186 14 230 174 11 210 150 10 163 110 8 +-104 69 6 10 10 10 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 6 6 6 91 60 6 167 114 7 +-206 145 10 230 174 11 242 186 14 246 190 14 +-246 190 14 246 190 14 246 186 14 242 186 14 +-239 182 13 230 174 11 224 166 10 213 154 11 +-180 133 36 124 112 88 86 86 86 58 58 58 +- 38 38 38 22 22 22 10 10 10 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 14 14 14 +- 34 34 34 70 70 70 138 110 50 158 118 36 +-167 114 7 180 123 7 192 133 9 197 138 11 +-200 144 11 206 145 10 213 154 11 219 162 10 +-224 166 10 230 174 11 239 182 13 242 186 14 +-246 186 14 246 186 14 246 186 14 246 186 14 +-239 182 13 216 158 10 185 133 11 152 99 6 +-104 69 6 18 14 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 2 2 6 2 2 6 2 2 6 +- 2 2 6 6 6 6 80 54 7 152 99 6 +-192 133 9 219 162 10 236 178 12 239 182 13 +-246 186 14 242 186 14 239 182 13 236 178 12 +-224 166 10 206 145 10 192 133 9 154 121 60 +- 94 94 94 62 62 62 42 42 42 22 22 22 +- 14 14 14 6 6 6 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 18 18 18 34 34 34 58 58 58 78 78 78 +-101 98 89 124 112 88 142 110 46 156 107 11 +-163 110 8 167 114 7 175 118 6 180 123 7 +-185 133 11 197 138 11 210 150 10 219 162 10 +-226 170 11 236 178 12 236 178 12 234 174 13 +-219 162 10 197 138 11 163 110 8 130 83 6 +- 91 60 6 10 10 10 2 2 6 2 2 6 +- 18 18 18 38 38 38 38 38 38 38 38 38 +- 38 38 38 38 38 38 38 38 38 38 38 38 +- 38 38 38 38 38 38 26 26 26 2 2 6 +- 2 2 6 6 6 6 70 47 6 137 92 6 +-175 118 6 200 144 11 219 162 10 230 174 11 +-234 174 13 230 174 11 219 162 10 210 150 10 +-192 133 9 163 110 8 124 112 88 82 82 82 +- 50 50 50 30 30 30 14 14 14 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 14 14 14 22 22 22 34 34 34 +- 42 42 42 58 58 58 74 74 74 86 86 86 +-101 98 89 122 102 70 130 98 46 121 87 25 +-137 92 6 152 99 6 163 110 8 180 123 7 +-185 133 11 197 138 11 206 145 10 200 144 11 +-180 123 7 156 107 11 130 83 6 104 69 6 +- 50 34 6 54 54 54 110 110 110 101 98 89 +- 86 86 86 82 82 82 78 78 78 78 78 78 +- 78 78 78 78 78 78 78 78 78 78 78 78 +- 78 78 78 82 82 82 86 86 86 94 94 94 +-106 106 106 101 101 101 86 66 34 124 80 6 +-156 107 11 180 123 7 192 133 9 200 144 11 +-206 145 10 200 144 11 192 133 9 175 118 6 +-139 102 15 109 106 95 70 70 70 42 42 42 +- 22 22 22 10 10 10 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 6 6 6 10 10 10 +- 14 14 14 22 22 22 30 30 30 38 38 38 +- 50 50 50 62 62 62 74 74 74 90 90 90 +-101 98 89 112 100 78 121 87 25 124 80 6 +-137 92 6 152 99 6 152 99 6 152 99 6 +-138 86 6 124 80 6 98 70 6 86 66 30 +-101 98 89 82 82 82 58 58 58 46 46 46 +- 38 38 38 34 34 34 34 34 34 34 34 34 +- 34 34 34 34 34 34 34 34 34 34 34 34 +- 34 34 34 34 34 34 38 38 38 42 42 42 +- 54 54 54 82 82 82 94 86 76 91 60 6 +-134 86 6 156 107 11 167 114 7 175 118 6 +-175 118 6 167 114 7 152 99 6 121 87 25 +-101 98 89 62 62 62 34 34 34 18 18 18 +- 6 6 6 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 6 6 6 10 10 10 +- 18 18 18 22 22 22 30 30 30 42 42 42 +- 50 50 50 66 66 66 86 86 86 101 98 89 +-106 86 58 98 70 6 104 69 6 104 69 6 +-104 69 6 91 60 6 82 62 34 90 90 90 +- 62 62 62 38 38 38 22 22 22 14 14 14 +- 10 10 10 10 10 10 10 10 10 10 10 10 +- 10 10 10 10 10 10 6 6 6 10 10 10 +- 10 10 10 10 10 10 10 10 10 14 14 14 +- 22 22 22 42 42 42 70 70 70 89 81 66 +- 80 54 7 104 69 6 124 80 6 137 92 6 +-134 86 6 116 81 8 100 82 52 86 86 86 +- 58 58 58 30 30 30 14 14 14 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 10 10 10 14 14 14 +- 18 18 18 26 26 26 38 38 38 54 54 54 +- 70 70 70 86 86 86 94 86 76 89 81 66 +- 89 81 66 86 86 86 74 74 74 50 50 50 +- 30 30 30 14 14 14 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 18 18 18 34 34 34 58 58 58 +- 82 82 82 89 81 66 89 81 66 89 81 66 +- 94 86 66 94 86 76 74 74 74 50 50 50 +- 26 26 26 14 14 14 6 6 6 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 6 6 6 6 6 6 14 14 14 18 18 18 +- 30 30 30 38 38 38 46 46 46 54 54 54 +- 50 50 50 42 42 42 30 30 30 18 18 18 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 6 6 6 14 14 14 26 26 26 +- 38 38 38 50 50 50 58 58 58 58 58 58 +- 54 54 54 42 42 42 30 30 30 18 18 18 +- 10 10 10 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 6 6 6 10 10 10 14 14 14 18 18 18 +- 18 18 18 14 14 14 10 10 10 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 6 6 6 +- 14 14 14 18 18 18 22 22 22 22 22 22 +- 18 18 18 14 14 14 10 10 10 6 6 6 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 +- 0 0 0 0 0 0 0 0 0 0 0 0 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 3 3 3 0 0 0 0 0 0 ++0 0 0 0 0 0 0 0 0 0 0 0 3 3 3 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 1 1 1 0 0 0 ++0 0 0 3 3 3 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 2 1 0 2 1 0 3 2 2 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 2 2 2 0 0 0 3 4 3 26 28 28 ++37 38 37 37 38 37 14 17 19 2 2 2 0 0 0 2 2 2 ++5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 3 3 3 0 0 0 1 1 1 6 6 6 ++2 2 2 0 0 0 3 3 3 4 4 4 4 4 4 4 4 4 ++4 4 5 3 3 3 1 0 0 0 0 0 1 0 0 0 0 0 ++1 1 1 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++2 2 2 0 0 0 0 0 0 14 17 19 60 74 84 137 136 137 ++153 152 153 137 136 137 125 124 125 60 73 81 6 6 6 3 1 0 ++0 0 0 3 3 3 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 0 0 0 4 4 4 41 54 63 125 124 125 ++60 73 81 6 6 6 4 0 0 3 3 3 4 4 4 4 4 4 ++4 4 4 0 0 0 6 9 11 41 54 63 41 65 82 22 30 35 ++2 2 2 2 1 0 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 5 5 5 5 5 5 2 2 2 0 0 0 ++4 0 0 6 6 6 41 54 63 137 136 137 174 174 174 167 166 167 ++165 164 165 165 164 165 163 162 163 163 162 163 125 124 125 41 54 63 ++1 1 1 0 0 0 0 0 0 3 3 3 5 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 ++3 3 3 2 0 0 4 0 0 60 73 81 156 155 156 167 166 167 ++163 162 163 85 115 134 5 7 8 0 0 0 4 4 4 5 5 5 ++0 0 0 2 5 5 55 98 126 90 154 193 90 154 193 72 125 159 ++37 51 59 2 0 0 1 1 1 4 5 5 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 4 4 4 1 1 1 0 0 0 3 3 3 ++37 38 37 125 124 125 163 162 163 174 174 174 158 157 158 158 157 158 ++156 155 156 156 155 156 158 157 158 165 164 165 174 174 174 166 165 166 ++125 124 125 16 19 21 1 0 0 0 0 0 0 0 0 4 4 4 ++5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 1 1 1 ++0 0 0 0 0 0 37 38 37 153 152 153 174 174 174 158 157 158 ++174 174 174 163 162 163 37 38 37 4 3 3 4 0 0 1 1 1 ++0 0 0 22 40 52 101 161 196 101 161 196 90 154 193 101 161 196 ++64 123 161 14 17 19 0 0 0 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 ++5 5 5 2 2 2 0 0 0 4 0 0 24 26 27 85 115 134 ++156 155 156 174 174 174 167 166 167 156 155 156 154 153 154 157 156 157 ++156 155 156 156 155 156 155 154 155 153 152 153 158 157 158 167 166 167 ++174 174 174 156 155 156 60 74 84 16 19 21 0 0 0 0 0 0 ++1 1 1 5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 6 6 6 3 3 3 0 0 0 4 0 0 ++13 16 17 60 73 81 137 136 137 165 164 165 156 155 156 153 152 153 ++174 174 174 177 184 187 60 73 81 3 1 0 0 0 0 1 1 2 ++22 30 35 64 123 161 136 185 209 90 154 193 90 154 193 90 154 193 ++90 154 193 21 29 34 0 0 0 3 2 2 4 4 5 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 3 3 3 ++0 0 0 0 0 0 10 13 16 60 74 84 157 156 157 174 174 174 ++174 174 174 158 157 158 153 152 153 154 153 154 156 155 156 155 154 155 ++156 155 156 155 154 155 154 153 154 157 156 157 154 153 154 153 152 153 ++163 162 163 174 174 174 177 184 187 137 136 137 60 73 81 13 16 17 ++4 0 0 0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 ++5 5 5 4 4 4 1 1 1 0 0 0 3 3 3 41 54 63 ++131 129 131 174 174 174 174 174 174 174 174 174 167 166 167 174 174 174 ++190 197 201 137 136 137 24 26 27 4 0 0 16 21 25 50 82 103 ++90 154 193 136 185 209 90 154 193 101 161 196 101 161 196 101 161 196 ++31 91 132 3 6 7 0 0 0 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 2 2 2 0 0 0 4 0 0 ++4 0 0 43 57 68 137 136 137 177 184 187 174 174 174 163 162 163 ++155 154 155 155 154 155 156 155 156 155 154 155 158 157 158 165 164 165 ++167 166 167 166 165 166 163 162 163 157 156 157 155 154 155 155 154 155 ++153 152 153 156 155 156 167 166 167 174 174 174 174 174 174 131 129 131 ++41 54 63 5 5 5 0 0 0 0 0 0 3 3 3 4 4 4 ++1 1 1 0 0 0 1 0 0 26 28 28 125 124 125 174 174 174 ++177 184 187 174 174 174 174 174 174 156 155 156 131 129 131 137 136 137 ++125 124 125 24 26 27 4 0 0 41 65 82 90 154 193 136 185 209 ++136 185 209 101 161 196 53 118 160 37 112 160 90 154 193 34 86 122 ++7 12 15 0 0 0 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 3 3 3 0 0 0 0 0 0 5 5 5 37 38 37 ++125 124 125 167 166 167 174 174 174 167 166 167 158 157 158 155 154 155 ++156 155 156 156 155 156 156 155 156 163 162 163 167 166 167 155 154 155 ++137 136 137 153 152 153 156 155 156 165 164 165 163 162 163 156 155 156 ++156 155 156 156 155 156 155 154 155 158 157 158 166 165 166 174 174 174 ++167 166 167 125 124 125 37 38 37 1 0 0 0 0 0 0 0 0 ++0 0 0 24 26 27 60 74 84 158 157 158 174 174 174 174 174 174 ++166 165 166 158 157 158 125 124 125 41 54 63 13 16 17 6 6 6 ++6 6 6 37 38 37 80 127 157 136 185 209 101 161 196 101 161 196 ++90 154 193 28 67 93 6 10 14 13 20 25 13 20 25 6 10 14 ++1 1 2 4 3 3 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++1 1 1 1 0 0 4 3 3 37 38 37 60 74 84 153 152 153 ++167 166 167 167 166 167 158 157 158 154 153 154 155 154 155 156 155 156 ++157 156 157 158 157 158 167 166 167 167 166 167 131 129 131 43 57 68 ++26 28 28 37 38 37 60 73 81 131 129 131 165 164 165 166 165 166 ++158 157 158 155 154 155 156 155 156 156 155 156 156 155 156 158 157 158 ++165 164 165 174 174 174 163 162 163 60 74 84 16 19 21 13 16 17 ++60 73 81 131 129 131 174 174 174 174 174 174 167 166 167 165 164 165 ++137 136 137 60 73 81 24 26 27 4 0 0 4 0 0 16 19 21 ++52 104 138 101 161 196 136 185 209 136 185 209 90 154 193 27 99 146 ++13 20 25 4 5 7 2 5 5 4 5 7 1 1 2 0 0 0 ++4 4 4 4 4 4 3 3 3 2 2 2 2 2 2 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 3 3 3 0 0 0 ++0 0 0 13 16 17 60 73 81 137 136 137 174 174 174 166 165 166 ++158 157 158 156 155 156 157 156 157 156 155 156 155 154 155 158 157 158 ++167 166 167 174 174 174 153 152 153 60 73 81 16 19 21 4 0 0 ++4 0 0 4 0 0 6 6 6 26 28 28 60 74 84 158 157 158 ++174 174 174 166 165 166 157 156 157 155 154 155 156 155 156 156 155 156 ++155 154 155 158 157 158 167 166 167 167 166 167 131 129 131 125 124 125 ++137 136 137 167 166 167 167 166 167 174 174 174 158 157 158 125 124 125 ++16 19 21 4 0 0 4 0 0 10 13 16 49 76 92 107 159 188 ++136 185 209 136 185 209 90 154 193 26 108 161 22 40 52 6 10 14 ++2 3 3 1 1 2 1 1 2 4 4 5 4 4 5 4 4 5 ++4 4 5 2 2 1 0 0 0 0 0 0 0 0 0 2 2 2 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 3 3 3 0 0 0 1 0 0 4 0 0 ++37 51 59 131 129 131 167 166 167 167 166 167 163 162 163 157 156 157 ++157 156 157 155 154 155 153 152 153 157 156 157 167 166 167 174 174 174 ++153 152 153 125 124 125 37 38 37 4 0 0 4 0 0 4 0 0 ++4 3 3 4 3 3 4 0 0 6 6 6 4 0 0 37 38 37 ++125 124 125 174 174 174 174 174 174 165 164 165 156 155 156 154 153 154 ++156 155 156 156 155 156 155 154 155 163 162 163 158 157 158 163 162 163 ++174 174 174 174 174 174 174 174 174 125 124 125 37 38 37 0 0 0 ++4 0 0 6 9 11 41 54 63 90 154 193 136 185 209 146 190 211 ++136 185 209 37 112 160 22 40 52 6 10 14 3 6 7 1 1 2 ++1 1 2 3 3 3 1 1 2 3 3 3 4 4 4 4 4 4 ++2 2 2 2 0 0 16 19 21 37 38 37 24 26 27 0 0 0 ++0 0 0 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 ++4 4 4 0 0 0 0 0 0 0 0 0 26 28 28 120 125 127 ++158 157 158 174 174 174 165 164 165 157 156 157 155 154 155 156 155 156 ++153 152 153 153 152 153 167 166 167 174 174 174 174 174 174 125 124 125 ++37 38 37 4 0 0 0 0 0 4 0 0 4 3 3 4 4 4 ++4 4 4 4 4 4 5 5 5 4 0 0 4 0 0 4 0 0 ++4 3 3 43 57 68 137 136 137 174 174 174 174 174 174 165 164 165 ++154 153 154 153 152 153 153 152 153 153 152 153 163 162 163 174 174 174 ++174 174 174 153 152 153 60 73 81 6 6 6 4 0 0 4 3 3 ++32 43 50 80 127 157 136 185 209 146 190 211 146 190 211 90 154 193 ++28 67 93 28 67 93 40 71 93 3 6 7 1 1 2 2 5 5 ++50 82 103 79 117 143 26 37 45 0 0 0 3 3 3 1 1 1 ++0 0 0 41 54 63 137 136 137 174 174 174 153 152 153 60 73 81 ++2 0 0 0 0 0 ++4 4 4 4 4 4 4 4 4 4 4 4 6 6 6 2 2 2 ++0 0 0 2 0 0 24 26 27 60 74 84 153 152 153 174 174 174 ++174 174 174 157 156 157 154 153 154 156 155 156 154 153 154 153 152 153 ++165 164 165 174 174 174 177 184 187 137 136 137 43 57 68 6 6 6 ++4 0 0 2 0 0 3 3 3 5 5 5 5 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 6 6 6 4 3 3 ++4 0 0 4 0 0 24 26 27 60 73 81 153 152 153 174 174 174 ++174 174 174 158 157 158 158 157 158 174 174 174 174 174 174 158 157 158 ++60 74 84 24 26 27 4 0 0 4 0 0 17 23 27 59 113 148 ++136 185 209 191 222 234 146 190 211 136 185 209 31 91 132 7 11 13 ++22 40 52 101 161 196 90 154 193 6 9 11 3 4 4 43 95 132 ++136 185 209 172 205 220 55 98 126 0 0 0 0 0 0 2 0 0 ++26 28 28 153 152 153 177 184 187 167 166 167 177 184 187 165 164 165 ++37 38 37 0 0 0 ++4 4 4 4 4 4 5 5 5 5 5 5 1 1 1 0 0 0 ++13 16 17 60 73 81 137 136 137 174 174 174 174 174 174 165 164 165 ++153 152 153 153 152 153 155 154 155 154 153 154 158 157 158 174 174 174 ++177 184 187 163 162 163 60 73 81 16 19 21 4 0 0 4 0 0 ++4 3 3 4 4 4 5 5 5 5 5 5 4 4 4 5 5 5 ++5 5 5 5 5 5 5 5 5 4 4 4 4 4 4 5 5 5 ++6 6 6 4 0 0 4 0 0 4 0 0 24 26 27 60 74 84 ++166 165 166 174 174 174 177 184 187 165 164 165 125 124 125 24 26 27 ++4 0 0 4 0 0 5 5 5 50 82 103 136 185 209 172 205 220 ++146 190 211 136 185 209 26 108 161 22 40 52 7 12 15 44 81 103 ++71 116 144 28 67 93 37 51 59 41 65 82 100 139 164 101 161 196 ++90 154 193 90 154 193 28 67 93 0 0 0 0 0 0 26 28 28 ++125 124 125 167 166 167 163 162 163 153 152 153 163 162 163 174 174 174 ++85 115 134 4 0 0 ++4 4 4 5 5 5 4 4 4 1 0 0 4 0 0 34 47 55 ++125 124 125 174 174 174 174 174 174 167 166 167 157 156 157 153 152 153 ++155 154 155 155 154 155 158 157 158 166 165 166 167 166 167 154 153 154 ++125 124 125 26 28 28 4 0 0 4 0 0 4 0 0 5 5 5 ++5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 1 1 1 ++0 0 0 0 0 0 1 1 1 4 4 4 4 4 4 4 4 4 ++5 5 5 5 5 5 4 3 3 4 0 0 4 0 0 6 6 6 ++37 38 37 131 129 131 137 136 137 37 38 37 0 0 0 4 0 0 ++4 5 5 43 61 72 90 154 193 172 205 220 146 190 211 136 185 209 ++90 154 193 28 67 93 13 20 25 43 61 72 71 116 144 44 81 103 ++2 5 5 7 11 13 59 113 148 101 161 196 90 154 193 28 67 93 ++13 20 25 6 10 14 0 0 0 13 16 17 60 73 81 137 136 137 ++166 165 166 158 157 158 156 155 156 154 153 154 167 166 167 174 174 174 ++60 73 81 4 0 0 ++4 4 4 4 4 4 0 0 0 3 3 3 60 74 84 174 174 174 ++174 174 174 167 166 167 163 162 163 155 154 155 157 156 157 155 154 155 ++156 155 156 163 162 163 167 166 167 158 157 158 125 124 125 37 38 37 ++4 3 3 4 0 0 4 0 0 6 6 6 6 6 6 5 5 5 ++4 4 4 4 4 4 4 4 4 1 1 1 0 0 0 2 3 3 ++10 13 16 7 11 13 1 0 0 0 0 0 2 2 1 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 4 3 3 4 0 0 ++4 0 0 7 11 13 13 16 17 4 0 0 3 3 3 34 47 55 ++80 127 157 146 190 211 172 205 220 136 185 209 136 185 209 136 185 209 ++28 67 93 22 40 52 55 98 126 55 98 126 21 29 34 7 11 13 ++50 82 103 101 161 196 101 161 196 35 83 115 13 20 25 2 2 1 ++1 1 2 1 1 2 37 51 59 131 129 131 174 174 174 174 174 174 ++167 166 167 163 162 163 163 162 163 167 166 167 174 174 174 125 124 125 ++16 19 21 4 0 0 ++4 4 4 4 0 0 4 0 0 60 74 84 174 174 174 174 174 174 ++158 157 158 155 154 155 155 154 155 156 155 156 155 154 155 158 157 158 ++167 166 167 165 164 165 131 129 131 60 73 81 13 16 17 4 0 0 ++4 0 0 4 3 3 6 6 6 4 3 3 5 5 5 4 4 4 ++4 4 4 3 2 2 0 0 0 0 0 0 7 11 13 45 69 86 ++80 127 157 71 116 144 43 61 72 7 11 13 0 0 0 1 1 1 ++4 3 3 4 4 4 4 4 4 4 4 4 6 6 6 5 5 5 ++3 2 2 4 0 0 1 0 0 21 29 34 59 113 148 136 185 209 ++146 190 211 136 185 209 136 185 209 136 185 209 136 185 209 136 185 209 ++68 124 159 44 81 103 22 40 52 13 16 17 43 61 72 90 154 193 ++136 185 209 59 113 148 21 29 34 3 4 3 1 1 1 0 0 0 ++24 26 27 125 124 125 163 162 163 174 174 174 166 165 166 165 164 165 ++163 162 163 125 124 125 125 124 125 125 124 125 125 124 125 26 28 28 ++4 0 0 4 3 3 ++3 3 3 0 0 0 24 26 27 153 152 153 177 184 187 158 157 158 ++156 155 156 156 155 156 155 154 155 155 154 155 165 164 165 174 174 174 ++155 154 155 60 74 84 26 28 28 4 0 0 4 0 0 3 1 0 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 3 3 ++2 0 0 0 0 0 0 0 0 32 43 50 72 125 159 101 161 196 ++136 185 209 101 161 196 101 161 196 79 117 143 32 43 50 0 0 0 ++0 0 0 2 2 2 4 4 4 4 4 4 3 3 3 1 0 0 ++0 0 0 4 5 5 49 76 92 101 161 196 146 190 211 146 190 211 ++136 185 209 136 185 209 136 185 209 136 185 209 136 185 209 90 154 193 ++28 67 93 13 16 17 37 51 59 80 127 157 136 185 209 90 154 193 ++22 40 52 6 9 11 3 4 3 2 2 1 16 19 21 60 73 81 ++137 136 137 163 162 163 158 157 158 166 165 166 167 166 167 153 152 153 ++60 74 84 37 38 37 6 6 6 13 16 17 4 0 0 1 0 0 ++3 2 2 4 4 4 ++3 2 2 4 0 0 37 38 37 137 136 137 167 166 167 158 157 158 ++157 156 157 154 153 154 157 156 157 167 166 167 174 174 174 125 124 125 ++37 38 37 4 0 0 4 0 0 4 0 0 4 3 3 4 4 4 ++4 4 4 4 4 4 5 5 5 5 5 5 1 1 1 0 0 0 ++0 0 0 16 21 25 55 98 126 90 154 193 136 185 209 101 161 196 ++101 161 196 101 161 196 136 185 209 136 185 209 101 161 196 55 98 126 ++14 17 19 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ++22 40 52 90 154 193 146 190 211 146 190 211 136 185 209 136 185 209 ++136 185 209 136 185 209 136 185 209 101 161 196 35 83 115 7 11 13 ++17 23 27 59 113 148 136 185 209 101 161 196 34 86 122 7 12 15 ++2 5 5 3 4 3 6 6 6 60 73 81 131 129 131 163 162 163 ++166 165 166 174 174 174 174 174 174 163 162 163 125 124 125 41 54 63 ++13 16 17 4 0 0 4 0 0 4 0 0 1 0 0 2 2 2 ++4 4 4 4 4 4 ++1 1 1 2 1 0 43 57 68 137 136 137 153 152 153 153 152 153 ++163 162 163 156 155 156 165 164 165 167 166 167 60 74 84 6 6 6 ++4 0 0 4 0 0 5 5 5 4 4 4 4 4 4 4 4 4 ++4 5 5 6 6 6 4 3 3 0 0 0 0 0 0 11 15 18 ++40 71 93 100 139 164 101 161 196 101 161 196 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 136 185 209 136 185 209 ++101 161 196 45 69 86 6 6 6 0 0 0 17 23 27 55 98 126 ++136 185 209 146 190 211 136 185 209 136 185 209 136 185 209 136 185 209 ++136 185 209 136 185 209 90 154 193 22 40 52 7 11 13 50 82 103 ++136 185 209 136 185 209 53 118 160 22 40 52 7 11 13 2 5 5 ++3 4 3 37 38 37 125 124 125 157 156 157 166 165 166 167 166 167 ++174 174 174 174 174 174 137 136 137 60 73 81 4 0 0 4 0 0 ++4 0 0 4 0 0 5 5 5 3 3 3 3 3 3 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 41 54 63 137 136 137 125 124 125 131 129 131 ++155 154 155 167 166 167 174 174 174 60 74 84 6 6 6 4 0 0 ++4 3 3 6 6 6 4 4 4 4 4 4 4 4 4 5 5 5 ++4 4 4 1 1 1 0 0 0 3 6 7 41 65 82 72 125 159 ++101 161 196 101 161 196 101 161 196 90 154 193 90 154 193 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 136 185 209 ++136 185 209 136 185 209 80 127 157 55 98 126 101 161 196 146 190 211 ++136 185 209 136 185 209 136 185 209 101 161 196 136 185 209 101 161 196 ++136 185 209 101 161 196 35 83 115 22 30 35 101 161 196 172 205 220 ++90 154 193 28 67 93 7 11 13 2 5 5 3 4 3 13 16 17 ++85 115 134 167 166 167 174 174 174 174 174 174 174 174 174 174 174 174 ++167 166 167 60 74 84 13 16 17 4 0 0 4 0 0 4 3 3 ++6 6 6 5 5 5 4 4 4 5 5 5 4 4 4 5 5 5 ++5 5 5 5 5 5 ++1 1 1 4 0 0 41 54 63 137 136 137 137 136 137 125 124 125 ++131 129 131 167 166 167 157 156 157 37 38 37 6 6 6 4 0 0 ++6 6 6 5 5 5 4 4 4 4 4 4 4 5 5 2 2 1 ++0 0 0 0 0 0 26 37 45 58 111 146 101 161 196 101 161 196 ++101 161 196 90 154 193 90 154 193 90 154 193 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 ++101 161 196 136 185 209 136 185 209 136 185 209 146 190 211 136 185 209 ++136 185 209 101 161 196 136 185 209 136 185 209 101 161 196 136 185 209 ++101 161 196 136 185 209 136 185 209 136 185 209 136 185 209 16 89 141 ++7 11 13 2 5 5 2 5 5 13 16 17 60 73 81 154 154 154 ++174 174 174 174 174 174 174 174 174 174 174 174 163 162 163 125 124 125 ++24 26 27 4 0 0 4 0 0 4 0 0 5 5 5 5 5 5 ++4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 5 5 5 ++5 5 5 4 4 4 ++4 0 0 6 6 6 37 38 37 137 136 137 137 136 137 131 129 131 ++131 129 131 153 152 153 131 129 131 26 28 28 4 0 0 4 3 3 ++6 6 6 4 4 4 4 4 4 4 4 4 0 0 0 0 0 0 ++13 20 25 51 88 114 90 154 193 101 161 196 101 161 196 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 136 185 209 101 161 196 ++101 161 196 136 185 209 101 161 196 136 185 209 136 185 209 101 161 196 ++136 185 209 101 161 196 136 185 209 101 161 196 101 161 196 101 161 196 ++136 185 209 136 185 209 136 185 209 37 112 160 21 29 34 5 7 8 ++2 5 5 13 16 17 43 57 68 131 129 131 174 174 174 174 174 174 ++174 174 174 167 166 167 157 156 157 125 124 125 37 38 37 4 0 0 ++4 0 0 4 0 0 5 5 5 5 5 5 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 41 54 63 153 152 153 137 136 137 137 136 137 ++137 136 137 153 152 153 125 124 125 24 26 27 4 0 0 3 2 2 ++4 4 4 4 4 4 4 3 3 4 0 0 3 6 7 43 61 72 ++64 123 161 101 161 196 90 154 193 90 154 193 90 154 193 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 101 161 196 90 154 193 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 ++136 185 209 101 161 196 101 161 196 136 185 209 136 185 209 101 161 196 ++101 161 196 90 154 193 28 67 93 13 16 17 7 11 13 3 6 7 ++37 51 59 125 124 125 163 162 163 174 174 174 167 166 167 166 165 166 ++167 166 167 131 129 131 60 73 81 4 0 0 4 0 0 4 0 0 ++3 3 3 5 5 5 6 6 6 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 41 54 63 137 136 137 153 152 153 137 136 137 ++153 152 153 157 156 157 125 124 125 24 26 27 0 0 0 2 2 2 ++4 4 4 4 4 4 2 0 0 0 0 0 28 67 93 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 64 123 161 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 101 161 196 ++90 154 193 101 161 196 101 161 196 101 161 196 90 154 193 136 185 209 ++101 161 196 101 161 196 136 185 209 101 161 196 136 185 209 101 161 196 ++101 161 196 101 161 196 136 185 209 101 161 196 101 161 196 90 154 193 ++35 83 115 13 16 17 3 6 7 2 5 5 13 16 17 60 74 84 ++154 154 154 166 165 166 165 164 165 158 157 158 163 162 163 157 156 157 ++60 74 84 13 16 17 4 0 0 4 0 0 3 2 2 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 41 54 63 157 156 157 155 154 155 137 136 137 ++153 152 153 158 157 158 137 136 137 26 28 28 2 0 0 2 2 2 ++4 4 4 4 4 4 1 0 0 6 10 14 34 86 122 90 154 193 ++64 123 161 90 154 193 64 123 161 90 154 193 90 154 193 90 154 193 ++64 123 161 90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 ++136 185 209 101 161 196 136 185 209 90 154 193 26 108 161 22 40 52 ++13 16 17 5 7 8 2 5 5 2 5 5 37 38 37 165 164 165 ++174 174 174 163 162 163 154 154 154 165 164 165 167 166 167 60 73 81 ++6 6 6 4 0 0 4 0 0 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 41 54 63 156 155 156 158 157 158 153 152 153 ++156 155 156 165 164 165 137 136 137 26 28 28 0 0 0 2 2 2 ++4 4 5 4 4 4 2 0 0 7 12 15 31 96 139 64 123 161 ++90 154 193 64 123 161 90 154 193 90 154 193 64 123 161 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 ++90 154 193 90 154 193 90 154 193 101 161 196 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 101 161 196 101 161 196 136 185 209 ++101 161 196 136 185 209 26 108 161 22 40 52 7 11 13 5 7 8 ++2 5 5 2 5 5 2 5 5 2 2 1 37 38 37 158 157 158 ++174 174 174 154 154 154 156 155 156 167 166 167 165 164 165 37 38 37 ++4 0 0 4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++3 1 0 4 0 0 60 73 81 157 156 157 163 162 163 153 152 153 ++158 157 158 167 166 167 137 136 137 26 28 28 2 0 0 2 2 2 ++4 5 5 4 4 4 4 0 0 7 12 15 24 86 132 26 108 161 ++37 112 160 64 123 161 90 154 193 64 123 161 90 154 193 90 154 193 ++90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 90 154 193 ++90 154 193 101 161 196 90 154 193 101 161 196 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 136 185 209 101 161 196 136 185 209 ++90 154 193 35 83 115 13 16 17 13 16 17 7 11 13 3 6 7 ++5 7 8 6 6 6 3 4 3 2 2 1 30 32 34 154 154 154 ++167 166 167 154 154 154 154 154 154 174 174 174 165 164 165 37 38 37 ++6 6 6 4 0 0 6 6 6 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 41 54 63 163 162 163 166 165 166 154 154 154 ++163 162 163 174 174 174 137 136 137 26 28 28 0 0 0 2 2 2 ++4 5 5 4 4 5 1 1 2 6 10 14 28 67 93 18 97 151 ++18 97 151 18 97 151 26 108 161 37 112 160 37 112 160 90 154 193 ++64 123 161 90 154 193 90 154 193 90 154 193 90 154 193 101 161 196 ++90 154 193 101 161 196 101 161 196 90 154 193 101 161 196 101 161 196 ++101 161 196 101 161 196 101 161 196 136 185 209 90 154 193 16 89 141 ++13 20 25 7 11 13 5 7 8 5 7 8 2 5 5 4 5 5 ++3 4 3 4 5 5 3 4 3 0 0 0 37 38 37 158 157 158 ++174 174 174 158 157 158 158 157 158 167 166 167 174 174 174 41 54 63 ++4 0 0 3 2 2 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 60 73 81 165 164 165 174 174 174 158 157 158 ++167 166 167 174 174 174 153 152 153 26 28 28 2 0 0 2 2 2 ++4 5 5 4 4 4 4 0 0 7 12 15 10 87 144 10 87 144 ++18 97 151 18 97 151 18 97 151 26 108 161 26 108 161 26 108 161 ++26 108 161 37 112 160 53 118 160 90 154 193 90 154 193 90 154 193 ++90 154 193 90 154 193 101 161 196 101 161 196 101 161 196 101 161 196 ++101 161 196 136 185 209 90 154 193 26 108 161 22 40 52 13 16 17 ++7 11 13 3 6 7 5 7 8 5 7 8 2 5 5 4 5 5 ++4 5 5 6 6 6 3 4 3 0 0 0 30 32 34 158 157 158 ++174 174 174 156 155 156 155 154 155 165 164 165 154 153 154 37 38 37 ++4 0 0 4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 167 166 167 174 174 174 163 162 163 ++174 174 174 174 174 174 153 152 153 26 28 28 0 0 0 3 3 3 ++5 5 5 4 4 4 1 1 2 7 12 15 28 67 93 18 97 151 ++18 97 151 18 97 151 18 97 151 18 97 151 18 97 151 26 108 161 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++90 154 193 26 108 161 90 154 193 90 154 193 90 154 193 101 161 196 ++101 161 196 26 108 161 22 40 52 13 16 17 7 11 13 2 5 5 ++2 5 5 6 6 6 2 5 5 4 5 5 4 5 5 4 5 5 ++3 4 3 5 5 5 3 4 3 2 0 0 30 32 34 137 136 137 ++153 152 153 137 136 137 131 129 131 137 136 137 131 129 131 37 38 37 ++4 0 0 4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 60 73 81 167 166 167 174 174 174 166 165 166 ++174 174 174 177 184 187 153 152 153 30 32 34 1 0 0 3 3 3 ++5 5 5 4 3 3 4 0 0 7 12 15 10 87 144 10 87 144 ++18 97 151 18 97 151 18 97 151 26 108 161 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 90 154 193 90 154 193 26 108 161 ++35 83 115 13 16 17 7 11 13 5 7 8 3 6 7 5 7 8 ++2 5 5 6 6 6 4 5 5 4 5 5 3 4 3 4 5 5 ++3 4 3 6 6 6 3 4 3 0 0 0 26 28 28 125 124 125 ++131 129 131 125 124 125 125 124 125 131 129 131 131 129 131 37 38 37 ++4 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++3 1 0 4 0 0 60 73 81 174 174 174 177 184 187 167 166 167 ++174 174 174 177 184 187 153 152 153 30 32 34 0 0 0 3 3 3 ++5 5 5 4 4 4 1 1 2 6 10 14 28 67 93 18 97 151 ++18 97 151 18 97 151 18 97 151 18 97 151 18 97 151 26 108 161 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++26 108 161 90 154 193 26 108 161 26 108 161 24 86 132 13 20 25 ++7 11 13 13 20 25 22 40 52 5 7 8 3 4 3 3 4 3 ++4 5 5 3 4 3 4 5 5 3 4 3 4 5 5 3 4 3 ++4 4 4 5 5 5 3 3 3 2 0 0 26 28 28 125 124 125 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 60 73 81 174 174 174 177 184 187 174 174 174 ++174 174 174 190 197 201 157 156 157 30 32 34 1 0 0 3 3 3 ++5 5 5 4 3 3 4 0 0 7 12 15 10 87 144 10 87 144 ++18 97 151 19 95 150 19 95 150 18 97 151 18 97 151 26 108 161 ++18 97 151 26 108 161 26 108 161 26 108 161 26 108 161 90 154 193 ++26 108 161 26 108 161 26 108 161 22 40 52 2 5 5 3 4 3 ++28 67 93 37 112 160 34 86 122 2 5 5 3 4 3 3 4 3 ++3 4 3 3 4 3 3 4 3 2 2 1 3 4 3 4 4 4 ++4 5 5 5 5 5 3 3 3 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 174 174 174 177 184 187 174 174 174 ++174 174 174 190 197 201 158 157 158 30 32 34 0 0 0 2 2 2 ++5 5 5 4 4 4 1 1 2 6 10 14 28 67 93 18 97 151 ++10 87 144 19 95 150 19 95 150 18 97 151 18 97 151 18 97 151 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++18 97 151 22 40 52 2 5 5 2 2 1 22 40 52 26 108 161 ++90 154 193 37 112 160 22 40 52 3 4 3 13 20 25 22 30 35 ++3 6 7 1 1 1 2 2 2 6 9 11 5 5 5 4 3 3 ++4 4 4 5 5 5 3 3 3 2 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++1 1 1 4 0 0 60 73 81 177 184 187 193 200 203 174 174 174 ++177 184 187 193 200 203 163 162 163 30 32 34 4 0 0 2 2 2 ++5 5 5 4 3 3 4 0 0 6 10 14 24 86 132 10 87 144 ++10 87 144 10 87 144 19 95 150 19 95 150 19 95 150 18 97 151 ++26 108 161 26 108 161 26 108 161 90 154 193 26 108 161 28 67 93 ++6 10 14 2 5 5 13 20 25 24 86 132 37 112 160 90 154 193 ++10 87 144 7 12 15 2 5 5 28 67 93 37 112 160 28 67 93 ++2 2 1 7 12 15 35 83 115 28 67 93 3 6 7 1 0 0 ++4 4 4 5 5 5 3 3 3 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 174 174 174 190 197 201 174 174 174 ++177 184 187 193 200 203 163 162 163 30 32 34 0 0 0 2 2 2 ++5 5 5 4 4 4 1 1 2 6 10 14 28 67 93 10 87 144 ++10 87 144 16 89 141 19 95 150 10 87 144 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 28 67 93 6 10 14 1 1 2 ++7 12 15 28 67 93 26 108 161 16 89 141 24 86 132 21 29 34 ++3 4 3 21 29 34 37 112 160 37 112 160 27 99 146 21 29 34 ++21 29 34 26 108 161 90 154 193 35 83 115 1 1 2 2 0 0 ++4 4 4 5 5 5 3 3 3 2 0 0 26 28 28 125 124 125 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++3 1 0 4 0 0 60 73 81 193 200 203 193 200 203 174 174 174 ++190 197 201 193 200 203 165 164 165 37 38 37 4 0 0 2 2 2 ++5 5 5 4 3 3 4 0 0 6 10 14 24 86 132 10 87 144 ++10 87 144 10 87 144 16 89 141 18 97 151 18 97 151 10 87 144 ++24 86 132 24 86 132 13 20 25 4 5 7 4 5 7 22 40 52 ++18 97 151 37 112 160 26 108 161 7 12 15 1 1 1 0 0 0 ++28 67 93 37 112 160 26 108 161 28 67 93 22 40 52 28 67 93 ++26 108 161 90 154 193 26 108 161 10 87 144 0 0 0 2 0 0 ++4 4 4 5 5 5 3 3 3 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 60 73 81 174 174 174 193 200 203 174 174 174 ++190 197 201 193 200 203 165 164 165 30 32 34 0 0 0 2 2 2 ++5 5 5 4 4 4 1 1 2 6 10 14 28 67 93 10 87 144 ++10 87 144 10 87 144 10 87 144 18 97 151 28 67 93 6 10 14 ++0 0 0 1 1 2 4 5 7 13 20 25 16 89 141 26 108 161 ++26 108 161 26 108 161 24 86 132 6 9 11 2 3 3 22 40 52 ++37 112 160 16 89 141 22 40 52 28 67 93 26 108 161 26 108 161 ++90 154 193 26 108 161 26 108 161 28 67 93 1 1 1 4 0 0 ++4 4 4 5 5 5 3 3 3 4 0 0 26 28 28 124 126 130 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 193 200 203 193 200 203 174 174 174 ++193 200 203 193 200 203 167 166 167 37 38 37 4 0 0 2 2 2 ++5 5 5 4 4 4 4 0 0 6 10 14 28 67 93 10 87 144 ++10 87 144 10 87 144 18 97 151 10 87 144 13 20 25 4 5 7 ++1 1 2 1 1 1 22 40 52 26 108 161 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 24 86 132 22 40 52 22 40 52 ++22 40 52 22 40 52 10 87 144 26 108 161 26 108 161 26 108 161 ++26 108 161 26 108 161 90 154 193 10 87 144 0 0 0 4 0 0 ++4 4 4 5 5 5 3 3 3 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 60 73 81 174 174 174 220 221 221 174 174 174 ++190 197 201 205 212 215 167 166 167 30 32 34 0 0 0 2 2 2 ++5 5 5 4 4 4 1 1 2 6 10 14 28 67 93 10 87 144 ++10 87 144 10 87 144 10 87 144 10 87 144 22 40 52 1 1 2 ++2 0 0 1 1 2 24 86 132 26 108 161 26 108 161 26 108 161 ++26 108 161 19 95 150 16 89 141 10 87 144 22 40 52 22 40 52 ++10 87 144 26 108 161 37 112 160 26 108 161 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 28 67 93 2 0 0 3 1 0 ++4 4 4 5 5 5 3 3 3 2 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 220 221 221 190 197 201 174 174 174 ++193 200 203 193 200 203 174 174 174 37 38 37 4 0 0 2 2 2 ++5 5 5 4 4 4 3 2 2 1 1 2 13 20 25 10 87 144 ++10 87 144 10 87 144 10 87 144 10 87 144 10 87 144 13 20 25 ++13 20 25 22 40 52 10 87 144 18 97 151 18 97 151 26 108 161 ++10 87 144 13 20 25 6 10 14 21 29 34 24 86 132 18 97 151 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++26 108 161 90 154 193 18 97 151 13 20 25 0 0 0 4 3 3 ++4 4 4 5 5 5 3 3 3 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 60 73 81 174 174 174 220 221 221 174 174 174 ++190 197 201 220 221 221 167 166 167 30 32 34 1 0 0 2 2 2 ++5 5 5 4 4 4 4 4 5 2 5 5 4 5 7 13 20 25 ++28 67 93 10 87 144 10 87 144 10 87 144 10 87 144 10 87 144 ++10 87 144 10 87 144 18 97 151 10 87 144 18 97 151 18 97 151 ++28 67 93 2 3 3 0 0 0 28 67 93 26 108 161 26 108 161 ++26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 26 108 161 ++26 108 161 10 87 144 13 20 25 1 1 2 3 2 2 4 4 4 ++4 4 4 5 5 5 3 3 3 2 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 220 221 221 190 197 201 174 174 174 ++193 200 203 193 200 203 174 174 174 26 28 28 4 0 0 4 3 3 ++5 5 5 4 4 4 4 4 4 4 4 5 1 1 2 2 5 5 ++4 5 7 22 40 52 10 87 144 10 87 144 18 97 151 10 87 144 ++10 87 144 10 87 144 10 87 144 10 87 144 10 87 144 18 97 151 ++10 87 144 28 67 93 22 40 52 10 87 144 26 108 161 18 97 151 ++18 97 151 18 97 151 26 108 161 26 108 161 26 108 161 26 108 161 ++22 40 52 1 1 2 0 0 0 2 3 3 4 4 4 4 4 4 ++4 4 4 5 5 5 4 4 4 0 0 0 26 28 28 131 129 131 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 60 73 81 174 174 174 220 221 221 174 174 174 ++190 197 201 220 221 221 190 197 201 41 54 63 4 0 0 2 2 2 ++6 6 6 4 4 4 4 4 4 4 4 5 4 4 5 3 3 3 ++1 1 2 1 1 2 6 10 14 22 40 52 10 87 144 18 97 151 ++18 97 151 10 87 144 10 87 144 10 87 144 18 97 151 10 87 144 ++10 87 144 18 97 151 26 108 161 18 97 151 18 97 151 10 87 144 ++26 108 161 26 108 161 26 108 161 10 87 144 28 67 93 6 10 14 ++1 1 2 1 1 2 4 3 3 4 4 5 4 4 4 4 4 4 ++5 5 5 5 5 5 1 1 1 4 0 0 37 51 59 137 136 137 ++137 136 137 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 4 0 0 60 73 81 220 221 221 193 200 203 174 174 174 ++193 200 203 193 200 203 220 221 221 137 136 137 13 16 17 4 0 0 ++2 2 2 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 ++4 4 5 4 3 3 1 1 2 4 5 7 13 20 25 28 67 93 ++10 87 144 10 87 144 10 87 144 10 87 144 10 87 144 10 87 144 ++10 87 144 18 97 151 18 97 151 10 87 144 18 97 151 26 108 161 ++26 108 161 18 97 151 28 67 93 6 10 14 0 0 0 0 0 0 ++2 3 3 4 5 5 4 4 5 4 4 4 4 4 4 5 5 5 ++3 3 3 1 1 1 0 0 0 16 19 21 125 124 125 137 136 137 ++131 129 131 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 60 73 81 174 174 174 220 221 221 174 174 174 ++193 200 203 190 197 201 220 221 221 220 221 221 153 152 153 30 32 34 ++0 0 0 0 0 0 2 2 2 4 4 4 4 4 4 4 4 4 ++4 4 4 4 5 5 4 5 7 1 1 2 1 1 2 4 5 7 ++13 20 25 28 67 93 10 87 144 18 97 151 10 87 144 10 87 144 ++10 87 144 10 87 144 10 87 144 18 97 151 26 108 161 18 97 151 ++28 67 93 7 12 15 0 0 0 0 0 0 2 2 1 4 4 4 ++4 5 5 4 5 5 4 4 4 4 4 4 3 3 3 0 0 0 ++0 0 0 0 0 0 37 38 37 125 124 125 158 157 158 131 129 131 ++125 124 125 125 124 125 125 124 125 137 136 137 131 129 131 37 38 37 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 3 3 4 0 0 41 54 63 193 200 203 220 221 221 174 174 174 ++193 200 203 193 200 203 193 200 203 220 221 221 244 246 246 193 200 203 ++120 125 127 5 5 5 1 0 0 0 0 0 1 1 1 4 4 4 ++4 4 4 4 4 4 4 5 5 4 5 5 4 4 5 1 1 2 ++4 5 7 4 5 7 22 40 52 10 87 144 10 87 144 10 87 144 ++10 87 144 10 87 144 18 97 151 10 87 144 10 87 144 13 20 25 ++4 5 7 2 3 3 1 1 2 4 4 4 4 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 1 1 1 0 0 0 1 1 2 ++24 26 27 60 74 84 153 152 153 163 162 163 137 136 137 125 124 125 ++125 124 125 125 124 125 125 124 125 137 136 137 125 124 125 26 28 28 ++0 0 0 3 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 0 0 6 6 6 26 28 28 156 155 156 220 221 221 220 221 221 ++174 174 174 193 200 203 193 200 203 193 200 203 205 212 215 220 221 221 ++220 221 221 167 166 167 60 73 81 7 11 13 0 0 0 0 0 0 ++3 3 3 4 4 4 4 4 4 4 4 4 4 4 5 4 4 5 ++4 4 5 1 1 2 1 1 2 4 5 7 22 40 52 10 87 144 ++10 87 144 10 87 144 10 87 144 22 40 52 4 5 7 1 1 2 ++1 1 2 4 4 5 4 4 4 4 4 4 4 4 4 4 4 4 ++5 5 5 2 2 2 0 0 0 4 0 0 16 19 21 60 73 81 ++137 136 137 167 166 167 158 157 158 137 136 137 131 129 131 131 129 131 ++125 124 125 125 124 125 131 129 131 155 154 155 60 74 84 5 7 8 ++0 0 0 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++5 5 5 4 0 0 4 0 0 60 73 81 193 200 203 220 221 221 ++193 200 203 193 200 203 193 200 203 193 200 203 205 212 215 220 221 221 ++220 221 221 220 221 221 220 221 221 137 136 137 43 57 68 6 6 6 ++4 0 0 1 1 1 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 5 4 4 5 3 2 2 1 1 2 2 5 5 13 20 25 ++22 40 52 22 40 52 13 20 25 2 3 3 1 1 2 3 3 3 ++4 5 7 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++1 1 1 0 0 0 2 3 3 41 54 63 131 129 131 166 165 166 ++166 165 166 155 154 155 153 152 153 137 136 137 137 136 137 125 124 125 ++125 124 125 137 136 137 137 136 137 125 124 125 37 38 37 4 3 3 ++4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 3 3 6 6 6 6 6 6 13 16 17 60 73 81 167 166 167 ++220 221 221 220 221 221 220 221 221 193 200 203 193 200 203 193 200 203 ++205 212 215 220 221 221 220 221 221 244 246 246 205 212 215 125 124 125 ++24 26 27 0 0 0 0 0 0 2 2 2 5 5 5 5 5 5 ++4 4 4 4 4 4 4 4 4 4 4 5 1 1 2 4 5 7 ++4 5 7 4 5 7 1 1 2 3 2 2 4 4 5 4 4 4 ++4 4 4 4 4 4 5 5 5 4 4 4 0 0 0 0 0 0 ++2 0 0 26 28 28 125 124 125 174 174 174 174 174 174 166 165 166 ++156 155 156 153 152 153 137 136 137 137 136 137 131 129 131 137 136 137 ++137 136 137 137 136 137 60 74 84 30 32 34 4 0 0 4 0 0 ++5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++5 5 5 6 6 6 4 0 0 4 0 0 6 6 6 26 28 28 ++125 124 125 174 174 174 220 221 221 220 221 221 220 221 221 193 200 203 ++205 212 215 220 221 221 205 212 215 220 221 221 220 221 221 244 246 246 ++193 200 203 60 74 84 13 16 17 4 0 0 0 0 0 3 3 3 ++5 5 5 5 5 5 4 4 4 4 4 4 4 4 5 3 3 3 ++1 1 2 3 3 3 4 4 5 4 4 5 4 4 4 4 4 4 ++5 5 5 5 5 5 2 2 2 0 0 0 0 0 0 13 16 17 ++60 74 84 174 174 174 193 200 203 174 174 174 167 166 167 163 162 163 ++153 152 153 153 152 153 137 136 137 137 136 137 153 152 153 137 136 137 ++125 124 125 41 54 63 24 26 27 4 0 0 4 0 0 5 5 5 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 3 3 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 ++6 6 6 37 38 37 131 129 131 220 221 221 220 221 221 220 221 221 ++193 200 203 193 200 203 220 221 221 205 212 215 220 221 221 244 246 246 ++244 246 246 244 246 246 174 174 174 41 54 63 0 0 0 0 0 0 ++0 0 0 4 4 4 5 5 5 5 5 5 4 4 4 4 4 5 ++4 4 5 4 4 5 4 4 4 4 4 4 6 6 6 6 6 6 ++3 3 3 0 0 0 2 0 0 13 16 17 60 73 81 156 155 156 ++220 221 221 193 200 203 174 174 174 165 164 165 163 162 163 154 153 154 ++153 152 153 153 152 153 158 157 158 163 162 163 137 136 137 60 73 81 ++13 16 17 4 0 0 4 0 0 4 3 3 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++5 5 5 4 3 3 4 3 3 6 6 6 6 6 6 6 6 6 ++6 6 6 6 6 6 6 6 6 37 38 37 167 166 167 244 246 246 ++244 246 246 220 221 221 205 212 215 205 212 215 220 221 221 193 200 203 ++220 221 221 244 246 246 244 246 246 244 246 246 137 136 137 37 38 37 ++3 2 2 0 0 0 1 1 1 5 5 5 5 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 4 4 4 1 1 1 ++0 0 0 5 5 5 43 57 68 153 152 153 193 200 203 220 221 221 ++177 184 187 174 174 174 167 166 167 166 165 166 158 157 158 157 156 157 ++158 157 158 166 165 166 156 155 156 85 115 134 13 16 17 4 0 0 ++4 0 0 4 0 0 5 5 5 5 5 5 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++5 5 5 4 3 3 6 6 6 6 6 6 4 0 0 6 6 6 ++6 6 6 6 6 6 6 6 6 6 6 6 13 16 17 60 73 81 ++177 184 187 220 221 221 220 221 221 220 221 221 205 212 215 220 221 221 ++220 221 221 205 212 215 220 221 221 244 246 246 244 246 246 205 212 215 ++125 124 125 30 32 34 0 0 0 0 0 0 2 2 2 5 5 5 ++4 4 4 4 4 4 4 4 4 1 1 1 0 0 0 1 0 0 ++37 38 37 131 129 131 205 212 215 220 221 221 193 200 203 174 174 174 ++174 174 174 174 174 174 167 166 167 165 164 165 166 165 166 167 166 167 ++158 157 158 125 124 125 37 38 37 4 0 0 4 0 0 4 0 0 ++4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 5 5 5 4 3 3 4 3 3 6 6 6 6 6 6 ++4 0 0 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 ++26 28 28 125 124 125 205 212 215 220 221 221 220 221 221 220 221 221 ++205 212 215 220 221 221 205 212 215 220 221 221 220 221 221 244 246 246 ++244 246 246 190 197 201 60 74 84 16 19 21 4 0 0 0 0 0 ++0 0 0 0 0 0 0 0 0 0 0 0 16 19 21 120 125 127 ++177 184 187 220 221 221 205 212 215 177 184 187 174 174 174 177 184 187 ++174 174 174 174 174 174 167 166 167 174 174 174 166 165 166 137 136 137 ++60 73 81 13 16 17 4 0 0 4 0 0 4 3 3 6 6 6 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++5 5 5 4 3 3 5 5 5 4 3 3 6 6 6 4 0 0 ++6 6 6 6 6 6 4 0 0 6 6 6 4 0 0 6 6 6 ++6 6 6 6 6 6 37 38 37 137 136 137 193 200 203 220 221 221 ++220 221 221 205 212 215 220 221 221 205 212 215 205 212 215 220 221 221 ++220 221 221 220 221 221 244 246 246 166 165 166 43 57 68 2 2 2 ++0 0 0 4 0 0 16 19 21 60 73 81 157 156 157 202 210 214 ++220 221 221 193 200 203 177 184 187 177 184 187 177 184 187 174 174 174 ++174 174 174 174 174 174 174 174 174 157 156 157 60 74 84 24 26 27 ++4 0 0 4 0 0 4 0 0 6 6 6 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 5 5 5 4 3 3 5 5 5 6 6 6 ++6 6 6 4 0 0 6 6 6 6 6 6 6 6 6 4 0 0 ++4 0 0 4 0 0 6 6 6 24 26 27 60 73 81 167 166 167 ++220 221 221 220 221 221 220 221 221 205 212 215 205 212 215 205 212 215 ++205 212 215 220 221 221 220 221 221 220 221 221 205 212 215 137 136 137 ++60 74 84 125 124 125 137 136 137 190 197 201 220 221 221 193 200 203 ++177 184 187 177 184 187 177 184 187 174 174 174 174 174 174 177 184 187 ++190 197 201 174 174 174 125 124 125 37 38 37 6 6 6 4 0 0 ++4 0 0 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 5 5 5 5 5 5 4 3 3 6 6 6 ++4 0 0 6 6 6 6 6 6 6 6 6 4 0 0 6 6 6 ++6 6 6 6 6 6 4 0 0 4 0 0 6 6 6 6 6 6 ++125 124 125 193 200 203 244 246 246 220 221 221 205 212 215 205 212 215 ++205 212 215 193 200 203 205 212 215 205 212 215 220 221 221 220 221 221 ++193 200 203 193 200 203 205 212 215 193 200 203 193 200 203 177 184 187 ++190 197 201 190 197 201 174 174 174 190 197 201 193 200 203 190 197 201 ++153 152 153 60 73 81 4 0 0 4 0 0 4 0 0 3 2 2 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 4 3 3 ++6 6 6 4 3 3 4 3 3 4 3 3 6 6 6 6 6 6 ++4 0 0 6 6 6 6 6 6 6 6 6 4 0 0 4 0 0 ++4 0 0 26 28 28 131 129 131 220 221 221 244 246 246 220 221 221 ++205 212 215 193 200 203 205 212 215 193 200 203 193 200 203 205 212 215 ++220 221 221 193 200 203 193 200 203 193 200 203 190 197 201 174 174 174 ++174 174 174 190 197 201 193 200 203 193 200 203 167 166 167 125 124 125 ++6 6 6 4 0 0 4 0 0 4 3 3 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 ++5 5 5 4 3 3 5 5 5 6 6 6 4 3 3 5 5 5 ++6 6 6 6 6 6 4 0 0 6 6 6 6 6 6 6 6 6 ++4 0 0 4 0 0 6 6 6 41 54 63 158 157 158 220 221 221 ++220 221 221 220 221 221 193 200 203 193 200 203 193 200 203 190 197 201 ++190 197 201 190 197 201 190 197 201 190 197 201 174 174 174 193 200 203 ++193 200 203 220 221 221 174 174 174 125 124 125 37 38 37 4 0 0 ++4 0 0 4 3 3 6 6 6 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 4 3 3 4 3 3 4 3 3 5 5 5 ++4 3 3 6 6 6 5 5 5 4 3 3 6 6 6 6 6 6 ++6 6 6 6 6 6 4 0 0 4 0 0 13 16 17 60 73 81 ++174 174 174 220 221 221 220 221 221 205 212 215 190 197 201 174 174 174 ++193 200 203 174 174 174 190 197 201 174 174 174 193 200 203 220 221 221 ++193 200 203 131 129 131 37 38 37 6 6 6 4 0 0 4 0 0 ++6 6 6 6 6 6 4 3 3 5 5 5 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 5 5 5 ++5 5 5 4 3 3 4 3 3 5 5 5 4 3 3 4 3 3 ++5 5 5 6 6 6 6 6 6 4 0 0 6 6 6 6 6 6 ++6 6 6 125 124 125 174 174 174 220 221 221 220 221 221 193 200 203 ++193 200 203 193 200 203 193 200 203 193 200 203 220 221 221 158 157 158 ++60 73 81 6 6 6 4 0 0 4 0 0 5 5 5 6 6 6 ++5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 5 5 5 4 3 3 5 5 5 4 3 3 ++5 5 5 5 5 5 6 6 6 6 6 6 4 0 0 4 0 0 ++4 0 0 4 0 0 26 28 28 125 124 125 174 174 174 193 200 203 ++193 200 203 174 174 174 193 200 203 167 166 167 125 124 125 6 6 6 ++6 6 6 6 6 6 4 0 0 6 6 6 6 6 6 5 5 5 ++4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 ++4 3 3 6 6 6 4 0 0 6 6 6 6 6 6 6 6 6 ++6 6 6 4 0 0 4 0 0 6 6 6 37 38 37 125 124 125 ++153 152 153 131 129 131 125 124 125 37 38 37 6 6 6 6 6 6 ++6 6 6 4 0 0 6 6 6 6 6 6 4 3 3 5 5 5 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 5 5 5 4 3 3 5 5 5 4 3 3 ++6 6 6 6 6 6 4 0 0 4 0 0 6 6 6 6 6 6 ++24 26 27 24 26 27 6 6 6 6 6 6 6 6 6 4 0 0 ++6 6 6 6 6 6 4 0 0 6 6 6 5 5 5 4 3 3 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 5 5 5 4 3 3 5 5 5 6 6 6 ++4 0 0 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 ++6 6 6 6 6 6 6 6 6 4 0 0 6 6 6 6 6 6 ++4 0 0 6 6 6 6 6 6 4 3 3 5 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 4 3 3 5 5 5 ++5 5 5 5 5 5 4 0 0 6 6 6 4 0 0 6 6 6 ++6 6 6 6 6 6 6 6 6 4 0 0 6 6 6 4 0 0 ++6 6 6 4 3 3 5 5 5 4 3 3 5 5 5 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 ++4 3 3 6 6 6 4 3 3 6 6 6 6 6 6 6 6 6 ++4 0 0 6 6 6 4 0 0 6 6 6 6 6 6 6 6 6 ++6 6 6 4 3 3 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 4 3 3 5 5 5 4 0 0 6 6 6 ++6 6 6 4 0 0 6 6 6 6 6 6 4 0 0 6 6 6 ++4 3 3 5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 5 5 5 4 3 3 5 5 5 6 6 6 4 3 3 ++4 3 3 6 6 6 6 6 6 4 3 3 6 6 6 4 3 3 ++5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 5 5 5 4 3 3 6 6 6 ++5 5 5 4 3 3 4 3 3 4 3 3 5 5 5 5 5 5 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 4 3 3 ++5 5 5 4 3 3 5 5 5 5 5 5 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 ++4 4 4 4 4 4 +diff --git a/drivers/video/nvidia/nv_backlight.c b/drivers/video/nvidia/nv_backlight.c +index 443e3c8..c443d6a 100644 +--- a/drivers/video/nvidia/nv_backlight.c ++++ b/drivers/video/nvidia/nv_backlight.c +@@ -87,7 +87,7 @@ static int nvidia_bl_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops nvidia_bl_ops = { ++static const struct backlight_ops nvidia_bl_ops = { + .get_brightness = nvidia_bl_get_brightness, + .update_status = nvidia_bl_update_status, + }; +diff --git a/drivers/video/riva/fbdev.c b/drivers/video/riva/fbdev.c +index d94c57f..912984c 100644 +--- a/drivers/video/riva/fbdev.c ++++ b/drivers/video/riva/fbdev.c +@@ -331,7 +331,7 @@ static int riva_bl_get_brightness(struct backlight_device *bd) + return bd->props.brightness; + } + +-static struct backlight_ops riva_bl_ops = { ++static const struct backlight_ops riva_bl_ops = { + .get_brightness = riva_bl_get_brightness, + .update_status = riva_bl_update_status, + }; +diff --git a/drivers/video/uvesafb.c b/drivers/video/uvesafb.c +index 54fbb29..2c108fc 100644 +--- a/drivers/video/uvesafb.c ++++ b/drivers/video/uvesafb.c +@@ -18,6 +18,7 @@ + #include + #include + #include ++#include + #include