kernel: fix plt fixup related crashes when loading kernel modules on mips, (happened when there was not enough physically contiguous memory available)

SVN-Revision: 18858
This commit is contained in:
Felix Fietkau 2009-12-20 04:41:10 +00:00
parent fa044a2d81
commit 6859128d3a
3 changed files with 228 additions and 117 deletions

View file

@ -16,10 +16,10 @@
const struct exception_table_entry *dbe_start; const struct exception_table_entry *dbe_start;
const struct exception_table_entry *dbe_end; const struct exception_table_entry *dbe_end;
+ +
+ void *plt_tbl; + void *phys_plt_tbl;
+ unsigned int core_plt_offset; + void *virt_plt_tbl;
+ unsigned int core_plt_size; + unsigned int phys_plt_offset;
+ unsigned int init_plt_offset; + unsigned int virt_plt_offset;
}; };
typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */ typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */
@ -142,7 +142,7 @@
void *module_alloc(unsigned long size) void *module_alloc(unsigned long size)
{ {
#ifdef MODULE_START #ifdef MODULE_START
@@ -58,16 +168,45 @@ void *module_alloc(unsigned long size) @@ -58,23 +168,101 @@ void *module_alloc(unsigned long size)
return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL); return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL);
#else #else
@ -190,7 +190,26 @@
/* FIXME: If module_region == mod->init_region, trim exception /* FIXME: If module_region == mod->init_region, trim exception
table entries. */ table entries. */
} }
@@ -75,6 +214,24 @@ void module_free(struct module *mod, voi
+static void *__module_alloc(int size, bool phys)
+{
+ void *ptr;
+
+ if (phys)
+ ptr = kmalloc(size, GFP_KERNEL);
+ else
+ ptr = vmalloc(size);
+ return ptr;
+}
+
+static void __module_free(void *ptr)
+{
+ if (is_phys_addr(ptr))
+ kfree(ptr);
+ else
+ vfree(ptr);
+}
+
int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs, int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
char *secstrings, struct module *mod) char *secstrings, struct module *mod)
{ {
@ -205,17 +224,29 @@
+ core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false); + core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false);
+ init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true); + init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true);
+ +
+ mod->arch.core_plt_offset = 0; + mod->arch.phys_plt_offset = 0;
+ mod->arch.core_plt_size = core_size; + mod->arch.virt_plt_offset = 0;
+ mod->arch.init_plt_offset = core_size; + mod->arch.phys_plt_tbl = NULL;
+ mod->arch.plt_tbl = kmalloc(core_size + init_size, GFP_KERNEL); + mod->arch.virt_plt_tbl = NULL;
+ if (!mod->arch.plt_tbl) +
+ if ((core_size + init_size) == 0)
+ return 0;
+
+ mod->arch.phys_plt_tbl = __module_alloc(core_size + init_size, 1);
+ if (!mod->arch.phys_plt_tbl)
+ return -ENOMEM; + return -ENOMEM;
+
+ mod->arch.virt_plt_tbl = __module_alloc(core_size + init_size, 0);
+ if (!mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ return -ENOMEM;
+ }
+ +
return 0; return 0;
} }
@@ -97,27 +254,41 @@ static int apply_r_mips_32_rela(struct m @@ -97,27 +285,37 @@ static int apply_r_mips_32_rela(struct m
return 0; return 0;
} }
@ -236,42 +267,39 @@
- return -ENOEXEC; - return -ENOEXEC;
- } - }
+ *plt_offset += 4 * sizeof(int); + *plt_offset += 4 * sizeof(int);
+
- *location = (*location & ~0x03ffffff) |
- ((*location + (v >> 2)) & 0x03ffffff);
+ /* adjust carry for addiu */ + /* adjust carry for addiu */
+ if (v & 0x00008000) + if (v & 0x00008000)
+ v += 0x10000; + v += 0x10000;
+
- return 0;
+ tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */ + tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */
+ tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */ + tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */
+ tramp[2] = 0x03200008; /* jr t9 */ + tramp[2] = 0x03200008; /* jr t9 */
+ tramp[3] = 0x00000000; /* nop */ + tramp[3] = 0x00000000; /* nop */
+
- *location = (*location & ~0x03ffffff) |
- ((*location + (v >> 2)) & 0x03ffffff);
+ return (Elf_Addr) tramp; + return (Elf_Addr) tramp;
+}
+
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (location >= me->module_core &&
+ location < me->module_core + me->core_size)
+ return add_plt_entry_to(&me->arch.core_plt_offset,
+ me->arch.plt_tbl, v);
+
+ if (location >= me->module_init &&
+ location < me->module_init + me->init_size)
+ return add_plt_entry_to(&me->arch.init_plt_offset,
+ me->arch.plt_tbl, v);
return 0;
} }
-static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v) -static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v)
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (is_phys_addr(location))
+ return add_plt_entry_to(&me->arch.phys_plt_offset,
+ me->arch.phys_plt_tbl, v);
+ else
+ return add_plt_entry_to(&me->arch.virt_plt_offset,
+ me->arch.virt_plt_tbl, v);
+
+}
+
+static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v) +static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v)
{ {
if (v % 4) { if (v % 4) {
printk(KERN_ERR "module %s: dangerous relocation\n", me->name); printk(KERN_ERR "module %s: dangerous relocation\n", me->name);
@@ -125,17 +296,31 @@ static int apply_r_mips_26_rela(struct m @@ -125,17 +323,31 @@ static int apply_r_mips_26_rela(struct m
} }
if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) { if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) {
@ -306,18 +334,20 @@
static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v) static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v)
{ {
struct mips_hi16 *n; struct mips_hi16 *n;
@@ -400,11 +585,23 @@ int module_finalize(const Elf_Ehdr *hdr, @@ -400,11 +612,32 @@ int module_finalize(const Elf_Ehdr *hdr,
list_add(&me->arch.dbe_list, &dbe_list); list_add(&me->arch.dbe_list, &dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);
} }
+ +
+ /* Get rid of the fixup trampoline if we're running the module + /* Get rid of the fixup trampoline if we're running the module
+ * from physically mapped address space */ + * from physically mapped address space */
+ if (me->arch.core_plt_offset == 0 && + if (me->arch.phys_plt_offset == 0) {
+ me->arch.init_plt_offset == me->arch.core_plt_size && + __module_free(me->arch.phys_plt_tbl);
+ is_phys_addr(me->module_core)) { + me->arch.phys_plt_tbl = NULL;
+ kfree(me->arch.plt_tbl); + }
+ me->arch.plt_tbl = NULL; + if (me->arch.virt_plt_offset == 0) {
+ __module_free(me->arch.virt_plt_tbl);
+ me->arch.virt_plt_tbl = NULL;
+ } + }
+ +
return 0; return 0;
@ -325,8 +355,15 @@
void module_arch_cleanup(struct module *mod) void module_arch_cleanup(struct module *mod)
{ {
+ if (mod->arch.plt_tbl) + if (mod->arch.phys_plt_tbl) {
+ kfree(mod->arch.plt_tbl); + __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ }
+ if (mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.virt_plt_tbl);
+ mod->arch.virt_plt_tbl = NULL;
+ }
+
spin_lock_irq(&dbe_lock); spin_lock_irq(&dbe_lock);
list_del(&mod->arch.dbe_list); list_del(&mod->arch.dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);

View file

@ -16,10 +16,10 @@
const struct exception_table_entry *dbe_start; const struct exception_table_entry *dbe_start;
const struct exception_table_entry *dbe_end; const struct exception_table_entry *dbe_end;
+ +
+ void *plt_tbl; + void *phys_plt_tbl;
+ unsigned int core_plt_offset; + void *virt_plt_tbl;
+ unsigned int core_plt_size; + unsigned int phys_plt_offset;
+ unsigned int init_plt_offset; + unsigned int virt_plt_offset;
}; };
typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */ typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */
@ -143,7 +143,7 @@
void *module_alloc(unsigned long size) void *module_alloc(unsigned long size)
{ {
#ifdef MODULE_START #ifdef MODULE_START
@@ -58,21 +169,68 @@ void *module_alloc(unsigned long size) @@ -58,21 +169,99 @@ void *module_alloc(unsigned long size)
return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL); return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL);
#else #else
@ -188,6 +188,25 @@
+ } else { + } else {
+ vfree(module_region); + vfree(module_region);
+ } + }
+}
+
+static void *__module_alloc(int size, bool phys)
+{
+ void *ptr;
+
+ if (phys)
+ ptr = kmalloc(size, GFP_KERNEL);
+ else
+ ptr = vmalloc(size);
+ return ptr;
+}
+
+static void __module_free(void *ptr)
+{
+ if (is_phys_addr(ptr))
+ kfree(ptr);
+ else
+ vfree(ptr);
} }
int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs, int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
@ -204,17 +223,29 @@
+ core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false); + core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false);
+ init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true); + init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true);
+ +
+ mod->arch.core_plt_offset = 0; + mod->arch.phys_plt_offset = 0;
+ mod->arch.core_plt_size = core_size; + mod->arch.virt_plt_offset = 0;
+ mod->arch.init_plt_offset = core_size; + mod->arch.phys_plt_tbl = NULL;
+ mod->arch.plt_tbl = kmalloc(core_size + init_size, GFP_KERNEL); + mod->arch.virt_plt_tbl = NULL;
+ if (!mod->arch.plt_tbl) +
+ if ((core_size + init_size) == 0)
+ return 0;
+
+ mod->arch.phys_plt_tbl = __module_alloc(core_size + init_size, 1);
+ if (!mod->arch.phys_plt_tbl)
+ return -ENOMEM; + return -ENOMEM;
+
+ mod->arch.virt_plt_tbl = __module_alloc(core_size + init_size, 0);
+ if (!mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ return -ENOMEM;
+ }
+ +
return 0; return 0;
} }
@@ -95,28 +253,40 @@ static int apply_r_mips_32_rela(struct m @@ -95,28 +284,36 @@ static int apply_r_mips_32_rela(struct m
return 0; return 0;
} }
@ -239,38 +270,35 @@
+ /* adjust carry for addiu */ + /* adjust carry for addiu */
+ if (v & 0x00008000) + if (v & 0x00008000)
+ v += 0x10000; + v += 0x10000;
+
- *location = (*location & ~0x03ffffff) |
- ((*location + (v >> 2)) & 0x03ffffff);
+ tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */ + tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */
+ tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */ + tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */
+ tramp[2] = 0x03200008; /* jr t9 */ + tramp[2] = 0x03200008; /* jr t9 */
+ tramp[3] = 0x00000000; /* nop */ + tramp[3] = 0x00000000; /* nop */
- *location = (*location & ~0x03ffffff) | - return 0;
- ((*location + (v >> 2)) & 0x03ffffff);
+ return (Elf_Addr) tramp; + return (Elf_Addr) tramp;
+}
+
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (location >= me->module_core &&
+ location < me->module_core + me->core_size)
+ return add_plt_entry_to(&me->arch.core_plt_offset,
+ me->arch.plt_tbl, v);
+
+ if (location >= me->module_init &&
+ location < me->module_init + me->init_size)
+ return add_plt_entry_to(&me->arch.init_plt_offset,
+ me->arch.plt_tbl, v);
return 0;
} }
-static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v) -static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v)
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (is_phys_addr(location))
+ return add_plt_entry_to(&me->arch.phys_plt_offset,
+ me->arch.phys_plt_tbl, v);
+ else
+ return add_plt_entry_to(&me->arch.virt_plt_offset,
+ me->arch.virt_plt_tbl, v);
+
+}
+
+static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v) +static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v)
{ {
if (v % 4) { if (v % 4) {
pr_err("module %s: dangerous R_MIPS_26 RELArelocation\n", pr_err("module %s: dangerous R_MIPS_26 RELArelocation\n",
@@ -125,17 +295,31 @@ static int apply_r_mips_26_rela(struct m @@ -125,17 +322,31 @@ static int apply_r_mips_26_rela(struct m
} }
if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) { if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) {
@ -305,18 +333,20 @@
static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v) static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v)
{ {
struct mips_hi16 *n; struct mips_hi16 *n;
@@ -400,11 +584,23 @@ int module_finalize(const Elf_Ehdr *hdr, @@ -400,11 +611,32 @@ int module_finalize(const Elf_Ehdr *hdr,
list_add(&me->arch.dbe_list, &dbe_list); list_add(&me->arch.dbe_list, &dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);
} }
+ +
+ /* Get rid of the fixup trampoline if we're running the module + /* Get rid of the fixup trampoline if we're running the module
+ * from physically mapped address space */ + * from physically mapped address space */
+ if (me->arch.core_plt_offset == 0 && + if (me->arch.phys_plt_offset == 0) {
+ me->arch.init_plt_offset == me->arch.core_plt_size && + __module_free(me->arch.phys_plt_tbl);
+ is_phys_addr(me->module_core)) { + me->arch.phys_plt_tbl = NULL;
+ kfree(me->arch.plt_tbl); + }
+ me->arch.plt_tbl = NULL; + if (me->arch.virt_plt_offset == 0) {
+ __module_free(me->arch.virt_plt_tbl);
+ me->arch.virt_plt_tbl = NULL;
+ } + }
+ +
return 0; return 0;
@ -324,8 +354,15 @@
void module_arch_cleanup(struct module *mod) void module_arch_cleanup(struct module *mod)
{ {
+ if (mod->arch.plt_tbl) + if (mod->arch.phys_plt_tbl) {
+ kfree(mod->arch.plt_tbl); + __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ }
+ if (mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.virt_plt_tbl);
+ mod->arch.virt_plt_tbl = NULL;
+ }
+
spin_lock_irq(&dbe_lock); spin_lock_irq(&dbe_lock);
list_del(&mod->arch.dbe_list); list_del(&mod->arch.dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);

View file

@ -16,10 +16,10 @@
const struct exception_table_entry *dbe_start; const struct exception_table_entry *dbe_start;
const struct exception_table_entry *dbe_end; const struct exception_table_entry *dbe_end;
+ +
+ void *plt_tbl; + void *phys_plt_tbl;
+ unsigned int core_plt_offset; + void *virt_plt_tbl;
+ unsigned int core_plt_size; + unsigned int phys_plt_offset;
+ unsigned int init_plt_offset; + unsigned int virt_plt_offset;
}; };
typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */ typedef uint8_t Elf64_Byte; /* Type for a 8-bit quantity. */
@ -143,7 +143,7 @@
void *module_alloc(unsigned long size) void *module_alloc(unsigned long size)
{ {
#ifdef MODULE_START #ifdef MODULE_START
@@ -58,21 +169,68 @@ void *module_alloc(unsigned long size) @@ -58,21 +169,99 @@ void *module_alloc(unsigned long size)
return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL); return __vmalloc_area(area, GFP_KERNEL, PAGE_KERNEL);
#else #else
@ -188,6 +188,25 @@
+ } else { + } else {
+ vfree(module_region); + vfree(module_region);
+ } + }
+}
+
+static void *__module_alloc(int size, bool phys)
+{
+ void *ptr;
+
+ if (phys)
+ ptr = kmalloc(size, GFP_KERNEL);
+ else
+ ptr = vmalloc(size);
+ return ptr;
+}
+
+static void __module_free(void *ptr)
+{
+ if (is_phys_addr(ptr))
+ kfree(ptr);
+ else
+ vfree(ptr);
} }
int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs, int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
@ -204,17 +223,29 @@
+ core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false); + core_size = get_plt_size(hdr, sechdrs, secstrings, symindex, false);
+ init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true); + init_size = get_plt_size(hdr, sechdrs, secstrings, symindex, true);
+ +
+ mod->arch.core_plt_offset = 0; + mod->arch.phys_plt_offset = 0;
+ mod->arch.core_plt_size = core_size; + mod->arch.virt_plt_offset = 0;
+ mod->arch.init_plt_offset = core_size; + mod->arch.phys_plt_tbl = NULL;
+ mod->arch.plt_tbl = kmalloc(core_size + init_size, GFP_KERNEL); + mod->arch.virt_plt_tbl = NULL;
+ if (!mod->arch.plt_tbl) +
+ if ((core_size + init_size) == 0)
+ return 0;
+
+ mod->arch.phys_plt_tbl = __module_alloc(core_size + init_size, 1);
+ if (!mod->arch.phys_plt_tbl)
+ return -ENOMEM; + return -ENOMEM;
+
+ mod->arch.virt_plt_tbl = __module_alloc(core_size + init_size, 0);
+ if (!mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ return -ENOMEM;
+ }
+ +
return 0; return 0;
} }
@@ -95,28 +253,40 @@ static int apply_r_mips_32_rela(struct m @@ -95,28 +284,36 @@ static int apply_r_mips_32_rela(struct m
return 0; return 0;
} }
@ -239,38 +270,35 @@
+ /* adjust carry for addiu */ + /* adjust carry for addiu */
+ if (v & 0x00008000) + if (v & 0x00008000)
+ v += 0x10000; + v += 0x10000;
+
- *location = (*location & ~0x03ffffff) |
- ((*location + (v >> 2)) & 0x03ffffff);
+ tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */ + tramp[0] = 0x3c190000 | (v >> 16); /* lui t9, hi16 */
+ tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */ + tramp[1] = 0x27390000 | (v & 0xffff); /* addiu t9, t9, lo16 */
+ tramp[2] = 0x03200008; /* jr t9 */ + tramp[2] = 0x03200008; /* jr t9 */
+ tramp[3] = 0x00000000; /* nop */ + tramp[3] = 0x00000000; /* nop */
- *location = (*location & ~0x03ffffff) | - return 0;
- ((*location + (v >> 2)) & 0x03ffffff);
+ return (Elf_Addr) tramp; + return (Elf_Addr) tramp;
+}
+
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (location >= me->module_core &&
+ location < me->module_core + me->core_size)
+ return add_plt_entry_to(&me->arch.core_plt_offset,
+ me->arch.plt_tbl, v);
+
+ if (location >= me->module_init &&
+ location < me->module_init + me->init_size)
+ return add_plt_entry_to(&me->arch.init_plt_offset,
+ me->arch.plt_tbl, v);
return 0;
} }
-static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v) -static int apply_r_mips_26_rela(struct module *me, u32 *location, Elf_Addr v)
+static Elf_Addr add_plt_entry(struct module *me, void *location, Elf_Addr v)
+{
+ if (is_phys_addr(location))
+ return add_plt_entry_to(&me->arch.phys_plt_offset,
+ me->arch.phys_plt_tbl, v);
+ else
+ return add_plt_entry_to(&me->arch.virt_plt_offset,
+ me->arch.virt_plt_tbl, v);
+
+}
+
+static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v) +static int set_r_mips_26(struct module *me, u32 *location, u32 ofs, Elf_Addr v)
{ {
if (v % 4) { if (v % 4) {
pr_err("module %s: dangerous R_MIPS_26 RELArelocation\n", pr_err("module %s: dangerous R_MIPS_26 RELArelocation\n",
@@ -125,17 +295,31 @@ static int apply_r_mips_26_rela(struct m @@ -125,17 +322,31 @@ static int apply_r_mips_26_rela(struct m
} }
if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) { if ((v & 0xf0000000) != (((unsigned long)location + 4) & 0xf0000000)) {
@ -305,18 +333,20 @@
static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v) static int apply_r_mips_hi16_rel(struct module *me, u32 *location, Elf_Addr v)
{ {
struct mips_hi16 *n; struct mips_hi16 *n;
@@ -400,11 +584,23 @@ int module_finalize(const Elf_Ehdr *hdr, @@ -400,11 +611,32 @@ int module_finalize(const Elf_Ehdr *hdr,
list_add(&me->arch.dbe_list, &dbe_list); list_add(&me->arch.dbe_list, &dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);
} }
+ +
+ /* Get rid of the fixup trampoline if we're running the module + /* Get rid of the fixup trampoline if we're running the module
+ * from physically mapped address space */ + * from physically mapped address space */
+ if (me->arch.core_plt_offset == 0 && + if (me->arch.phys_plt_offset == 0) {
+ me->arch.init_plt_offset == me->arch.core_plt_size && + __module_free(me->arch.phys_plt_tbl);
+ is_phys_addr(me->module_core)) { + me->arch.phys_plt_tbl = NULL;
+ kfree(me->arch.plt_tbl); + }
+ me->arch.plt_tbl = NULL; + if (me->arch.virt_plt_offset == 0) {
+ __module_free(me->arch.virt_plt_tbl);
+ me->arch.virt_plt_tbl = NULL;
+ } + }
+ +
return 0; return 0;
@ -324,8 +354,15 @@
void module_arch_cleanup(struct module *mod) void module_arch_cleanup(struct module *mod)
{ {
+ if (mod->arch.plt_tbl) + if (mod->arch.phys_plt_tbl) {
+ kfree(mod->arch.plt_tbl); + __module_free(mod->arch.phys_plt_tbl);
+ mod->arch.phys_plt_tbl = NULL;
+ }
+ if (mod->arch.virt_plt_tbl) {
+ __module_free(mod->arch.virt_plt_tbl);
+ mod->arch.virt_plt_tbl = NULL;
+ }
+
spin_lock_irq(&dbe_lock); spin_lock_irq(&dbe_lock);
list_del(&mod->arch.dbe_list); list_del(&mod->arch.dbe_list);
spin_unlock_irq(&dbe_lock); spin_unlock_irq(&dbe_lock);