<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD>
<META content="text/html; charset=utf-8" http-equiv=Content-Type>
<STYLE>
BLOCKQUOTE {
MARGIN-TOP: 0px; MARGIN-BOTTOM: 0px; MARGIN-LEFT: 2em
}
OL {
MARGIN-TOP: 0px; MARGIN-BOTTOM: 0px
}
UL {
MARGIN-TOP: 0px; MARGIN-BOTTOM: 0px
}
P {
MARGIN-TOP: 0px; MARGIN-BOTTOM: 0px
}
BODY {
LINE-HEIGHT: 1.5; FONT-FAMILY: 微软雅黑; COLOR: #000000; FONT-SIZE: 10.5pt
}
</STYLE>
<META name=GENERATOR content="MSHTML 9.00.8112.16545"></HEAD>
<BODY style="MARGIN: 10px">
<DIV>
<DIV>
<DIV>
<DIV>When we test the cpu and memory hotplug feature in the server with x86
architecture and kernel4.0-rc4,we met the similar problem.<BR><BR></DIV>
<DIV></DIV>
<DIV>The situation is that when memory in node0 is offline,the system is down
during booting.</DIV></DIV>
<DIV><BR>Following is the bug information:<BR>
<DIV>[ 0.335176] BUG: unable to handle kernel paging request
at 0000000000001b08<BR>[ 0.342164] IP:
[<ffffffff81182587>]
__alloc_pages_nodemask+0xb7/0x940<BR>[ 0.348706] PGD 0
<BR>[ 0.350735] Oops: 0000 [#1] SMP <BR>[
0.353993] Modules linked in:<BR>[ 0.357063] CPU: 0 PID: 1
Comm: swapper/0 Not tainted 4.0.0-rc4 #1<BR>[ 0.363232]
Hardware name: Inspur TS860/TS860, BIOS TS860_2.0.0
2015/03/24<BR>[ 0.370095] task: ffff88085b1e0000 ti:
ffff88085b1e8000 task.ti: ffff88085b1e8000<BR>[ 0.377564] RIP:
0010:[<ffffffff81182587>] [<ffffffff81182587>]
__alloc_pages_nodemask+0xb7/0x940<BR>[ 0.386524] RSP:
0000:ffff88085b1ebac8 EFLAGS: 00010246<BR>[ 0.391828]
RAX: 0000000000001b00 RBX: 0000000000000010 RCX:
0000000000000000<BR>[ 0.398953] RDX: 0000000000000000 RSI:
0000000000000000 RDI: 00000000002052d0<BR>[ 0.406075] RBP:
ffff88085b1ebbb8 R08: ffff88085b13fec0 R09:
000000005b13fe01<BR>[ 0.413198] R10: ffff88085e807300 R11:
ffffffff810d4bc1 R12: 000000000001002a<BR>[ 0.420321] R13:
00000000002052d0 R14: 0000000000000001 R15:
00000000000040d0<BR>[ 0.427446] FS:
0000000000000000(0000) GS:ffff88085ee00000(0000)
knlGS:0000000000000000<BR>[ 0.435522] CS: 0010 DS: 0000
ES: 0000 CR0: 0000000080050033<BR>[ 0.441259] CR2:
0000000000001b08 CR3: 00000000019ae000 CR4:
00000000001406f0<BR>[ 0.448382] Stack:<BR>[
0.450392] ffff88085b1e0000 0000000000000400 ffff88085b1effff
ffff88085b1ebb68<BR>[ 0.457846] 000000000000007b
ffff88085b12d140 ffff88085b249000 000000000000007b<BR>[
0.465298] ffff88085b1ebb28 ffffffff81af2900 0000000000000000
002052d05b12d140<BR>[ 0.472750] Call
Trace:<BR>[ 0.475206] [<ffffffff811d27b3>] ?
deactivate_slab+0x383/0x400<BR>[ 0.481123]
[<ffffffff811d3947>] new_slab+0xa7/0x460<BR>[
0.486174] [<ffffffff816789e5>]
__slab_alloc+0x310/0x470<BR>[ 0.491655]
[<ffffffff8105304f>] ?
dmar_msi_set_affinity+0x8f/0xc0<BR>[ 0.497921]
[<ffffffff810d4bc1>] ? __irq_domain_add+0x41/0x100<BR>[
0.503838] [<ffffffff810d0fee>] ?
irq_do_set_affinity+0x5e/0x70<BR>[ 0.509920]
[<ffffffff811d571d>] __kmalloc_node+0xad/0x2e0<BR>[
0.515483] [<ffffffff810d4bc1>] ?
__irq_domain_add+0x41/0x100<BR>[ 0.521392]
[<ffffffff810d4bc1>] __irq_domain_add+0x41/0x100<BR>[
0.527133] [<ffffffff8105102e>]
mp_irqdomain_create+0x9e/0x120<BR>[ 0.533140]
[<ffffffff81b2fb14>] setup_IO_APIC+0x64/0x1be<BR>[
0.538622] [<ffffffff81b2e226>]
apic_bsp_setup+0xa2/0xae<BR>[ 0.544099]
[<ffffffff81b2bc70>]
native_smp_prepare_cpus+0x267/0x2b2<BR>[ 0.550531]
[<ffffffff81b1927b>]
kernel_init_freeable+0xf2/0x253<BR>[ 0.556625]
[<ffffffff8166b960>] ? rest_init+0x80/0x80<BR>[
0.561845] [<ffffffff8166b96e>]
kernel_init+0xe/0xf0<BR>[ 0.566979]
[<ffffffff81681bd8>] ret_from_fork+0x58/0x90<BR>[
0.572374] [<ffffffff8166b960>] ?
rest_init+0x80/0x80<BR>[ 0.577591] Code: 30 97 00 89 45 bc 83
e1 0f b8 22 01 32 01 01 c9 d3 f8 83 e0 03 89 9d 6c ff ff ff 83 e3 10 89 45 c0 0f
85 6d 01 00 00 48 8b 45 88 <48> 83 78 08 00 0f 84 51 01 00 00 b8 01 00 00
00 44 89 f1 d3 e0 <BR>[ 0.597537] RIP
[<ffffffff81182587>]
__alloc_pages_nodemask+0xb7/0x940<BR>[ 0.604158] RSP
<ffff88085b1ebac8><BR>[ 0.607643] CR2:
0000000000001b08<BR>[ 0.610962] ---[ end trace
0a600c0841386992 ]---<BR>[ 0.615573] Kernel panic - not
syncing: Fatal exception<BR>[ 0.620792] ---[ end Kernel panic
- not syncing: Fatal exception<BR></DIV></DIV></DIV></DIV>
<BLOCKQUOTE cite=mid:2015042818254949058834@inspur.com type="cite">
<DIV></DIV></BLOCKQUOTE>
<BLOCKQUOTE cite=mid:2015042818254949058834@inspur.com type="cite">
<DIV>
<DIV> </DIV></DIV></BLOCKQUOTE>
<DIV> </DIV>
<DIV> </DIV>
<DIV
style="BORDER-BOTTOM: medium none; BORDER-LEFT: medium none; PADDING-BOTTOM: 0cm; PADDING-LEFT: 0cm; PADDING-RIGHT: 0cm; BORDER-TOP: #b5c4df 1pt solid; BORDER-RIGHT: medium none; PADDING-TOP: 3pt">
<DIV
style="PADDING-BOTTOM: 8px; PADDING-LEFT: 8px; PADDING-RIGHT: 8px; FONT-FAMILY: tahoma; BACKGROUND: #efefef; COLOR: #000000; FONT-SIZE: 12px; PADDING-TOP: 8px">
<DIV><B>From:</B> <A href="mailto:robherring2@gmail.com">Rob
Herring</A></DIV>
<DIV><B>Date:</B> 2015-04-14 00:49</DIV>
<DIV><B>To:</B> <A href="mailto:khlebnikov@yandex-team.ru">Konstantin
Khlebnikov</A></DIV>
<DIV><B>CC:</B> <A href="mailto:grant.likely@linaro.org">Grant Likely</A>;
<A href="mailto:devicetree@vger.kernel.org">devicetree@vger.kernel.org</A>; <A
href="mailto:robh+dt@kernel.org">Rob Herring</A>; <A
href="mailto:linux-kernel@vger.kernel.org">linux-kernel@vger.kernel.org</A>; <A
href="mailto:sparclinux@vger.kernel.org">sparclinux@vger.kernel.org</A>; <A
href="mailto:linux-mm@kvack.org">linux-mm@kvack.org</A>; <A
href="mailto:linuxppc-dev@lists.ozlabs.org">linuxppc-dev</A></DIV>
<DIV><B>Subject:</B> Re: [PATCH] of: return NUMA_NO_NODE from fallback
of_node_to_nid()</DIV></DIV></DIV>
<DIV>
<DIV>On Mon, Apr 13, 2015 at 8:38 AM, Konstantin Khlebnikov</DIV>
<DIV><khlebnikov@yandex-team.ru> wrote:</DIV>
<DIV>> On 13.04.2015 16:22, Rob Herring wrote:</DIV>
<DIV>>></DIV>
<DIV>>> On Wed, Apr 8, 2015 at 11:59 AM, Konstantin Khlebnikov</DIV>
<DIV>>> <khlebnikov@yandex-team.ru> wrote:</DIV>
<DIV>>>></DIV>
<DIV>>>> Node 0 might be offline as well as any other numa node,</DIV>
<DIV>>>> in this case kernel cannot handle memory allocation and
crashes.</DIV>
<DIV>>>></DIV>
<DIV>>>> Signed-off-by: Konstantin Khlebnikov
<khlebnikov@yandex-team.ru></DIV>
<DIV>>>> Fixes: 0c3f061c195c ("of: implement of_node_to_nid as a weak
function")</DIV>
<DIV>>>> ---</DIV>
<DIV>>>> drivers/of/base.c | 2
+-</DIV>
<DIV>>>> include/linux/of.h | 5
++++-</DIV>
<DIV>>>> 2 files changed, 5 insertions(+), 2
deletions(-)</DIV>
<DIV>>>></DIV>
<DIV>>>> diff --git a/drivers/of/base.c b/drivers/of/base.c</DIV>
<DIV>>>> index 8f165b112e03..51f4bd16e613 100644</DIV>
<DIV>>>> --- a/drivers/of/base.c</DIV>
<DIV>>>> +++ b/drivers/of/base.c</DIV>
<DIV>>>> @@ -89,7 +89,7 @@ EXPORT_SYMBOL(of_n_size_cells);</DIV>
<DIV>>>> #ifdef CONFIG_NUMA</DIV>
<DIV>>>> int __weak of_node_to_nid(struct device_node
*np)</DIV>
<DIV>>>> {</DIV>
<DIV>>>> - return
numa_node_id();</DIV>
<DIV>>>> + return
NUMA_NO_NODE;</DIV>
<DIV>>></DIV>
<DIV>>></DIV>
<DIV>>> This is going to break any NUMA machine that enables OF and
expects</DIV>
<DIV>>> the weak function to work.</DIV>
<DIV>></DIV>
<DIV>></DIV>
<DIV>> Why? NUMA_NO_NODE == -1 -- this's standard "no-affinity" signal.</DIV>
<DIV>> As I see powerpc/sparc versions of of_node_to_nid returns -1 if
they</DIV>
<DIV>> cannot find out which node should be used.</DIV>
<DIV> </DIV>
<DIV>Ah, I was thinking those platforms were relying on the default</DIV>
<DIV>implementation. I guess any real NUMA support is going to need to</DIV>
<DIV>override this function. The arm64 patch series does that as well. We</DIV>
<DIV>need to be sure this change is correct for metag which appears to be</DIV>
<DIV>the only other OF enabled platform with NUMA support.</DIV>
<DIV> </DIV>
<DIV>In that case, then there is little reason to keep the inline and we</DIV>
<DIV>can just always enable the weak function (with your change). It is</DIV>
<DIV>slightly less optimal, but the few callers hardly appear to be hot</DIV>
<DIV>paths.</DIV>
<DIV> </DIV>
<DIV>Rob</DIV>
<DIV>--</DIV>
<DIV>To unsubscribe from this list: send the line "unsubscribe linux-kernel"
in</DIV>
<DIV>the body of a message to majordomo@vger.kernel.org</DIV>
<DIV>More majordomo info at
http://vger.kernel.org/majordomo-info.html</DIV>
<DIV>Please read the FAQ at http://www.tux.org/lkml/</DIV>
<DIV> </DIV></DIV></BODY></HTML>