From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id D783CC433EF for ; Tue, 30 Nov 2021 21:08:19 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S240434AbhK3VLi (ORCPT ); Tue, 30 Nov 2021 16:11:38 -0500 Received: from sin.source.kernel.org ([145.40.73.55]:50962 "EHLO sin.source.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235904AbhK3VLf (ORCPT ); Tue, 30 Nov 2021 16:11:35 -0500 Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sin.source.kernel.org (Postfix) with ESMTPS id C5101CE1C4F; Tue, 30 Nov 2021 21:08:13 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 78438C53FC7; Tue, 30 Nov 2021 21:07:46 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1638306491; bh=CGQDQlRxlHpJq0xzSuKQi7AU9OYm37j7CJ7xfuMMlQY=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=MQT0554slXBNgzV6dz6ATkc3EGBq02EcnyIjFHm4NC/TQW+Ti/5+ZwpYdUyXTQSX0 FP/Mg9FHX1WkJdgEmvH+Sjh6chDp0qa4rdvwceLtLU4mrx9YqTjWtdVLiOuUAqqTr9 JxBBF8wiCf2msoHxAESTMHrwoozuA53LUQYu0nChlx4mPkmdPvsjcuZz6ld/73Q1EP 49cP1v8fZ5yU4pp+jiMhNoi6TZsD6QEaZ+s7211uBewrAF+q/t4jJF4C5l0VvizKZK UqaJcABTKX49+eBLI8ebfD/p5ZMIh9Z6ly7YavV63rHaytjZkBuc5OrtUPGVxLkKwd 8IDw15RIkFr7Q== Date: Tue, 30 Nov 2021 23:07:40 +0200 From: Mike Rapoport To: Rob Herring Cc: Calvin Zhang , Vineet Gupta , Russell King , Catalin Marinas , Will Deacon , Guo Ren , Yoshinori Sato , Thomas Bogendoerfer , Nick Hu , Greentime Hu , Vincent Chen , Dinh Nguyen , Jonas Bonn , Stefan Kristiansson , Stafford Horne , Michael Ellerman , Benjamin Herrenschmidt , Paul Mackerras , Paul Walmsley , Palmer Dabbelt , Albert Ou , Rich Felker , Chris Zankel , Max Filippov , Frank Rowand , Andrew Morton , David Hildenbrand , Arnd Bergmann , Kefeng Wang , Vladimir Isaev , "Russell King (Oracle)" , "Kirill A. Shutemov" , Guenter Roeck , Marc Zyngier , David Brazdil , Anshuman Khandual , Andrey Konovalov , Mark Rutland , Souptick Joarder , Jinyang He , "Rafael J. Wysocki" , Serge Semin , Tiezhu Yang , Geert Uytterhoeven , Randy Dunlap , Ley Foon Tan , Andreas Oetken , Christophe JAILLET , Christophe Leroy , Zhang Yunkai , Andy Shevchenko , Markus Elfring , Ganesh Goudar , "Aneesh Kumar K.V" , Atish Patra , Anup Patel , Nick Kossifidis , Alexandre Ghiti , Vitaly Wool , Thierry Reding , Lee Jones , Guo Ren , Alexander Sverdlin , Greg Kroah-Hartman , Mauri Sandberg , Palmer Dabbelt , linux-snps-arc@lists.infradead.org, linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-csky@vger.kernel.org, uclinux-h8-devel@lists.sourceforge.jp, linux-mips@vger.kernel.org, openrisc@lists.librecores.org, linuxppc-dev@lists.ozlabs.org, linux-riscv@lists.infradead.org, linux-sh@vger.kernel.org, linux-xtensa@linux-xtensa.org, devicetree@vger.kernel.org Subject: Re: [PATCH 0/2] of: remove reserved regions count restriction Message-ID: References: <20211119075844.2902592-1-calvinzhang.cool@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Precedence: bulk List-ID: X-Mailing-List: linux-csky@vger.kernel.org On Mon, Nov 29, 2021 at 06:08:10PM -0600, Rob Herring wrote: > On Sun, Nov 21, 2021 at 08:43:47AM +0200, Mike Rapoport wrote: > > On Fri, Nov 19, 2021 at 03:58:17PM +0800, Calvin Zhang wrote: > > > The count of reserved regions in /reserved-memory was limited because > > > the struct reserved_mem array was defined statically. This series sorts > > > out reserved memory code and allocates that array from early allocator. > > > > > > Note: reserved region with fixed location must be reserved before any > > > memory allocation. While struct reserved_mem array should be allocated > > > after allocator is activated. We make early_init_fdt_scan_reserved_mem() > > > do reservation only and add another call to initialize reserved memory. > > > So arch code have to change for it. > > > > I think much simpler would be to use the same constant for sizing > > memblock.reserved and reserved_mem arrays. > > Do those arrays get shrunk? Or do we waste the memory forever? Memblock arrays don't get shrunk, but they are __init unless an architecture chose to keep them after boot, but most architectures that use device tree actually keep memblock structures. > Maybe we can copy and shrink the initial array? Though I suspect struct > reserved_mem pointers have already been given out. I'm not sure. It seems that reserved_mem pointers are given out at initcall time and AFAIU the reserved_mem array is created somewhere around setup_arch(). So maybe it's possible to copy and shrink the initial array. > > > > If there is too much reserved regions in the device tree, reserving them in > > memblock will fail anyway because memblock also starts with static array > > for memblock.reserved, so doing one pass with memblock_reserve() and > > another to set up reserved_mem wouldn't help anyway. > > > > > I'm only familiar with arm and arm64 architectures. Approvals from arch > > > maintainers are required. Thank you all. -- Sincerely yours, Mike.