mirror of
https://git.proxmox.com/git/mirror_ubuntu-kernels.git
synced 2025-11-22 09:21:56 +00:00
are included in this merge do the following:
- Peng Zhang has done some mapletree maintainance work in the
series
"maple_tree: add mt_free_one() and mt_attr() helpers"
"Some cleanups of maple tree"
- In the series "mm: use memmap_on_memory semantics for dax/kmem"
Vishal Verma has altered the interworking between memory-hotplug
and dax/kmem so that newly added 'device memory' can more easily
have its memmap placed within that newly added memory.
- Matthew Wilcox continues folio-related work (including a few
fixes) in the patch series
"Add folio_zero_tail() and folio_fill_tail()"
"Make folio_start_writeback return void"
"Fix fault handler's handling of poisoned tail pages"
"Convert aops->error_remove_page to ->error_remove_folio"
"Finish two folio conversions"
"More swap folio conversions"
- Kefeng Wang has also contributed folio-related work in the series
"mm: cleanup and use more folio in page fault"
- Jim Cromie has improved the kmemleak reporting output in the
series "tweak kmemleak report format".
- In the series "stackdepot: allow evicting stack traces" Andrey
Konovalov to permits clients (in this case KASAN) to cause
eviction of no longer needed stack traces.
- Charan Teja Kalla has fixed some accounting issues in the page
allocator's atomic reserve calculations in the series "mm:
page_alloc: fixes for high atomic reserve caluculations".
- Dmitry Rokosov has added to the samples/ dorectory some sample
code for a userspace memcg event listener application. See the
series "samples: introduce cgroup events listeners".
- Some mapletree maintanance work from Liam Howlett in the series
"maple_tree: iterator state changes".
- Nhat Pham has improved zswap's approach to writeback in the
series "workload-specific and memory pressure-driven zswap
writeback".
- DAMON/DAMOS feature and maintenance work from SeongJae Park in
the series
"mm/damon: let users feed and tame/auto-tune DAMOS"
"selftests/damon: add Python-written DAMON functionality tests"
"mm/damon: misc updates for 6.8"
- Yosry Ahmed has improved memcg's stats flushing in the series
"mm: memcg: subtree stats flushing and thresholds".
- In the series "Multi-size THP for anonymous memory" Ryan Roberts
has added a runtime opt-in feature to transparent hugepages which
improves performance by allocating larger chunks of memory during
anonymous page faults.
- Matthew Wilcox has also contributed some cleanup and maintenance
work against eh buffer_head code int he series "More buffer_head
cleanups".
- Suren Baghdasaryan has done work on Andrea Arcangeli's series
"userfaultfd move option". UFFDIO_MOVE permits userspace heap
compaction algorithms to move userspace's pages around rather than
UFFDIO_COPY'a alloc/copy/free.
- Stefan Roesch has developed a "KSM Advisor", in the series
"mm/ksm: Add ksm advisor". This is a governor which tunes KSM's
scanning aggressiveness in response to userspace's current needs.
- Chengming Zhou has optimized zswap's temporary working memory
use in the series "mm/zswap: dstmem reuse optimizations and
cleanups".
- Matthew Wilcox has performed some maintenance work on the
writeback code, both code and within filesystems. The series is
"Clean up the writeback paths".
- Andrey Konovalov has optimized KASAN's handling of alloc and
free stack traces for secondary-level allocators, in the series
"kasan: save mempool stack traces".
- Andrey also performed some KASAN maintenance work in the series
"kasan: assorted clean-ups".
- David Hildenbrand has gone to town on the rmap code. Cleanups,
more pte batching, folio conversions and more. See the series
"mm/rmap: interface overhaul".
- Kinsey Ho has contributed some maintenance work on the MGLRU
code in the series "mm/mglru: Kconfig cleanup".
- Matthew Wilcox has contributed lruvec page accounting code
cleanups in the series "Remove some lruvec page accounting
functions".
-----BEGIN PGP SIGNATURE-----
iHUEABYIAB0WIQTTMBEPP41GrTpTJgfdBJ7gKXxAjgUCZZyF2wAKCRDdBJ7gKXxA
jjWjAP42LHvGSjp5M+Rs2rKFL0daBQsrlvy6/jCHUequSdWjSgEAmOx7bc5fbF27
Oa8+DxGM9C+fwqZ/7YxU2w/WuUmLPgU=
=0NHs
-----END PGP SIGNATURE-----
Merge tag 'mm-stable-2024-01-08-15-31' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm
Pull MM updates from Andrew Morton:
"Many singleton patches against the MM code. The patch series which are
included in this merge do the following:
- Peng Zhang has done some mapletree maintainance work in the series
'maple_tree: add mt_free_one() and mt_attr() helpers'
'Some cleanups of maple tree'
- In the series 'mm: use memmap_on_memory semantics for dax/kmem'
Vishal Verma has altered the interworking between memory-hotplug
and dax/kmem so that newly added 'device memory' can more easily
have its memmap placed within that newly added memory.
- Matthew Wilcox continues folio-related work (including a few fixes)
in the patch series
'Add folio_zero_tail() and folio_fill_tail()'
'Make folio_start_writeback return void'
'Fix fault handler's handling of poisoned tail pages'
'Convert aops->error_remove_page to ->error_remove_folio'
'Finish two folio conversions'
'More swap folio conversions'
- Kefeng Wang has also contributed folio-related work in the series
'mm: cleanup and use more folio in page fault'
- Jim Cromie has improved the kmemleak reporting output in the series
'tweak kmemleak report format'.
- In the series 'stackdepot: allow evicting stack traces' Andrey
Konovalov to permits clients (in this case KASAN) to cause eviction
of no longer needed stack traces.
- Charan Teja Kalla has fixed some accounting issues in the page
allocator's atomic reserve calculations in the series 'mm:
page_alloc: fixes for high atomic reserve caluculations'.
- Dmitry Rokosov has added to the samples/ dorectory some sample code
for a userspace memcg event listener application. See the series
'samples: introduce cgroup events listeners'.
- Some mapletree maintanance work from Liam Howlett in the series
'maple_tree: iterator state changes'.
- Nhat Pham has improved zswap's approach to writeback in the series
'workload-specific and memory pressure-driven zswap writeback'.
- DAMON/DAMOS feature and maintenance work from SeongJae Park in the
series
'mm/damon: let users feed and tame/auto-tune DAMOS'
'selftests/damon: add Python-written DAMON functionality tests'
'mm/damon: misc updates for 6.8'
- Yosry Ahmed has improved memcg's stats flushing in the series 'mm:
memcg: subtree stats flushing and thresholds'.
- In the series 'Multi-size THP for anonymous memory' Ryan Roberts
has added a runtime opt-in feature to transparent hugepages which
improves performance by allocating larger chunks of memory during
anonymous page faults.
- Matthew Wilcox has also contributed some cleanup and maintenance
work against eh buffer_head code int he series 'More buffer_head
cleanups'.
- Suren Baghdasaryan has done work on Andrea Arcangeli's series
'userfaultfd move option'. UFFDIO_MOVE permits userspace heap
compaction algorithms to move userspace's pages around rather than
UFFDIO_COPY'a alloc/copy/free.
- Stefan Roesch has developed a 'KSM Advisor', in the series 'mm/ksm:
Add ksm advisor'. This is a governor which tunes KSM's scanning
aggressiveness in response to userspace's current needs.
- Chengming Zhou has optimized zswap's temporary working memory use
in the series 'mm/zswap: dstmem reuse optimizations and cleanups'.
- Matthew Wilcox has performed some maintenance work on the writeback
code, both code and within filesystems. The series is 'Clean up the
writeback paths'.
- Andrey Konovalov has optimized KASAN's handling of alloc and free
stack traces for secondary-level allocators, in the series 'kasan:
save mempool stack traces'.
- Andrey also performed some KASAN maintenance work in the series
'kasan: assorted clean-ups'.
- David Hildenbrand has gone to town on the rmap code. Cleanups, more
pte batching, folio conversions and more. See the series 'mm/rmap:
interface overhaul'.
- Kinsey Ho has contributed some maintenance work on the MGLRU code
in the series 'mm/mglru: Kconfig cleanup'.
- Matthew Wilcox has contributed lruvec page accounting code cleanups
in the series 'Remove some lruvec page accounting functions'"
* tag 'mm-stable-2024-01-08-15-31' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm: (361 commits)
mm, treewide: rename MAX_ORDER to MAX_PAGE_ORDER
mm, treewide: introduce NR_PAGE_ORDERS
selftests/mm: add separate UFFDIO_MOVE test for PMD splitting
selftests/mm: skip test if application doesn't has root privileges
selftests/mm: conform test to TAP format output
selftests: mm: hugepage-mmap: conform to TAP format output
selftests/mm: gup_test: conform test to TAP format output
mm/selftests: hugepage-mremap: conform test to TAP format output
mm/vmstat: move pgdemote_* out of CONFIG_NUMA_BALANCING
mm: zsmalloc: return -ENOSPC rather than -EINVAL in zs_malloc while size is too large
mm/memcontrol: remove __mod_lruvec_page_state()
mm/khugepaged: use a folio more in collapse_file()
slub: use a folio in __kmalloc_large_node
slub: use folio APIs in free_large_kmalloc()
slub: use alloc_pages_node() in alloc_slab_page()
mm: remove inc/dec lruvec page state functions
mm: ratelimit stat flush from workingset shrinker
kasan: stop leaking stack trace handles
mm/mglru: remove CONFIG_TRANSPARENT_HUGEPAGE
mm/mglru: add dummy pmd_dirty()
...
524 lines
17 KiB
C
524 lines
17 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __CPUHOTPLUG_H
|
|
#define __CPUHOTPLUG_H
|
|
|
|
#include <linux/types.h>
|
|
|
|
/*
|
|
* CPU-up CPU-down
|
|
*
|
|
* BP AP BP AP
|
|
*
|
|
* OFFLINE OFFLINE
|
|
* | ^
|
|
* v |
|
|
* BRINGUP_CPU->AP_OFFLINE BRINGUP_CPU <- AP_IDLE_DEAD (idle thread/play_dead)
|
|
* | AP_OFFLINE
|
|
* v (IRQ-off) ,---------------^
|
|
* AP_ONLNE | (stop_machine)
|
|
* | TEARDOWN_CPU <- AP_ONLINE_IDLE
|
|
* | ^
|
|
* v |
|
|
* AP_ACTIVE AP_ACTIVE
|
|
*/
|
|
|
|
/*
|
|
* CPU hotplug states. The state machine invokes the installed state
|
|
* startup callbacks sequentially from CPUHP_OFFLINE + 1 to CPUHP_ONLINE
|
|
* during a CPU online operation. During a CPU offline operation the
|
|
* installed teardown callbacks are invoked in the reverse order from
|
|
* CPU_ONLINE - 1 down to CPUHP_OFFLINE.
|
|
*
|
|
* The state space has three sections: PREPARE, STARTING and ONLINE.
|
|
*
|
|
* PREPARE: The callbacks are invoked on a control CPU before the
|
|
* hotplugged CPU is started up or after the hotplugged CPU has died.
|
|
*
|
|
* STARTING: The callbacks are invoked on the hotplugged CPU from the low level
|
|
* hotplug startup/teardown code with interrupts disabled.
|
|
*
|
|
* ONLINE: The callbacks are invoked on the hotplugged CPU from the per CPU
|
|
* hotplug thread with interrupts and preemption enabled.
|
|
*
|
|
* Adding explicit states to this enum is only necessary when:
|
|
*
|
|
* 1) The state is within the STARTING section
|
|
*
|
|
* 2) The state has ordering constraints vs. other states in the
|
|
* same section.
|
|
*
|
|
* If neither #1 nor #2 apply, please use the dynamic state space when
|
|
* setting up a state by using CPUHP_BP_PREPARE_DYN or CPUHP_AP_ONLINE_DYN
|
|
* for the @state argument of the setup function.
|
|
*
|
|
* See Documentation/core-api/cpu_hotplug.rst for further information and
|
|
* examples.
|
|
*/
|
|
enum cpuhp_state {
|
|
CPUHP_INVALID = -1,
|
|
|
|
/* PREPARE section invoked on a control CPU */
|
|
CPUHP_OFFLINE = 0,
|
|
CPUHP_CREATE_THREADS,
|
|
CPUHP_PERF_PREPARE,
|
|
CPUHP_PERF_X86_PREPARE,
|
|
CPUHP_PERF_X86_AMD_UNCORE_PREP,
|
|
CPUHP_PERF_POWER,
|
|
CPUHP_PERF_SUPERH,
|
|
CPUHP_X86_HPET_DEAD,
|
|
CPUHP_X86_MCE_DEAD,
|
|
CPUHP_VIRT_NET_DEAD,
|
|
CPUHP_IBMVNIC_DEAD,
|
|
CPUHP_SLUB_DEAD,
|
|
CPUHP_DEBUG_OBJ_DEAD,
|
|
CPUHP_MM_WRITEBACK_DEAD,
|
|
CPUHP_MM_VMSTAT_DEAD,
|
|
CPUHP_SOFTIRQ_DEAD,
|
|
CPUHP_NET_MVNETA_DEAD,
|
|
CPUHP_CPUIDLE_DEAD,
|
|
CPUHP_ARM64_FPSIMD_DEAD,
|
|
CPUHP_ARM_OMAP_WAKE_DEAD,
|
|
CPUHP_IRQ_POLL_DEAD,
|
|
CPUHP_BLOCK_SOFTIRQ_DEAD,
|
|
CPUHP_BIO_DEAD,
|
|
CPUHP_ACPI_CPUDRV_DEAD,
|
|
CPUHP_S390_PFAULT_DEAD,
|
|
CPUHP_BLK_MQ_DEAD,
|
|
CPUHP_FS_BUFF_DEAD,
|
|
CPUHP_PRINTK_DEAD,
|
|
CPUHP_MM_MEMCQ_DEAD,
|
|
CPUHP_PERCPU_CNT_DEAD,
|
|
CPUHP_RADIX_DEAD,
|
|
CPUHP_PAGE_ALLOC,
|
|
CPUHP_NET_DEV_DEAD,
|
|
CPUHP_PCI_XGENE_DEAD,
|
|
CPUHP_IOMMU_IOVA_DEAD,
|
|
CPUHP_AP_ARM_CACHE_B15_RAC_DEAD,
|
|
CPUHP_PADATA_DEAD,
|
|
CPUHP_AP_DTPM_CPU_DEAD,
|
|
CPUHP_RANDOM_PREPARE,
|
|
CPUHP_WORKQUEUE_PREP,
|
|
CPUHP_POWER_NUMA_PREPARE,
|
|
CPUHP_HRTIMERS_PREPARE,
|
|
CPUHP_PROFILE_PREPARE,
|
|
CPUHP_X2APIC_PREPARE,
|
|
CPUHP_SMPCFD_PREPARE,
|
|
CPUHP_RELAY_PREPARE,
|
|
CPUHP_MD_RAID5_PREPARE,
|
|
CPUHP_RCUTREE_PREP,
|
|
CPUHP_CPUIDLE_COUPLED_PREPARE,
|
|
CPUHP_POWERPC_PMAC_PREPARE,
|
|
CPUHP_POWERPC_MMU_CTX_PREPARE,
|
|
CPUHP_XEN_PREPARE,
|
|
CPUHP_XEN_EVTCHN_PREPARE,
|
|
CPUHP_ARM_SHMOBILE_SCU_PREPARE,
|
|
CPUHP_SH_SH3X_PREPARE,
|
|
CPUHP_TOPOLOGY_PREPARE,
|
|
CPUHP_NET_IUCV_PREPARE,
|
|
CPUHP_ARM_BL_PREPARE,
|
|
CPUHP_TRACE_RB_PREPARE,
|
|
CPUHP_MM_ZS_PREPARE,
|
|
CPUHP_MM_ZSWP_POOL_PREPARE,
|
|
CPUHP_KVM_PPC_BOOK3S_PREPARE,
|
|
CPUHP_ZCOMP_PREPARE,
|
|
CPUHP_TIMERS_PREPARE,
|
|
CPUHP_MIPS_SOC_PREPARE,
|
|
CPUHP_BP_PREPARE_DYN,
|
|
CPUHP_BP_PREPARE_DYN_END = CPUHP_BP_PREPARE_DYN + 20,
|
|
CPUHP_BP_KICK_AP,
|
|
CPUHP_BRINGUP_CPU,
|
|
|
|
/*
|
|
* STARTING section invoked on the hotplugged CPU in low level
|
|
* bringup and teardown code.
|
|
*/
|
|
CPUHP_AP_IDLE_DEAD,
|
|
CPUHP_AP_OFFLINE,
|
|
CPUHP_AP_CACHECTRL_STARTING,
|
|
CPUHP_AP_SCHED_STARTING,
|
|
CPUHP_AP_RCUTREE_DYING,
|
|
CPUHP_AP_CPU_PM_STARTING,
|
|
CPUHP_AP_IRQ_GIC_STARTING,
|
|
CPUHP_AP_IRQ_HIP04_STARTING,
|
|
CPUHP_AP_IRQ_APPLE_AIC_STARTING,
|
|
CPUHP_AP_IRQ_ARMADA_XP_STARTING,
|
|
CPUHP_AP_IRQ_BCM2836_STARTING,
|
|
CPUHP_AP_IRQ_MIPS_GIC_STARTING,
|
|
CPUHP_AP_IRQ_LOONGARCH_STARTING,
|
|
CPUHP_AP_IRQ_SIFIVE_PLIC_STARTING,
|
|
CPUHP_AP_ARM_MVEBU_COHERENCY,
|
|
CPUHP_AP_PERF_X86_AMD_UNCORE_STARTING,
|
|
CPUHP_AP_PERF_X86_STARTING,
|
|
CPUHP_AP_PERF_X86_AMD_IBS_STARTING,
|
|
CPUHP_AP_PERF_X86_CSTATE_STARTING,
|
|
CPUHP_AP_PERF_XTENSA_STARTING,
|
|
CPUHP_AP_ARM_VFP_STARTING,
|
|
CPUHP_AP_ARM64_DEBUG_MONITORS_STARTING,
|
|
CPUHP_AP_PERF_ARM_HW_BREAKPOINT_STARTING,
|
|
CPUHP_AP_PERF_ARM_ACPI_STARTING,
|
|
CPUHP_AP_PERF_ARM_STARTING,
|
|
CPUHP_AP_PERF_RISCV_STARTING,
|
|
CPUHP_AP_ARM_L2X0_STARTING,
|
|
CPUHP_AP_EXYNOS4_MCT_TIMER_STARTING,
|
|
CPUHP_AP_ARM_ARCH_TIMER_STARTING,
|
|
CPUHP_AP_ARM_ARCH_TIMER_EVTSTRM_STARTING,
|
|
CPUHP_AP_ARM_GLOBAL_TIMER_STARTING,
|
|
CPUHP_AP_JCORE_TIMER_STARTING,
|
|
CPUHP_AP_ARM_TWD_STARTING,
|
|
CPUHP_AP_QCOM_TIMER_STARTING,
|
|
CPUHP_AP_TEGRA_TIMER_STARTING,
|
|
CPUHP_AP_ARMADA_TIMER_STARTING,
|
|
CPUHP_AP_MIPS_GIC_TIMER_STARTING,
|
|
CPUHP_AP_ARC_TIMER_STARTING,
|
|
CPUHP_AP_RISCV_TIMER_STARTING,
|
|
CPUHP_AP_CLINT_TIMER_STARTING,
|
|
CPUHP_AP_CSKY_TIMER_STARTING,
|
|
CPUHP_AP_TI_GP_TIMER_STARTING,
|
|
CPUHP_AP_HYPERV_TIMER_STARTING,
|
|
/* Must be the last timer callback */
|
|
CPUHP_AP_DUMMY_TIMER_STARTING,
|
|
CPUHP_AP_ARM_XEN_STARTING,
|
|
CPUHP_AP_ARM_XEN_RUNSTATE_STARTING,
|
|
CPUHP_AP_ARM_CORESIGHT_STARTING,
|
|
CPUHP_AP_ARM_CORESIGHT_CTI_STARTING,
|
|
CPUHP_AP_ARM64_ISNDEP_STARTING,
|
|
CPUHP_AP_SMPCFD_DYING,
|
|
CPUHP_AP_HRTIMERS_DYING,
|
|
CPUHP_AP_X86_TBOOT_DYING,
|
|
CPUHP_AP_ARM_CACHE_B15_RAC_DYING,
|
|
CPUHP_AP_ONLINE,
|
|
CPUHP_TEARDOWN_CPU,
|
|
|
|
/* Online section invoked on the hotplugged CPU from the hotplug thread */
|
|
CPUHP_AP_ONLINE_IDLE,
|
|
CPUHP_AP_HYPERV_ONLINE,
|
|
CPUHP_AP_KVM_ONLINE,
|
|
CPUHP_AP_SCHED_WAIT_EMPTY,
|
|
CPUHP_AP_SMPBOOT_THREADS,
|
|
CPUHP_AP_IRQ_AFFINITY_ONLINE,
|
|
CPUHP_AP_BLK_MQ_ONLINE,
|
|
CPUHP_AP_ARM_MVEBU_SYNC_CLOCKS,
|
|
CPUHP_AP_X86_INTEL_EPB_ONLINE,
|
|
CPUHP_AP_PERF_ONLINE,
|
|
CPUHP_AP_PERF_X86_ONLINE,
|
|
CPUHP_AP_PERF_X86_UNCORE_ONLINE,
|
|
CPUHP_AP_PERF_X86_AMD_UNCORE_ONLINE,
|
|
CPUHP_AP_PERF_X86_AMD_POWER_ONLINE,
|
|
CPUHP_AP_PERF_X86_RAPL_ONLINE,
|
|
CPUHP_AP_PERF_X86_CSTATE_ONLINE,
|
|
CPUHP_AP_PERF_S390_CF_ONLINE,
|
|
CPUHP_AP_PERF_S390_SF_ONLINE,
|
|
CPUHP_AP_PERF_ARM_CCI_ONLINE,
|
|
CPUHP_AP_PERF_ARM_CCN_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_CPA_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_DDRC_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_HHA_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_L3_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_PA_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_SLLC_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HISI_PCIE_PMU_ONLINE,
|
|
CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE,
|
|
CPUHP_AP_PERF_ARM_L2X0_ONLINE,
|
|
CPUHP_AP_PERF_ARM_QCOM_L2_ONLINE,
|
|
CPUHP_AP_PERF_ARM_QCOM_L3_ONLINE,
|
|
CPUHP_AP_PERF_ARM_APM_XGENE_ONLINE,
|
|
CPUHP_AP_PERF_ARM_CAVIUM_TX2_UNCORE_ONLINE,
|
|
CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_NEST_IMC_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_CORE_IMC_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_THREAD_IMC_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_TRACE_IMC_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_HV_24x7_ONLINE,
|
|
CPUHP_AP_PERF_POWERPC_HV_GPCI_ONLINE,
|
|
CPUHP_AP_PERF_CSKY_ONLINE,
|
|
CPUHP_AP_WATCHDOG_ONLINE,
|
|
CPUHP_AP_WORKQUEUE_ONLINE,
|
|
CPUHP_AP_RANDOM_ONLINE,
|
|
CPUHP_AP_RCUTREE_ONLINE,
|
|
CPUHP_AP_BASE_CACHEINFO_ONLINE,
|
|
CPUHP_AP_ONLINE_DYN,
|
|
CPUHP_AP_ONLINE_DYN_END = CPUHP_AP_ONLINE_DYN + 40,
|
|
CPUHP_AP_X86_HPET_ONLINE,
|
|
CPUHP_AP_X86_KVM_CLK_ONLINE,
|
|
CPUHP_AP_ACTIVE,
|
|
CPUHP_ONLINE,
|
|
};
|
|
|
|
int __cpuhp_setup_state(enum cpuhp_state state, const char *name, bool invoke,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu), bool multi_instance);
|
|
|
|
int __cpuhp_setup_state_cpuslocked(enum cpuhp_state state, const char *name,
|
|
bool invoke,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu),
|
|
bool multi_instance);
|
|
/**
|
|
* cpuhp_setup_state - Setup hotplug state callbacks with calling the @startup
|
|
* callback
|
|
* @state: The state for which the calls are installed
|
|
* @name: Name of the callback (will be used in debug output)
|
|
* @startup: startup callback function or NULL if not required
|
|
* @teardown: teardown callback function or NULL if not required
|
|
*
|
|
* Installs the callback functions and invokes the @startup callback on
|
|
* the online cpus which have already reached the @state.
|
|
*/
|
|
static inline int cpuhp_setup_state(enum cpuhp_state state,
|
|
const char *name,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu))
|
|
{
|
|
return __cpuhp_setup_state(state, name, true, startup, teardown, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_setup_state_cpuslocked - Setup hotplug state callbacks with calling
|
|
* @startup callback from a cpus_read_lock()
|
|
* held region
|
|
* @state: The state for which the calls are installed
|
|
* @name: Name of the callback (will be used in debug output)
|
|
* @startup: startup callback function or NULL if not required
|
|
* @teardown: teardown callback function or NULL if not required
|
|
*
|
|
* Same as cpuhp_setup_state() except that it must be invoked from within a
|
|
* cpus_read_lock() held region.
|
|
*/
|
|
static inline int cpuhp_setup_state_cpuslocked(enum cpuhp_state state,
|
|
const char *name,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu))
|
|
{
|
|
return __cpuhp_setup_state_cpuslocked(state, name, true, startup,
|
|
teardown, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_setup_state_nocalls - Setup hotplug state callbacks without calling the
|
|
* @startup callback
|
|
* @state: The state for which the calls are installed
|
|
* @name: Name of the callback.
|
|
* @startup: startup callback function or NULL if not required
|
|
* @teardown: teardown callback function or NULL if not required
|
|
*
|
|
* Same as cpuhp_setup_state() except that the @startup callback is not
|
|
* invoked during installation. NOP if SMP=n or HOTPLUG_CPU=n.
|
|
*/
|
|
static inline int cpuhp_setup_state_nocalls(enum cpuhp_state state,
|
|
const char *name,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu))
|
|
{
|
|
return __cpuhp_setup_state(state, name, false, startup, teardown,
|
|
false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_setup_state_nocalls_cpuslocked - Setup hotplug state callbacks without
|
|
* invoking the @startup callback from
|
|
* a cpus_read_lock() held region
|
|
* callbacks
|
|
* @state: The state for which the calls are installed
|
|
* @name: Name of the callback.
|
|
* @startup: startup callback function or NULL if not required
|
|
* @teardown: teardown callback function or NULL if not required
|
|
*
|
|
* Same as cpuhp_setup_state_nocalls() except that it must be invoked from
|
|
* within a cpus_read_lock() held region.
|
|
*/
|
|
static inline int cpuhp_setup_state_nocalls_cpuslocked(enum cpuhp_state state,
|
|
const char *name,
|
|
int (*startup)(unsigned int cpu),
|
|
int (*teardown)(unsigned int cpu))
|
|
{
|
|
return __cpuhp_setup_state_cpuslocked(state, name, false, startup,
|
|
teardown, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_setup_state_multi - Add callbacks for multi state
|
|
* @state: The state for which the calls are installed
|
|
* @name: Name of the callback.
|
|
* @startup: startup callback function or NULL if not required
|
|
* @teardown: teardown callback function or NULL if not required
|
|
*
|
|
* Sets the internal multi_instance flag and prepares a state to work as a multi
|
|
* instance callback. No callbacks are invoked at this point. The callbacks are
|
|
* invoked once an instance for this state are registered via
|
|
* cpuhp_state_add_instance() or cpuhp_state_add_instance_nocalls()
|
|
*/
|
|
static inline int cpuhp_setup_state_multi(enum cpuhp_state state,
|
|
const char *name,
|
|
int (*startup)(unsigned int cpu,
|
|
struct hlist_node *node),
|
|
int (*teardown)(unsigned int cpu,
|
|
struct hlist_node *node))
|
|
{
|
|
return __cpuhp_setup_state(state, name, false,
|
|
(void *) startup,
|
|
(void *) teardown, true);
|
|
}
|
|
|
|
int __cpuhp_state_add_instance(enum cpuhp_state state, struct hlist_node *node,
|
|
bool invoke);
|
|
int __cpuhp_state_add_instance_cpuslocked(enum cpuhp_state state,
|
|
struct hlist_node *node, bool invoke);
|
|
|
|
/**
|
|
* cpuhp_state_add_instance - Add an instance for a state and invoke startup
|
|
* callback.
|
|
* @state: The state for which the instance is installed
|
|
* @node: The node for this individual state.
|
|
*
|
|
* Installs the instance for the @state and invokes the registered startup
|
|
* callback on the online cpus which have already reached the @state. The
|
|
* @state must have been earlier marked as multi-instance by
|
|
* cpuhp_setup_state_multi().
|
|
*/
|
|
static inline int cpuhp_state_add_instance(enum cpuhp_state state,
|
|
struct hlist_node *node)
|
|
{
|
|
return __cpuhp_state_add_instance(state, node, true);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_state_add_instance_nocalls - Add an instance for a state without
|
|
* invoking the startup callback.
|
|
* @state: The state for which the instance is installed
|
|
* @node: The node for this individual state.
|
|
*
|
|
* Installs the instance for the @state. The @state must have been earlier
|
|
* marked as multi-instance by cpuhp_setup_state_multi. NOP if SMP=n or
|
|
* HOTPLUG_CPU=n.
|
|
*/
|
|
static inline int cpuhp_state_add_instance_nocalls(enum cpuhp_state state,
|
|
struct hlist_node *node)
|
|
{
|
|
return __cpuhp_state_add_instance(state, node, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_state_add_instance_nocalls_cpuslocked - Add an instance for a state
|
|
* without invoking the startup
|
|
* callback from a cpus_read_lock()
|
|
* held region.
|
|
* @state: The state for which the instance is installed
|
|
* @node: The node for this individual state.
|
|
*
|
|
* Same as cpuhp_state_add_instance_nocalls() except that it must be
|
|
* invoked from within a cpus_read_lock() held region.
|
|
*/
|
|
static inline int
|
|
cpuhp_state_add_instance_nocalls_cpuslocked(enum cpuhp_state state,
|
|
struct hlist_node *node)
|
|
{
|
|
return __cpuhp_state_add_instance_cpuslocked(state, node, false);
|
|
}
|
|
|
|
void __cpuhp_remove_state(enum cpuhp_state state, bool invoke);
|
|
void __cpuhp_remove_state_cpuslocked(enum cpuhp_state state, bool invoke);
|
|
|
|
/**
|
|
* cpuhp_remove_state - Remove hotplug state callbacks and invoke the teardown
|
|
* @state: The state for which the calls are removed
|
|
*
|
|
* Removes the callback functions and invokes the teardown callback on
|
|
* the online cpus which have already reached the @state.
|
|
*/
|
|
static inline void cpuhp_remove_state(enum cpuhp_state state)
|
|
{
|
|
__cpuhp_remove_state(state, true);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_remove_state_nocalls - Remove hotplug state callbacks without invoking
|
|
* the teardown callback
|
|
* @state: The state for which the calls are removed
|
|
*/
|
|
static inline void cpuhp_remove_state_nocalls(enum cpuhp_state state)
|
|
{
|
|
__cpuhp_remove_state(state, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_remove_state_nocalls_cpuslocked - Remove hotplug state callbacks without invoking
|
|
* teardown from a cpus_read_lock() held region.
|
|
* @state: The state for which the calls are removed
|
|
*
|
|
* Same as cpuhp_remove_state nocalls() except that it must be invoked
|
|
* from within a cpus_read_lock() held region.
|
|
*/
|
|
static inline void cpuhp_remove_state_nocalls_cpuslocked(enum cpuhp_state state)
|
|
{
|
|
__cpuhp_remove_state_cpuslocked(state, false);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_remove_multi_state - Remove hotplug multi state callback
|
|
* @state: The state for which the calls are removed
|
|
*
|
|
* Removes the callback functions from a multi state. This is the reverse of
|
|
* cpuhp_setup_state_multi(). All instances should have been removed before
|
|
* invoking this function.
|
|
*/
|
|
static inline void cpuhp_remove_multi_state(enum cpuhp_state state)
|
|
{
|
|
__cpuhp_remove_state(state, false);
|
|
}
|
|
|
|
int __cpuhp_state_remove_instance(enum cpuhp_state state,
|
|
struct hlist_node *node, bool invoke);
|
|
|
|
/**
|
|
* cpuhp_state_remove_instance - Remove hotplug instance from state and invoke
|
|
* the teardown callback
|
|
* @state: The state from which the instance is removed
|
|
* @node: The node for this individual state.
|
|
*
|
|
* Removes the instance and invokes the teardown callback on the online cpus
|
|
* which have already reached @state.
|
|
*/
|
|
static inline int cpuhp_state_remove_instance(enum cpuhp_state state,
|
|
struct hlist_node *node)
|
|
{
|
|
return __cpuhp_state_remove_instance(state, node, true);
|
|
}
|
|
|
|
/**
|
|
* cpuhp_state_remove_instance_nocalls - Remove hotplug instance from state
|
|
* without invoking the teardown callback
|
|
* @state: The state from which the instance is removed
|
|
* @node: The node for this individual state.
|
|
*
|
|
* Removes the instance without invoking the teardown callback.
|
|
*/
|
|
static inline int cpuhp_state_remove_instance_nocalls(enum cpuhp_state state,
|
|
struct hlist_node *node)
|
|
{
|
|
return __cpuhp_state_remove_instance(state, node, false);
|
|
}
|
|
|
|
#ifdef CONFIG_SMP
|
|
void cpuhp_online_idle(enum cpuhp_state state);
|
|
#else
|
|
static inline void cpuhp_online_idle(enum cpuhp_state state) { }
|
|
#endif
|
|
|
|
struct task_struct;
|
|
|
|
void cpuhp_ap_sync_alive(void);
|
|
void arch_cpuhp_sync_state_poll(void);
|
|
void arch_cpuhp_cleanup_kick_cpu(unsigned int cpu);
|
|
int arch_cpuhp_kick_ap_alive(unsigned int cpu, struct task_struct *tidle);
|
|
bool arch_cpuhp_init_parallel_bringup(void);
|
|
|
|
#ifdef CONFIG_HOTPLUG_CORE_SYNC_DEAD
|
|
void cpuhp_ap_report_dead(void);
|
|
void arch_cpuhp_cleanup_dead_cpu(unsigned int cpu);
|
|
#else
|
|
static inline void cpuhp_ap_report_dead(void) { }
|
|
static inline void arch_cpuhp_cleanup_dead_cpu(unsigned int cpu) { }
|
|
#endif
|
|
|
|
#endif
|