From: Linus Torvalds Date: Tue, 15 Sep 2009 16:39:44 +0000 (-0700) Subject: Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/percpu X-Git-Tag: firefly_0821_release~12956 X-Git-Url: http://plrg.eecs.uci.edu/git/?a=commitdiff_plain;h=ada3fa15057205b7d3f727bba5cd26b5912e350f;p=firefly-linux-kernel-4.4.55.git Merge branch 'for-linus' of git://git./linux/kernel/git/tj/percpu * 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/percpu: (46 commits) powerpc64: convert to dynamic percpu allocator sparc64: use embedding percpu first chunk allocator percpu: kill lpage first chunk allocator x86,percpu: use embedding for 64bit NUMA and page for 32bit NUMA percpu: update embedding first chunk allocator to handle sparse units percpu: use group information to allocate vmap areas sparsely vmalloc: implement pcpu_get_vm_areas() vmalloc: separate out insert_vmalloc_vm() percpu: add chunk->base_addr percpu: add pcpu_unit_offsets[] percpu: introduce pcpu_alloc_info and pcpu_group_info percpu: move pcpu_lpage_build_unit_map() and pcpul_lpage_dump_cfg() upward percpu: add @align to pcpu_fc_alloc_fn_t percpu: make @dyn_size mandatory for pcpu_setup_first_chunk() percpu: drop @static_size from first chunk allocators percpu: generalize first chunk allocator selection percpu: build first chunk allocators selectively percpu: rename 4k first chunk allocator to page percpu: improve boot messages percpu: fix pcpu_reclaim() locking ... Fix trivial conflict as by Tejun Heo in kernel/sched.c --- ada3fa15057205b7d3f727bba5cd26b5912e350f diff --cc kernel/sched.c index e27a53685ed9,d3d7e7694da6..d9db3fb17573 --- a/kernel/sched.c +++ b/kernel/sched.c @@@ -295,7 -318,7 +295,7 @@@ struct task_group root_task_group /* Default task group's sched entity on each cpu */ static DEFINE_PER_CPU(struct sched_entity, init_sched_entity); /* Default task group's cfs_rq on each cpu */ - static DEFINE_PER_CPU(struct cfs_rq, init_tg_cfs_rq) ____cacheline_aligned_in_smp; -static DEFINE_PER_CPU_SHARED_ALIGNED(struct cfs_rq, init_cfs_rq); ++static DEFINE_PER_CPU_SHARED_ALIGNED(struct cfs_rq, init_tg_cfs_rq); #endif /* CONFIG_FAIR_GROUP_SCHED */ #ifdef CONFIG_RT_GROUP_SCHED diff --cc kernel/trace/trace_events.c index 78b1ed230177,0db0a41e0079..97e2c4d2e9eb --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c @@@ -1465,10 -1365,10 +1465,10 @@@ function_test_events_call(unsigned lon entry->ip = ip; entry->parent_ip = parent_ip; - trace_nowake_buffer_unlock_commit(event, flags, pc); + trace_nowake_buffer_unlock_commit(buffer, event, flags, pc); out: - atomic_dec(&per_cpu(test_event_disable, cpu)); + atomic_dec(&per_cpu(ftrace_test_event_disable, cpu)); ftrace_preempt_enable(resched); } diff --cc net/rds/ib_stats.c index 8d8488306fe4,301ae51ae409..d2c904dd6fbc --- a/net/rds/ib_stats.c +++ b/net/rds/ib_stats.c @@@ -37,9 -37,9 +37,9 @@@ #include "rds.h" #include "ib.h" - DEFINE_PER_CPU(struct rds_ib_statistics, rds_ib_stats) ____cacheline_aligned; + DEFINE_PER_CPU_SHARED_ALIGNED(struct rds_ib_statistics, rds_ib_stats); -static char *rds_ib_stat_names[] = { +static const char *const rds_ib_stat_names[] = { "ib_connect_raced", "ib_listen_closed_stale", "ib_tx_cq_call", diff --cc net/rds/iw_stats.c index d33ea790484e,fafea3cc92d7..5fe67f6a1d80 --- a/net/rds/iw_stats.c +++ b/net/rds/iw_stats.c @@@ -37,9 -37,9 +37,9 @@@ #include "rds.h" #include "iw.h" - DEFINE_PER_CPU(struct rds_iw_statistics, rds_iw_stats) ____cacheline_aligned; + DEFINE_PER_CPU_SHARED_ALIGNED(struct rds_iw_statistics, rds_iw_stats); -static char *rds_iw_stat_names[] = { +static const char *const rds_iw_stat_names[] = { "iw_connect_raced", "iw_listen_closed_stale", "iw_tx_cq_call",