diff options
Diffstat (limited to 'kernel/cpuset.c')
| -rw-r--r-- | kernel/cpuset.c | 15 | 
1 files changed, 5 insertions, 10 deletions
diff --git a/kernel/cpuset.c b/kernel/cpuset.c index 39e52175f4af..bf69717325b4 100644 --- a/kernel/cpuset.c +++ b/kernel/cpuset.c @@ -119,12 +119,6 @@ static inline struct cpuset *css_cs(struct cgroup_subsys_state *css)  	return css ? container_of(css, struct cpuset, css) : NULL;  } -/* Retrieve the cpuset for a cgroup */ -static inline struct cpuset *cgroup_cs(struct cgroup *cgrp) -{ -	return css_cs(cgroup_css(cgrp, cpuset_subsys_id)); -} -  /* Retrieve the cpuset for a task */  static inline struct cpuset *task_cs(struct task_struct *task)  { @@ -1459,7 +1453,7 @@ static int cpuset_can_attach(struct cgroup_subsys_state *css,  	    (cpumask_empty(cs->cpus_allowed) || nodes_empty(cs->mems_allowed)))  		goto out_unlock; -	cgroup_taskset_for_each(task, css->cgroup, tset) { +	cgroup_taskset_for_each(task, css, tset) {  		/*  		 * Kthreads which disallow setaffinity shouldn't be moved  		 * to a new cpuset; we don't want to change their cpu @@ -1511,9 +1505,10 @@ static void cpuset_attach(struct cgroup_subsys_state *css,  	struct mm_struct *mm;  	struct task_struct *task;  	struct task_struct *leader = cgroup_taskset_first(tset); -	struct cgroup *oldcgrp = cgroup_taskset_cur_cgroup(tset); +	struct cgroup_subsys_state *oldcss = cgroup_taskset_cur_css(tset, +							cpuset_subsys_id);  	struct cpuset *cs = css_cs(css); -	struct cpuset *oldcs = cgroup_cs(oldcgrp); +	struct cpuset *oldcs = css_cs(oldcss);  	struct cpuset *cpus_cs = effective_cpumask_cpuset(cs);  	struct cpuset *mems_cs = effective_nodemask_cpuset(cs); @@ -1527,7 +1522,7 @@ static void cpuset_attach(struct cgroup_subsys_state *css,  	guarantee_online_mems(mems_cs, &cpuset_attach_nodemask_to); -	cgroup_taskset_for_each(task, css->cgroup, tset) { +	cgroup_taskset_for_each(task, css, tset) {  		/*  		 * can_attach beforehand should guarantee that this doesn't  		 * fail.  TODO: have a better way to handle failure here  | 
