[Devel] [PATCH rh7 v3 4/8] oom: introduce oom timeout
Vladimir Davydov
vdavydov at parallels.com
Sun Sep 13 06:42:06 PDT 2015
Currently, we won't select a new oom victim until the previous one has
passed away. This might lead to a deadlock if an allocating task holds a
lock needed by the victim to complete. To cope with this problem, this
patch introduced oom timeout, after which a new task will be selected
even if the previous victim hasn't died. The timeout is hard-coded,
equals 5 seconds.
https://jira.sw.ru/browse/PSBM-38581
Signed-off-by: Vladimir Davydov <vdavydov at parallels.com>
---
include/linux/oom.h | 2 ++
mm/oom_kill.c | 60 ++++++++++++++++++++++++++++++++++++++++++++++-------
2 files changed, 54 insertions(+), 8 deletions(-)
diff --git a/include/linux/oom.h b/include/linux/oom.h
index e19385dd29aa..f804551c0a5d 100644
--- a/include/linux/oom.h
+++ b/include/linux/oom.h
@@ -34,6 +34,8 @@ enum oom_scan_t {
struct oom_context {
struct task_struct *owner;
struct task_struct *victim;
+ bool marked;
+ unsigned long oom_start;
wait_queue_head_t waitq;
};
diff --git a/mm/oom_kill.c b/mm/oom_kill.c
index ef7773f67c3c..2fab831187d2 100644
--- a/mm/oom_kill.c
+++ b/mm/oom_kill.c
@@ -45,6 +45,8 @@ int sysctl_oom_dump_tasks;
static DEFINE_SPINLOCK(oom_context_lock);
+#define OOM_TIMEOUT (5 * HZ)
+
#ifndef CONFIG_MEMCG
struct oom_context oom_ctx = {
.waitq = __WAIT_QUEUE_HEAD_INITIALIZER(oom_ctx.waitq),
@@ -55,6 +57,8 @@ void init_oom_context(struct oom_context *ctx)
{
ctx->owner = NULL;
ctx->victim = NULL;
+ ctx->marked = false;
+ ctx->oom_start = 0;
init_waitqueue_head(&ctx->waitq);
}
@@ -62,6 +66,7 @@ static void __release_oom_context(struct oom_context *ctx)
{
ctx->owner = NULL;
ctx->victim = NULL;
+ ctx->marked = false;
wake_up_all(&ctx->waitq);
}
@@ -291,11 +296,14 @@ enum oom_scan_t oom_scan_process_thread(struct task_struct *task,
/*
* This task already has access to memory reserves and is being killed.
- * Don't allow any other task to have access to the reserves.
+ * Try to select another one.
+ *
+ * This can only happen if oom_trylock timeout-ed, which most probably
+ * means that the victim had dead-locked.
*/
if (test_tsk_thread_flag(task, TIF_MEMDIE)) {
if (!force_kill)
- return OOM_SCAN_ABORT;
+ return OOM_SCAN_CONTINUE;
}
if (!task->mm)
return OOM_SCAN_CONTINUE;
@@ -463,8 +471,10 @@ void mark_oom_victim(struct task_struct *tsk)
memcg = try_get_mem_cgroup_from_mm(tsk->mm);
ctx = mem_cgroup_oom_context(memcg);
spin_lock(&oom_context_lock);
- if (!ctx->victim)
+ if (!ctx->victim) {
ctx->victim = tsk;
+ ctx->marked = true;
+ }
spin_unlock(&oom_context_lock);
mem_cgroup_put(memcg);
}
@@ -499,21 +509,26 @@ void exit_oom_victim(void)
static void __wait_oom_context(struct oom_context *ctx)
{
+ unsigned long now = jiffies;
+ unsigned long timeout;
DEFINE_WAIT(wait);
- if (ctx->victim == current) {
+ if (ctx->victim == current ||
+ time_after_eq(now, ctx->oom_start + OOM_TIMEOUT)) {
spin_unlock(&oom_context_lock);
return;
}
prepare_to_wait(&ctx->waitq, &wait, TASK_KILLABLE);
+ timeout = ctx->oom_start + OOM_TIMEOUT - now;
spin_unlock(&oom_context_lock);
- schedule();
+ schedule_timeout(timeout);
finish_wait(&ctx->waitq, &wait);
}
bool oom_trylock(struct mem_cgroup *memcg)
{
+ unsigned long now = jiffies;
struct mem_cgroup *iter;
struct oom_context *ctx;
@@ -528,10 +543,32 @@ bool oom_trylock(struct mem_cgroup *memcg)
iter = mem_cgroup_iter(memcg, NULL, NULL);
do {
ctx = mem_cgroup_oom_context(iter);
- if (ctx->owner || ctx->victim) {
+ if ((ctx->owner || ctx->victim) &&
+ time_before(now, ctx->oom_start + OOM_TIMEOUT)) {
__wait_oom_context(ctx);
mem_cgroup_iter_break(memcg, iter);
return false;
+ } else if (ctx->owner || ctx->victim) {
+ /*
+ * Timeout. Release the context and dump stack
+ * trace of the stuck process.
+ *
+ * To avoid dumping stack trace of the same task
+ * more than once, we mark the context that
+ * contained the victim when it was killed (see
+ * mark_oom_victim).
+ */
+ struct task_struct *p = ctx->victim;
+
+ if (p && ctx->marked) {
+ task_lock(p);
+ pr_err("OOM kill timeout: %d (%s)\n",
+ task_pid_nr(p), p->comm);
+ task_unlock(p);
+ show_stack(p, NULL);
+ }
+
+ __release_oom_context(ctx);
}
} while ((iter = mem_cgroup_iter(memcg, iter, NULL)));
@@ -544,6 +581,7 @@ bool oom_trylock(struct mem_cgroup *memcg)
BUG_ON(ctx->owner);
BUG_ON(ctx->victim);
ctx->owner = current;
+ ctx->oom_start = now;
} while ((iter = mem_cgroup_iter(memcg, iter, NULL)));
spin_unlock(&oom_context_lock);
@@ -565,7 +603,11 @@ void oom_unlock(struct mem_cgroup *memcg)
iter = mem_cgroup_iter(memcg, NULL, NULL);
do {
ctx = mem_cgroup_oom_context(iter);
- BUG_ON(ctx->owner != current);
+ if (ctx->owner != current) {
+ /* Lost ownership on timeout */
+ mem_cgroup_iter_break(memcg, iter);
+ break;
+ }
if (ctx->victim) {
victim = ctx->victim;
/*
@@ -598,7 +640,9 @@ void oom_unlock(struct mem_cgroup *memcg)
iter = mem_cgroup_iter(memcg, NULL, NULL);
do {
ctx = mem_cgroup_oom_context(iter);
- BUG_ON(ctx->owner != current);
+ if (ctx->owner != current)
+ /* Lost ownership on timeout */
+ continue;
if (!ctx->victim)
/*
* Victim already exited or nobody was killed in
--
2.1.4
More information about the Devel
mailing list