e9a05bb4d3
Modify spdx tag for all CAF authored files for Kona. Change-Id: I9308c7189412b1e428a7f67ded0dc076b0e38254 Signed-off-by: Rishabh Bhatnagar <rishabhb@codeaurora.org>
192 lines
4.4 KiB
C
192 lines
4.4 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (c) 2015-2018, The Linux Foundation. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/mm_types.h>
|
|
#include <linux/swapfile.h>
|
|
#include <linux/swap.h>
|
|
|
|
#define SWAP_RATIO_GROUP_START (SWAP_FLAG_PRIO_MASK - 9) /* 32758 */
|
|
#define SWAP_RATIO_GROUP_END (SWAP_FLAG_PRIO_MASK) /* 32767 */
|
|
#define SWAP_FAST_WRITES \
|
|
((SWAPFILE_CLUSTER * (SWAP_CLUSTER_MAX / 8)) / SWAP_BATCH)
|
|
#define SWAP_SLOW_WRITES (SWAPFILE_CLUSTER / SWAP_BATCH)
|
|
|
|
/*
|
|
* The fast/slow swap write ratio.
|
|
* 100 indicates that all writes should
|
|
* go to fast swap device.
|
|
*/
|
|
int sysctl_swap_ratio = 100;
|
|
|
|
/* Enable the swap ratio feature */
|
|
int sysctl_swap_ratio_enable;
|
|
|
|
static bool is_same_group(struct swap_info_struct *a,
|
|
struct swap_info_struct *b)
|
|
{
|
|
if (!sysctl_swap_ratio_enable)
|
|
return false;
|
|
|
|
if (!is_swap_ratio_group(a->prio))
|
|
return false;
|
|
|
|
if (a->prio == b->prio)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
/* Caller must hold swap_avail_lock */
|
|
static int calculate_write_pending(struct swap_info_struct *si,
|
|
struct swap_info_struct *n)
|
|
{
|
|
int ratio = sysctl_swap_ratio;
|
|
|
|
if ((ratio < 0) || (ratio > 100))
|
|
return -EINVAL;
|
|
|
|
if (WARN_ON(!(si->flags & SWP_SYNCHRONOUS_IO)))
|
|
return -ENODEV;
|
|
|
|
if ((n->flags & SWP_SYNCHRONOUS_IO) || !is_same_group(si, n))
|
|
return -ENODEV;
|
|
|
|
si->max_writes = ratio ? SWAP_FAST_WRITES : 0;
|
|
n->max_writes = ratio ? (SWAP_FAST_WRITES * 100) /
|
|
ratio - SWAP_FAST_WRITES : SWAP_SLOW_WRITES;
|
|
|
|
si->write_pending = si->max_writes;
|
|
n->write_pending = n->max_writes;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int swap_ratio_slow(struct swap_info_struct **si, int node)
|
|
{
|
|
struct swap_info_struct *n = NULL;
|
|
int ret = 0;
|
|
|
|
spin_lock(&(*si)->lock);
|
|
spin_lock(&swap_avail_lock);
|
|
if (&(*si)->avail_lists[node] == plist_last(&swap_avail_heads[node])) {
|
|
/* just to make skip work */
|
|
n = *si;
|
|
ret = -ENODEV;
|
|
goto skip;
|
|
}
|
|
n = plist_next_entry(&(*si)->avail_lists[node],
|
|
struct swap_info_struct,
|
|
avail_lists[node]);
|
|
if (n == *si) {
|
|
/* No other swap device */
|
|
ret = -ENODEV;
|
|
goto skip;
|
|
}
|
|
|
|
spin_unlock(&swap_avail_lock);
|
|
spin_lock(&n->lock);
|
|
spin_lock(&swap_avail_lock);
|
|
|
|
if ((*si)->flags & SWP_SYNCHRONOUS_IO) {
|
|
if ((*si)->write_pending) {
|
|
(*si)->write_pending--;
|
|
goto exit;
|
|
} else {
|
|
if ((n->flags & SWP_SYNCHRONOUS_IO) ||
|
|
!is_same_group(*si, n)) {
|
|
/* Should never happen */
|
|
ret = -ENODEV;
|
|
} else if (n->write_pending) {
|
|
/*
|
|
* Requeue fast device, since there are pending
|
|
* writes for slow device.
|
|
*/
|
|
plist_requeue(&(*si)->avail_lists[node],
|
|
&swap_avail_heads[node]);
|
|
n->write_pending--;
|
|
spin_unlock(&(*si)->lock);
|
|
*si = n;
|
|
goto skip;
|
|
} else {
|
|
if (calculate_write_pending(*si, n) < 0) {
|
|
ret = -ENODEV;
|
|
goto exit;
|
|
}
|
|
/* Restart from fast device */
|
|
(*si)->write_pending--;
|
|
}
|
|
}
|
|
} else {
|
|
if (!(n->flags & SWP_SYNCHRONOUS_IO) ||
|
|
!is_same_group(*si, n)) {
|
|
/* Should never happen */
|
|
ret = -ENODEV;
|
|
} else if (n->write_pending) {
|
|
/*
|
|
* Pending writes for fast device.
|
|
* We reach here when slow device is swapped on first,
|
|
* before fast device.
|
|
*/
|
|
/* requeue slow device to the end */
|
|
plist_requeue(&(*si)->avail_lists[node],
|
|
&swap_avail_heads[node]);
|
|
n->write_pending--;
|
|
spin_unlock(&(*si)->lock);
|
|
*si = n;
|
|
goto skip;
|
|
} else {
|
|
if ((*si)->write_pending) {
|
|
(*si)->write_pending--;
|
|
} else {
|
|
if (calculate_write_pending(n, *si) < 0) {
|
|
ret = -ENODEV;
|
|
goto exit;
|
|
}
|
|
n->write_pending--;
|
|
plist_requeue(&(*si)->avail_lists[node],
|
|
&swap_avail_heads[node]);
|
|
spin_unlock(&(*si)->lock);
|
|
*si = n;
|
|
goto skip;
|
|
}
|
|
}
|
|
}
|
|
exit:
|
|
spin_unlock(&(*si)->lock);
|
|
skip:
|
|
spin_unlock(&swap_avail_lock);
|
|
/* n and si would have got interchanged */
|
|
spin_unlock(&n->lock);
|
|
return ret;
|
|
}
|
|
|
|
bool is_swap_ratio_group(int prio)
|
|
{
|
|
return ((prio >= SWAP_RATIO_GROUP_START) &&
|
|
(prio <= SWAP_RATIO_GROUP_END)) ? true : false;
|
|
}
|
|
|
|
void setup_swap_ratio(struct swap_info_struct *p, int prio)
|
|
{
|
|
/* Used only if sysctl_swap_ratio_enable is set */
|
|
if (is_swap_ratio_group(prio)) {
|
|
if (p->flags & SWP_SYNCHRONOUS_IO)
|
|
p->write_pending = SWAP_FAST_WRITES;
|
|
else
|
|
p->write_pending = SWAP_SLOW_WRITES;
|
|
p->max_writes = p->write_pending;
|
|
}
|
|
}
|
|
|
|
int swap_ratio(struct swap_info_struct **si, int node)
|
|
{
|
|
if (!sysctl_swap_ratio_enable)
|
|
return -ENODEV;
|
|
|
|
if (is_swap_ratio_group((*si)->prio))
|
|
return swap_ratio_slow(si, node);
|
|
else
|
|
return -ENODEV;
|
|
}
|