148 lines
3.3 KiB
C
148 lines
3.3 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* MCE event pool management in MCE context
|
|
*
|
|
* Copyright (C) 2015 Intel Corp.
|
|
* Author: Chen, Gong <gong.chen@linux.intel.com>
|
|
*/
|
|
#include <linux/smp.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/genalloc.h>
|
|
#include <linux/llist.h>
|
|
#include "internal.h"
|
|
|
|
/*
|
|
* printk() is not safe in MCE context. This is a lock-less memory allocator
|
|
* used to save error information organized in a lock-less list.
|
|
*
|
|
* This memory pool is only to be used to save MCE records in MCE context.
|
|
* MCE events are rare, so a fixed size memory pool should be enough. Use
|
|
* 2 pages to save MCE events for now (~80 MCE records at most).
|
|
*/
|
|
#define MCE_POOLSZ (2 * PAGE_SIZE)
|
|
|
|
static struct gen_pool *mce_evt_pool;
|
|
static LLIST_HEAD(mce_event_llist);
|
|
static char gen_pool_buf[MCE_POOLSZ];
|
|
|
|
/*
|
|
* Compare the record "t" with each of the records on list "l" to see if
|
|
* an equivalent one is present in the list.
|
|
*/
|
|
static bool is_duplicate_mce_record(struct mce_evt_llist *t, struct mce_evt_llist *l)
|
|
{
|
|
struct mce_evt_llist *node;
|
|
struct mce *m1, *m2;
|
|
|
|
m1 = &t->mce;
|
|
|
|
llist_for_each_entry(node, &l->llnode, llnode) {
|
|
m2 = &node->mce;
|
|
|
|
if (!mce_cmp(m1, m2))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/*
|
|
* The system has panicked - we'd like to peruse the list of MCE records
|
|
* that have been queued, but not seen by anyone yet. The list is in
|
|
* reverse time order, so we need to reverse it. While doing that we can
|
|
* also drop duplicate records (these were logged because some banks are
|
|
* shared between cores or by all threads on a socket).
|
|
*/
|
|
struct llist_node *mce_gen_pool_prepare_records(void)
|
|
{
|
|
struct llist_node *head;
|
|
LLIST_HEAD(new_head);
|
|
struct mce_evt_llist *node, *t;
|
|
|
|
head = llist_del_all(&mce_event_llist);
|
|
if (!head)
|
|
return NULL;
|
|
|
|
/* squeeze out duplicates while reversing order */
|
|
llist_for_each_entry_safe(node, t, head, llnode) {
|
|
if (!is_duplicate_mce_record(node, t))
|
|
llist_add(&node->llnode, &new_head);
|
|
}
|
|
|
|
return new_head.first;
|
|
}
|
|
|
|
void mce_gen_pool_process(struct work_struct *__unused)
|
|
{
|
|
struct llist_node *head;
|
|
struct mce_evt_llist *node, *tmp;
|
|
struct mce *mce;
|
|
|
|
head = llist_del_all(&mce_event_llist);
|
|
if (!head)
|
|
return;
|
|
|
|
head = llist_reverse_order(head);
|
|
llist_for_each_entry_safe(node, tmp, head, llnode) {
|
|
mce = &node->mce;
|
|
blocking_notifier_call_chain(&x86_mce_decoder_chain, 0, mce);
|
|
gen_pool_free(mce_evt_pool, (unsigned long)node, sizeof(*node));
|
|
}
|
|
}
|
|
|
|
bool mce_gen_pool_empty(void)
|
|
{
|
|
return llist_empty(&mce_event_llist);
|
|
}
|
|
|
|
int mce_gen_pool_add(struct mce *mce)
|
|
{
|
|
struct mce_evt_llist *node;
|
|
|
|
if (filter_mce(mce))
|
|
return -EINVAL;
|
|
|
|
if (!mce_evt_pool)
|
|
return -EINVAL;
|
|
|
|
node = (void *)gen_pool_alloc(mce_evt_pool, sizeof(*node));
|
|
if (!node) {
|
|
pr_warn_ratelimited("MCE records pool full!\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memcpy(&node->mce, mce, sizeof(*mce));
|
|
llist_add(&node->llnode, &mce_event_llist);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mce_gen_pool_create(void)
|
|
{
|
|
struct gen_pool *tmpp;
|
|
int ret = -ENOMEM;
|
|
|
|
tmpp = gen_pool_create(ilog2(sizeof(struct mce_evt_llist)), -1);
|
|
if (!tmpp)
|
|
goto out;
|
|
|
|
ret = gen_pool_add(tmpp, (unsigned long)gen_pool_buf, MCE_POOLSZ, -1);
|
|
if (ret) {
|
|
gen_pool_destroy(tmpp);
|
|
goto out;
|
|
}
|
|
|
|
mce_evt_pool = tmpp;
|
|
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
int mce_gen_pool_init(void)
|
|
{
|
|
/* Just init mce_gen_pool once. */
|
|
if (mce_evt_pool)
|
|
return 0;
|
|
|
|
return mce_gen_pool_create();
|
|
}
|