2026-01-29 22:25:33 +08:00

335 lines
8.0 KiB
C

// SPDX-License-Identifier: GPL-2.0
/* Copyright (C) 2021 - 2023, Shanghai Yunsilicon Technology Co., Ltd.
* All rights reserved.
*/
#include <linux/errno.h>
#include <linux/slab.h>
#include <linux/mm.h>
#include <linux/export.h>
#include <linux/bitmap.h>
#include <linux/dma-mapping.h>
#include <linux/vmalloc.h>
#include "common/driver.h"
/* Handling for queue buffers -- we allocate a bunch of memory and
* register it in a memory region at HCA virtual address 0. If the
* requested size is > max_direct, we split the allocation into
* multiple pages, so we don't require too much contiguous memory.
*/
int xsc_buf_alloc(struct xsc_core_device *xdev, int size, int max_direct,
struct xsc_buf *buf)
{
dma_addr_t t;
buf->size = size;
if (size <= max_direct) {
buf->nbufs = 1;
buf->npages = 1;
buf->page_shift = get_order(size) + PAGE_SHIFT;
buf->direct.buf = dma_alloc_coherent(&xdev->pdev->dev,
size, &t, GFP_KERNEL | __GFP_ZERO);
if (!buf->direct.buf)
return -ENOMEM;
buf->direct.map = t;
while (t & ((1 << buf->page_shift) - 1)) {
--buf->page_shift;
buf->npages *= 2;
}
} else {
int i;
buf->direct.buf = NULL;
buf->nbufs = (size + PAGE_SIZE - 1) / PAGE_SIZE;
buf->npages = buf->nbufs;
buf->page_shift = PAGE_SHIFT;
buf->page_list = kcalloc(buf->nbufs, sizeof(*buf->page_list),
GFP_KERNEL);
if (!buf->page_list)
return -ENOMEM;
for (i = 0; i < buf->nbufs; i++) {
buf->page_list[i].buf =
dma_alloc_coherent(&xdev->pdev->dev, PAGE_SIZE,
&t, GFP_KERNEL | __GFP_ZERO);
if (!buf->page_list[i].buf)
goto err_free;
buf->page_list[i].map = t;
}
if (BITS_PER_LONG == 64) {
struct page **pages;
pages = kmalloc_array(buf->nbufs, sizeof(*pages), GFP_KERNEL);
if (!pages)
goto err_free;
for (i = 0; i < buf->nbufs; i++)
pages[i] = virt_to_page(buf->page_list[i].buf);
buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP, PAGE_KERNEL);
kfree(pages);
if (!buf->direct.buf)
goto err_free;
}
}
return 0;
err_free:
xsc_buf_free(xdev, buf);
return -ENOMEM;
}
EXPORT_SYMBOL_GPL(xsc_buf_alloc);
void xsc_buf_free(struct xsc_core_device *xdev, struct xsc_buf *buf)
{
int i;
if (buf->nbufs == 1) {
dma_free_coherent(&xdev->pdev->dev, buf->size, buf->direct.buf,
buf->direct.map);
} else {
if (BITS_PER_LONG == 64 && buf->direct.buf)
vunmap(buf->direct.buf);
for (i = 0; i < buf->nbufs; i++)
if (buf->page_list[i].buf)
dma_free_coherent(&xdev->pdev->dev, PAGE_SIZE,
buf->page_list[i].buf,
buf->page_list[i].map);
kfree(buf->page_list);
}
}
EXPORT_SYMBOL_GPL(xsc_buf_free);
void xsc_fill_page_array(struct xsc_buf *buf, __be64 *pas, int npages)
{
u64 addr;
int i;
int shift = PAGE_SHIFT - PAGE_SHIFT_4K;
int mask = (1 << shift) - 1;
for (i = 0; i < npages; i++) {
if (buf->nbufs == 1)
addr = buf->direct.map + (i << PAGE_SHIFT_4K);
else
addr = buf->page_list[i >> shift].map + ((i & mask) << PAGE_SHIFT_4K);
pas[i] = cpu_to_be64(addr);
}
}
EXPORT_SYMBOL_GPL(xsc_fill_page_array);
void xsc_fill_page_frag_array(struct xsc_frag_buf *buf, __be64 *pas, int npages)
{
int i;
dma_addr_t addr;
int shift = PAGE_SHIFT - PAGE_SHIFT_4K;
int mask = (1 << shift) - 1;
for (i = 0; i < npages; i++) {
addr = buf->frags[i >> shift].map + ((i & mask) << PAGE_SHIFT_4K);
pas[i] = cpu_to_be64(addr);
}
}
EXPORT_SYMBOL_GPL(xsc_fill_page_frag_array);
static void *xsc_dma_zalloc_coherent_node(struct xsc_core_device *xdev,
size_t size, dma_addr_t *dma_handle,
int node)
{
struct xsc_dev_resource *dev_res = xdev->dev_res;
struct device *device = &xdev->pdev->dev;
int original_node;
void *cpu_handle;
/* WA for kernels that don't use numa_mem_id in alloc_pages_node */
if (node == NUMA_NO_NODE)
#ifdef HAVE_NUMA_MEM_ID
node = numa_mem_id();
#else
node = first_memory_node;
#endif
mutex_lock(&dev_res->alloc_mutex);
original_node = dev_to_node(device);
set_dev_node(device, node);
cpu_handle = dma_alloc_coherent(device, size, dma_handle,
GFP_KERNEL);
set_dev_node(device, original_node);
mutex_unlock(&dev_res->alloc_mutex);
return cpu_handle;
}
int xsc_frag_buf_alloc_node(struct xsc_core_device *xdev, int size,
struct xsc_frag_buf *buf, int node)
{
int i;
buf->size = size;
buf->npages = DIV_ROUND_UP(size, PAGE_SIZE);
buf->page_shift = PAGE_SHIFT;
buf->frags = kcalloc(buf->npages, sizeof(struct xsc_buf_list),
GFP_KERNEL);
if (!buf->frags)
goto err_out;
for (i = 0; i < buf->npages; i++) {
struct xsc_buf_list *frag = &buf->frags[i];
int frag_sz = min_t(int, size, PAGE_SIZE);
frag->buf = xsc_dma_zalloc_coherent_node(xdev, frag_sz,
&frag->map, node);
if (!frag->buf)
goto err_free_buf;
if (frag->map & ((1 << buf->page_shift) - 1)) {
dma_free_coherent(&xdev->pdev->dev, frag_sz,
buf->frags[i].buf, buf->frags[i].map);
xsc_core_warn(xdev, "unexpected map alignment: %pad, page_shift=%d\n",
&frag->map, buf->page_shift);
goto err_free_buf;
}
size -= frag_sz;
}
return 0;
err_free_buf:
while (i--)
dma_free_coherent(&xdev->pdev->dev, PAGE_SIZE, buf->frags[i].buf,
buf->frags[i].map);
kfree(buf->frags);
err_out:
return -ENOMEM;
}
EXPORT_SYMBOL_GPL(xsc_frag_buf_alloc_node);
void xsc_frag_buf_free(struct xsc_core_device *xdev, struct xsc_frag_buf *buf)
{
int size = buf->size;
int i;
for (i = 0; i < buf->npages; i++) {
int frag_sz = min_t(int, size, PAGE_SIZE);
dma_free_coherent(&xdev->pdev->dev, frag_sz, buf->frags[i].buf,
buf->frags[i].map);
size -= frag_sz;
}
kfree(buf->frags);
}
EXPORT_SYMBOL_GPL(xsc_frag_buf_free);
static struct xsc_db_pgdir *xsc_alloc_db_pgdir(struct xsc_core_device *xdev,
int node)
{
u32 db_per_page = PAGE_SIZE / cache_line_size();
struct xsc_db_pgdir *pgdir;
pgdir = kzalloc(sizeof(*pgdir), GFP_KERNEL);
if (!pgdir)
return NULL;
pgdir->bitmap = bitmap_zalloc(db_per_page, GFP_KERNEL);
if (!pgdir->bitmap) {
kfree(pgdir);
return NULL;
}
bitmap_fill(pgdir->bitmap, db_per_page);
pgdir->db_page = xsc_dma_zalloc_coherent_node(xdev, PAGE_SIZE,
&pgdir->db_dma, node);
if (!pgdir->db_page) {
bitmap_free(pgdir->bitmap);
kfree(pgdir);
return NULL;
}
return pgdir;
}
static int xsc_alloc_db_from_pgdir(struct xsc_db_pgdir *pgdir,
struct xsc_db *db)
{
u32 db_per_page = PAGE_SIZE / cache_line_size();
int offset;
int i;
i = find_first_bit(pgdir->bitmap, db_per_page);
if (i >= db_per_page)
return -ENOMEM;
__clear_bit(i, pgdir->bitmap);
db->u.pgdir = pgdir;
db->index = i;
offset = db->index * cache_line_size();
db->db = pgdir->db_page + offset / sizeof(*pgdir->db_page);
db->dma = pgdir->db_dma + offset;
db->db[0] = 0;
db->db[1] = 0;
return 0;
}
int xsc_db_alloc_node(struct xsc_core_device *xdev, struct xsc_db *db, int node)
{
struct xsc_db_pgdir *pgdir;
int ret = 0;
mutex_lock(&xdev->dev_res->pgdir_mutex);
list_for_each_entry(pgdir, &xdev->dev_res->pgdir_list, list)
if (!xsc_alloc_db_from_pgdir(pgdir, db))
goto out;
pgdir = xsc_alloc_db_pgdir(xdev, node);
if (!pgdir) {
ret = -ENOMEM;
goto out;
}
list_add(&pgdir->list, &xdev->dev_res->pgdir_list);
/* This should never fail -- we just allocated an empty page: */
WARN_ON(xsc_alloc_db_from_pgdir(pgdir, db));
out:
mutex_unlock(&xdev->dev_res->pgdir_mutex);
return ret;
}
EXPORT_SYMBOL_GPL(xsc_db_alloc_node);
int xsc_db_alloc(struct xsc_core_device *xdev, struct xsc_db *db)
{
return xsc_db_alloc_node(xdev, db, xdev->priv.numa_node);
}
EXPORT_SYMBOL_GPL(xsc_db_alloc);
void xsc_db_free(struct xsc_core_device *xdev, struct xsc_db *db)
{
u32 db_per_page = PAGE_SIZE / cache_line_size();
mutex_lock(&xdev->dev_res->pgdir_mutex);
__set_bit(db->index, db->u.pgdir->bitmap);
if (bitmap_full(db->u.pgdir->bitmap, db_per_page)) {
dma_free_coherent(&xdev->pdev->dev, PAGE_SIZE,
db->u.pgdir->db_page, db->u.pgdir->db_dma);
list_del(&db->u.pgdir->list);
bitmap_free(db->u.pgdir->bitmap);
kfree(db->u.pgdir);
}
mutex_unlock(&xdev->dev_res->pgdir_mutex);
}
EXPORT_SYMBOL_GPL(xsc_db_free);