This commit is contained in:
Zhengyi Chen 2024-01-18 22:06:24 +00:00
parent eae77aa719
commit f9282a627f

View file

@ -1,3 +1,11 @@
#include "asm-generic/cacheflush.h"
#include "asm/cacheflush.h"
#include "asm/page-def.h"
#include "linux/atomic/atomic-long.h"
#include "linux/device.h"
#include "linux/device/class.h"
#include "linux/mutex.h"
#include "linux/pid.h"
#include <linux/rcupdate.h> #include <linux/rcupdate.h>
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/list.h> #include <linux/list.h>
@ -10,6 +18,10 @@
#include <linux/module.h> #include <linux/module.h>
#include <linux/kernel.h> #include <linux/kernel.h>
MODULE_AUTHOR("Zk.");
MODULE_DESCRIPTION("4.2.1: mmap for point of coherency");
MODULE_LICENSE("GPL");
struct my_shmem_page { struct my_shmem_page {
struct page *page; struct page *page;
struct list_head list; struct list_head list;
@ -22,48 +34,56 @@ static LIST_HEAD(my_shmem_pages);
/* [!] READ/WRITE UNDER LOCK */ /* [!] READ/WRITE UNDER LOCK */
static size_t my_shmem_page_count = 0; static size_t my_shmem_page_count = 0;
static int major;
static struct class* class;
static struct device* dev;
const char* DEV_NAME = "my_shmem";
/* Virtual Memory Area Operations... /* Virtual Memory Area Operations...
* ============================================================================ * ============================================================================
*/ */
static void my_shmem_vmops_close(struct vm_area_struct *vma)
{
// [TODO] Flush dcache at close.
// `dcache_clean_poc` writebacks D-cache region s.t. PoC. Period.
// This should? work on all ARM64 CPUs w/ no-alias VIPT dcache.
// The addresses are VAs (obv., as opposed to PAs).
// dcache_clean_poc(unsigned long start, unsigned long end)
dcache_clean_poc(vma->vm_start, vma->vm_end);
}
static vm_fault_t my_shmem_vmops_fault(struct vm_fault *vmf) static vm_fault_t my_shmem_vmops_fault(struct vm_fault *vmf)
{ {
pgoff_t page_offset = vmf->pgoff; ulong nr_pages_from_vm_start =
phys_addr_t _phys; (vmf->address - vmf->vma->vm_start) >> PAGE_SHIFT;
size_t old_shmem_page_count; const pgoff_t _dbg_offset_from_page = vmf->pgoff;
phys_addr_t _dbg_phys;
mutex_lock(&my_shmem_pages_mtx); mutex_lock(&my_shmem_pages_mtx);
old_shmem_page_count = READ_ONCE(my_shmem_page_count); if (nr_pages_from_vm_start < my_shmem_page_count) {
if (unlikely(page_offset > old_shmem_page_count)) {
/* IMPOSSIBLE -- programming error or wrong assumption... */
mutex_unlock(&my_shmem_pages_mtx);
goto err_ret_impossible_count;
}
if (page_offset < old_shmem_page_count) {
/* Offset in range, return existing page */ /* Offset in range, return existing page */
pr_info("[%s] Found remappable page offset %lu.", pr_info("[%s] Found remappable page nr: %lu, offset: %lu.\n",
__func__, page_offset); __func__, nr_pages_from_vm_start, _dbg_offset_from_page);
// We won't delete elements from list here! // We won't delete elements from list here!
struct my_shmem_page *page_entry; struct my_shmem_page *page_entry;
list_for_each_entry(page_entry, &my_shmem_pages, list) { list_for_each_entry(page_entry, &my_shmem_pages, list) {
if (!page_offset) if (!nr_pages_from_vm_start)
break; break;
page_offset--; nr_pages_from_vm_start--;
} }
// Found correct page entry, remap // Found correct page entry, remap
get_page(page_entry->page); get_page(page_entry->page);
vmf->page = page_entry->page; vmf->page = page_entry->page;
_phys = page_to_phys(page_entry->page); _dbg_phys = page_to_phys(page_entry->page);
mutex_unlock(&my_shmem_pages_mtx); mutex_unlock(&my_shmem_pages_mtx);
goto ok_ret_remapped; goto ok_ret_remapped;
} }
/* Otherwise, allocate the new page */ /* Otherwise, allocate the new page(s) */
for (int i = 0; i < page_offset - old_shmem_page_count; i++) for (int i = 0; i <= nr_pages_from_vm_start - my_shmem_page_count; i++)
{ {
/* The loop is misleading -- this loops exactly once! */
// Allocate page handle in kernel // Allocate page handle in kernel
struct my_shmem_page *new_page = kzalloc( struct my_shmem_page *new_page = kzalloc(
sizeof(struct my_shmem_page), GFP_KERNEL); sizeof(struct my_shmem_page), GFP_KERNEL);
@ -73,6 +93,9 @@ static vm_fault_t my_shmem_vmops_fault(struct vm_fault *vmf)
} }
// Allocate page in virtual memory // Allocate page in virtual memory
// [!] We specifically WANT to allocate cachable memory to
// create cache incoherence btwn procs. Synchronization
// (e.g., on arm64) is done via calling fsync for now.
void *addr = vmalloc_user(PAGE_SIZE); void *addr = vmalloc_user(PAGE_SIZE);
if (!addr) { if (!addr) {
mutex_unlock(&my_shmem_pages_mtx); mutex_unlock(&my_shmem_pages_mtx);
@ -87,30 +110,25 @@ static vm_fault_t my_shmem_vmops_fault(struct vm_fault *vmf)
// Fill in allocated page entry // Fill in allocated page entry
get_page(new_page->page); get_page(new_page->page);
vmf->page = new_page->page; vmf->page = new_page->page;
_phys = page_to_phys(new_page->page); _dbg_phys = page_to_phys(new_page->page);
} }
mutex_unlock(&my_shmem_pages_mtx); mutex_unlock(&my_shmem_pages_mtx);
goto ok_ret_allocated; goto ok_ret_allocated;
err_ret_impossible_count:
pr_crit("[%s] IMPOSSIBLE list count %ld > %ld "
"-- no way one fault services multiple page!!!",
__func__, page_offset, old_shmem_page_count);
return VM_FAULT_ERROR;
err_ret_no_kmem: err_ret_no_kmem:
pr_err("[%s] Cannot allocate `struct my_shmem_page` in kernel memory.", pr_err("[%s] Cannot allocate `struct my_shmem_page` in kernel memory.\n",
__func__); __func__);
return VM_FAULT_OOM; return VM_FAULT_OOM;
err_ret_no_vmem: err_ret_no_vmem:
pr_err("[%s] Cannot allocate requested page for virtual memory.", pr_err("[%s] Cannot allocate requested page for virtual memory.\n",
__func__); __func__);
return VM_FAULT_OOM; return VM_FAULT_OOM;
ok_ret_remapped: ok_ret_remapped:
if (vmf->vma->vm_mm) { if (vmf->vma->vm_mm) {
rcu_read_lock(); rcu_read_lock();
struct task_struct *fault_owner = vmf->vma->vm_mm->owner; struct task_struct *fault_owner = vmf->vma->vm_mm->owner;
pr_info("[%s] Remapped phys: 0x%llx -> virt@PID(%d): 0x%lx.", pr_info("[%s] Remapped phys: 0x%llx -> virt@PID(%d): 0x%lx.\n",
__func__, _phys, fault_owner->pid, vmf->address); __func__, _dbg_phys, fault_owner->pid, vmf->address);
rcu_read_unlock(); rcu_read_unlock();
} }
return 0; return 0;
@ -118,13 +136,15 @@ ok_ret_allocated:
if (vmf->vma->vm_mm){ if (vmf->vma->vm_mm){
rcu_read_lock(); rcu_read_lock();
struct task_struct *fault_owner = vmf->vma->vm_mm->owner; struct task_struct *fault_owner = vmf->vma->vm_mm->owner;
pr_info("[%s] Allocated phys: 0x%llx -> virt@PID(%d): 0x%lx.", pr_info("[%s] Allocated phys: 0x%llx -> virt@PID(%d): 0x%lx.\n",
__func__, _phys, fault_owner->pid, vmf->address); __func__, _dbg_phys, fault_owner->pid, vmf->address);
rcu_read_unlock(); rcu_read_unlock();
} }
return 0; return 0;
} }
static const struct vm_operations_struct my_shmem_vmops = { static const struct vm_operations_struct my_shmem_vmops = {
.fault = my_shmem_vmops_fault, .fault = my_shmem_vmops_fault,
}; };
@ -134,16 +154,42 @@ static const struct vm_operations_struct my_shmem_vmops = {
*/ */
// static int my_shmem_fops_open(struct inode *inode, struct file *filp); // static int my_shmem_fops_open(struct inode *inode, struct file *filp);
static const struct file_operations my_shmem_fops;
static int my_shmem_fops_mmap(struct file *filp, struct vm_area_struct *vma) static int my_shmem_fops_mmap(struct file *filp, struct vm_area_struct *vma)
{ {
vma->vm_ops = &my_shmem_vmops; vma->vm_ops = &my_shmem_vmops;
pr_info("[%s] Device file '%s' mmapped for vma: [0x%lx - 0x%lx].\n",
__func__, file_dentry(filp)->d_name.name, vma->vm_start, vma->vm_end);
return 0;
}
static int my_shmem_fops_open(struct inode *inode, struct file *filp)
{
filp->f_op = &my_shmem_fops;
pr_info("[%s] Device file '%s' opened.\n",
__func__, file_dentry(filp)->d_name.name);
return 0;
}
static int my_shmem_fops_release(struct inode *inode, struct file *filp)
{
pr_info("[%s] Device file '%s' released.\n",
__func__, file_dentry(filp)->d_name.name);
/* Garbage collection requires knowing who references which page...
* Ideally this is stored in filp->private_data but ah well, oversight.
*/
return 0; return 0;
} }
static const struct file_operations my_shmem_fops = { static const struct file_operations my_shmem_fops = {
.owner = THIS_MODULE, .owner = THIS_MODULE,
.open = my_shmem_fops_open,
.mmap = my_shmem_fops_mmap, .mmap = my_shmem_fops_mmap,
.fsync = noop_fsync, .release = my_shmem_fops_release,
}; };
/* Module init & exit... /* Module init & exit...
@ -151,13 +197,68 @@ static const struct file_operations my_shmem_fops = {
*/ */
static int __init my_shmem_init(void) static int __init my_shmem_init(void)
{ {
int reg_cdev_ret = register_chrdev(0, "my_shmem", &my_shmem_fops); /* Register cdev */
if (reg_cdev_ret != 0) major = register_chrdev(
0, DEV_NAME, &my_shmem_fops);
if (major < 0)
goto err_ret_cdev_reg_failed; goto err_ret_cdev_reg_failed;
/* Create device class */
class = class_create(DEV_NAME);
if (IS_ERR(class)) {
unregister_chrdev(major, DEV_NAME);
goto err_ret_class_crea_failed;
}
/* Create one device */
dev_t dev_nr = MKDEV(major, 0);
dev = device_create(
class, NULL, dev_nr, NULL, DEV_NAME);
if (IS_ERR(dev)) {
class_destroy(class);
unregister_chrdev(major, DEV_NAME);
goto err_ret_dev_crea_failed;
}
pr_info("[%s] Device `%s` built successfully.\n",
__func__, dev->init_name);
return 0; return 0;
err_ret_cdev_reg_failed: err_ret_cdev_reg_failed:
pr_err("[%s] Cannot register character dev -- error code %d.", pr_err("[%s] Cannot register character dev -- error code %d.\n",
__func__, reg_cdev_ret); __func__, major);
return reg_cdev_ret; return major;
err_ret_class_crea_failed:
pr_err("[%s] Cannot create device class -- error code %ld.\n",
__func__, PTR_ERR(class));
return (int) PTR_ERR(class);
err_ret_dev_crea_failed:
pr_err("[%s] Cannot create device -- error code %ld.\n",
__func__, PTR_ERR(dev));
return (int) PTR_ERR(dev);
} }
static void __exit my_shmem_exit(void)
{
/* Destroy device */
device_destroy(class, dev->devt);
class_destroy(class);
unregister_chrdev(major, DEV_NAME);
pr_info("[%s] Device destroyed.\n", __func__);
/* Free all pages -- I'm not compacting in runtime!!! */
struct my_shmem_page *page_entry, *tmp;
mutex_lock(&my_shmem_pages_mtx);
list_for_each_entry_safe(page_entry, tmp, &my_shmem_pages, list) {
vfree(page_to_virt(page_entry->page));
put_page(page_entry->page);
my_shmem_page_count--;
list_del(&page_entry->list);
kfree(page_entry);
}
mutex_unlock(&my_shmem_pages_mtx);
}
module_init(my_shmem_init);
module_exit(my_shmem_exit);