Skip to content
Snippets Groups Projects
Commit bd1f6606 authored by Asahi Lina's avatar Asahi Lina :sparkles: Committed by Lyude Paul
Browse files

rust: drm: gem: shmem: Add DRM shmem helper abstraction


The DRM shmem helper includes common code useful for drivers which
allocate GEM objects as anonymous shmem. Add a Rust abstraction for
this. Drivers can choose the raw GEM implementation or the shmem layer,
depending on their needs.

Lyude changelog:
* Rebase
* Squash "rust: drm: device: Convert Device to AlwaysRefCounted" as this
  commit doesn't compile without it anyway

Signed-off-by: default avatarAsahi Lina <lina@asahilina.net>
parent a093c757
Loading
......@@ -10,6 +10,7 @@
#include <drm/drm_drv.h>
#include <drm/drm_file.h>
#include <drm/drm_gem.h>
#include <drm/drm_gem_shmem_helper.h>
#include <drm/drm_ioctl.h>
#include <kunit/test.h>
#include <linux/blk-mq.h>
......@@ -25,6 +26,7 @@
#include <linux/fs.h>
#include <linux/jiffies.h>
#include <linux/jump_label.h>
#include <linux/iosys-map.h>
#include <linux/mdio.h>
#include <linux/miscdevice.h>
#include <linux/of_device.h>
......
// SPDX-License-Identifier: GPL-2.0
#include "gem.c"
#ifdef CONFIG_DRM_GEM_SHMEM_HELPER
#include "gem_shmem_helper.c"
#endif
#include "vma_manager.c"
#include <drm/drm_gem_shmem_helper.h>
void rust_helper_drm_gem_shmem_object_free(struct drm_gem_object *obj)
{
return drm_gem_shmem_object_free(obj);
}
void rust_helper_drm_gem_shmem_object_print_info(struct drm_printer *p, unsigned int indent,
const struct drm_gem_object *obj)
{
drm_gem_shmem_object_print_info(p, indent, obj);
}
int rust_helper_drm_gem_shmem_object_pin(struct drm_gem_object *obj)
{
return drm_gem_shmem_object_pin(obj);
}
void rust_helper_drm_gem_shmem_object_unpin(struct drm_gem_object *obj)
{
drm_gem_shmem_object_unpin(obj);
}
struct sg_table *rust_helper_drm_gem_shmem_object_get_sg_table(struct drm_gem_object *obj)
{
return drm_gem_shmem_object_get_sg_table(obj);
}
int rust_helper_drm_gem_shmem_object_vmap(struct drm_gem_object *obj,
struct iosys_map *map)
{
return drm_gem_shmem_object_vmap(obj, map);
}
void rust_helper_drm_gem_shmem_object_vunmap(struct drm_gem_object *obj,
struct iosys_map *map)
{
drm_gem_shmem_object_vunmap(obj, map);
}
int rust_helper_drm_gem_shmem_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
{
return drm_gem_shmem_object_mmap(obj, vma);
}
......@@ -31,6 +31,7 @@
#include "rcu.c"
#include "refcount.c"
#include "security.c"
#include "scatterlist.c"
#include "signal.c"
#include "slab.c"
#include "spinlock.c"
......
// SPDX-License-Identifier: GPL-2.0
#include <linux/scatterlist.h>
dma_addr_t rust_helper_sg_dma_address(const struct scatterlist *sg)
{
return sg_dma_address(sg);
}
int rust_helper_sg_dma_len(const struct scatterlist *sg)
{
return sg_dma_len(sg);
}
......@@ -4,6 +4,9 @@
//!
//! C header: [`include/linux/drm/drm_gem.h`](srctree/include/linux/drm/drm_gem.h)
#[cfg(CONFIG_DRM_GEM_SHMEM_HELPER = "y")]
pub mod shmem;
use crate::{
alloc::flags::*,
bindings,
......@@ -197,8 +200,6 @@ impl<T: IntoGEMObject> BaseObject for T {}
#[pin_data]
pub struct Object<T: DriverObject> {
obj: bindings::drm_gem_object,
// The DRM core ensures the Device exists as long as its objects exist, so we don't need to
// manage the reference count here.
dev: *const bindings::drm_device,
#[pin]
inner: T,
......
// SPDX-License-Identifier: GPL-2.0
//! DRM GEM shmem helper objects
//!
//! C header: [`include/linux/drm/drm_gem_shmem_helper.h`](../../../../include/linux/drm/drm_gem_shmem_helper.h)
use crate::drm::{device, drv, gem};
use crate::{
error::{from_err_ptr, to_result},
prelude::*,
};
use core::{
marker::{PhantomData, PhantomPinned},
mem,
mem::MaybeUninit,
ops::{Deref, DerefMut},
slice,
};
use gem::BaseObject;
/// Trait which must be implemented by drivers using shmem-backed GEM objects.
pub trait DriverObject: gem::BaseDriverObject<Object<Self>> {
/// Parent `Driver` for this object.
type Driver: drv::Driver;
}
// FIXME: This is terrible and I don't know how to avoid it
#[cfg(CONFIG_NUMA)]
macro_rules! vm_numa_fields {
( $($field:ident: $val:expr),* $(,)? ) => {
bindings::vm_operations_struct {
$( $field: $val ),*,
set_policy: None,
get_policy: None,
}
}
}
#[cfg(not(CONFIG_NUMA))]
macro_rules! vm_numa_fields {
( $($field:ident: $val:expr),* $(,)? ) => {
bindings::vm_operations_struct {
$( $field: $val ),*
}
}
}
const SHMEM_VM_OPS: bindings::vm_operations_struct = vm_numa_fields! {
open: Some(bindings::drm_gem_shmem_vm_open),
close: Some(bindings::drm_gem_shmem_vm_close),
may_split: None,
mremap: None,
mprotect: None,
fault: Some(bindings::drm_gem_shmem_fault),
huge_fault: None,
map_pages: None,
pagesize: None,
page_mkwrite: None,
pfn_mkwrite: None,
access: None,
name: None,
find_special_page: None,
};
/// A shmem-backed GEM object.
#[repr(C)]
#[pin_data]
pub struct Object<T: DriverObject> {
obj: bindings::drm_gem_shmem_object,
// The DRM core ensures the Device exists as long as its objects exist, so we don't need to
// manage the reference count here.
dev: *const bindings::drm_device,
#[pin]
inner: T,
}
// SAFETY: This struct is safe to zero-initialize
unsafe impl init::Zeroable for bindings::drm_gem_shmem_object {}
unsafe extern "C" fn gem_create_object<T: DriverObject>(
dev: *mut bindings::drm_device,
size: usize,
) -> *mut bindings::drm_gem_object {
let p = unsafe {
bindings::krealloc(
core::ptr::null(),
Object::<T>::SIZE,
bindings::GFP_KERNEL | bindings::__GFP_ZERO,
) as *mut Object<T>
};
if p.is_null() {
return ENOMEM.to_ptr();
}
let init = try_pin_init!(Object {
obj <- init::zeroed(),
// SAFETY: GEM ensures the device lives as long as its objects live
inner <- T::new(unsafe { device::Device::borrow(dev)}, size),
dev,
});
// SAFETY: p is a valid pointer to an uninitialized Object<T>.
if let Err(e) = unsafe { init.__pinned_init(p) } {
// SAFETY: p is a valid pointer from `krealloc` and __pinned_init guarantees we can dealloc it.
unsafe { bindings::kfree(p as *mut _) };
return e.to_ptr();
}
// SAFETY: drm_gem_shmem_object is safe to zero-init, and
// the rest of Object has been initialized
let new: &mut Object<T> = unsafe { &mut *(p as *mut _) };
new.obj.base.funcs = &Object::<T>::VTABLE;
&mut new.obj.base
}
unsafe extern "C" fn free_callback<T: DriverObject>(obj: *mut bindings::drm_gem_object) {
// SAFETY: All of our objects are Object<T>.
let p = unsafe {
let shmem = crate::container_of!(obj, bindings::drm_gem_shmem_object, base)
as *mut bindings::drm_gem_shmem_object;
crate::container_of!(shmem, Object<T>, obj) as *mut Object<T>
};
// SAFETY: p is never used after this
unsafe {
core::ptr::drop_in_place(&mut (*p).inner);
}
// SAFETY: This pointer has to be valid, since p is valid
unsafe {
bindings::drm_gem_shmem_free(&mut (*p).obj);
}
}
impl<T: DriverObject> Object<T> {
/// The size of this object's structure.
const SIZE: usize = mem::size_of::<Self>();
/// `drm_gem_object_funcs` vtable suitable for GEM shmem objects.
const VTABLE: bindings::drm_gem_object_funcs = bindings::drm_gem_object_funcs {
free: Some(free_callback::<T>),
open: Some(super::open_callback::<T, Object<T>>),
close: Some(super::close_callback::<T, Object<T>>),
print_info: Some(bindings::drm_gem_shmem_object_print_info),
export: None,
pin: Some(bindings::drm_gem_shmem_object_pin),
unpin: Some(bindings::drm_gem_shmem_object_unpin),
get_sg_table: Some(bindings::drm_gem_shmem_object_get_sg_table),
vmap: Some(bindings::drm_gem_shmem_object_vmap),
vunmap: Some(bindings::drm_gem_shmem_object_vunmap),
mmap: Some(bindings::drm_gem_shmem_object_mmap),
status: None,
vm_ops: &SHMEM_VM_OPS,
evict: None,
rss: None,
};
// SAFETY: Must only be used with DRM functions that are thread-safe
unsafe fn mut_shmem(&self) -> *mut bindings::drm_gem_shmem_object {
&self.obj as *const _ as *mut _
}
/// Create a new shmem-backed DRM object of the given size.
pub fn new(dev: &device::Device<T::Driver>, size: usize) -> Result<gem::UniqueObjectRef<Self>> {
// SAFETY: This function can be called as long as the ALLOC_OPS are set properly
// for this driver, and the gem_create_object is called.
let p = unsafe { bindings::drm_gem_shmem_create(dev.as_raw(), size) };
// SAFETY: `p` was created by the last call which uses our ALLOC_OPS to create the gem, so
// we're guaranteed `p` resides within an `Object<T>`
let p = unsafe { crate::container_of!(p, Object<T>, obj) as *mut _ };
// SAFETY: The gem_create_object callback ensures this is a valid Object<T>,
// so we can take a unique reference to it.
let obj_ref = gem::UniqueObjectRef { ptr: p, _p: PhantomPinned };
Ok(obj_ref)
}
/// Returns the `Device` that owns this GEM object.
pub fn dev(&self) -> &device::Device<T::Driver> {
// SAFETY: GEM ensures that the device outlives its objects, so we can
// just borrow here.
unsafe { device::Device::borrow(self.dev) }
}
/// Creates (if necessary) and returns a scatter-gather table of DMA pages for this object.
///
/// This will pin the object in memory.
pub fn sg_table(&self) -> Result<SGTable<T>> {
// SAFETY: drm_gem_shmem_get_pages_sgt is thread-safe.
let sgt = from_err_ptr(unsafe { bindings::drm_gem_shmem_get_pages_sgt(self.mut_shmem()) })?;
Ok(SGTable {
sgt,
_owner: self.reference(),
})
}
/// Creates and returns a virtual kernel memory mapping for this object.
pub fn vmap(&self) -> Result<VMap<T>> {
let mut map: MaybeUninit<bindings::iosys_map> = MaybeUninit::uninit();
// SAFETY: drm_gem_shmem_vmap can be called with the DMA reservation lock held
to_result(unsafe {
let resv = self.obj.base.resv as *const _ as *mut _;
bindings::dma_resv_lock(resv, core::ptr::null_mut());
let ret = bindings::drm_gem_shmem_vmap(self.mut_shmem(), map.as_mut_ptr());
bindings::dma_resv_unlock(resv);
ret
})?;
// SAFETY: if drm_gem_shmem_vmap did not fail, map is initialized now
let map = unsafe { map.assume_init() };
Ok(VMap {
map,
owner: self.reference(),
})
}
/// Set the write-combine flag for this object.
///
/// Should be called before any mappings are made.
pub fn set_wc(&mut self, map_wc: bool) {
unsafe { (*self.mut_shmem()).set_map_wc(map_wc) };
}
}
impl<T: DriverObject> Deref for Object<T> {
type Target = T;
fn deref(&self) -> &Self::Target {
&self.inner
}
}
impl<T: DriverObject> DerefMut for Object<T> {
fn deref_mut(&mut self) -> &mut Self::Target {
&mut self.inner
}
}
impl<T: DriverObject> crate::private::Sealed for Object<T> {}
impl<T: DriverObject> gem::IntoGEMObject for Object<T> {
type Driver = T::Driver;
fn gem_obj(&self) -> &bindings::drm_gem_object {
&self.obj.base
}
fn from_gem_obj(obj: *mut bindings::drm_gem_object) -> *mut Object<T> {
// SAFETY: Guaranteed by type invariance
let shmem = unsafe {
crate::container_of!(obj, bindings::drm_gem_shmem_object, base)
as *mut bindings::drm_gem_shmem_object;
};
// SAFETY: Guaranteed by type invariance
unsafe { crate::container_of!(&shmem, Object<T>, obj) as *mut Object<T> }
}
}
impl<T: DriverObject> drv::AllocImpl for Object<T> {
const ALLOC_OPS: drv::AllocOps = drv::AllocOps {
gem_create_object: Some(gem_create_object::<T>),
prime_handle_to_fd: None,
prime_fd_to_handle: None,
gem_prime_import: None,
gem_prime_import_sg_table: Some(bindings::drm_gem_shmem_prime_import_sg_table),
dumb_create: Some(bindings::drm_gem_shmem_dumb_create),
dumb_map_offset: None,
};
}
/// A virtual mapping for a shmem-backed GEM object in kernel address space.
pub struct VMap<T: DriverObject> {
map: bindings::iosys_map,
owner: gem::ObjectRef<Object<T>>,
}
impl<T: DriverObject> VMap<T> {
/// Returns a const raw pointer to the start of the mapping.
pub fn as_ptr(&self) -> *const core::ffi::c_void {
// SAFETY: The shmem helpers always return non-iomem maps
unsafe { self.map.__bindgen_anon_1.vaddr }
}
/// Returns a mutable raw pointer to the start of the mapping.
pub fn as_mut_ptr(&mut self) -> *mut core::ffi::c_void {
// SAFETY: The shmem helpers always return non-iomem maps
unsafe { self.map.__bindgen_anon_1.vaddr }
}
/// Returns a byte slice view of the mapping.
pub fn as_slice(&self) -> &[u8] {
// SAFETY: The vmap maps valid memory up to the owner size
unsafe { slice::from_raw_parts(self.as_ptr() as *const u8, self.owner.size()) }
}
/// Returns mutable a byte slice view of the mapping.
pub fn as_mut_slice(&mut self) -> &mut [u8] {
// SAFETY: The vmap maps valid memory up to the owner size
unsafe { slice::from_raw_parts_mut(self.as_mut_ptr() as *mut u8, self.owner.size()) }
}
/// Borrows a reference to the object that owns this virtual mapping.
pub fn owner(&self) -> &gem::ObjectRef<Object<T>> {
&self.owner
}
}
impl<T: DriverObject> Drop for VMap<T> {
fn drop(&mut self) {
// SAFETY: This function is safe to call with the DMA reservation lock held
unsafe {
let resv = self.owner.obj.base.resv as *const _ as *mut _;
bindings::dma_resv_lock(resv, core::ptr::null_mut());
bindings::drm_gem_shmem_vunmap(self.owner.mut_shmem(), &mut self.map);
bindings::dma_resv_unlock(resv);
}
}
}
/// SAFETY: `iosys_map` objects are safe to send across threads.
unsafe impl<T: DriverObject> Send for VMap<T> {}
unsafe impl<T: DriverObject> Sync for VMap<T> {}
/// A single scatter-gather entry, representing a span of pages in the device's DMA address space.
///
/// For devices not behind a standalone IOMMU, this corresponds to physical addresses.
#[repr(transparent)]
pub struct SGEntry(bindings::scatterlist);
impl SGEntry {
/// Returns the starting DMA address of this span
pub fn dma_address(&self) -> usize {
(unsafe { bindings::sg_dma_address(&self.0) }) as usize
}
/// Returns the length of this span in bytes
pub fn dma_len(&self) -> usize {
(unsafe { bindings::sg_dma_len(&self.0) }) as usize
}
}
/// A scatter-gather table of DMA address spans for a GEM shmem object.
///
/// # Invariants
/// `sgt` must be a valid pointer to the `sg_table`, which must correspond to the owned
/// object in `_owner` (which ensures it remains valid).
pub struct SGTable<T: DriverObject> {
sgt: *const bindings::sg_table,
_owner: gem::ObjectRef<Object<T>>,
}
impl<T: DriverObject> SGTable<T> {
/// Returns an iterator through the SGTable's entries
pub fn iter(&'_ self) -> SGTableIter<'_> {
SGTableIter {
left: unsafe { (*self.sgt).nents } as usize,
sg: unsafe { (*self.sgt).sgl },
_p: PhantomData,
}
}
}
impl<'a, T: DriverObject> IntoIterator for &'a SGTable<T> {
type Item = &'a SGEntry;
type IntoIter = SGTableIter<'a>;
fn into_iter(self) -> Self::IntoIter {
self.iter()
}
}
/// SAFETY: `sg_table` objects are safe to send across threads.
unsafe impl<T: DriverObject> Send for SGTable<T> {}
unsafe impl<T: DriverObject> Sync for SGTable<T> {}
/// An iterator through `SGTable` entries.
///
/// # Invariants
/// `sg` must be a valid pointer to the scatterlist, which must outlive our lifetime.
pub struct SGTableIter<'a> {
sg: *mut bindings::scatterlist,
left: usize,
_p: PhantomData<&'a ()>,
}
impl<'a> Iterator for SGTableIter<'a> {
type Item = &'a SGEntry;
fn next(&mut self) -> Option<Self::Item> {
if self.left == 0 {
None
} else {
let sg = self.sg;
self.sg = unsafe { bindings::sg_next(self.sg) };
self.left -= 1;
Some(unsafe { &(*(sg as *const SGEntry)) })
}
}
}
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment