| // SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note |
| /* |
| * |
| * (C) COPYRIGHT 2020-2021 ARM Limited. All rights reserved. |
| * |
| * This program is free software and is provided to you under the terms of the |
| * GNU General Public License version 2 as published by the Free Software |
| * Foundation, and any use by you of this program is subject to the terms |
| * of such GNU license. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, you can access it online at |
| * http://www.gnu.org/licenses/gpl-2.0.html. |
| * |
| */ |
| |
| #include "mali_kbase_csf_cpu_queue_debugfs.h" |
| #include <mali_kbase.h> |
| #include <linux/seq_file.h> |
| |
| #if IS_ENABLED(CONFIG_DEBUG_FS) |
| |
| bool kbase_csf_cpu_queue_read_dump_req(struct kbase_context *kctx, |
| struct base_csf_notification *req) |
| { |
| if (atomic_cmpxchg(&kctx->csf.cpu_queue.dump_req_status, |
| BASE_CSF_CPU_QUEUE_DUMP_ISSUED, |
| BASE_CSF_CPU_QUEUE_DUMP_PENDING) != |
| BASE_CSF_CPU_QUEUE_DUMP_ISSUED) { |
| return false; |
| } |
| |
| req->type = BASE_CSF_NOTIFICATION_CPU_QUEUE_DUMP; |
| return true; |
| } |
| |
| /** |
| * kbasep_csf_cpu_queue_debugfs_show() - Print cpu queue information for per context |
| * |
| * @file: The seq_file for printing to |
| * @data: The debugfs dentry private data, a pointer to kbase_context |
| * |
| * Return: Negative error code or 0 on success. |
| */ |
| static int kbasep_csf_cpu_queue_debugfs_show(struct seq_file *file, void *data) |
| { |
| struct kbase_context *kctx = file->private; |
| |
| mutex_lock(&kctx->csf.lock); |
| if (atomic_read(&kctx->csf.cpu_queue.dump_req_status) != |
| BASE_CSF_CPU_QUEUE_DUMP_COMPLETE) { |
| seq_printf(file, "Dump request already started! (try again)\n"); |
| mutex_unlock(&kctx->csf.lock); |
| return -EBUSY; |
| } |
| |
| atomic_set(&kctx->csf.cpu_queue.dump_req_status, BASE_CSF_CPU_QUEUE_DUMP_ISSUED); |
| init_completion(&kctx->csf.cpu_queue.dump_cmp); |
| kbase_event_wakeup(kctx); |
| mutex_unlock(&kctx->csf.lock); |
| |
| seq_printf(file, "CPU Queues table (version:v%u):\n", MALI_CSF_CPU_QUEUE_DEBUGFS_VERSION); |
| |
| wait_for_completion_timeout(&kctx->csf.cpu_queue.dump_cmp, |
| msecs_to_jiffies(3000)); |
| |
| mutex_lock(&kctx->csf.lock); |
| if (kctx->csf.cpu_queue.buffer) { |
| WARN_ON(atomic_read(&kctx->csf.cpu_queue.dump_req_status) != |
| BASE_CSF_CPU_QUEUE_DUMP_PENDING); |
| |
| seq_printf(file, "%s\n", kctx->csf.cpu_queue.buffer); |
| |
| kfree(kctx->csf.cpu_queue.buffer); |
| kctx->csf.cpu_queue.buffer = NULL; |
| kctx->csf.cpu_queue.buffer_size = 0; |
| } |
| else |
| seq_printf(file, "Dump error! (time out)\n"); |
| |
| atomic_set(&kctx->csf.cpu_queue.dump_req_status, |
| BASE_CSF_CPU_QUEUE_DUMP_COMPLETE); |
| |
| mutex_unlock(&kctx->csf.lock); |
| return 0; |
| } |
| |
| static int kbasep_csf_cpu_queue_debugfs_open(struct inode *in, struct file *file) |
| { |
| return single_open(file, kbasep_csf_cpu_queue_debugfs_show, in->i_private); |
| } |
| |
| static const struct file_operations kbasep_csf_cpu_queue_debugfs_fops = { |
| .open = kbasep_csf_cpu_queue_debugfs_open, |
| .read = seq_read, |
| .llseek = seq_lseek, |
| .release = single_release, |
| }; |
| |
| void kbase_csf_cpu_queue_debugfs_init(struct kbase_context *kctx) |
| { |
| struct dentry *file; |
| |
| if (WARN_ON(!kctx || IS_ERR_OR_NULL(kctx->kctx_dentry))) |
| return; |
| |
| file = debugfs_create_file("cpu_queue", 0444, kctx->kctx_dentry, |
| kctx, &kbasep_csf_cpu_queue_debugfs_fops); |
| |
| if (IS_ERR_OR_NULL(file)) { |
| dev_warn(kctx->kbdev->dev, |
| "Unable to create cpu queue debugfs entry"); |
| } |
| |
| kctx->csf.cpu_queue.buffer = NULL; |
| kctx->csf.cpu_queue.buffer_size = 0; |
| atomic_set(&kctx->csf.cpu_queue.dump_req_status, |
| BASE_CSF_CPU_QUEUE_DUMP_COMPLETE); |
| } |
| |
| int kbase_csf_cpu_queue_dump(struct kbase_context *kctx, |
| u64 buffer, size_t buf_size) |
| { |
| int err = 0; |
| |
| size_t alloc_size = buf_size; |
| char *dump_buffer; |
| |
| if (!buffer || !alloc_size) |
| goto done; |
| |
| alloc_size = (alloc_size + PAGE_SIZE) & ~(PAGE_SIZE - 1); |
| dump_buffer = kzalloc(alloc_size, GFP_KERNEL); |
| if (ZERO_OR_NULL_PTR(dump_buffer)) { |
| err = -ENOMEM; |
| goto done; |
| } |
| |
| WARN_ON(kctx->csf.cpu_queue.buffer != NULL); |
| |
| err = copy_from_user(dump_buffer, |
| u64_to_user_ptr(buffer), |
| buf_size); |
| if (err) { |
| kfree(dump_buffer); |
| err = -EFAULT; |
| goto done; |
| } |
| |
| mutex_lock(&kctx->csf.lock); |
| |
| kfree(kctx->csf.cpu_queue.buffer); |
| |
| if (atomic_read(&kctx->csf.cpu_queue.dump_req_status) == |
| BASE_CSF_CPU_QUEUE_DUMP_PENDING) { |
| kctx->csf.cpu_queue.buffer = dump_buffer; |
| kctx->csf.cpu_queue.buffer_size = buf_size; |
| complete_all(&kctx->csf.cpu_queue.dump_cmp); |
| } else { |
| kfree(dump_buffer); |
| } |
| |
| mutex_unlock(&kctx->csf.lock); |
| done: |
| return err; |
| } |
| #else |
| /* |
| * Stub functions for when debugfs is disabled |
| */ |
| void kbase_csf_cpu_queue_debugfs_init(struct kbase_context *kctx) |
| { |
| } |
| |
| bool kbase_csf_cpu_queue_read_dump_req(struct kbase_context *kctx, |
| struct base_csf_notification *req) |
| { |
| return false; |
| } |
| |
| int kbase_csf_cpu_queue_dump(struct kbase_context *kctx, |
| u64 buffer, size_t buf_size) |
| { |
| return 0; |
| } |
| #endif /* CONFIG_DEBUG_FS */ |