| /* Copyright (c) 2016, The Linux Foundation. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 and |
| * only version 2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| */ |
| |
| #undef TRACE_SYSTEM |
| #define TRACE_SYSTEM kmem |
| |
| #if !defined(_TRACE_KMEM_H) || defined(TRACE_HEADER_MULTI_READ) |
| #define _TRACE_KMEM_H |
| |
| #include <linux/types.h> |
| #include <linux/tracepoint.h> |
| #include <trace/events/mmflags.h> |
| |
| DECLARE_EVENT_CLASS(kmem_alloc, |
| |
| TP_PROTO(unsigned long call_site, |
| const void *ptr, |
| size_t bytes_req, |
| size_t bytes_alloc, |
| gfp_t gfp_flags), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, call_site ) |
| __field( const void *, ptr ) |
| __field( size_t, bytes_req ) |
| __field( size_t, bytes_alloc ) |
| __field( gfp_t, gfp_flags ) |
| ), |
| |
| TP_fast_assign( |
| __entry->call_site = call_site; |
| __entry->ptr = ptr; |
| __entry->bytes_req = bytes_req; |
| __entry->bytes_alloc = bytes_alloc; |
| __entry->gfp_flags = gfp_flags; |
| ), |
| |
| TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s", |
| __entry->call_site, |
| __entry->ptr, |
| __entry->bytes_req, |
| __entry->bytes_alloc, |
| show_gfp_flags(__entry->gfp_flags)) |
| ); |
| |
| DEFINE_EVENT(kmem_alloc, kmalloc, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr, |
| size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags) |
| ); |
| |
| DEFINE_EVENT(kmem_alloc, kmem_cache_alloc, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr, |
| size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags) |
| ); |
| |
| DECLARE_EVENT_CLASS(kmem_alloc_node, |
| |
| TP_PROTO(unsigned long call_site, |
| const void *ptr, |
| size_t bytes_req, |
| size_t bytes_alloc, |
| gfp_t gfp_flags, |
| int node), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, call_site ) |
| __field( const void *, ptr ) |
| __field( size_t, bytes_req ) |
| __field( size_t, bytes_alloc ) |
| __field( gfp_t, gfp_flags ) |
| __field( int, node ) |
| ), |
| |
| TP_fast_assign( |
| __entry->call_site = call_site; |
| __entry->ptr = ptr; |
| __entry->bytes_req = bytes_req; |
| __entry->bytes_alloc = bytes_alloc; |
| __entry->gfp_flags = gfp_flags; |
| __entry->node = node; |
| ), |
| |
| TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d", |
| __entry->call_site, |
| __entry->ptr, |
| __entry->bytes_req, |
| __entry->bytes_alloc, |
| show_gfp_flags(__entry->gfp_flags), |
| __entry->node) |
| ); |
| |
| DEFINE_EVENT(kmem_alloc_node, kmalloc_node, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr, |
| size_t bytes_req, size_t bytes_alloc, |
| gfp_t gfp_flags, int node), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) |
| ); |
| |
| DEFINE_EVENT(kmem_alloc_node, kmem_cache_alloc_node, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr, |
| size_t bytes_req, size_t bytes_alloc, |
| gfp_t gfp_flags, int node), |
| |
| TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) |
| ); |
| |
| DECLARE_EVENT_CLASS(kmem_free, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr), |
| |
| TP_ARGS(call_site, ptr), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, call_site ) |
| __field( const void *, ptr ) |
| ), |
| |
| TP_fast_assign( |
| __entry->call_site = call_site; |
| __entry->ptr = ptr; |
| ), |
| |
| TP_printk("call_site=%lx ptr=%p", __entry->call_site, __entry->ptr) |
| ); |
| |
| DEFINE_EVENT(kmem_free, kfree, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr), |
| |
| TP_ARGS(call_site, ptr) |
| ); |
| |
| DEFINE_EVENT(kmem_free, kmem_cache_free, |
| |
| TP_PROTO(unsigned long call_site, const void *ptr), |
| |
| TP_ARGS(call_site, ptr) |
| ); |
| |
| TRACE_EVENT(mm_page_free, |
| |
| TP_PROTO(struct page *page, unsigned int order), |
| |
| TP_ARGS(page, order), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( unsigned int, order ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page_to_pfn(page); |
| __entry->order = order; |
| ), |
| |
| TP_printk("page=%p pfn=%lu order=%d", |
| pfn_to_page(__entry->pfn), |
| __entry->pfn, |
| __entry->order) |
| ); |
| |
| TRACE_EVENT(mm_page_free_batched, |
| |
| TP_PROTO(struct page *page, int cold), |
| |
| TP_ARGS(page, cold), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( int, cold ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page_to_pfn(page); |
| __entry->cold = cold; |
| ), |
| |
| TP_printk("page=%p pfn=%lu order=0 cold=%d", |
| pfn_to_page(__entry->pfn), |
| __entry->pfn, |
| __entry->cold) |
| ); |
| |
| TRACE_EVENT(mm_page_alloc, |
| |
| TP_PROTO(struct page *page, unsigned int order, |
| gfp_t gfp_flags, int migratetype), |
| |
| TP_ARGS(page, order, gfp_flags, migratetype), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( unsigned int, order ) |
| __field( gfp_t, gfp_flags ) |
| __field( int, migratetype ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page ? page_to_pfn(page) : -1UL; |
| __entry->order = order; |
| __entry->gfp_flags = gfp_flags; |
| __entry->migratetype = migratetype; |
| ), |
| |
| TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s", |
| __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL, |
| __entry->pfn != -1UL ? __entry->pfn : 0, |
| __entry->order, |
| __entry->migratetype, |
| show_gfp_flags(__entry->gfp_flags)) |
| ); |
| |
| DECLARE_EVENT_CLASS(mm_page, |
| |
| TP_PROTO(struct page *page, unsigned int order, int migratetype), |
| |
| TP_ARGS(page, order, migratetype), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( unsigned int, order ) |
| __field( int, migratetype ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page ? page_to_pfn(page) : -1UL; |
| __entry->order = order; |
| __entry->migratetype = migratetype; |
| ), |
| |
| TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d", |
| __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL, |
| __entry->pfn != -1UL ? __entry->pfn : 0, |
| __entry->order, |
| __entry->migratetype, |
| __entry->order == 0) |
| ); |
| |
| DEFINE_EVENT(mm_page, mm_page_alloc_zone_locked, |
| |
| TP_PROTO(struct page *page, unsigned int order, int migratetype), |
| |
| TP_ARGS(page, order, migratetype) |
| ); |
| |
| TRACE_EVENT(mm_page_pcpu_drain, |
| |
| TP_PROTO(struct page *page, unsigned int order, int migratetype), |
| |
| TP_ARGS(page, order, migratetype), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( unsigned int, order ) |
| __field( int, migratetype ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page ? page_to_pfn(page) : -1UL; |
| __entry->order = order; |
| __entry->migratetype = migratetype; |
| ), |
| |
| TP_printk("page=%p pfn=%lu order=%d migratetype=%d", |
| pfn_to_page(__entry->pfn), __entry->pfn, |
| __entry->order, __entry->migratetype) |
| ); |
| |
| TRACE_EVENT(mm_page_alloc_extfrag, |
| |
| TP_PROTO(struct page *page, |
| int alloc_order, int fallback_order, |
| int alloc_migratetype, int fallback_migratetype), |
| |
| TP_ARGS(page, |
| alloc_order, fallback_order, |
| alloc_migratetype, fallback_migratetype), |
| |
| TP_STRUCT__entry( |
| __field( unsigned long, pfn ) |
| __field( int, alloc_order ) |
| __field( int, fallback_order ) |
| __field( int, alloc_migratetype ) |
| __field( int, fallback_migratetype ) |
| __field( int, change_ownership ) |
| ), |
| |
| TP_fast_assign( |
| __entry->pfn = page_to_pfn(page); |
| __entry->alloc_order = alloc_order; |
| __entry->fallback_order = fallback_order; |
| __entry->alloc_migratetype = alloc_migratetype; |
| __entry->fallback_migratetype = fallback_migratetype; |
| __entry->change_ownership = (alloc_migratetype == |
| get_pageblock_migratetype(page)); |
| ), |
| |
| TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d", |
| pfn_to_page(__entry->pfn), |
| __entry->pfn, |
| __entry->alloc_order, |
| __entry->fallback_order, |
| pageblock_order, |
| __entry->alloc_migratetype, |
| __entry->fallback_migratetype, |
| __entry->fallback_order < pageblock_order, |
| __entry->change_ownership) |
| ); |
| |
| |
| DECLARE_EVENT_CLASS(ion_alloc, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags), |
| |
| TP_STRUCT__entry( |
| __array(char, client_name, 64) |
| __field(const char *, heap_name) |
| __field(size_t, len) |
| __field(unsigned int, mask) |
| __field(unsigned int, flags) |
| ), |
| |
| TP_fast_assign( |
| strlcpy(__entry->client_name, client_name, 64); |
| __entry->heap_name = heap_name; |
| __entry->len = len; |
| __entry->mask = mask; |
| __entry->flags = flags; |
| ), |
| |
| TP_printk("client_name=%s heap_name=%s len=%zu mask=0x%x flags=0x%x", |
| __entry->client_name, |
| __entry->heap_name, |
| __entry->len, |
| __entry->mask, |
| __entry->flags) |
| ); |
| |
| DEFINE_EVENT(ion_alloc, ion_alloc_buffer_start, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags) |
| ); |
| |
| DEFINE_EVENT(ion_alloc, ion_alloc_buffer_end, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_alloc_error, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags, |
| long error), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags, error), |
| |
| TP_STRUCT__entry( |
| __field(const char *, client_name) |
| __field(const char *, heap_name) |
| __field(size_t, len) |
| __field(unsigned int, mask) |
| __field(unsigned int, flags) |
| __field(long, error) |
| ), |
| |
| TP_fast_assign( |
| __entry->client_name = client_name; |
| __entry->heap_name = heap_name; |
| __entry->len = len; |
| __entry->mask = mask; |
| __entry->flags = flags; |
| __entry->error = error; |
| ), |
| |
| TP_printk( |
| "client_name=%s heap_name=%s len=%zu mask=0x%x flags=0x%x error=%ld", |
| __entry->client_name, |
| __entry->heap_name, |
| __entry->len, |
| __entry->mask, |
| __entry->flags, |
| __entry->error) |
| ); |
| |
| |
| DEFINE_EVENT(ion_alloc_error, ion_alloc_buffer_fallback, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags, |
| long error), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags, error) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_error, ion_alloc_buffer_fail, |
| |
| TP_PROTO(const char *client_name, |
| const char *heap_name, |
| size_t len, |
| unsigned int mask, |
| unsigned int flags, |
| long error), |
| |
| TP_ARGS(client_name, heap_name, len, mask, flags, error) |
| ); |
| |
| |
| DECLARE_EVENT_CLASS(alloc_retry, |
| |
| TP_PROTO(int tries), |
| |
| TP_ARGS(tries), |
| |
| TP_STRUCT__entry( |
| __field(int, tries) |
| ), |
| |
| TP_fast_assign( |
| __entry->tries = tries; |
| ), |
| |
| TP_printk("tries=%d", |
| __entry->tries) |
| ); |
| |
| DEFINE_EVENT(alloc_retry, ion_cp_alloc_retry, |
| |
| TP_PROTO(int tries), |
| |
| TP_ARGS(tries) |
| ); |
| |
| DEFINE_EVENT(alloc_retry, migrate_retry, |
| |
| TP_PROTO(int tries), |
| |
| TP_ARGS(tries) |
| ); |
| |
| DEFINE_EVENT(alloc_retry, dma_alloc_contiguous_retry, |
| |
| TP_PROTO(int tries), |
| |
| TP_ARGS(tries) |
| ); |
| |
| DECLARE_EVENT_CLASS(migrate_pages, |
| |
| TP_PROTO(int mode), |
| |
| TP_ARGS(mode), |
| |
| TP_STRUCT__entry( |
| __field(int, mode) |
| ), |
| |
| TP_fast_assign( |
| __entry->mode = mode; |
| ), |
| |
| TP_printk("mode=%d", |
| __entry->mode) |
| ); |
| |
| DEFINE_EVENT(migrate_pages, migrate_pages_start, |
| |
| TP_PROTO(int mode), |
| |
| TP_ARGS(mode) |
| ); |
| |
| DEFINE_EVENT(migrate_pages, migrate_pages_end, |
| |
| TP_PROTO(int mode), |
| |
| TP_ARGS(mode) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_alloc_pages, |
| |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order), |
| |
| TP_STRUCT__entry( |
| __field(gfp_t, gfp_flags) |
| __field(unsigned int, order) |
| ), |
| |
| TP_fast_assign( |
| __entry->gfp_flags = gfp_flags; |
| __entry->order = order; |
| ), |
| |
| TP_printk("gfp_flags=%s order=%d", |
| show_gfp_flags(__entry->gfp_flags), |
| __entry->order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_iommu_start, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_iommu_end, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_iommu_fail, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_sys_start, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_sys_end, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| ); |
| |
| DEFINE_EVENT(ion_alloc_pages, alloc_pages_sys_fail, |
| TP_PROTO(gfp_t gfp_flags, |
| unsigned int order), |
| |
| TP_ARGS(gfp_flags, order) |
| |
| ); |
| |
| DECLARE_EVENT_CLASS(smmu_map, |
| |
| TP_PROTO(unsigned long va, |
| phys_addr_t pa, |
| unsigned long chunk_size, |
| size_t len), |
| |
| TP_ARGS(va, pa, chunk_size, len), |
| |
| TP_STRUCT__entry( |
| __field(unsigned long, va) |
| __field(phys_addr_t, pa) |
| __field(unsigned long, chunk_size) |
| __field(size_t, len) |
| ), |
| |
| TP_fast_assign( |
| __entry->va = va; |
| __entry->pa = pa; |
| __entry->chunk_size = chunk_size; |
| __entry->len = len; |
| ), |
| |
| TP_printk("v_addr=%p p_addr=%pa chunk_size=0x%lx len=%zu", |
| (void *)__entry->va, |
| &__entry->pa, |
| __entry->chunk_size, |
| __entry->len) |
| ); |
| |
| DEFINE_EVENT(smmu_map, iommu_map_range, |
| TP_PROTO(unsigned long va, |
| phys_addr_t pa, |
| unsigned long chunk_size, |
| size_t len), |
| |
| TP_ARGS(va, pa, chunk_size, len) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_secure_cma_add_to_pool, |
| |
| TP_PROTO(unsigned long len, |
| int pool_total, |
| bool is_prefetch), |
| |
| TP_ARGS(len, pool_total, is_prefetch), |
| |
| TP_STRUCT__entry( |
| __field(unsigned long, len) |
| __field(int, pool_total) |
| __field(bool, is_prefetch) |
| ), |
| |
| TP_fast_assign( |
| __entry->len = len; |
| __entry->pool_total = pool_total; |
| __entry->is_prefetch = is_prefetch; |
| ), |
| |
| TP_printk("len %lx, pool total %x is_prefetch %d", |
| __entry->len, |
| __entry->pool_total, |
| __entry->is_prefetch) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_add_to_pool, ion_secure_cma_add_to_pool_start, |
| TP_PROTO(unsigned long len, |
| int pool_total, |
| bool is_prefetch), |
| |
| TP_ARGS(len, pool_total, is_prefetch) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_add_to_pool, ion_secure_cma_add_to_pool_end, |
| TP_PROTO(unsigned long len, |
| int pool_total, |
| bool is_prefetch), |
| |
| TP_ARGS(len, pool_total, is_prefetch) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_secure_cma_shrink_pool, |
| |
| TP_PROTO(unsigned long drained_size, |
| unsigned long skipped_size), |
| |
| TP_ARGS(drained_size, skipped_size), |
| |
| TP_STRUCT__entry( |
| __field(unsigned long, drained_size) |
| __field(unsigned long, skipped_size) |
| ), |
| |
| TP_fast_assign( |
| __entry->drained_size = drained_size; |
| __entry->skipped_size = skipped_size; |
| ), |
| |
| TP_printk("drained size %lx, skipped size %lx", |
| __entry->drained_size, |
| __entry->skipped_size) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_shrink_pool, ion_secure_cma_shrink_pool_start, |
| TP_PROTO(unsigned long drained_size, |
| unsigned long skipped_size), |
| |
| TP_ARGS(drained_size, skipped_size) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_shrink_pool, ion_secure_cma_shrink_pool_end, |
| TP_PROTO(unsigned long drained_size, |
| unsigned long skipped_size), |
| |
| TP_ARGS(drained_size, skipped_size) |
| ); |
| |
| TRACE_EVENT(ion_prefetching, |
| |
| TP_PROTO(unsigned long len), |
| |
| TP_ARGS(len), |
| |
| TP_STRUCT__entry( |
| __field(unsigned long, len) |
| ), |
| |
| TP_fast_assign( |
| __entry->len = len; |
| ), |
| |
| TP_printk("prefetch size %lx", |
| __entry->len) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_secure_cma_allocate, |
| |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags), |
| |
| TP_STRUCT__entry( |
| __field(const char *, heap_name) |
| __field(unsigned long, len) |
| __field(unsigned long, align) |
| __field(unsigned long, flags) |
| ), |
| |
| TP_fast_assign( |
| __entry->heap_name = heap_name; |
| __entry->len = len; |
| __entry->align = align; |
| __entry->flags = flags; |
| ), |
| |
| TP_printk("heap_name=%s len=%lx align=%lx flags=%lx", |
| __entry->heap_name, |
| __entry->len, |
| __entry->align, |
| __entry->flags) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_allocate, ion_secure_cma_allocate_start, |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags) |
| ); |
| |
| DEFINE_EVENT(ion_secure_cma_allocate, ion_secure_cma_allocate_end, |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags) |
| ); |
| |
| DECLARE_EVENT_CLASS(ion_cp_secure_buffer, |
| |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags), |
| |
| TP_STRUCT__entry( |
| __field(const char *, heap_name) |
| __field(unsigned long, len) |
| __field(unsigned long, align) |
| __field(unsigned long, flags) |
| ), |
| |
| TP_fast_assign( |
| __entry->heap_name = heap_name; |
| __entry->len = len; |
| __entry->align = align; |
| __entry->flags = flags; |
| ), |
| |
| TP_printk("heap_name=%s len=%lx align=%lx flags=%lx", |
| __entry->heap_name, |
| __entry->len, |
| __entry->align, |
| __entry->flags) |
| ); |
| |
| DEFINE_EVENT(ion_cp_secure_buffer, ion_cp_secure_buffer_start, |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags) |
| ); |
| |
| DEFINE_EVENT(ion_cp_secure_buffer, ion_cp_secure_buffer_end, |
| TP_PROTO(const char *heap_name, |
| unsigned long len, |
| unsigned long align, |
| unsigned long flags), |
| |
| TP_ARGS(heap_name, len, align, flags) |
| ); |
| |
| DECLARE_EVENT_CLASS(iommu_sec_ptbl_map_range, |
| |
| TP_PROTO(int sec_id, |
| int num, |
| unsigned long va, |
| unsigned int pa, |
| size_t len), |
| |
| TP_ARGS(sec_id, num, va, pa, len), |
| |
| TP_STRUCT__entry( |
| __field(int, sec_id) |
| __field(int, num) |
| __field(unsigned long, va) |
| __field(unsigned int, pa) |
| __field(size_t, len) |
| ), |
| |
| TP_fast_assign( |
| __entry->sec_id = sec_id; |
| __entry->num = num; |
| __entry->va = va; |
| __entry->pa = pa; |
| __entry->len = len; |
| ), |
| |
| TP_printk("sec_id=%d num=%d va=%lx pa=%u len=%zu", |
| __entry->sec_id, |
| __entry->num, |
| __entry->va, |
| __entry->pa, |
| __entry->len) |
| ); |
| |
| DEFINE_EVENT(iommu_sec_ptbl_map_range, iommu_sec_ptbl_map_range_start, |
| |
| TP_PROTO(int sec_id, |
| int num, |
| unsigned long va, |
| unsigned int pa, |
| size_t len), |
| |
| TP_ARGS(sec_id, num, va, pa, len) |
| ); |
| |
| DEFINE_EVENT(iommu_sec_ptbl_map_range, iommu_sec_ptbl_map_range_end, |
| |
| TP_PROTO(int sec_id, |
| int num, |
| unsigned long va, |
| unsigned int pa, |
| size_t len), |
| |
| TP_ARGS(sec_id, num, va, pa, len) |
| ); |
| #endif /* _TRACE_KMEM_H */ |
| |
| /* This part must be outside protection */ |
| #include <trace/define_trace.h> |