Quellcodebibliothek Statistik Leitseite products/sources/formale Sprachen/C/Linux/drivers/cxl/core/   (Open Source Betriebssystem Version 6.17.9©)  Datei vom 24.10.2025 mit Größe 40 kB image not shown  

Quelle  trace.h   Sprache: C

 
// SPDX-License-Identifier: GPL-2.0
/* Copyright(c) 2022 Intel Corporation. All rights reserved. */
#undef TRACE_SYSTEM
#define TRACE_SYSTEM cxl

#if !defined(_CXL_EVENTS_H) || defined(TRACE_HEADER_MULTI_READ)
#define _CXL_EVENTS_H

#include <linux/tracepoint.h>
#include <linux/pci.h>
#include <linux/unaligned.h>

#include <cxl.h>
#include <cxlmem.h>
#include "core.h"

#define CXL_RAS_UC_CACHE_DATA_PARITY BIT(0)
#define CXL_RAS_UC_CACHE_ADDR_PARITY BIT(1)
#define CXL_RAS_UC_CACHE_BE_PARITY BIT(2)
#define CXL_RAS_UC_CACHE_DATA_ECC BIT(3)
#define CXL_RAS_UC_MEM_DATA_PARITY BIT(4)
#define CXL_RAS_UC_MEM_ADDR_PARITY BIT(5)
#define CXL_RAS_UC_MEM_BE_PARITY BIT(6)
#define CXL_RAS_UC_MEM_DATA_ECC  BIT(7)
#define CXL_RAS_UC_REINIT_THRESH BIT(8)
#define CXL_RAS_UC_RSVD_ENCODE  BIT(9)
#define CXL_RAS_UC_POISON  BIT(10)
#define CXL_RAS_UC_RECV_OVERFLOW BIT(11)
#define CXL_RAS_UC_INTERNAL_ERR  BIT(14)
#define CXL_RAS_UC_IDE_TX_ERR  BIT(15)
#define CXL_RAS_UC_IDE_RX_ERR  BIT(16)

#define show_uc_errs(status) __print_flags(status, " | ",    \
 { CXL_RAS_UC_CACHE_DATA_PARITY, "Cache Data Parity Error" },   \
 { CXL_RAS_UC_CACHE_ADDR_PARITY, "Cache Address Parity Error" },   \
 { CXL_RAS_UC_CACHE_BE_PARITY, "Cache Byte Enable Parity Error" }, \
 { CXL_RAS_UC_CACHE_DATA_ECC, "Cache Data ECC Error" },    \
 { CXL_RAS_UC_MEM_DATA_PARITY, "Memory Data Parity Error" },   \
 { CXL_RAS_UC_MEM_ADDR_PARITY, "Memory Address Parity Error" },   \
 { CXL_RAS_UC_MEM_BE_PARITY, "Memory Byte Enable Parity Error" },  \
 { CXL_RAS_UC_MEM_DATA_ECC, "Memory Data ECC Error" },    \
 { CXL_RAS_UC_REINIT_THRESH, "REINIT Threshold Hit" },    \
 { CXL_RAS_UC_RSVD_ENCODE, "Received Unrecognized Encoding" },   \
 { CXL_RAS_UC_POISON, "Received Poison From Peer" },    \
 { CXL_RAS_UC_RECV_OVERFLOW, "Receiver Overflow" },    \
 { CXL_RAS_UC_INTERNAL_ERR, "Component Specific Error" },   \
 { CXL_RAS_UC_IDE_TX_ERR, "IDE Tx Error" },     \
 { CXL_RAS_UC_IDE_RX_ERR, "IDE Rx Error" }     \
)

TRACE_EVENT(cxl_port_aer_uncorrectable_error,
 TP_PROTO(struct device *dev, u32 status, u32 fe, u32 *hl),
 TP_ARGS(dev, status, fe, hl),
 TP_STRUCT__entry(
  __string(device, dev_name(dev))
  __string(host, dev_name(dev->parent))
  __field(u32, status)
  __field(u32, first_error)
  __array(u32, header_log, CXL_HEADERLOG_SIZE_U32)
 ),
 TP_fast_assign(
  __assign_str(device);
  __assign_str(host);
  __entry->status = status;
  __entry->first_error = fe;
  /*
 * Embed the 512B headerlog data for user app retrieval and
 * parsing, but no need to print this in the trace buffer.
 */

  memcpy(__entry->header_log, hl, CXL_HEADERLOG_SIZE);
 ),
 TP_printk("device=%s host=%s status: '%s' first_error: '%s'",
    __get_str(device), __get_str(host),
    show_uc_errs(__entry->status),
    show_uc_errs(__entry->first_error)
 )
);

TRACE_EVENT(cxl_aer_uncorrectable_error,
 TP_PROTO(const struct cxl_memdev *cxlmd, u32 status, u32 fe, u32 *hl),
 TP_ARGS(cxlmd, status, fe, hl),
 TP_STRUCT__entry(
  __string(memdev, dev_name(&cxlmd->dev))
  __string(host, dev_name(cxlmd->dev.parent))
  __field(u64, serial)
  __field(u32, status)
  __field(u32, first_error)
  __array(u32, header_log, CXL_HEADERLOG_SIZE_U32)
 ),
 TP_fast_assign(
  __assign_str(memdev);
  __assign_str(host);
  __entry->serial = cxlmd->cxlds->serial;
  __entry->status = status;
  __entry->first_error = fe;
  /*
 * Embed the 512B headerlog data for user app retrieval and
 * parsing, but no need to print this in the trace buffer.
 */

  memcpy(__entry->header_log, hl, CXL_HEADERLOG_SIZE);
 ),
 TP_printk("memdev=%s host=%s serial=%lld: status: '%s' first_error: '%s'",
    __get_str(memdev), __get_str(host), __entry->serial,
    show_uc_errs(__entry->status),
    show_uc_errs(__entry->first_error)
 )
);

#define CXL_RAS_CE_CACHE_DATA_ECC BIT(0)
#define CXL_RAS_CE_MEM_DATA_ECC  BIT(1)
#define CXL_RAS_CE_CRC_THRESH  BIT(2)
#define CLX_RAS_CE_RETRY_THRESH  BIT(3)
#define CXL_RAS_CE_CACHE_POISON  BIT(4)
#define CXL_RAS_CE_MEM_POISON  BIT(5)
#define CXL_RAS_CE_PHYS_LAYER_ERR BIT(6)

#define show_ce_errs(status) __print_flags(status, " | ",   \
 { CXL_RAS_CE_CACHE_DATA_ECC, "Cache Data ECC Error" },   \
 { CXL_RAS_CE_MEM_DATA_ECC, "Memory Data ECC Error" },   \
 { CXL_RAS_CE_CRC_THRESH, "CRC Threshold Hit" },    \
 { CLX_RAS_CE_RETRY_THRESH, "Retry Threshold" },    \
 { CXL_RAS_CE_CACHE_POISON, "Received Cache Poison From Peer" },  \
 { CXL_RAS_CE_MEM_POISON, "Received Memory Poison From Peer" },  \
 { CXL_RAS_CE_PHYS_LAYER_ERR, "Received Error From Physical Layer" } \
)

TRACE_EVENT(cxl_port_aer_correctable_error,
 TP_PROTO(struct device *dev, u32 status),
 TP_ARGS(dev, status),
 TP_STRUCT__entry(
  __string(device, dev_name(dev))
  __string(host, dev_name(dev->parent))
  __field(u32, status)
 ),
 TP_fast_assign(
  __assign_str(device);
  __assign_str(host);
  __entry->status = status;
 ),
 TP_printk("device=%s host=%s status='%s'",
    __get_str(device), __get_str(host),
    show_ce_errs(__entry->status)
 )
);

TRACE_EVENT(cxl_aer_correctable_error,
 TP_PROTO(const struct cxl_memdev *cxlmd, u32 status),
 TP_ARGS(cxlmd, status),
 TP_STRUCT__entry(
  __string(memdev, dev_name(&cxlmd->dev))
  __string(host, dev_name(cxlmd->dev.parent))
  __field(u64, serial)
  __field(u32, status)
 ),
 TP_fast_assign(
  __assign_str(memdev);
  __assign_str(host);
  __entry->serial = cxlmd->cxlds->serial;
  __entry->status = status;
 ),
 TP_printk("memdev=%s host=%s serial=%lld: status: '%s'",
    __get_str(memdev), __get_str(host), __entry->serial,
    show_ce_errs(__entry->status)
 )
);

#define cxl_event_log_type_str(type)    \
 __print_symbolic(type,     \
  { CXL_EVENT_TYPE_INFO, "Informational" }, \
  { CXL_EVENT_TYPE_WARN, "Warning" },  \
  { CXL_EVENT_TYPE_FAIL, "Failure" },  \
  { CXL_EVENT_TYPE_FATAL, "Fatal" })

TRACE_EVENT(cxl_overflow,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   struct cxl_get_event_payload *payload),

 TP_ARGS(cxlmd, log, payload),

 TP_STRUCT__entry(
  __string(memdev, dev_name(&cxlmd->dev))
  __string(host, dev_name(cxlmd->dev.parent))
  __field(int, log)
  __field(u64, serial)
  __field(u64, first_ts)
  __field(u64, last_ts)
  __field(u16, count)
 ),

 TP_fast_assign(
  __assign_str(memdev);
  __assign_str(host);
  __entry->serial = cxlmd->cxlds->serial;
  __entry->log = log;
  __entry->count = le16_to_cpu(payload->overflow_err_count);
  __entry->first_ts = le64_to_cpu(payload->first_overflow_timestamp);
  __entry->last_ts = le64_to_cpu(payload->last_overflow_timestamp);
 ),

 TP_printk("memdev=%s host=%s serial=%lld: log=%s : %u records from %llu to %llu",
  __get_str(memdev), __get_str(host), __entry->serial,
  cxl_event_log_type_str(__entry->log), __entry->count,
  __entry->first_ts, __entry->last_ts)

);

/*
 * Common Event Record Format
 * CXL 3.0 section 8.2.9.2.1; Table 8-42
 */

#define CXL_EVENT_RECORD_FLAG_PERMANENT  BIT(2)
#define CXL_EVENT_RECORD_FLAG_MAINT_NEEDED BIT(3)
#define CXL_EVENT_RECORD_FLAG_PERF_DEGRADED BIT(4)
#define CXL_EVENT_RECORD_FLAG_HW_REPLACE BIT(5)
#define CXL_EVENT_RECORD_FLAG_MAINT_OP_SUB_CLASS_VALID BIT(6)
#define CXL_EVENT_RECORD_FLAG_LD_ID_VALID BIT(7)
#define CXL_EVENT_RECORD_FLAG_HEAD_ID_VALID BIT(8)
#define show_hdr_flags(flags) __print_flags(flags, " | ",      \
 { CXL_EVENT_RECORD_FLAG_PERMANENT, "PERMANENT_CONDITION"  }, \
 { CXL_EVENT_RECORD_FLAG_MAINT_NEEDED, "MAINTENANCE_NEEDED"  }, \
 { CXL_EVENT_RECORD_FLAG_PERF_DEGRADED, "PERFORMANCE_DEGRADED"  }, \
 { CXL_EVENT_RECORD_FLAG_HW_REPLACE, "HARDWARE_REPLACEMENT_NEEDED" },  \
 { CXL_EVENT_RECORD_FLAG_MAINT_OP_SUB_CLASS_VALID, "MAINT_OP_SUB_CLASS_VALID" }, \
 { CXL_EVENT_RECORD_FLAG_LD_ID_VALID, "LD_ID_VALID" }, \
 { CXL_EVENT_RECORD_FLAG_HEAD_ID_VALID, "HEAD_ID_VALID" } \
)

/*
 * Define macros for the common header of each CXL event.
 *
 * Tracepoints using these macros must do 3 things:
 *
 * 1) Add CXL_EVT_TP_entry to TP_STRUCT__entry
 * 2) Use CXL_EVT_TP_fast_assign within TP_fast_assign;
 *    pass the dev, log, and CXL event header
 *    NOTE: The uuid must be assigned by the specific trace event
 * 3) Use CXL_EVT_TP_printk() instead of TP_printk()
 *
 * See the generic_event tracepoint as an example.
 */

#define CXL_EVT_TP_entry     \
 __string(memdev, dev_name(&cxlmd->dev))   \
 __string(host, dev_name(cxlmd->dev.parent))  \
 __field(int, log)     \
 __field_struct(uuid_t, hdr_uuid)   \
 __field(u64, serial)     \
 __field(u32, hdr_flags)     \
 __field(u16, hdr_handle)    \
 __field(u16, hdr_related_handle)   \
 __field(u64, hdr_timestamp)    \
 __field(u8, hdr_length)     \
 __field(u8, hdr_maint_op_class)    \
 __field(u8, hdr_maint_op_sub_class)   \
 __field(u16, hdr_ld_id)     \
 __field(u8, hdr_head_id)

#define CXL_EVT_TP_fast_assign(cxlmd, l, hdr)     \
 __assign_str(memdev);    \
 __assign_str(host);   \
 __entry->log = (l);       \
 __entry->serial = (cxlmd)->cxlds->serial;    \
 __entry->hdr_length = (hdr).length;     \
 __entry->hdr_flags = get_unaligned_le24((hdr).flags);   \
 __entry->hdr_handle = le16_to_cpu((hdr).handle);   \
 __entry->hdr_related_handle = le16_to_cpu((hdr).related_handle); \
 __entry->hdr_timestamp = le64_to_cpu((hdr).timestamp);   \
 __entry->hdr_maint_op_class = (hdr).maint_op_class;   \
 __entry->hdr_maint_op_sub_class = (hdr).maint_op_sub_class;  \
 __entry->hdr_ld_id = le16_to_cpu((hdr).ld_id);    \
 __entry->hdr_head_id = (hdr).head_id

#define CXL_EVT_TP_printk(fmt, ...) \
 TP_printk("memdev=%s host=%s serial=%lld log=%s : time=%llu uuid=%pUb " \
  "len=%d flags='%s' handle=%x related_handle=%x "  \
  "maint_op_class=%u maint_op_sub_class=%u "   \
  "ld_id=%x head_id=%x : " fmt,     \
  __get_str(memdev), __get_str(host), __entry->serial,  \
  cxl_event_log_type_str(__entry->log),    \
  __entry->hdr_timestamp, &__entry->hdr_uuid, __entry->hdr_length,\
  show_hdr_flags(__entry->hdr_flags), __entry->hdr_handle, \
  __entry->hdr_related_handle, __entry->hdr_maint_op_class, \
  __entry->hdr_maint_op_sub_class, \
  __entry->hdr_ld_id, __entry->hdr_head_id,   \
  ##__VA_ARGS__)

TRACE_EVENT(cxl_generic_event,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   const uuid_t *uuid, struct cxl_event_generic *gen_rec),

 TP_ARGS(cxlmd, log, uuid, gen_rec),

 TP_STRUCT__entry(
  CXL_EVT_TP_entry
  __array(u8, data, CXL_EVENT_RECORD_DATA_LENGTH)
 ),

 TP_fast_assign(
  CXL_EVT_TP_fast_assign(cxlmd, log, gen_rec->hdr);
  memcpy(&__entry->hdr_uuid, uuid, sizeof(uuid_t));
  memcpy(__entry->data, gen_rec->data, CXL_EVENT_RECORD_DATA_LENGTH);
 ),

 CXL_EVT_TP_printk("%s",
  __print_hex(__entry->data, CXL_EVENT_RECORD_DATA_LENGTH))
);

/*
 * Physical Address field masks
 *
 * General Media Event Record
 * CXL rev 3.0 Section 8.2.9.2.1.1; Table 8-43
 *
 * DRAM Event Record
 * CXL rev 3.0 section 8.2.9.2.1.2; Table 8-44
 */

#define CXL_DPA_FLAGS_MASK   GENMASK(1, 0)
#define CXL_DPA_MASK    GENMASK_ULL(63, 6)

#define CXL_DPA_VOLATILE   BIT(0)
#define CXL_DPA_NOT_REPAIRABLE   BIT(1)
#define show_dpa_flags(flags) __print_flags(flags, "|",     \
 { CXL_DPA_VOLATILE,   "VOLATILE"  }, \
 { CXL_DPA_NOT_REPAIRABLE,  "NOT_REPAIRABLE" }  \
)

/*
 * Component ID Format
 * CXL 3.1 section 8.2.9.2.1; Table 8-44
 */

#define CXL_PLDM_COMPONENT_ID_ENTITY_VALID BIT(0)
#define CXL_PLDM_COMPONENT_ID_RES_VALID  BIT(1)

#define show_comp_id_pldm_flags(flags)  __print_flags(flags, " | ", \
 { CXL_PLDM_COMPONENT_ID_ENTITY_VALID,   "PLDM Entity ID" }, \
 { CXL_PLDM_COMPONENT_ID_RES_VALID,      "Resource ID" }  \
)

#define show_pldm_entity_id(flags, valid_comp_id, valid_id_format, comp_id) \
 (flags & valid_comp_id && flags & valid_id_format) ?   \
 (comp_id[0] & CXL_PLDM_COMPONENT_ID_ENTITY_VALID) ?   \
 __print_hex(&comp_id[1], 6) : "0x00" : "0x00"

#define show_pldm_resource_id(flags, valid_comp_id, valid_id_format, comp_id) \
 (flags & valid_comp_id && flags & valid_id_format) ?   \
 (comp_id[0] & CXL_PLDM_COMPONENT_ID_RES_VALID) ?   \
 __print_hex(&comp_id[7], 4) : "0x00" : "0x00"

/*
 * General Media Event Record - GMER
 * CXL rev 3.1 Section 8.2.9.2.1.1; Table 8-45
 */

#define CXL_GMER_EVT_DESC_UNCORECTABLE_EVENT  BIT(0)
#define CXL_GMER_EVT_DESC_THRESHOLD_EVENT  BIT(1)
#define CXL_GMER_EVT_DESC_POISON_LIST_OVERFLOW  BIT(2)
#define show_event_desc_flags(flags) __print_flags(flags, "|",     \
 { CXL_GMER_EVT_DESC_UNCORECTABLE_EVENT,  "UNCORRECTABLE_EVENT" }, \
 { CXL_GMER_EVT_DESC_THRESHOLD_EVENT,  "THRESHOLD_EVENT" }, \
 { CXL_GMER_EVT_DESC_POISON_LIST_OVERFLOW, "POISON_LIST_OVERFLOW" }  \
)

#define CXL_GMER_MEM_EVT_TYPE_ECC_ERROR   0x00
#define CXL_GMER_MEM_EVT_TYPE_INV_ADDR   0x01
#define CXL_GMER_MEM_EVT_TYPE_DATA_PATH_ERROR  0x02
#define CXL_GMER_MEM_EVT_TYPE_TE_STATE_VIOLATION 0x03
#define CXL_GMER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR 0x04
#define CXL_GMER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE 0x05
#define CXL_GMER_MEM_EVT_TYPE_CKID_VIOLATION  0x06
#define show_gmer_mem_event_type(type) __print_symbolic(type,    \
 { CXL_GMER_MEM_EVT_TYPE_ECC_ERROR,  "ECC Error" },   \
 { CXL_GMER_MEM_EVT_TYPE_INV_ADDR,  "Invalid Address" },  \
 { CXL_GMER_MEM_EVT_TYPE_DATA_PATH_ERROR, "Data Path Error" },  \
 { CXL_GMER_MEM_EVT_TYPE_TE_STATE_VIOLATION, "TE State Violation" },  \
 { CXL_GMER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR, "Scrub Media ECC Error" }, \
 { CXL_GMER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE, "Adv Prog CME Counter Expiration" }, \
 { CXL_GMER_MEM_EVT_TYPE_CKID_VIOLATION,  "CKID Violation" }  \
)

#define CXL_GMER_TRANS_UNKNOWN    0x00
#define CXL_GMER_TRANS_HOST_READ   0x01
#define CXL_GMER_TRANS_HOST_WRITE   0x02
#define CXL_GMER_TRANS_HOST_SCAN_MEDIA   0x03
#define CXL_GMER_TRANS_HOST_INJECT_POISON  0x04
#define CXL_GMER_TRANS_INTERNAL_MEDIA_SCRUB  0x05
#define CXL_GMER_TRANS_INTERNAL_MEDIA_MANAGEMENT 0x06
#define CXL_GMER_TRANS_INTERNAL_MEDIA_ECS  0x07
#define CXL_GMER_TRANS_MEDIA_INITIALIZATION  0x08
#define show_trans_type(type) __print_symbolic(type,     \
 { CXL_GMER_TRANS_UNKNOWN,   "Unknown" },   \
 { CXL_GMER_TRANS_HOST_READ,   "Host Read" },   \
 { CXL_GMER_TRANS_HOST_WRITE,   "Host Write" },   \
 { CXL_GMER_TRANS_HOST_SCAN_MEDIA,  "Host Scan Media" },  \
 { CXL_GMER_TRANS_HOST_INJECT_POISON,  "Host Inject Poison" },  \
 { CXL_GMER_TRANS_INTERNAL_MEDIA_SCRUB,  "Internal Media Scrub" }, \
 { CXL_GMER_TRANS_INTERNAL_MEDIA_MANAGEMENT, "Internal Media Management" }, \
 { CXL_GMER_TRANS_INTERNAL_MEDIA_ECS,  "Internal Media Error Check Scrub" }, \
 { CXL_GMER_TRANS_MEDIA_INITIALIZATION,  "Media Initialization" } \
)

#define CXL_GMER_VALID_CHANNEL    BIT(0)
#define CXL_GMER_VALID_RANK    BIT(1)
#define CXL_GMER_VALID_DEVICE    BIT(2)
#define CXL_GMER_VALID_COMPONENT   BIT(3)
#define CXL_GMER_VALID_COMPONENT_ID_FORMAT  BIT(4)
#define show_valid_flags(flags) __print_flags(flags, "|",     \
 { CXL_GMER_VALID_CHANNEL,   "CHANNEL" }, \
 { CXL_GMER_VALID_RANK,    "RANK"  }, \
 { CXL_GMER_VALID_DEVICE,   "DEVICE" }, \
 { CXL_GMER_VALID_COMPONENT,   "COMPONENT" }, \
 { CXL_GMER_VALID_COMPONENT_ID_FORMAT,  "COMPONENT PLDM FORMAT" } \
)

#define CXL_GMER_CME_EV_FLAG_CME_MULTIPLE_MEDIA  BIT(0)
#define CXL_GMER_CME_EV_FLAG_THRESHOLD_EXCEEDED  BIT(1)
#define show_cme_threshold_ev_flags(flags) __print_flags(flags, "|", \
 {         \
  CXL_GMER_CME_EV_FLAG_CME_MULTIPLE_MEDIA,   \
  "Corrected Memory Errors in Multiple Media Components"  \
 }, {         \
  CXL_GMER_CME_EV_FLAG_THRESHOLD_EXCEEDED,   \
  "Exceeded Programmable Threshold"    \
 }         \
)

#define CXL_GMER_MEM_EVT_SUB_TYPE_NOT_REPORTED    0x00
#define CXL_GMER_MEM_EVT_SUB_TYPE_INTERNAL_DATAPATH_ERROR  0x01
#define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_COMMAND_TRAINING_ERROR 0x02
#define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CONTROL_TRAINING_ERROR 0x03
#define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_DATA_TRAINING_ERROR 0x04
#define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CRC_ERROR   0x05
#define show_mem_event_sub_type(sub_type) __print_symbolic(sub_type,   \
 { CXL_GMER_MEM_EVT_SUB_TYPE_NOT_REPORTED, "Not Reported" },    \
 { CXL_GMER_MEM_EVT_SUB_TYPE_INTERNAL_DATAPATH_ERROR, "Internal Datapath Error" }, \
 {           \
  CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_COMMAND_TRAINING_ERROR,   \
  "Media Link Command Training Error"      \
 }, {           \
  CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CONTROL_TRAINING_ERROR,   \
  "Media Link Control Training Error"      \
 }, {           \
  CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_DATA_TRAINING_ERROR,   \
  "Media Link Data Training Error"      \
 }, {           \
  CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CRC_ERROR, "Media Link CRC Error"  \
 }           \
)

TRACE_EVENT(cxl_general_media,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   struct cxl_region *cxlr, u64 hpa, u64 hpa_alias0,
   struct cxl_event_gen_media *rec),

 TP_ARGS(cxlmd, log, cxlr, hpa, hpa_alias0, rec),

 TP_STRUCT__entry(
  CXL_EVT_TP_entry
  /* General Media */
  __field(u64, dpa)
  __field(u8, descriptor)
  __field(u8, type)
  __field(u8, transaction_type)
  __field(u8, channel)
  __field(u32, device)
  __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE)
  /* Following are out of order to pack trace record */
  __field(u64, hpa)
  __field(u64, hpa_alias0)
  __field_struct(uuid_t, region_uuid)
  __field(u16, validity_flags)
  __field(u8, rank)
  __field(u8, dpa_flags)
  __field(u32, cme_count)
  __field(u8, sub_type)
  __field(u8, cme_threshold_ev_flags)
  __string(region_name, cxlr ? dev_name(&cxlr->dev) : "")
 ),

 TP_fast_assign(
  CXL_EVT_TP_fast_assign(cxlmd, log, rec->media_hdr.hdr);
  __entry->hdr_uuid = CXL_EVENT_GEN_MEDIA_UUID;

  /* General Media */
  __entry->dpa = le64_to_cpu(rec->media_hdr.phys_addr);
  __entry->dpa_flags = __entry->dpa & CXL_DPA_FLAGS_MASK;
  /* Mask after flags have been parsed */
  __entry->dpa &= CXL_DPA_MASK;
  __entry->descriptor = rec->media_hdr.descriptor;
  __entry->type = rec->media_hdr.type;
  __entry->sub_type = rec->sub_type;
  __entry->transaction_type = rec->media_hdr.transaction_type;
  __entry->channel = rec->media_hdr.channel;
  __entry->rank = rec->media_hdr.rank;
  __entry->device = get_unaligned_le24(rec->device);
  memcpy(__entry->comp_id, &rec->component_id,
   CXL_EVENT_GEN_MED_COMP_ID_SIZE);
  __entry->validity_flags = get_unaligned_le16(&rec->media_hdr.validity_flags);
  __entry->hpa = hpa;
  __entry->hpa_alias0 = hpa_alias0;
  if (cxlr) {
   __assign_str(region_name);
   uuid_copy(&__entry->region_uuid, &cxlr->params.uuid);
  } else {
   __assign_str(region_name);
   uuid_copy(&__entry->region_uuid, &uuid_null);
  }
  __entry->cme_threshold_ev_flags = rec->cme_threshold_ev_flags;
  if (rec->media_hdr.descriptor & CXL_GMER_EVT_DESC_THRESHOLD_EVENT)
   __entry->cme_count = get_unaligned_le24(rec->cme_count);
  else
   __entry->cme_count = 0;
 ),

 CXL_EVT_TP_printk("dpa=%llx dpa_flags='%s' " \
  "descriptor='%s' type='%s' sub_type='%s' " \
  "transaction_type='%s' channel=%u rank=%u " \
  "device=%x validity_flags='%s' " \
  "comp_id=%s comp_id_pldm_valid_flags='%s' " \
  "pldm_entity_id=%s pldm_resource_id=%s " \
  "hpa=%llx hpa_alias0=%llx region=%s region_uuid=%pUb " \
  "cme_threshold_ev_flags='%s' cme_count=%u",
  __entry->dpa, show_dpa_flags(__entry->dpa_flags),
  show_event_desc_flags(__entry->descriptor),
  show_gmer_mem_event_type(__entry->type),
  show_mem_event_sub_type(__entry->sub_type),
  show_trans_type(__entry->transaction_type),
  __entry->channel, __entry->rank, __entry->device,
  show_valid_flags(__entry->validity_flags),
  __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE),
  show_comp_id_pldm_flags(__entry->comp_id[0]),
  show_pldm_entity_id(__entry->validity_flags, CXL_GMER_VALID_COMPONENT,
        CXL_GMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  show_pldm_resource_id(__entry->validity_flags, CXL_GMER_VALID_COMPONENT,
          CXL_GMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  __entry->hpa, __entry->hpa_alias0, __get_str(region_name), &__entry->region_uuid,
  show_cme_threshold_ev_flags(__entry->cme_threshold_ev_flags), __entry->cme_count
 )
);

/*
 * DRAM Event Record - DER
 *
 * CXL rev 3.1 section 8.2.9.2.1.2; Table 8-46
 */

/*
 * DRAM Event Record defines many fields the same as the General Media Event
 * Record.  Reuse those definitions as appropriate.
 */

#define CXL_DER_MEM_EVT_TYPE_ECC_ERROR   0x00
#define CXL_DER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR 0x01
#define CXL_DER_MEM_EVT_TYPE_INV_ADDR   0x02
#define CXL_DER_MEM_EVT_TYPE_DATA_PATH_ERROR  0x03
#define CXL_DER_MEM_EVT_TYPE_TE_STATE_VIOLATION 0x04
#define CXL_DER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE 0x05
#define CXL_DER_MEM_EVT_TYPE_CKID_VIOLATION  0x06
#define show_dram_mem_event_type(type) __print_symbolic(type,     \
 { CXL_DER_MEM_EVT_TYPE_ECC_ERROR,  "ECC Error" },    \
 { CXL_DER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR, "Scrub Media ECC Error" },  \
 { CXL_DER_MEM_EVT_TYPE_INV_ADDR,  "Invalid Address" },   \
 { CXL_DER_MEM_EVT_TYPE_DATA_PATH_ERROR,  "Data Path Error" },   \
 { CXL_DER_MEM_EVT_TYPE_TE_STATE_VIOLATION, "TE State Violation" },   \
 { CXL_DER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE, "Adv Prog CME Counter Expiration" }, \
 { CXL_DER_MEM_EVT_TYPE_CKID_VIOLATION,  "CKID Violation" }   \
)

#define CXL_DER_VALID_CHANNEL    BIT(0)
#define CXL_DER_VALID_RANK    BIT(1)
#define CXL_DER_VALID_NIBBLE    BIT(2)
#define CXL_DER_VALID_BANK_GROUP   BIT(3)
#define CXL_DER_VALID_BANK    BIT(4)
#define CXL_DER_VALID_ROW    BIT(5)
#define CXL_DER_VALID_COLUMN    BIT(6)
#define CXL_DER_VALID_CORRECTION_MASK   BIT(7)
#define CXL_DER_VALID_COMPONENT    BIT(8)
#define CXL_DER_VALID_COMPONENT_ID_FORMAT  BIT(9)
#define CXL_DER_VALID_SUB_CHANNEL   BIT(10)
#define show_dram_valid_flags(flags) __print_flags(flags, "|",   \
 { CXL_DER_VALID_CHANNEL,   "CHANNEL"  }, \
 { CXL_DER_VALID_RANK,    "RANK"   }, \
 { CXL_DER_VALID_NIBBLE,    "NIBBLE"  }, \
 { CXL_DER_VALID_BANK_GROUP,   "BANK GROUP"  }, \
 { CXL_DER_VALID_BANK,    "BANK"   }, \
 { CXL_DER_VALID_ROW,    "ROW"   }, \
 { CXL_DER_VALID_COLUMN,    "COLUMN"  }, \
 { CXL_DER_VALID_CORRECTION_MASK,  "CORRECTION MASK" }, \
 { CXL_DER_VALID_COMPONENT,   "COMPONENT"  }, \
 { CXL_DER_VALID_COMPONENT_ID_FORMAT,  "COMPONENT PLDM FORMAT" }, \
 { CXL_DER_VALID_SUB_CHANNEL,   "SUB CHANNEL"  } \
)

TRACE_EVENT(cxl_dram,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   struct cxl_region *cxlr, u64 hpa, u64 hpa_alias0,
   struct cxl_event_dram *rec),

 TP_ARGS(cxlmd, log, cxlr, hpa, hpa_alias0, rec),

 TP_STRUCT__entry(
  CXL_EVT_TP_entry
  /* DRAM */
  __field(u64, dpa)
  __field(u8, descriptor)
  __field(u8, type)
  __field(u8, transaction_type)
  __field(u8, channel)
  __field(u16, validity_flags)
  __field(u16, column) /* Out of order to pack trace record */
  __field(u32, nibble_mask)
  __field(u32, row)
  __array(u8, cor_mask, CXL_EVENT_DER_CORRECTION_MASK_SIZE)
  __field(u64, hpa)
  __field(u64, hpa_alias0)
  __field_struct(uuid_t, region_uuid)
  __field(u8, rank) /* Out of order to pack trace record */
  __field(u8, bank_group) /* Out of order to pack trace record */
  __field(u8, bank) /* Out of order to pack trace record */
  __field(u8, dpa_flags) /* Out of order to pack trace record */
  /* Following are out of order to pack trace record */
  __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE)
  __field(u32, cvme_count)
  __field(u8, sub_type)
  __field(u8, sub_channel)
  __field(u8, cme_threshold_ev_flags)
  __string(region_name, cxlr ? dev_name(&cxlr->dev) : "")
 ),

 TP_fast_assign(
  CXL_EVT_TP_fast_assign(cxlmd, log, rec->media_hdr.hdr);
  __entry->hdr_uuid = CXL_EVENT_DRAM_UUID;

  /* DRAM */
  __entry->dpa = le64_to_cpu(rec->media_hdr.phys_addr);
  __entry->dpa_flags = __entry->dpa & CXL_DPA_FLAGS_MASK;
  __entry->dpa &= CXL_DPA_MASK;
  __entry->descriptor = rec->media_hdr.descriptor;
  __entry->type = rec->media_hdr.type;
  __entry->sub_type = rec->sub_type;
  __entry->transaction_type = rec->media_hdr.transaction_type;
  __entry->validity_flags = get_unaligned_le16(rec->media_hdr.validity_flags);
  __entry->channel = rec->media_hdr.channel;
  __entry->rank = rec->media_hdr.rank;
  __entry->nibble_mask = get_unaligned_le24(rec->nibble_mask);
  __entry->bank_group = rec->bank_group;
  __entry->bank = rec->bank;
  __entry->row = get_unaligned_le24(rec->row);
  __entry->column = get_unaligned_le16(rec->column);
  memcpy(__entry->cor_mask, &rec->correction_mask,
   CXL_EVENT_DER_CORRECTION_MASK_SIZE);
  __entry->hpa = hpa;
  __entry->hpa_alias0 = hpa_alias0;
  if (cxlr) {
   __assign_str(region_name);
   uuid_copy(&__entry->region_uuid, &cxlr->params.uuid);
  } else {
   __assign_str(region_name);
   uuid_copy(&__entry->region_uuid, &uuid_null);
  }
  memcpy(__entry->comp_id, &rec->component_id,
         CXL_EVENT_GEN_MED_COMP_ID_SIZE);
  __entry->sub_channel = rec->sub_channel;
  __entry->cme_threshold_ev_flags = rec->cme_threshold_ev_flags;
  if (rec->media_hdr.descriptor & CXL_GMER_EVT_DESC_THRESHOLD_EVENT)
   __entry->cvme_count = get_unaligned_le24(rec->cvme_count);
  else
   __entry->cvme_count = 0;
 ),

 CXL_EVT_TP_printk("dpa=%llx dpa_flags='%s' descriptor='%s' type='%s' sub_type='%s' " \
  "transaction_type='%s' channel=%u rank=%u nibble_mask=%x " \
  "bank_group=%u bank=%u row=%u column=%u cor_mask=%s " \
  "validity_flags='%s' " \
  "comp_id=%s comp_id_pldm_valid_flags='%s' " \
  "pldm_entity_id=%s pldm_resource_id=%s " \
  "hpa=%llx hpa_alias0=%llx region=%s region_uuid=%pUb " \
  "sub_channel=%u cme_threshold_ev_flags='%s' cvme_count=%u",
  __entry->dpa, show_dpa_flags(__entry->dpa_flags),
  show_event_desc_flags(__entry->descriptor),
  show_dram_mem_event_type(__entry->type),
  show_mem_event_sub_type(__entry->sub_type),
  show_trans_type(__entry->transaction_type),
  __entry->channel, __entry->rank, __entry->nibble_mask,
  __entry->bank_group, __entry->bank,
  __entry->row, __entry->column,
  __print_hex(__entry->cor_mask, CXL_EVENT_DER_CORRECTION_MASK_SIZE),
  show_dram_valid_flags(__entry->validity_flags),
  __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE),
  show_comp_id_pldm_flags(__entry->comp_id[0]),
  show_pldm_entity_id(__entry->validity_flags, CXL_DER_VALID_COMPONENT,
        CXL_DER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  show_pldm_resource_id(__entry->validity_flags, CXL_DER_VALID_COMPONENT,
          CXL_DER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  __entry->hpa, __entry->hpa_alias0, __get_str(region_name), &__entry->region_uuid,
  __entry->sub_channel, show_cme_threshold_ev_flags(__entry->cme_threshold_ev_flags),
  __entry->cvme_count
 )
);

/*
 * Memory Module Event Record - MMER
 *
 * CXL res 3.1 section 8.2.9.2.1.3; Table 8-47
 */

#define CXL_MMER_HEALTH_STATUS_CHANGE  0x00
#define CXL_MMER_MEDIA_STATUS_CHANGE  0x01
#define CXL_MMER_LIFE_USED_CHANGE  0x02
#define CXL_MMER_TEMP_CHANGE   0x03
#define CXL_MMER_DATA_PATH_ERROR  0x04
#define CXL_MMER_LSA_ERROR   0x05
#define CXL_MMER_UNRECOV_SIDEBAND_BUS_ERROR 0x06
#define CXL_MMER_MEMORY_MEDIA_FRU_ERROR  0x07
#define CXL_MMER_POWER_MANAGEMENT_FAULT  0x08
#define show_dev_evt_type(type) __print_symbolic(type,      \
 { CXL_MMER_HEALTH_STATUS_CHANGE, "Health Status Change" }, \
 { CXL_MMER_MEDIA_STATUS_CHANGE,  "Media Status Change" }, \
 { CXL_MMER_LIFE_USED_CHANGE,  "Life Used Change" }, \
 { CXL_MMER_TEMP_CHANGE,   "Temperature Change" }, \
 { CXL_MMER_DATA_PATH_ERROR,  "Data Path Error" }, \
 { CXL_MMER_LSA_ERROR,   "LSA Error"  }, \
 { CXL_MMER_UNRECOV_SIDEBAND_BUS_ERROR, "Unrecoverable Internal Sideband Bus Error" }, \
 { CXL_MMER_MEMORY_MEDIA_FRU_ERROR, "Memory Media FRU Error" }, \
 { CXL_MMER_POWER_MANAGEMENT_FAULT, "Power Management Fault" }  \
)

/*
 * Device Health Information - DHI
 *
 * CXL res 3.1 section 8.2.9.9.3.1; Table 8-133
 */

#define CXL_DHI_HS_MAINTENANCE_NEEDED    BIT(0)
#define CXL_DHI_HS_PERFORMANCE_DEGRADED    BIT(1)
#define CXL_DHI_HS_HW_REPLACEMENT_NEEDED   BIT(2)
#define CXL_DHI_HS_MEM_CAPACITY_DEGRADED   BIT(3)
#define show_health_status_flags(flags) __print_flags(flags, "|",    \
 { CXL_DHI_HS_MAINTENANCE_NEEDED, "MAINTENANCE_NEEDED" }, \
 { CXL_DHI_HS_PERFORMANCE_DEGRADED, "PERFORMANCE_DEGRADED" }, \
 { CXL_DHI_HS_HW_REPLACEMENT_NEEDED, "REPLACEMENT_NEEDED" }, \
 { CXL_DHI_HS_MEM_CAPACITY_DEGRADED, "MEM_CAPACITY_DEGRADED" }  \
)

#define CXL_DHI_MS_NORMAL       0x00
#define CXL_DHI_MS_NOT_READY       0x01
#define CXL_DHI_MS_WRITE_PERSISTENCY_LOST     0x02
#define CXL_DHI_MS_ALL_DATA_LOST      0x03
#define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_POWER_LOSS   0x04
#define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_SHUTDOWN   0x05
#define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_IMMINENT    0x06
#define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_POWER_LOSS    0x07
#define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_SHUTDOWN    0x08
#define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_IMMINENT     0x09
#define show_media_status(ms) __print_symbolic(ms,      \
 { CXL_DHI_MS_NORMAL,         \
  "Normal"      }, \
 { CXL_DHI_MS_NOT_READY,         \
  "Not Ready"      }, \
 { CXL_DHI_MS_WRITE_PERSISTENCY_LOST,       \
  "Write Persistency Lost"    }, \
 { CXL_DHI_MS_ALL_DATA_LOST,        \
  "All Data Lost"      }, \
 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_POWER_LOSS,     \
  "Write Persistency Loss in the Event of Power Loss" }, \
 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_SHUTDOWN,     \
  "Write Persistency Loss in Event of Shutdown"  }, \
 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_IMMINENT,      \
  "Write Persistency Loss Imminent"   }, \
 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_POWER_LOSS,     \
  "All Data Loss in Event of Power Loss"   }, \
 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_SHUTDOWN,     \
  "All Data loss in the Event of Shutdown"  }, \
 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_IMMINENT,      \
  "All Data Loss Imminent"    }  \
)

#define CXL_DHI_AS_NORMAL  0x0
#define CXL_DHI_AS_WARNING  0x1
#define CXL_DHI_AS_CRITICAL  0x2
#define show_two_bit_status(as) __print_symbolic(as,    \
 { CXL_DHI_AS_NORMAL,  "Normal" }, \
 { CXL_DHI_AS_WARNING,  "Warning" }, \
 { CXL_DHI_AS_CRITICAL,  "Critical" }  \
)
#define show_one_bit_status(as) __print_symbolic(as,    \
 { CXL_DHI_AS_NORMAL,  "Normal" }, \
 { CXL_DHI_AS_WARNING,  "Warning" }  \
)

#define CXL_DHI_AS_LIFE_USED(as)   (as & 0x3)
#define CXL_DHI_AS_DEV_TEMP(as)    ((as & 0xC) >> 2)
#define CXL_DHI_AS_COR_VOL_ERR_CNT(as)   ((as & 0x10) >> 4)
#define CXL_DHI_AS_COR_PER_ERR_CNT(as)   ((as & 0x20) >> 5)

#define CXL_MMER_VALID_COMPONENT   BIT(0)
#define CXL_MMER_VALID_COMPONENT_ID_FORMAT  BIT(1)
#define show_mem_module_valid_flags(flags) __print_flags(flags, "|", \
 { CXL_MMER_VALID_COMPONENT,  "COMPONENT" },   \
 { CXL_MMER_VALID_COMPONENT_ID_FORMAT, "COMPONENT PLDM FORMAT" } \
)
#define CXL_MMER_DEV_EVT_SUB_TYPE_NOT_REPORTED   0x00
#define CXL_MMER_DEV_EVT_SUB_TYPE_INVALID_CONFIG_DATA  0x01
#define CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_CONFIG_DATA  0x02
#define CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_MEM_MEDIA_FRU  0x03
#define show_dev_event_sub_type(sub_type) __print_symbolic(sub_type,   \
 { CXL_MMER_DEV_EVT_SUB_TYPE_NOT_REPORTED,  "Not Reported" },  \
 { CXL_MMER_DEV_EVT_SUB_TYPE_INVALID_CONFIG_DATA, "Invalid Config Data" }, \
 { CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_CONFIG_DATA,  "Unsupported Config Data" }, \
 {           \
  CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_MEM_MEDIA_FRU,     \
  "Unsupported Memory Media FRU"       \
 }           \
)

TRACE_EVENT(cxl_memory_module,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   struct cxl_event_mem_module *rec),

 TP_ARGS(cxlmd, log, rec),

 TP_STRUCT__entry(
  CXL_EVT_TP_entry

  /* Memory Module Event */
  __field(u8, event_type)

  /* Device Health Info */
  __field(u8, health_status)
  __field(u8, media_status)
  __field(u8, life_used)
  __field(u32, dirty_shutdown_cnt)
  __field(u32, cor_vol_err_cnt)
  __field(u32, cor_per_err_cnt)
  __field(s16, device_temp)
  __field(u8, add_status)
  __field(u8, event_sub_type)
  __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE)
  __field(u16, validity_flags)
 ),

 TP_fast_assign(
  CXL_EVT_TP_fast_assign(cxlmd, log, rec->hdr);
  __entry->hdr_uuid = CXL_EVENT_MEM_MODULE_UUID;

  /* Memory Module Event */
  __entry->event_type = rec->event_type;
  __entry->event_sub_type = rec->event_sub_type;

  /* Device Health Info */
  __entry->health_status = rec->info.health_status;
  __entry->media_status = rec->info.media_status;
  __entry->life_used = rec->info.life_used;
  __entry->dirty_shutdown_cnt = get_unaligned_le32(rec->info.dirty_shutdown_cnt);
  __entry->cor_vol_err_cnt = get_unaligned_le32(rec->info.cor_vol_err_cnt);
  __entry->cor_per_err_cnt = get_unaligned_le32(rec->info.cor_per_err_cnt);
  __entry->device_temp = get_unaligned_le16(rec->info.device_temp);
  __entry->add_status = rec->info.add_status;
  __entry->validity_flags = get_unaligned_le16(rec->validity_flags);
  memcpy(__entry->comp_id, &rec->component_id,
         CXL_EVENT_GEN_MED_COMP_ID_SIZE);
 ),

 CXL_EVT_TP_printk("event_type='%s' event_sub_type='%s' health_status='%s' " \
  "media_status='%s' as_life_used=%s as_dev_temp=%s as_cor_vol_err_cnt=%s " \
  "as_cor_per_err_cnt=%s life_used=%u device_temp=%d " \
  "dirty_shutdown_cnt=%u cor_vol_err_cnt=%u cor_per_err_cnt=%u " \
  "validity_flags='%s' " \
  "comp_id=%s comp_id_pldm_valid_flags='%s' " \
  "pldm_entity_id=%s pldm_resource_id=%s",
  show_dev_evt_type(__entry->event_type),
  show_dev_event_sub_type(__entry->event_sub_type),
  show_health_status_flags(__entry->health_status),
  show_media_status(__entry->media_status),
  show_two_bit_status(CXL_DHI_AS_LIFE_USED(__entry->add_status)),
  show_two_bit_status(CXL_DHI_AS_DEV_TEMP(__entry->add_status)),
  show_one_bit_status(CXL_DHI_AS_COR_VOL_ERR_CNT(__entry->add_status)),
  show_one_bit_status(CXL_DHI_AS_COR_PER_ERR_CNT(__entry->add_status)),
  __entry->life_used, __entry->device_temp,
  __entry->dirty_shutdown_cnt, __entry->cor_vol_err_cnt,
  __entry->cor_per_err_cnt,
  show_mem_module_valid_flags(__entry->validity_flags),
  __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE),
  show_comp_id_pldm_flags(__entry->comp_id[0]),
  show_pldm_entity_id(__entry->validity_flags, CXL_MMER_VALID_COMPONENT,
        CXL_MMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  show_pldm_resource_id(__entry->validity_flags, CXL_MMER_VALID_COMPONENT,
          CXL_MMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id)
 )
);

/*
 * Memory Sparing Event Record - MSER
 *
 * CXL rev 3.2 section 8.2.10.2.1.4; Table 8-60
 */

#define CXL_MSER_QUERY_RESOURCE_FLAG   BIT(0)
#define CXL_MSER_HARD_SPARING_FLAG   BIT(1)
#define CXL_MSER_DEV_INITED_FLAG   BIT(2)
#define show_mem_sparing_flags(flags) __print_flags(flags, "|", \
 { CXL_MSER_QUERY_RESOURCE_FLAG,  "Query Resources" }, \
 { CXL_MSER_HARD_SPARING_FLAG,  "Hard Sparing" }, \
 { CXL_MSER_DEV_INITED_FLAG, "Device Initiated Sparing" } \
)

#define CXL_MSER_VALID_CHANNEL    BIT(0)
#define CXL_MSER_VALID_RANK    BIT(1)
#define CXL_MSER_VALID_NIBBLE    BIT(2)
#define CXL_MSER_VALID_BANK_GROUP   BIT(3)
#define CXL_MSER_VALID_BANK    BIT(4)
#define CXL_MSER_VALID_ROW    BIT(5)
#define CXL_MSER_VALID_COLUMN    BIT(6)
#define CXL_MSER_VALID_COMPONENT_ID   BIT(7)
#define CXL_MSER_VALID_COMPONENT_ID_FORMAT  BIT(8)
#define CXL_MSER_VALID_SUB_CHANNEL   BIT(9)
#define show_mem_sparing_valid_flags(flags) __print_flags(flags, "|",  \
 { CXL_MSER_VALID_CHANNEL,   "CHANNEL" },   \
 { CXL_MSER_VALID_RANK,    "RANK" },   \
 { CXL_MSER_VALID_NIBBLE,   "NIBBLE" },   \
 { CXL_MSER_VALID_BANK_GROUP,   "BANK GROUP" },   \
 { CXL_MSER_VALID_BANK,    "BANK" },   \
 { CXL_MSER_VALID_ROW,    "ROW" },   \
 { CXL_MSER_VALID_COLUMN,   "COLUMN" },   \
 { CXL_MSER_VALID_COMPONENT_ID,   "COMPONENT ID" },  \
 { CXL_MSER_VALID_COMPONENT_ID_FORMAT,  "COMPONENT ID PLDM FORMAT" }, \
 { CXL_MSER_VALID_SUB_CHANNEL,   "SUB CHANNEL" }   \
)

TRACE_EVENT(cxl_memory_sparing,

 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log,
   struct cxl_event_mem_sparing *rec),

 TP_ARGS(cxlmd, log, rec),

 TP_STRUCT__entry(
  CXL_EVT_TP_entry

  /* Memory Sparing Event */
  __field(u8, flags)
  __field(u8, result)
  __field(u16, validity_flags)
  __field(u16, res_avail)
  __field(u8, channel)
  __field(u8, rank)
  __field(u32, nibble_mask)
  __field(u8, bank_group)
  __field(u8, bank)
  __field(u32, row)
  __field(u16, column)
  __field(u8, sub_channel)
  __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE)
 ),

 TP_fast_assign(
  CXL_EVT_TP_fast_assign(cxlmd, log, rec->hdr);
  __entry->hdr_uuid = CXL_EVENT_MEM_SPARING_UUID;

  /* Memory Sparing Event */
  __entry->flags = rec->flags;
  __entry->result = rec->result;
  __entry->validity_flags = le16_to_cpu(rec->validity_flags);
  __entry->res_avail = le16_to_cpu(rec->res_avail);
  __entry->channel = rec->channel;
  __entry->rank = rec->rank;
  __entry->nibble_mask = get_unaligned_le24(rec->nibble_mask);
  __entry->bank_group = rec->bank_group;
  __entry->bank = rec->bank;
  __entry->row = get_unaligned_le24(rec->row);
  __entry->column = le16_to_cpu(rec->column);
  __entry->sub_channel = rec->sub_channel;
  memcpy(__entry->comp_id, &rec->component_id,
         CXL_EVENT_GEN_MED_COMP_ID_SIZE);
 ),

 CXL_EVT_TP_printk("flags='%s' result=%u validity_flags='%s' " \
  "spare resource avail=%u channel=%u rank=%u " \
  "nibble_mask=%x bank_group=%u bank=%u " \
  "row=%u column=%u sub_channel=%u " \
  "comp_id=%s comp_id_pldm_valid_flags='%s' " \
  "pldm_entity_id=%s pldm_resource_id=%s",
  show_mem_sparing_flags(__entry->flags),
  __entry->result,
  show_mem_sparing_valid_flags(__entry->validity_flags),
  __entry->res_avail, __entry->channel, __entry->rank,
  __entry->nibble_mask, __entry->bank_group, __entry->bank,
  __entry->row, __entry->column, __entry->sub_channel,
  __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE),
  show_comp_id_pldm_flags(__entry->comp_id[0]),
  show_pldm_entity_id(__entry->validity_flags, CXL_MSER_VALID_COMPONENT_ID,
        CXL_MSER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id),
  show_pldm_resource_id(__entry->validity_flags, CXL_MSER_VALID_COMPONENT_ID,
          CXL_MSER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id)
 )
);

#define show_poison_trace_type(type)   \
 __print_symbolic(type,    \
 { CXL_POISON_TRACE_LIST, "List"   }, \
 { CXL_POISON_TRACE_INJECT, "Inject" }, \
 { CXL_POISON_TRACE_CLEAR, "Clear"  })

#define __show_poison_source(source)                          \
 __print_symbolic(source,                              \
  { CXL_POISON_SOURCE_UNKNOWN,   "Unknown"  },  \
  { CXL_POISON_SOURCE_EXTERNAL,  "External" },  \
  { CXL_POISON_SOURCE_INTERNAL,  "Internal" },  \
  { CXL_POISON_SOURCE_INJECTED,  "Injected" },  \
  { CXL_POISON_SOURCE_VENDOR,    "Vendor"   })

#define show_poison_source(source)        \
 (((source > CXL_POISON_SOURCE_INJECTED) &&      \
  (source != CXL_POISON_SOURCE_VENDOR)) ? "Reserved"  \
  : __show_poison_source(source))

#define show_poison_flags(flags)                             \
 __print_flags(flags, "|",                            \
  { CXL_POISON_FLAG_MORE,      "More"     },   \
  { CXL_POISON_FLAG_OVERFLOW,  "Overflow"  },  \
  { CXL_POISON_FLAG_SCANNING,  "Scanning"  })

#define __cxl_poison_addr(record)     \
 (le64_to_cpu(record->address))
#define cxl_poison_record_dpa(record)     \
 (__cxl_poison_addr(record) & CXL_POISON_START_MASK)
#define cxl_poison_record_source(record)    \
 (__cxl_poison_addr(record)  & CXL_POISON_SOURCE_MASK)
#define cxl_poison_record_dpa_length(record)    \
 (le32_to_cpu(record->length) * CXL_POISON_LEN_MULT)
#define cxl_poison_overflow(flags, time)    \
 (flags & CXL_POISON_FLAG_OVERFLOW ? le64_to_cpu(time) : 0)

TRACE_EVENT(cxl_poison,

 TP_PROTO(struct cxl_memdev *cxlmd, struct cxl_region *cxlr,
   const struct cxl_poison_record *record, u8 flags,
   __le64 overflow_ts, enum cxl_poison_trace_type trace_type),

 TP_ARGS(cxlmd, cxlr, record, flags, overflow_ts, trace_type),

 TP_STRUCT__entry(
  __string(memdev, dev_name(&cxlmd->dev))
  __string(host, dev_name(cxlmd->dev.parent))
  __field(u64, serial)
  __field(u8, trace_type)
  __string(region, cxlr ? dev_name(&cxlr->dev) : "")
  __field(u64, overflow_ts)
  __field(u64, hpa)
  __field(u64, hpa_alias0)
  __field(u64, dpa)
  __field(u32, dpa_length)
  __array(char, uuid, 16)
  __field(u8, source)
  __field(u8, flags)
     ),

 TP_fast_assign(
  __assign_str(memdev);
  __assign_str(host);
  __entry->serial = cxlmd->cxlds->serial;
  __entry->overflow_ts = cxl_poison_overflow(flags, overflow_ts);
  __entry->dpa = cxl_poison_record_dpa(record);
  __entry->dpa_length = cxl_poison_record_dpa_length(record);
  __entry->source = cxl_poison_record_source(record);
  __entry->trace_type = trace_type;
  __entry->flags = flags;
  if (cxlr) {
   __assign_str(region);
   memcpy(__entry->uuid, &cxlr->params.uuid, 16);
   __entry->hpa = cxl_dpa_to_hpa(cxlr, cxlmd,
            __entry->dpa);
   if (__entry->hpa != ULLONG_MAX && cxlr->params.cache_size)
    __entry->hpa_alias0 = __entry->hpa -
            cxlr->params.cache_size;
   else
    __entry->hpa_alias0 = ULLONG_MAX;
  } else {
   __assign_str(region);
   memset(__entry->uuid, 0, 16);
   __entry->hpa = ULLONG_MAX;
   __entry->hpa_alias0 = ULLONG_MAX;
  }
     ),

 TP_printk("memdev=%s host=%s serial=%lld trace_type=%s region=%s "  \
  "region_uuid=%pU hpa=0x%llx hpa_alias0=0x%llx dpa=0x%llx " \
  "dpa_length=0x%x source=%s flags=%s overflow_time=%llu",
  __get_str(memdev),
  __get_str(host),
  __entry->serial,
  show_poison_trace_type(__entry->trace_type),
  __get_str(region),
  __entry->uuid,
  __entry->hpa,
  __entry->hpa_alias0,
  __entry->dpa,
  __entry->dpa_length,
  show_poison_source(__entry->source),
  show_poison_flags(__entry->flags),
  __entry->overflow_ts
 )
);

#endif /* _CXL_EVENTS_H */

#define TRACE_INCLUDE_FILE trace
#include <trace/define_trace.h>

Messung V0.5
C=95 H=96 G=95

¤ Dauer der Verarbeitung: 0.14 Sekunden  (vorverarbeitet)  ¤

*© Formatika GbR, Deutschland






Wurzel

Suchen

Beweissystem der NASA

Beweissystem Isabelle

NIST Cobol Testsuite

Cephes Mathematical Library

Wiener Entwicklungsmethode

Haftungshinweis

Die Informationen auf dieser Webseite wurden nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit, noch Qualität der bereit gestellten Informationen zugesichert.

Bemerkung:

Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.