Murali Karicheri | 0e7623b | 2019-04-05 13:31:34 -0400 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 2 | /* Copyright 2011-2014 Autronica Fire and Security AS |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 3 | * |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 4 | * Author(s): |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 5 | * 2011-2014 Arvid Brodin, arvid.brodin@alten.se |
Murali Karicheri | 8f4c0e0 | 2020-07-22 10:40:16 -0400 | [diff] [blame] | 6 | * |
| 7 | * include file for HSR and PRP. |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 8 | */ |
| 9 | |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 10 | #ifndef __HSR_PRIVATE_H |
| 11 | #define __HSR_PRIVATE_H |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 12 | |
| 13 | #include <linux/netdevice.h> |
| 14 | #include <linux/list.h> |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 15 | #include <linux/if_vlan.h> |
George McCollister | dcf0cd1 | 2021-02-09 19:02:11 -0600 | [diff] [blame] | 16 | #include <linux/if_hsr.h> |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 17 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 18 | /* Time constants as specified in the HSR specification (IEC-62439-3 2010) |
| 19 | * Table 8. |
| 20 | * All values in milliseconds. |
| 21 | */ |
| 22 | #define HSR_LIFE_CHECK_INTERVAL 2000 /* ms */ |
| 23 | #define HSR_NODE_FORGET_TIME 60000 /* ms */ |
| 24 | #define HSR_ANNOUNCE_INTERVAL 100 /* ms */ |
Marco Wenzel | f176411 | 2021-02-24 10:46:49 +0100 | [diff] [blame] | 25 | #define HSR_ENTRY_FORGET_TIME 400 /* ms */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 26 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 27 | /* By how much may slave1 and slave2 timestamps of latest received frame from |
| 28 | * each node differ before we notify of communication problem? |
| 29 | */ |
| 30 | #define MAX_SLAVE_DIFF 3000 /* ms */ |
Arvid Brodin | f266a68 | 2014-07-04 23:41:03 +0200 | [diff] [blame] | 31 | #define HSR_SEQNR_START (USHRT_MAX - 1024) |
Peter Heise | ee1c279 | 2016-04-13 13:52:22 +0200 | [diff] [blame] | 32 | #define HSR_SUP_SEQNR_START (HSR_SEQNR_START / 2) |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 33 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 34 | /* How often shall we check for broken ring and remove node entries older than |
| 35 | * HSR_NODE_FORGET_TIME? |
| 36 | */ |
| 37 | #define PRUNE_PERIOD 3000 /* ms */ |
Andreas Oetken | eafaa88 | 2021-10-25 20:56:18 +0200 | [diff] [blame] | 38 | #define HSR_TLV_EOT 0 /* End of TLVs */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 39 | #define HSR_TLV_ANNOUNCE 22 |
| 40 | #define HSR_TLV_LIFE_CHECK 23 |
Murali Karicheri | c643ff0 | 2020-07-22 10:40:19 -0400 | [diff] [blame] | 41 | /* PRP V1 life check for Duplicate discard */ |
| 42 | #define PRP_TLV_LIFE_CHECK_DD 20 |
| 43 | /* PRP V1 life check for Duplicate Accept */ |
| 44 | #define PRP_TLV_LIFE_CHECK_DA 21 |
Andreas Oetken | eafaa88 | 2021-10-25 20:56:18 +0200 | [diff] [blame] | 45 | /* PRP V1 life redundancy box MAC address */ |
| 46 | #define PRP_TLV_REDBOX_MAC 30 |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 47 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 48 | /* HSR Tag. |
| 49 | * As defined in IEC-62439-3:2010, the HSR tag is really { ethertype = 0x88FB, |
| 50 | * path, LSDU_size, sequence Nr }. But we let eth_header() create { h_dest, |
| 51 | * h_source, h_proto = 0x88FB }, and add { path, LSDU_size, sequence Nr, |
| 52 | * encapsulated protocol } instead. |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 53 | * |
| 54 | * Field names as defined in the IEC:2010 standard for HSR. |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 55 | */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 56 | struct hsr_tag { |
| 57 | __be16 path_and_LSDU_size; |
| 58 | __be16 sequence_nr; |
| 59 | __be16 encap_proto; |
| 60 | } __packed; |
| 61 | |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 62 | #define HSR_HLEN 6 |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 63 | |
Peter Heise | ee1c279 | 2016-04-13 13:52:22 +0200 | [diff] [blame] | 64 | #define HSR_V1_SUP_LSDUSIZE 52 |
| 65 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 66 | /* The helper functions below assumes that 'path' occupies the 4 most |
| 67 | * significant bits of the 16-bit field shared by 'path' and 'LSDU_size' (or |
| 68 | * equivalently, the 4 most significant bits of HSR tag byte 14). |
| 69 | * |
| 70 | * This is unclear in the IEC specification; its definition of MAC addresses |
| 71 | * indicates the spec is written with the least significant bit first (to the |
| 72 | * left). This, however, would mean that the LSDU field would be split in two |
| 73 | * with the path field in-between, which seems strange. I'm guessing the MAC |
| 74 | * address definition is in error. |
| 75 | */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 76 | |
| 77 | static inline void set_hsr_tag_path(struct hsr_tag *ht, u16 path) |
| 78 | { |
Murali Karicheri | d595b85 | 2019-04-05 13:31:23 -0400 | [diff] [blame] | 79 | ht->path_and_LSDU_size = |
| 80 | htons((ntohs(ht->path_and_LSDU_size) & 0x0FFF) | (path << 12)); |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 81 | } |
| 82 | |
| 83 | static inline void set_hsr_tag_LSDU_size(struct hsr_tag *ht, u16 LSDU_size) |
| 84 | { |
Murali Karicheri | 0525fc0 | 2019-04-05 13:31:27 -0400 | [diff] [blame] | 85 | ht->path_and_LSDU_size = htons((ntohs(ht->path_and_LSDU_size) & |
| 86 | 0xF000) | (LSDU_size & 0x0FFF)); |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 87 | } |
| 88 | |
| 89 | struct hsr_ethhdr { |
| 90 | struct ethhdr ethhdr; |
| 91 | struct hsr_tag hsr_tag; |
| 92 | } __packed; |
| 93 | |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 94 | struct hsr_vlan_ethhdr { |
| 95 | struct vlan_ethhdr vlanhdr; |
| 96 | struct hsr_tag hsr_tag; |
| 97 | } __packed; |
| 98 | |
Andreas Oetken | eafaa88 | 2021-10-25 20:56:18 +0200 | [diff] [blame] | 99 | struct hsr_sup_tlv { |
| 100 | u8 HSR_TLV_type; |
| 101 | u8 HSR_TLV_length; |
| 102 | }; |
| 103 | |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 104 | /* HSR/PRP Supervision Frame data types. |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 105 | * Field names as defined in the IEC:2010 standard for HSR. |
| 106 | */ |
| 107 | struct hsr_sup_tag { |
Andreas Oetken | eafaa88 | 2021-10-25 20:56:18 +0200 | [diff] [blame] | 108 | __be16 path_and_HSR_ver; |
| 109 | __be16 sequence_nr; |
| 110 | struct hsr_sup_tlv tlv; |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 111 | } __packed; |
| 112 | |
| 113 | struct hsr_sup_payload { |
Murali Karicheri | b1b4aa9 | 2019-04-05 13:31:32 -0400 | [diff] [blame] | 114 | unsigned char macaddress_A[ETH_ALEN]; |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 115 | } __packed; |
| 116 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 117 | static inline void set_hsr_stag_path(struct hsr_sup_tag *hst, u16 path) |
| 118 | { |
Murali Karicheri | 5fa9677 | 2019-04-05 13:31:29 -0400 | [diff] [blame] | 119 | set_hsr_tag_path((struct hsr_tag *)hst, path); |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 120 | } |
| 121 | |
Murali Karicheri | b1b4aa9 | 2019-04-05 13:31:32 -0400 | [diff] [blame] | 122 | static inline void set_hsr_stag_HSR_ver(struct hsr_sup_tag *hst, u16 HSR_ver) |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 123 | { |
Murali Karicheri | b1b4aa9 | 2019-04-05 13:31:32 -0400 | [diff] [blame] | 124 | set_hsr_tag_LSDU_size((struct hsr_tag *)hst, HSR_ver); |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 125 | } |
| 126 | |
Peter Heise | ee1c279 | 2016-04-13 13:52:22 +0200 | [diff] [blame] | 127 | struct hsrv0_ethhdr_sp { |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 128 | struct ethhdr ethhdr; |
| 129 | struct hsr_sup_tag hsr_sup; |
| 130 | } __packed; |
| 131 | |
Peter Heise | ee1c279 | 2016-04-13 13:52:22 +0200 | [diff] [blame] | 132 | struct hsrv1_ethhdr_sp { |
| 133 | struct ethhdr ethhdr; |
| 134 | struct hsr_tag hsr; |
| 135 | struct hsr_sup_tag hsr_sup; |
| 136 | } __packed; |
| 137 | |
Arvid Brodin | c5a7591 | 2014-07-04 23:38:05 +0200 | [diff] [blame] | 138 | enum hsr_port_type { |
| 139 | HSR_PT_NONE = 0, /* Must be 0, used by framereg */ |
| 140 | HSR_PT_SLAVE_A, |
| 141 | HSR_PT_SLAVE_B, |
| 142 | HSR_PT_INTERLINK, |
| 143 | HSR_PT_MASTER, |
| 144 | HSR_PT_PORTS, /* This must be the last item in the enum */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 145 | }; |
Arvid Brodin | c5a7591 | 2014-07-04 23:38:05 +0200 | [diff] [blame] | 146 | |
Murali Karicheri | c643ff0 | 2020-07-22 10:40:19 -0400 | [diff] [blame] | 147 | /* PRP Redunancy Control Trailor (RCT). |
| 148 | * As defined in IEC-62439-4:2012, the PRP RCT is really { sequence Nr, |
| 149 | * Lan indentifier (LanId), LSDU_size and PRP_suffix = 0x88FB }. |
| 150 | * |
| 151 | * Field names as defined in the IEC:2012 standard for PRP. |
| 152 | */ |
| 153 | struct prp_rct { |
| 154 | __be16 sequence_nr; |
| 155 | __be16 lan_id_and_LSDU_size; |
| 156 | __be16 PRP_suffix; |
| 157 | } __packed; |
| 158 | |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 159 | static inline u16 get_prp_LSDU_size(struct prp_rct *rct) |
| 160 | { |
| 161 | return ntohs(rct->lan_id_and_LSDU_size) & 0x0FFF; |
| 162 | } |
| 163 | |
| 164 | static inline void set_prp_lan_id(struct prp_rct *rct, u16 lan_id) |
| 165 | { |
| 166 | rct->lan_id_and_LSDU_size = htons((ntohs(rct->lan_id_and_LSDU_size) & |
| 167 | 0x0FFF) | (lan_id << 12)); |
| 168 | } |
Murali Karicheri | c643ff0 | 2020-07-22 10:40:19 -0400 | [diff] [blame] | 169 | static inline void set_prp_LSDU_size(struct prp_rct *rct, u16 LSDU_size) |
| 170 | { |
| 171 | rct->lan_id_and_LSDU_size = htons((ntohs(rct->lan_id_and_LSDU_size) & |
| 172 | 0xF000) | (LSDU_size & 0x0FFF)); |
| 173 | } |
| 174 | |
Arvid Brodin | c5a7591 | 2014-07-04 23:38:05 +0200 | [diff] [blame] | 175 | struct hsr_port { |
| 176 | struct list_head port_list; |
| 177 | struct net_device *dev; |
| 178 | struct hsr_priv *hsr; |
| 179 | enum hsr_port_type type; |
| 180 | }; |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 181 | |
Murali Karicheri | fa4dc89 | 2020-07-22 10:40:20 -0400 | [diff] [blame] | 182 | struct hsr_frame_info; |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 183 | struct hsr_node; |
Murali Karicheri | fa4dc89 | 2020-07-22 10:40:20 -0400 | [diff] [blame] | 184 | |
Murali Karicheri | 28e458e | 2020-07-22 10:40:18 -0400 | [diff] [blame] | 185 | struct hsr_proto_ops { |
| 186 | /* format and send supervision frame */ |
| 187 | void (*send_sv_frame)(struct hsr_port *port, unsigned long *interval); |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 188 | void (*handle_san_frame)(bool san, enum hsr_port_type port, |
| 189 | struct hsr_node *node); |
| 190 | bool (*drop_frame)(struct hsr_frame_info *frame, struct hsr_port *port); |
Murali Karicheri | fa4dc89 | 2020-07-22 10:40:20 -0400 | [diff] [blame] | 191 | struct sk_buff * (*get_untagged_frame)(struct hsr_frame_info *frame, |
| 192 | struct hsr_port *port); |
| 193 | struct sk_buff * (*create_tagged_frame)(struct hsr_frame_info *frame, |
| 194 | struct hsr_port *port); |
George McCollister | 48b491a | 2021-05-24 13:50:54 -0500 | [diff] [blame] | 195 | int (*fill_frame_info)(__be16 proto, struct sk_buff *skb, |
| 196 | struct hsr_frame_info *frame); |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 197 | bool (*invalid_dan_ingress_frame)(__be16 protocol); |
| 198 | void (*update_san_info)(struct hsr_node *node, bool is_sup); |
Murali Karicheri | 28e458e | 2020-07-22 10:40:18 -0400 | [diff] [blame] | 199 | }; |
| 200 | |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 201 | struct hsr_priv { |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 202 | struct rcu_head rcu_head; |
Arvid Brodin | c5a7591 | 2014-07-04 23:38:05 +0200 | [diff] [blame] | 203 | struct list_head ports; |
Arvid Brodin | f266a68 | 2014-07-04 23:41:03 +0200 | [diff] [blame] | 204 | struct list_head node_db; /* Known HSR nodes */ |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 205 | struct list_head self_node_db; /* MACs of slaves */ |
| 206 | struct timer_list announce_timer; /* Supervision frame dispatch */ |
Arvid Brodin | abff716 | 2014-07-04 23:35:47 +0200 | [diff] [blame] | 207 | struct timer_list prune_timer; |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 208 | int announce_count; |
| 209 | u16 sequence_nr; |
Murali Karicheri | d595b85 | 2019-04-05 13:31:23 -0400 | [diff] [blame] | 210 | u16 sup_sequence_nr; /* For HSRv1 separate seq_nr for supervision */ |
Murali Karicheri | 8f4c0e0 | 2020-07-22 10:40:16 -0400 | [diff] [blame] | 211 | enum hsr_version prot_version; /* Indicate if HSRv0, HSRv1 or PRPv1 */ |
Taehee Yoo | 92a3567 | 2019-12-22 11:26:54 +0000 | [diff] [blame] | 212 | spinlock_t seqnr_lock; /* locking for sequence_nr */ |
| 213 | spinlock_t list_lock; /* locking for node list */ |
Murali Karicheri | 28e458e | 2020-07-22 10:40:18 -0400 | [diff] [blame] | 214 | struct hsr_proto_ops *proto_ops; |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 215 | #define PRP_LAN_ID 0x5 /* 0x1010 for A and 0x1011 for B. Bit 0 is set |
| 216 | * based on SLAVE_A or SLAVE_B |
| 217 | */ |
| 218 | u8 net_id; /* for PRP, it occupies most significant 3 bits |
| 219 | * of lan_id |
| 220 | */ |
Andreas Oetken | 6c9f18f | 2021-02-02 10:03:04 +0100 | [diff] [blame] | 221 | unsigned char sup_multicast_addr[ETH_ALEN] __aligned(sizeof(u16)); |
| 222 | /* Align to u16 boundary to avoid unaligned access |
| 223 | * in ether_addr_equal |
| 224 | */ |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 225 | #ifdef CONFIG_DEBUG_FS |
| 226 | struct dentry *node_tbl_root; |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 227 | #endif |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 228 | }; |
| 229 | |
Arvid Brodin | f266a68 | 2014-07-04 23:41:03 +0200 | [diff] [blame] | 230 | #define hsr_for_each_port(hsr, port) \ |
| 231 | list_for_each_entry_rcu((port), &(hsr)->ports, port_list) |
| 232 | |
Arvid Brodin | c5a7591 | 2014-07-04 23:38:05 +0200 | [diff] [blame] | 233 | struct hsr_port *hsr_port_get_hsr(struct hsr_priv *hsr, enum hsr_port_type pt); |
Arvid Brodin | f421436 | 2013-10-30 21:10:47 +0100 | [diff] [blame] | 234 | |
Arvid Brodin | f266a68 | 2014-07-04 23:41:03 +0200 | [diff] [blame] | 235 | /* Caller must ensure skb is a valid HSR frame */ |
| 236 | static inline u16 hsr_get_skb_sequence_nr(struct sk_buff *skb) |
| 237 | { |
| 238 | struct hsr_ethhdr *hsr_ethhdr; |
| 239 | |
Murali Karicheri | 5fa9677 | 2019-04-05 13:31:29 -0400 | [diff] [blame] | 240 | hsr_ethhdr = (struct hsr_ethhdr *)skb_mac_header(skb); |
Arvid Brodin | f266a68 | 2014-07-04 23:41:03 +0200 | [diff] [blame] | 241 | return ntohs(hsr_ethhdr->hsr_tag.sequence_nr); |
| 242 | } |
| 243 | |
Murali Karicheri | 451d812 | 2020-07-22 10:40:21 -0400 | [diff] [blame] | 244 | static inline struct prp_rct *skb_get_PRP_rct(struct sk_buff *skb) |
| 245 | { |
| 246 | unsigned char *tail = skb_tail_pointer(skb) - HSR_HLEN; |
| 247 | |
| 248 | struct prp_rct *rct = (struct prp_rct *)tail; |
| 249 | |
| 250 | if (rct->PRP_suffix == htons(ETH_P_PRP)) |
| 251 | return rct; |
| 252 | |
| 253 | return NULL; |
| 254 | } |
| 255 | |
| 256 | /* Assume caller has confirmed this skb is PRP suffixed */ |
| 257 | static inline u16 prp_get_skb_sequence_nr(struct prp_rct *rct) |
| 258 | { |
| 259 | return ntohs(rct->sequence_nr); |
| 260 | } |
| 261 | |
| 262 | static inline u16 get_prp_lan_id(struct prp_rct *rct) |
| 263 | { |
| 264 | return ntohs(rct->lan_id_and_LSDU_size) >> 12; |
| 265 | } |
| 266 | |
| 267 | /* assume there is a valid rct */ |
| 268 | static inline bool prp_check_lsdu_size(struct sk_buff *skb, |
| 269 | struct prp_rct *rct, |
| 270 | bool is_sup) |
| 271 | { |
| 272 | struct ethhdr *ethhdr; |
| 273 | int expected_lsdu_size; |
| 274 | |
| 275 | if (is_sup) { |
| 276 | expected_lsdu_size = HSR_V1_SUP_LSDUSIZE; |
| 277 | } else { |
| 278 | ethhdr = (struct ethhdr *)skb_mac_header(skb); |
| 279 | expected_lsdu_size = skb->len - 14; |
| 280 | if (ethhdr->h_proto == htons(ETH_P_8021Q)) |
| 281 | expected_lsdu_size -= 4; |
| 282 | } |
| 283 | |
| 284 | return (expected_lsdu_size == get_prp_LSDU_size(rct)); |
| 285 | } |
| 286 | |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 287 | #if IS_ENABLED(CONFIG_DEBUG_FS) |
Taehee Yoo | 4c2d5e3 | 2019-12-22 11:26:39 +0000 | [diff] [blame] | 288 | void hsr_debugfs_rename(struct net_device *dev); |
Taehee Yoo | 1d19e2d | 2019-12-22 11:26:15 +0000 | [diff] [blame] | 289 | void hsr_debugfs_init(struct hsr_priv *priv, struct net_device *hsr_dev); |
Murali Karicheri | 9c5f8a19 | 2019-04-15 11:36:01 -0400 | [diff] [blame] | 290 | void hsr_debugfs_term(struct hsr_priv *priv); |
Taehee Yoo | c6c4ccd | 2019-12-22 11:26:27 +0000 | [diff] [blame] | 291 | void hsr_debugfs_create_root(void); |
| 292 | void hsr_debugfs_remove_root(void); |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 293 | #else |
xiaofeng.yan | 8089277 | 2020-01-20 14:26:39 +0800 | [diff] [blame] | 294 | static inline void hsr_debugfs_rename(struct net_device *dev) |
Taehee Yoo | 4c2d5e3 | 2019-12-22 11:26:39 +0000 | [diff] [blame] | 295 | { |
| 296 | } |
Taehee Yoo | 1d19e2d | 2019-12-22 11:26:15 +0000 | [diff] [blame] | 297 | static inline void hsr_debugfs_init(struct hsr_priv *priv, |
| 298 | struct net_device *hsr_dev) |
| 299 | {} |
Murali Karicheri | 9c5f8a19 | 2019-04-15 11:36:01 -0400 | [diff] [blame] | 300 | static inline void hsr_debugfs_term(struct hsr_priv *priv) |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 301 | {} |
Taehee Yoo | c6c4ccd | 2019-12-22 11:26:27 +0000 | [diff] [blame] | 302 | static inline void hsr_debugfs_create_root(void) |
| 303 | {} |
| 304 | static inline void hsr_debugfs_remove_root(void) |
| 305 | {} |
Murali Karicheri | fc4ecae | 2019-04-05 13:31:35 -0400 | [diff] [blame] | 306 | #endif |
| 307 | |
Arvid Brodin | 70ebe4a | 2014-07-04 23:34:38 +0200 | [diff] [blame] | 308 | #endif /* __HSR_PRIVATE_H */ |