Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    3)  * IPv4 over IEEE 1394, per RFC 2734
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    4)  * IPv6 over IEEE 1394, per RFC 3146
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    5)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    6)  * Copyright (C) 2009 Jay Fenlason <fenlason@redhat.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    7)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    8)  * based on eth1394 by Ben Collins et al
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    9)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   10) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   11) #include <linux/bug.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   12) #include <linux/compiler.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   13) #include <linux/delay.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   14) #include <linux/device.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   15) #include <linux/ethtool.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   16) #include <linux/firewire.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   17) #include <linux/firewire-constants.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   18) #include <linux/highmem.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   19) #include <linux/in.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   20) #include <linux/ip.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   21) #include <linux/jiffies.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   22) #include <linux/mod_devicetable.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   23) #include <linux/module.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   24) #include <linux/moduleparam.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   25) #include <linux/mutex.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   26) #include <linux/netdevice.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   27) #include <linux/skbuff.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   28) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   29) #include <linux/spinlock.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   30) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   31) #include <asm/unaligned.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   32) #include <net/arp.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   33) #include <net/firewire.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   34) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   35) /* rx limits */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   36) #define FWNET_MAX_FRAGMENTS		30 /* arbitrary, > TX queue depth */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   37) #define FWNET_ISO_PAGE_COUNT		(PAGE_SIZE < 16*1024 ? 4 : 2)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   38) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   39) /* tx limits */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   40) #define FWNET_MAX_QUEUED_DATAGRAMS	20 /* < 64 = number of tlabels */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   41) #define FWNET_MIN_QUEUED_DATAGRAMS	10 /* should keep AT DMA busy enough */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   42) #define FWNET_TX_QUEUE_LEN		FWNET_MAX_QUEUED_DATAGRAMS /* ? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   43) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   44) #define IEEE1394_BROADCAST_CHANNEL	31
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   45) #define IEEE1394_ALL_NODES		(0xffc0 | 0x003f)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   46) #define IEEE1394_MAX_PAYLOAD_S100	512
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   47) #define FWNET_NO_FIFO_ADDR		(~0ULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   48) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   49) #define IANA_SPECIFIER_ID		0x00005eU
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   50) #define RFC2734_SW_VERSION		0x000001U
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   51) #define RFC3146_SW_VERSION		0x000002U
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   52) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   53) #define IEEE1394_GASP_HDR_SIZE	8
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   54) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   55) #define RFC2374_UNFRAG_HDR_SIZE	4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   56) #define RFC2374_FRAG_HDR_SIZE	8
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   57) #define RFC2374_FRAG_OVERHEAD	4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   58) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   59) #define RFC2374_HDR_UNFRAG	0	/* unfragmented		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   60) #define RFC2374_HDR_FIRSTFRAG	1	/* first fragment	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   61) #define RFC2374_HDR_LASTFRAG	2	/* last fragment	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   62) #define RFC2374_HDR_INTFRAG	3	/* interior fragment	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   63) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   64) static bool fwnet_hwaddr_is_multicast(u8 *ha)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   65) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   66) 	return !!(*ha & 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   67) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   68) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   69) /* IPv4 and IPv6 encapsulation header */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   70) struct rfc2734_header {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   71) 	u32 w0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   72) 	u32 w1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   73) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   74) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   75) #define fwnet_get_hdr_lf(h)		(((h)->w0 & 0xc0000000) >> 30)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   76) #define fwnet_get_hdr_ether_type(h)	(((h)->w0 & 0x0000ffff))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   77) #define fwnet_get_hdr_dg_size(h)	((((h)->w0 & 0x0fff0000) >> 16) + 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   78) #define fwnet_get_hdr_fg_off(h)		(((h)->w0 & 0x00000fff))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   79) #define fwnet_get_hdr_dgl(h)		(((h)->w1 & 0xffff0000) >> 16)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   80) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   81) #define fwnet_set_hdr_lf(lf)		((lf) << 30)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   82) #define fwnet_set_hdr_ether_type(et)	(et)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   83) #define fwnet_set_hdr_dg_size(dgs)	(((dgs) - 1) << 16)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   84) #define fwnet_set_hdr_fg_off(fgo)	(fgo)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   85) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   86) #define fwnet_set_hdr_dgl(dgl)		((dgl) << 16)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   87) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   88) static inline void fwnet_make_uf_hdr(struct rfc2734_header *hdr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   89) 		unsigned ether_type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   90) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   91) 	hdr->w0 = fwnet_set_hdr_lf(RFC2374_HDR_UNFRAG)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   92) 		  | fwnet_set_hdr_ether_type(ether_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   93) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   94) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   95) static inline void fwnet_make_ff_hdr(struct rfc2734_header *hdr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   96) 		unsigned ether_type, unsigned dg_size, unsigned dgl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   97) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   98) 	hdr->w0 = fwnet_set_hdr_lf(RFC2374_HDR_FIRSTFRAG)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   99) 		  | fwnet_set_hdr_dg_size(dg_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  100) 		  | fwnet_set_hdr_ether_type(ether_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  101) 	hdr->w1 = fwnet_set_hdr_dgl(dgl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  103) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  104) static inline void fwnet_make_sf_hdr(struct rfc2734_header *hdr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  105) 		unsigned lf, unsigned dg_size, unsigned fg_off, unsigned dgl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  106) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  107) 	hdr->w0 = fwnet_set_hdr_lf(lf)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  108) 		  | fwnet_set_hdr_dg_size(dg_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  109) 		  | fwnet_set_hdr_fg_off(fg_off);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  110) 	hdr->w1 = fwnet_set_hdr_dgl(dgl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  111) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  112) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  113) /* This list keeps track of what parts of the datagram have been filled in */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  114) struct fwnet_fragment_info {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  115) 	struct list_head fi_link;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  116) 	u16 offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  117) 	u16 len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  118) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  119) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  120) struct fwnet_partial_datagram {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  121) 	struct list_head pd_link;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  122) 	struct list_head fi_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  123) 	struct sk_buff *skb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  124) 	/* FIXME Why not use skb->data? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  125) 	char *pbuf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  126) 	u16 datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  127) 	u16 ether_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  128) 	u16 datagram_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  129) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  130) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  131) static DEFINE_MUTEX(fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  132) static LIST_HEAD(fwnet_device_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  133) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  134) struct fwnet_device {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  135) 	struct list_head dev_link;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  136) 	spinlock_t lock;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  137) 	enum {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  138) 		FWNET_BROADCAST_ERROR,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  139) 		FWNET_BROADCAST_RUNNING,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  140) 		FWNET_BROADCAST_STOPPED,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  141) 	} broadcast_state;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  142) 	struct fw_iso_context *broadcast_rcv_context;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  143) 	struct fw_iso_buffer broadcast_rcv_buffer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  144) 	void **broadcast_rcv_buffer_ptrs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  145) 	unsigned broadcast_rcv_next_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  146) 	unsigned num_broadcast_rcv_ptrs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  147) 	unsigned rcv_buffer_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  148) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  149) 	 * This value is the maximum unfragmented datagram size that can be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  150) 	 * sent by the hardware.  It already has the GASP overhead and the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  151) 	 * unfragmented datagram header overhead calculated into it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  152) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  153) 	unsigned broadcast_xmt_max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  154) 	u16 broadcast_xmt_datagramlabel;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  155) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  156) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  157) 	 * The CSR address that remote nodes must send datagrams to for us to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  158) 	 * receive them.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  159) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  160) 	struct fw_address_handler handler;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  161) 	u64 local_fifo;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  162) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  163) 	/* Number of tx datagrams that have been queued but not yet acked */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  164) 	int queued_datagrams;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  165) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  166) 	int peer_count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  167) 	struct list_head peer_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  168) 	struct fw_card *card;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  169) 	struct net_device *netdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  170) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  171) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  172) struct fwnet_peer {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  173) 	struct list_head peer_link;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  174) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  175) 	u64 guid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  176) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  177) 	/* guarded by dev->lock */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  178) 	struct list_head pd_list; /* received partial datagrams */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  179) 	unsigned pdg_size;        /* pd_list size */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  180) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  181) 	u16 datagram_label;       /* outgoing datagram label */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  182) 	u16 max_payload;          /* includes RFC2374_FRAG_HDR_SIZE overhead */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  183) 	int node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  184) 	int generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  185) 	unsigned speed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  186) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  187) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  188) /* This is our task struct. It's used for the packet complete callback.  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  189) struct fwnet_packet_task {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  190) 	struct fw_transaction transaction;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  191) 	struct rfc2734_header hdr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  192) 	struct sk_buff *skb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  193) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  194) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  195) 	int outstanding_pkts;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  196) 	u64 fifo_addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  197) 	u16 dest_node;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  198) 	u16 max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  199) 	u8 generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  200) 	u8 speed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  201) 	u8 enqueued;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  202) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  203) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  204) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  205)  * Get fifo address embedded in hwaddr
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  206)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  207) static __u64 fwnet_hwaddr_fifo(union fwnet_hwaddr *ha)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  208) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  209) 	return (u64)get_unaligned_be16(&ha->uc.fifo_hi) << 32
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  210) 	       | get_unaligned_be32(&ha->uc.fifo_lo);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  211) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  212) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  213) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  214)  * saddr == NULL means use device source address.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  215)  * daddr == NULL means leave destination address (eg unresolved arp).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  216)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  217) static int fwnet_header_create(struct sk_buff *skb, struct net_device *net,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  218) 			unsigned short type, const void *daddr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  219) 			const void *saddr, unsigned len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  220) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  221) 	struct fwnet_header *h;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  222) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  223) 	h = skb_push(skb, sizeof(*h));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  224) 	put_unaligned_be16(type, &h->h_proto);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  225) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  226) 	if (net->flags & (IFF_LOOPBACK | IFF_NOARP)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  227) 		memset(h->h_dest, 0, net->addr_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  228) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  229) 		return net->hard_header_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  230) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  231) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  232) 	if (daddr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  233) 		memcpy(h->h_dest, daddr, net->addr_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  234) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  235) 		return net->hard_header_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  236) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  237) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  238) 	return -net->hard_header_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  239) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  240) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  241) static int fwnet_header_cache(const struct neighbour *neigh,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  242) 			      struct hh_cache *hh, __be16 type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  243) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  244) 	struct net_device *net;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  245) 	struct fwnet_header *h;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  246) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  247) 	if (type == cpu_to_be16(ETH_P_802_3))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  248) 		return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  249) 	net = neigh->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  250) 	h = (struct fwnet_header *)((u8 *)hh->hh_data + HH_DATA_OFF(sizeof(*h)));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  251) 	h->h_proto = type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  252) 	memcpy(h->h_dest, neigh->ha, net->addr_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  253) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  254) 	/* Pairs with the READ_ONCE() in neigh_resolve_output(),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  255) 	 * neigh_hh_output() and neigh_update_hhs().
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  256) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  257) 	smp_store_release(&hh->hh_len, FWNET_HLEN);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  258) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  259) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  260) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  261) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  262) /* Called by Address Resolution module to notify changes in address. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  263) static void fwnet_header_cache_update(struct hh_cache *hh,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  264) 		const struct net_device *net, const unsigned char *haddr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  265) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  266) 	memcpy((u8 *)hh->hh_data + HH_DATA_OFF(FWNET_HLEN), haddr, net->addr_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  267) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  268) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  269) static int fwnet_header_parse(const struct sk_buff *skb, unsigned char *haddr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  270) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  271) 	memcpy(haddr, skb->dev->dev_addr, FWNET_ALEN);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  272) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  273) 	return FWNET_ALEN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  274) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  275) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  276) static const struct header_ops fwnet_header_ops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  277) 	.create         = fwnet_header_create,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  278) 	.cache		= fwnet_header_cache,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  279) 	.cache_update	= fwnet_header_cache_update,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  280) 	.parse          = fwnet_header_parse,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  281) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  282) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  283) /* FIXME: is this correct for all cases? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  284) static bool fwnet_frag_overlap(struct fwnet_partial_datagram *pd,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  285) 			       unsigned offset, unsigned len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  286) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  287) 	struct fwnet_fragment_info *fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  288) 	unsigned end = offset + len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  289) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  290) 	list_for_each_entry(fi, &pd->fi_list, fi_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  291) 		if (offset < fi->offset + fi->len && end > fi->offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  292) 			return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  294) 	return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  295) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  296) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  297) /* Assumes that new fragment does not overlap any existing fragments */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  298) static struct fwnet_fragment_info *fwnet_frag_new(
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  299) 	struct fwnet_partial_datagram *pd, unsigned offset, unsigned len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  300) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  301) 	struct fwnet_fragment_info *fi, *fi2, *new;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  302) 	struct list_head *list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  303) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  304) 	list = &pd->fi_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  305) 	list_for_each_entry(fi, &pd->fi_list, fi_link) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  306) 		if (fi->offset + fi->len == offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  307) 			/* The new fragment can be tacked on to the end */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  308) 			/* Did the new fragment plug a hole? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  309) 			fi2 = list_entry(fi->fi_link.next,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  310) 					 struct fwnet_fragment_info, fi_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  311) 			if (fi->offset + fi->len == fi2->offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  312) 				/* glue fragments together */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  313) 				fi->len += len + fi2->len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  314) 				list_del(&fi2->fi_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  315) 				kfree(fi2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  316) 			} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  317) 				fi->len += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  318) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  319) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  320) 			return fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  321) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  322) 		if (offset + len == fi->offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  323) 			/* The new fragment can be tacked on to the beginning */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  324) 			/* Did the new fragment plug a hole? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  325) 			fi2 = list_entry(fi->fi_link.prev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  326) 					 struct fwnet_fragment_info, fi_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  327) 			if (fi2->offset + fi2->len == fi->offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  328) 				/* glue fragments together */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  329) 				fi2->len += fi->len + len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  330) 				list_del(&fi->fi_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  331) 				kfree(fi);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  332) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  333) 				return fi2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  334) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  335) 			fi->offset = offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  336) 			fi->len += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  337) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  338) 			return fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  339) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  340) 		if (offset > fi->offset + fi->len) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  341) 			list = &fi->fi_link;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  342) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  343) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  344) 		if (offset + len < fi->offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  345) 			list = fi->fi_link.prev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  346) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  347) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  348) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  349) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  350) 	new = kmalloc(sizeof(*new), GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  351) 	if (!new)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  352) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  353) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  354) 	new->offset = offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  355) 	new->len = len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  356) 	list_add(&new->fi_link, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  358) 	return new;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  359) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  360) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  361) static struct fwnet_partial_datagram *fwnet_pd_new(struct net_device *net,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  362) 		struct fwnet_peer *peer, u16 datagram_label, unsigned dg_size,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  363) 		void *frag_buf, unsigned frag_off, unsigned frag_len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  364) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  365) 	struct fwnet_partial_datagram *new;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  366) 	struct fwnet_fragment_info *fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  367) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  368) 	new = kmalloc(sizeof(*new), GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  369) 	if (!new)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  370) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  371) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  372) 	INIT_LIST_HEAD(&new->fi_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  373) 	fi = fwnet_frag_new(new, frag_off, frag_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  374) 	if (fi == NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  375) 		goto fail_w_new;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  376) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  377) 	new->datagram_label = datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  378) 	new->datagram_size = dg_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  379) 	new->skb = dev_alloc_skb(dg_size + LL_RESERVED_SPACE(net));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  380) 	if (new->skb == NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  381) 		goto fail_w_fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  382) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  383) 	skb_reserve(new->skb, LL_RESERVED_SPACE(net));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  384) 	new->pbuf = skb_put(new->skb, dg_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  385) 	memcpy(new->pbuf + frag_off, frag_buf, frag_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  386) 	list_add_tail(&new->pd_link, &peer->pd_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  387) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  388) 	return new;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  389) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  390) fail_w_fi:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  391) 	kfree(fi);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  392) fail_w_new:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  393) 	kfree(new);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  394) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  395) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  396) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  397) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  398) static struct fwnet_partial_datagram *fwnet_pd_find(struct fwnet_peer *peer,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  399) 						    u16 datagram_label)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  400) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  401) 	struct fwnet_partial_datagram *pd;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  402) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  403) 	list_for_each_entry(pd, &peer->pd_list, pd_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  404) 		if (pd->datagram_label == datagram_label)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  405) 			return pd;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  406) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  407) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  408) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  409) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  410) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  411) static void fwnet_pd_delete(struct fwnet_partial_datagram *old)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  412) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  413) 	struct fwnet_fragment_info *fi, *n;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  414) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  415) 	list_for_each_entry_safe(fi, n, &old->fi_list, fi_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  416) 		kfree(fi);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  417) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  418) 	list_del(&old->pd_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  419) 	dev_kfree_skb_any(old->skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  420) 	kfree(old);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  421) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  422) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  423) static bool fwnet_pd_update(struct fwnet_peer *peer,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  424) 		struct fwnet_partial_datagram *pd, void *frag_buf,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  425) 		unsigned frag_off, unsigned frag_len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  426) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  427) 	if (fwnet_frag_new(pd, frag_off, frag_len) == NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  428) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  429) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  430) 	memcpy(pd->pbuf + frag_off, frag_buf, frag_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  431) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  432) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  433) 	 * Move list entry to beginning of list so that oldest partial
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  434) 	 * datagrams percolate to the end of the list
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  435) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  436) 	list_move_tail(&pd->pd_link, &peer->pd_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  437) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  438) 	return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  439) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  440) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  441) static bool fwnet_pd_is_complete(struct fwnet_partial_datagram *pd)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  442) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  443) 	struct fwnet_fragment_info *fi;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  444) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  445) 	fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  446) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  447) 	return fi->len == pd->datagram_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  448) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  449) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  450) /* caller must hold dev->lock */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  451) static struct fwnet_peer *fwnet_peer_find_by_guid(struct fwnet_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  452) 						  u64 guid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  453) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  454) 	struct fwnet_peer *peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  455) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  456) 	list_for_each_entry(peer, &dev->peer_list, peer_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  457) 		if (peer->guid == guid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  458) 			return peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  459) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  460) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  461) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  462) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  463) /* caller must hold dev->lock */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  464) static struct fwnet_peer *fwnet_peer_find_by_node_id(struct fwnet_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  465) 						int node_id, int generation)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  466) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  467) 	struct fwnet_peer *peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  468) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  469) 	list_for_each_entry(peer, &dev->peer_list, peer_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  470) 		if (peer->node_id    == node_id &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  471) 		    peer->generation == generation)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  472) 			return peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  473) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  474) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  475) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  476) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  477) /* See IEEE 1394-2008 table 6-4, table 8-8, table 16-18. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  478) static unsigned fwnet_max_payload(unsigned max_rec, unsigned speed)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  479) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  480) 	max_rec = min(max_rec, speed + 8);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  481) 	max_rec = clamp(max_rec, 8U, 11U); /* 512...4096 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  482) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  483) 	return (1 << (max_rec + 1)) - RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  484) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  485) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  486) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  487) static int fwnet_finish_incoming_packet(struct net_device *net,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  488) 					struct sk_buff *skb, u16 source_node_id,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  489) 					bool is_broadcast, u16 ether_type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  490) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  491) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  492) 	int status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  493) 	__be64 guid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  494) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  495) 	switch (ether_type) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  496) 	case ETH_P_ARP:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  497) 	case ETH_P_IP:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  498) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  499) 	case ETH_P_IPV6:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  500) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  501) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  502) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  503) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  504) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  505) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  506) 	dev = netdev_priv(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  507) 	/* Write metadata, and then pass to the receive level */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  508) 	skb->dev = net;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  509) 	skb->ip_summed = CHECKSUM_NONE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  510) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  511) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  512) 	 * Parse the encapsulation header. This actually does the job of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  513) 	 * converting to an ethernet-like pseudo frame header.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  514) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  515) 	guid = cpu_to_be64(dev->card->guid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  516) 	if (dev_hard_header(skb, net, ether_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  517) 			   is_broadcast ? net->broadcast : net->dev_addr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  518) 			   NULL, skb->len) >= 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  519) 		struct fwnet_header *eth;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  520) 		u16 *rawp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  521) 		__be16 protocol;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  522) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  523) 		skb_reset_mac_header(skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  524) 		skb_pull(skb, sizeof(*eth));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  525) 		eth = (struct fwnet_header *)skb_mac_header(skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  526) 		if (fwnet_hwaddr_is_multicast(eth->h_dest)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  527) 			if (memcmp(eth->h_dest, net->broadcast,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  528) 				   net->addr_len) == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  529) 				skb->pkt_type = PACKET_BROADCAST;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  530) #if 0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  531) 			else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  532) 				skb->pkt_type = PACKET_MULTICAST;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  533) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  534) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  535) 			if (memcmp(eth->h_dest, net->dev_addr, net->addr_len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  536) 				skb->pkt_type = PACKET_OTHERHOST;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  537) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  538) 		if (ntohs(eth->h_proto) >= ETH_P_802_3_MIN) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  539) 			protocol = eth->h_proto;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  540) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  541) 			rawp = (u16 *)skb->data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  542) 			if (*rawp == 0xffff)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  543) 				protocol = htons(ETH_P_802_3);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  544) 			else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  545) 				protocol = htons(ETH_P_802_2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  546) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  547) 		skb->protocol = protocol;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  548) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  549) 	status = netif_rx(skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  550) 	if (status == NET_RX_DROP) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  551) 		net->stats.rx_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  552) 		net->stats.rx_dropped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  553) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  554) 		net->stats.rx_packets++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  555) 		net->stats.rx_bytes += skb->len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  556) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  557) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  558) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  559) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  560)  err:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  561) 	net->stats.rx_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  562) 	net->stats.rx_dropped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  563) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  564) 	dev_kfree_skb_any(skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  565) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  566) 	return -ENOENT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  567) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  568) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  569) static int fwnet_incoming_packet(struct fwnet_device *dev, __be32 *buf, int len,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  570) 				 int source_node_id, int generation,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  571) 				 bool is_broadcast)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  572) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  573) 	struct sk_buff *skb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  574) 	struct net_device *net = dev->netdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  575) 	struct rfc2734_header hdr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  576) 	unsigned lf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  577) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  578) 	struct fwnet_peer *peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  579) 	struct fwnet_partial_datagram *pd;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  580) 	int fg_off;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  581) 	int dg_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  582) 	u16 datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  583) 	int retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  584) 	u16 ether_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  585) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  586) 	if (len <= RFC2374_UNFRAG_HDR_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  587) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  588) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  589) 	hdr.w0 = be32_to_cpu(buf[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  590) 	lf = fwnet_get_hdr_lf(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  591) 	if (lf == RFC2374_HDR_UNFRAG) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  592) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  593) 		 * An unfragmented datagram has been received by the ieee1394
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  594) 		 * bus. Build an skbuff around it so we can pass it to the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  595) 		 * high level network layer.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  596) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  597) 		ether_type = fwnet_get_hdr_ether_type(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  598) 		buf++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  599) 		len -= RFC2374_UNFRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  600) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  601) 		skb = dev_alloc_skb(len + LL_RESERVED_SPACE(net));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  602) 		if (unlikely(!skb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  603) 			net->stats.rx_dropped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  604) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  605) 			return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  606) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  607) 		skb_reserve(skb, LL_RESERVED_SPACE(net));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  608) 		skb_put_data(skb, buf, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  609) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  610) 		return fwnet_finish_incoming_packet(net, skb, source_node_id,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  611) 						    is_broadcast, ether_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  612) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  613) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  614) 	/* A datagram fragment has been received, now the fun begins. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  615) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  616) 	if (len <= RFC2374_FRAG_HDR_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  617) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  618) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  619) 	hdr.w1 = ntohl(buf[1]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  620) 	buf += 2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  621) 	len -= RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  622) 	if (lf == RFC2374_HDR_FIRSTFRAG) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  623) 		ether_type = fwnet_get_hdr_ether_type(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  624) 		fg_off = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  625) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  626) 		ether_type = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  627) 		fg_off = fwnet_get_hdr_fg_off(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  628) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  629) 	datagram_label = fwnet_get_hdr_dgl(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  630) 	dg_size = fwnet_get_hdr_dg_size(&hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  631) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  632) 	if (fg_off + len > dg_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  633) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  634) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  635) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  636) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  637) 	peer = fwnet_peer_find_by_node_id(dev, source_node_id, generation);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  638) 	if (!peer) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  639) 		retval = -ENOENT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  640) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  641) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  642) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  643) 	pd = fwnet_pd_find(peer, datagram_label);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  644) 	if (pd == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  645) 		while (peer->pdg_size >= FWNET_MAX_FRAGMENTS) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  646) 			/* remove the oldest */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  647) 			fwnet_pd_delete(list_first_entry(&peer->pd_list,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  648) 				struct fwnet_partial_datagram, pd_link));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  649) 			peer->pdg_size--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  650) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  651) 		pd = fwnet_pd_new(net, peer, datagram_label,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  652) 				  dg_size, buf, fg_off, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  653) 		if (pd == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  654) 			retval = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  655) 			goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  656) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  657) 		peer->pdg_size++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  658) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  659) 		if (fwnet_frag_overlap(pd, fg_off, len) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  660) 		    pd->datagram_size != dg_size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  661) 			/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  662) 			 * Differing datagram sizes or overlapping fragments,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  663) 			 * discard old datagram and start a new one.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  664) 			 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  665) 			fwnet_pd_delete(pd);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  666) 			pd = fwnet_pd_new(net, peer, datagram_label,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  667) 					  dg_size, buf, fg_off, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  668) 			if (pd == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  669) 				peer->pdg_size--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  670) 				retval = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  671) 				goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  672) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  673) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  674) 			if (!fwnet_pd_update(peer, pd, buf, fg_off, len)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  675) 				/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  676) 				 * Couldn't save off fragment anyway
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  677) 				 * so might as well obliterate the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  678) 				 * datagram now.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  679) 				 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  680) 				fwnet_pd_delete(pd);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  681) 				peer->pdg_size--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  682) 				retval = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  683) 				goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  684) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  685) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  686) 	} /* new datagram or add to existing one */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  687) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  688) 	if (lf == RFC2374_HDR_FIRSTFRAG)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  689) 		pd->ether_type = ether_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  690) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  691) 	if (fwnet_pd_is_complete(pd)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  692) 		ether_type = pd->ether_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  693) 		peer->pdg_size--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  694) 		skb = skb_get(pd->skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  695) 		fwnet_pd_delete(pd);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  696) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  697) 		spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  698) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  699) 		return fwnet_finish_incoming_packet(net, skb, source_node_id,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  700) 						    false, ether_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  701) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  702) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  703) 	 * Datagram is not complete, we're done for the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  704) 	 * moment.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  705) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  706) 	retval = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  707)  fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  708) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  709) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  710) 	return retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  711) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  712) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  713) static void fwnet_receive_packet(struct fw_card *card, struct fw_request *r,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  714) 		int tcode, int destination, int source, int generation,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  715) 		unsigned long long offset, void *payload, size_t length,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  716) 		void *callback_data)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  717) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  718) 	struct fwnet_device *dev = callback_data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  719) 	int rcode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  720) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  721) 	if (destination == IEEE1394_ALL_NODES) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  722) 		kfree(r);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  723) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  724) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  725) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  726) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  727) 	if (offset != dev->handler.offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  728) 		rcode = RCODE_ADDRESS_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  729) 	else if (tcode != TCODE_WRITE_BLOCK_REQUEST)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  730) 		rcode = RCODE_TYPE_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  731) 	else if (fwnet_incoming_packet(dev, payload, length,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  732) 				       source, generation, false) != 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  733) 		dev_err(&dev->netdev->dev, "incoming packet failure\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  734) 		rcode = RCODE_CONFLICT_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  735) 	} else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  736) 		rcode = RCODE_COMPLETE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  737) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  738) 	fw_send_response(card, r, rcode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  739) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  740) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  741) static int gasp_source_id(__be32 *p)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  742) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  743) 	return be32_to_cpu(p[0]) >> 16;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  744) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  745) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  746) static u32 gasp_specifier_id(__be32 *p)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  747) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  748) 	return (be32_to_cpu(p[0]) & 0xffff) << 8 |
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  749) 	       (be32_to_cpu(p[1]) & 0xff000000) >> 24;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  750) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  751) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  752) static u32 gasp_version(__be32 *p)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  753) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  754) 	return be32_to_cpu(p[1]) & 0xffffff;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  755) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  756) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  757) static void fwnet_receive_broadcast(struct fw_iso_context *context,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  758) 		u32 cycle, size_t header_length, void *header, void *data)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  759) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  760) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  761) 	struct fw_iso_packet packet;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  762) 	__be16 *hdr_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  763) 	__be32 *buf_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  764) 	int retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  765) 	u32 length;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  766) 	unsigned long offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  767) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  768) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  769) 	dev = data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  770) 	hdr_ptr = header;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  771) 	length = be16_to_cpup(hdr_ptr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  772) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  773) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  774) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  775) 	offset = dev->rcv_buffer_size * dev->broadcast_rcv_next_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  776) 	buf_ptr = dev->broadcast_rcv_buffer_ptrs[dev->broadcast_rcv_next_ptr++];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  777) 	if (dev->broadcast_rcv_next_ptr == dev->num_broadcast_rcv_ptrs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  778) 		dev->broadcast_rcv_next_ptr = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  779) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  780) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  781) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  782) 	if (length > IEEE1394_GASP_HDR_SIZE &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  783) 	    gasp_specifier_id(buf_ptr) == IANA_SPECIFIER_ID &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  784) 	    (gasp_version(buf_ptr) == RFC2734_SW_VERSION
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  785) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  786) 	     || gasp_version(buf_ptr) == RFC3146_SW_VERSION
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  787) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  788) 	    ))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  789) 		fwnet_incoming_packet(dev, buf_ptr + 2,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  790) 				      length - IEEE1394_GASP_HDR_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  791) 				      gasp_source_id(buf_ptr),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  792) 				      context->card->generation, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  793) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  794) 	packet.payload_length = dev->rcv_buffer_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  795) 	packet.interrupt = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  796) 	packet.skip = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  797) 	packet.tag = 3;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  798) 	packet.sy = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  799) 	packet.header_length = IEEE1394_GASP_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  800) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  801) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  802) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  803) 	retval = fw_iso_context_queue(dev->broadcast_rcv_context, &packet,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  804) 				      &dev->broadcast_rcv_buffer, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  805) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  806) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  807) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  808) 	if (retval >= 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  809) 		fw_iso_context_queue_flush(dev->broadcast_rcv_context);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  810) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  811) 		dev_err(&dev->netdev->dev, "requeue failed\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  812) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  813) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  814) static struct kmem_cache *fwnet_packet_task_cache;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  815) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  816) static void fwnet_free_ptask(struct fwnet_packet_task *ptask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  817) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  818) 	dev_kfree_skb_any(ptask->skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  819) 	kmem_cache_free(fwnet_packet_task_cache, ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  820) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  821) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  822) /* Caller must hold dev->lock. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  823) static void dec_queued_datagrams(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  824) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  825) 	if (--dev->queued_datagrams == FWNET_MIN_QUEUED_DATAGRAMS)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  826) 		netif_wake_queue(dev->netdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  827) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  828) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  829) static int fwnet_send_packet(struct fwnet_packet_task *ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  830) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  831) static void fwnet_transmit_packet_done(struct fwnet_packet_task *ptask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  832) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  833) 	struct fwnet_device *dev = ptask->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  834) 	struct sk_buff *skb = ptask->skb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  835) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  836) 	bool free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  837) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  838) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  839) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  840) 	ptask->outstanding_pkts--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  841) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  842) 	/* Check whether we or the networking TX soft-IRQ is last user. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  843) 	free = (ptask->outstanding_pkts == 0 && ptask->enqueued);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  844) 	if (free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  845) 		dec_queued_datagrams(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  846) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  847) 	if (ptask->outstanding_pkts == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  848) 		dev->netdev->stats.tx_packets++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  849) 		dev->netdev->stats.tx_bytes += skb->len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  850) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  851) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  852) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  853) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  854) 	if (ptask->outstanding_pkts > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  855) 		u16 dg_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  856) 		u16 fg_off;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  857) 		u16 datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  858) 		u16 lf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  859) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  860) 		/* Update the ptask to point to the next fragment and send it */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  861) 		lf = fwnet_get_hdr_lf(&ptask->hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  862) 		switch (lf) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  863) 		case RFC2374_HDR_LASTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  864) 		case RFC2374_HDR_UNFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  865) 		default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  866) 			dev_err(&dev->netdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  867) 				"outstanding packet %x lf %x, header %x,%x\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  868) 				ptask->outstanding_pkts, lf, ptask->hdr.w0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  869) 				ptask->hdr.w1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  870) 			BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  871) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  872) 		case RFC2374_HDR_FIRSTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  873) 			/* Set frag type here for future interior fragments */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  874) 			dg_size = fwnet_get_hdr_dg_size(&ptask->hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  875) 			fg_off = ptask->max_payload - RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  876) 			datagram_label = fwnet_get_hdr_dgl(&ptask->hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  877) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  878) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  879) 		case RFC2374_HDR_INTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  880) 			dg_size = fwnet_get_hdr_dg_size(&ptask->hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  881) 			fg_off = fwnet_get_hdr_fg_off(&ptask->hdr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  882) 				  + ptask->max_payload - RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  883) 			datagram_label = fwnet_get_hdr_dgl(&ptask->hdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  884) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  885) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  886) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  887) 		if (ptask->dest_node == IEEE1394_ALL_NODES) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  888) 			skb_pull(skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  889) 				 ptask->max_payload + IEEE1394_GASP_HDR_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  890) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  891) 			skb_pull(skb, ptask->max_payload);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  892) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  893) 		if (ptask->outstanding_pkts > 1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  894) 			fwnet_make_sf_hdr(&ptask->hdr, RFC2374_HDR_INTFRAG,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  895) 					  dg_size, fg_off, datagram_label);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  896) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  897) 			fwnet_make_sf_hdr(&ptask->hdr, RFC2374_HDR_LASTFRAG,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  898) 					  dg_size, fg_off, datagram_label);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  899) 			ptask->max_payload = skb->len + RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  900) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  901) 		fwnet_send_packet(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  902) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  903) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  904) 	if (free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  905) 		fwnet_free_ptask(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  906) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  907) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  908) static void fwnet_transmit_packet_failed(struct fwnet_packet_task *ptask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  909) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  910) 	struct fwnet_device *dev = ptask->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  911) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  912) 	bool free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  913) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  914) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  915) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  916) 	/* One fragment failed; don't try to send remaining fragments. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  917) 	ptask->outstanding_pkts = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  918) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  919) 	/* Check whether we or the networking TX soft-IRQ is last user. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  920) 	free = ptask->enqueued;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  921) 	if (free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  922) 		dec_queued_datagrams(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  923) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  924) 	dev->netdev->stats.tx_dropped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  925) 	dev->netdev->stats.tx_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  926) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  927) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  928) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  929) 	if (free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  930) 		fwnet_free_ptask(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  931) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  932) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  933) static void fwnet_write_complete(struct fw_card *card, int rcode,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  934) 				 void *payload, size_t length, void *data)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  935) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  936) 	struct fwnet_packet_task *ptask = data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  937) 	static unsigned long j;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  938) 	static int last_rcode, errors_skipped;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  939) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  940) 	if (rcode == RCODE_COMPLETE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  941) 		fwnet_transmit_packet_done(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  942) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  943) 		if (printk_timed_ratelimit(&j,  1000) || rcode != last_rcode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  944) 			dev_err(&ptask->dev->netdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  945) 				"fwnet_write_complete failed: %x (skipped %d)\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  946) 				rcode, errors_skipped);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  947) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  948) 			errors_skipped = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  949) 			last_rcode = rcode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  950) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  951) 			errors_skipped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  952) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  953) 		fwnet_transmit_packet_failed(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  954) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  955) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  956) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  957) static int fwnet_send_packet(struct fwnet_packet_task *ptask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  958) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  959) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  960) 	unsigned tx_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  961) 	struct rfc2734_header *bufhdr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  962) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  963) 	bool free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  964) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  965) 	dev = ptask->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  966) 	tx_len = ptask->max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  967) 	switch (fwnet_get_hdr_lf(&ptask->hdr)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  968) 	case RFC2374_HDR_UNFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  969) 		bufhdr = skb_push(ptask->skb, RFC2374_UNFRAG_HDR_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  970) 		put_unaligned_be32(ptask->hdr.w0, &bufhdr->w0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  971) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  972) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  973) 	case RFC2374_HDR_FIRSTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  974) 	case RFC2374_HDR_INTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  975) 	case RFC2374_HDR_LASTFRAG:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  976) 		bufhdr = skb_push(ptask->skb, RFC2374_FRAG_HDR_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  977) 		put_unaligned_be32(ptask->hdr.w0, &bufhdr->w0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  978) 		put_unaligned_be32(ptask->hdr.w1, &bufhdr->w1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  979) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  980) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  981) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  982) 		BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  983) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  984) 	if (ptask->dest_node == IEEE1394_ALL_NODES) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  985) 		u8 *p;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  986) 		int generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  987) 		int node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  988) 		unsigned int sw_version;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  989) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  990) 		/* ptask->generation may not have been set yet */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  991) 		generation = dev->card->generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  992) 		smp_rmb();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  993) 		node_id = dev->card->node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  994) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  995) 		switch (ptask->skb->protocol) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  996) 		default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  997) 			sw_version = RFC2734_SW_VERSION;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  998) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  999) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1000) 		case htons(ETH_P_IPV6):
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1001) 			sw_version = RFC3146_SW_VERSION;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1002) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1003) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1004) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1005) 		p = skb_push(ptask->skb, IEEE1394_GASP_HDR_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1006) 		put_unaligned_be32(node_id << 16 | IANA_SPECIFIER_ID >> 8, p);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1007) 		put_unaligned_be32((IANA_SPECIFIER_ID & 0xff) << 24
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1008) 						| sw_version, &p[4]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1009) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1010) 		/* We should not transmit if broadcast_channel.valid == 0. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1011) 		fw_send_request(dev->card, &ptask->transaction,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1012) 				TCODE_STREAM_DATA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1013) 				fw_stream_packet_destination_id(3,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1014) 						IEEE1394_BROADCAST_CHANNEL, 0),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1015) 				generation, SCODE_100, 0ULL, ptask->skb->data,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1016) 				tx_len + 8, fwnet_write_complete, ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1017) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1018) 		spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1019) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1020) 		/* If the AT tasklet already ran, we may be last user. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1021) 		free = (ptask->outstanding_pkts == 0 && !ptask->enqueued);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1022) 		if (!free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1023) 			ptask->enqueued = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1024) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1025) 			dec_queued_datagrams(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1026) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1027) 		spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1028) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1029) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1030) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1031) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1032) 	fw_send_request(dev->card, &ptask->transaction,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1033) 			TCODE_WRITE_BLOCK_REQUEST, ptask->dest_node,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1034) 			ptask->generation, ptask->speed, ptask->fifo_addr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1035) 			ptask->skb->data, tx_len, fwnet_write_complete, ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1036) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1037) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1038) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1039) 	/* If the AT tasklet already ran, we may be last user. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1040) 	free = (ptask->outstanding_pkts == 0 && !ptask->enqueued);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1041) 	if (!free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1042) 		ptask->enqueued = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1043) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1044) 		dec_queued_datagrams(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1045) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1046) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1047) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1048) 	netif_trans_update(dev->netdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1049)  out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1050) 	if (free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1051) 		fwnet_free_ptask(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1052) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1053) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1054) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1055) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1056) static void fwnet_fifo_stop(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1057) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1058) 	if (dev->local_fifo == FWNET_NO_FIFO_ADDR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1059) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1060) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1061) 	fw_core_remove_address_handler(&dev->handler);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1062) 	dev->local_fifo = FWNET_NO_FIFO_ADDR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1063) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1064) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1065) static int fwnet_fifo_start(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1066) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1067) 	int retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1068) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1069) 	if (dev->local_fifo != FWNET_NO_FIFO_ADDR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1070) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1071) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1072) 	dev->handler.length = 4096;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1073) 	dev->handler.address_callback = fwnet_receive_packet;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1074) 	dev->handler.callback_data = dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1075) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1076) 	retval = fw_core_add_address_handler(&dev->handler,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1077) 					     &fw_high_memory_region);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1078) 	if (retval < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1079) 		return retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1080) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1081) 	dev->local_fifo = dev->handler.offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1082) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1083) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1084) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1085) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1086) static void __fwnet_broadcast_stop(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1087) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1088) 	unsigned u;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1089) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1090) 	if (dev->broadcast_state != FWNET_BROADCAST_ERROR) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1091) 		for (u = 0; u < FWNET_ISO_PAGE_COUNT; u++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1092) 			kunmap(dev->broadcast_rcv_buffer.pages[u]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1093) 		fw_iso_buffer_destroy(&dev->broadcast_rcv_buffer, dev->card);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1094) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1095) 	if (dev->broadcast_rcv_context) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1096) 		fw_iso_context_destroy(dev->broadcast_rcv_context);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1097) 		dev->broadcast_rcv_context = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1098) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1099) 	kfree(dev->broadcast_rcv_buffer_ptrs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1100) 	dev->broadcast_rcv_buffer_ptrs = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1101) 	dev->broadcast_state = FWNET_BROADCAST_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1103) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1104) static void fwnet_broadcast_stop(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1105) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1106) 	if (dev->broadcast_state == FWNET_BROADCAST_ERROR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1107) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1108) 	fw_iso_context_stop(dev->broadcast_rcv_context);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1109) 	__fwnet_broadcast_stop(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1110) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1111) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1112) static int fwnet_broadcast_start(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1113) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1114) 	struct fw_iso_context *context;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1115) 	int retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1116) 	unsigned num_packets;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1117) 	unsigned max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1118) 	struct fw_iso_packet packet;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1119) 	unsigned long offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1120) 	void **ptrptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1121) 	unsigned u;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1122) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1123) 	if (dev->broadcast_state != FWNET_BROADCAST_ERROR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1124) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1125) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1126) 	max_receive = 1U << (dev->card->max_receive + 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1127) 	num_packets = (FWNET_ISO_PAGE_COUNT * PAGE_SIZE) / max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1128) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1129) 	ptrptr = kmalloc_array(num_packets, sizeof(void *), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1130) 	if (!ptrptr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1131) 		retval = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1132) 		goto failed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1133) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1134) 	dev->broadcast_rcv_buffer_ptrs = ptrptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1135) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1136) 	context = fw_iso_context_create(dev->card, FW_ISO_CONTEXT_RECEIVE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1137) 					IEEE1394_BROADCAST_CHANNEL,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1138) 					dev->card->link_speed, 8,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1139) 					fwnet_receive_broadcast, dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1140) 	if (IS_ERR(context)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1141) 		retval = PTR_ERR(context);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1142) 		goto failed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1143) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1144) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1145) 	retval = fw_iso_buffer_init(&dev->broadcast_rcv_buffer, dev->card,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1146) 				    FWNET_ISO_PAGE_COUNT, DMA_FROM_DEVICE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1147) 	if (retval < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1148) 		goto failed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1149) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1150) 	dev->broadcast_state = FWNET_BROADCAST_STOPPED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1151) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1152) 	for (u = 0; u < FWNET_ISO_PAGE_COUNT; u++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1153) 		void *ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1154) 		unsigned v;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1155) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1156) 		ptr = kmap(dev->broadcast_rcv_buffer.pages[u]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1157) 		for (v = 0; v < num_packets / FWNET_ISO_PAGE_COUNT; v++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1158) 			*ptrptr++ = (void *) ((char *)ptr + v * max_receive);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1159) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1160) 	dev->broadcast_rcv_context = context;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1161) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1162) 	packet.payload_length = max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1163) 	packet.interrupt = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1164) 	packet.skip = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1165) 	packet.tag = 3;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1166) 	packet.sy = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1167) 	packet.header_length = IEEE1394_GASP_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1168) 	offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1169) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1170) 	for (u = 0; u < num_packets; u++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1171) 		retval = fw_iso_context_queue(context, &packet,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1172) 				&dev->broadcast_rcv_buffer, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1173) 		if (retval < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1174) 			goto failed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1175) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1176) 		offset += max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1177) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1178) 	dev->num_broadcast_rcv_ptrs = num_packets;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1179) 	dev->rcv_buffer_size = max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1180) 	dev->broadcast_rcv_next_ptr = 0U;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1181) 	retval = fw_iso_context_start(context, -1, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1182) 			FW_ISO_CONTEXT_MATCH_ALL_TAGS); /* ??? sync */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1183) 	if (retval < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1184) 		goto failed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1185) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1186) 	/* FIXME: adjust it according to the min. speed of all known peers? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1187) 	dev->broadcast_xmt_max_payload = IEEE1394_MAX_PAYLOAD_S100
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1188) 			- IEEE1394_GASP_HDR_SIZE - RFC2374_UNFRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1189) 	dev->broadcast_state = FWNET_BROADCAST_RUNNING;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1190) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1191) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1192) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1193)  failed:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1194) 	__fwnet_broadcast_stop(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1195) 	return retval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1196) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1197) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1198) static void set_carrier_state(struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1199) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1200) 	if (dev->peer_count > 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1201) 		netif_carrier_on(dev->netdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1202) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1203) 		netif_carrier_off(dev->netdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1204) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1205) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1206) /* ifup */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1207) static int fwnet_open(struct net_device *net)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1208) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1209) 	struct fwnet_device *dev = netdev_priv(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1210) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1211) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1212) 	ret = fwnet_broadcast_start(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1213) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1214) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1215) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1216) 	netif_start_queue(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1217) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1218) 	spin_lock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1219) 	set_carrier_state(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1220) 	spin_unlock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1221) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1222) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1223) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1224) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1225) /* ifdown */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1226) static int fwnet_stop(struct net_device *net)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1227) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1228) 	struct fwnet_device *dev = netdev_priv(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1229) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1230) 	netif_stop_queue(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1231) 	fwnet_broadcast_stop(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1232) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1233) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1234) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1235) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1236) static netdev_tx_t fwnet_tx(struct sk_buff *skb, struct net_device *net)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1237) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1238) 	struct fwnet_header hdr_buf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1239) 	struct fwnet_device *dev = netdev_priv(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1240) 	__be16 proto;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1241) 	u16 dest_node;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1242) 	unsigned max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1243) 	u16 dg_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1244) 	u16 *datagram_label_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1245) 	struct fwnet_packet_task *ptask;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1246) 	struct fwnet_peer *peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1247) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1248) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1249) 	spin_lock_irqsave(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1250) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1251) 	/* Can this happen? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1252) 	if (netif_queue_stopped(dev->netdev)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1253) 		spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1254) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1255) 		return NETDEV_TX_BUSY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1256) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1257) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1258) 	ptask = kmem_cache_alloc(fwnet_packet_task_cache, GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1259) 	if (ptask == NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1260) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1261) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1262) 	skb = skb_share_check(skb, GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1263) 	if (!skb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1264) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1265) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1266) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1267) 	 * Make a copy of the driver-specific header.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1268) 	 * We might need to rebuild the header on tx failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1269) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1270) 	memcpy(&hdr_buf, skb->data, sizeof(hdr_buf));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1271) 	proto = hdr_buf.h_proto;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1272) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1273) 	switch (proto) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1274) 	case htons(ETH_P_ARP):
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1275) 	case htons(ETH_P_IP):
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1276) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1277) 	case htons(ETH_P_IPV6):
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1278) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1279) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1280) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1281) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1282) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1283) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1284) 	skb_pull(skb, sizeof(hdr_buf));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1285) 	dg_size = skb->len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1286) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1287) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1288) 	 * Set the transmission type for the packet.  ARP packets and IP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1289) 	 * broadcast packets are sent via GASP.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1290) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1291) 	if (fwnet_hwaddr_is_multicast(hdr_buf.h_dest)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1292) 		max_payload        = dev->broadcast_xmt_max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1293) 		datagram_label_ptr = &dev->broadcast_xmt_datagramlabel;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1294) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1295) 		ptask->fifo_addr   = FWNET_NO_FIFO_ADDR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1296) 		ptask->generation  = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1297) 		ptask->dest_node   = IEEE1394_ALL_NODES;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1298) 		ptask->speed       = SCODE_100;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1299) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1300) 		union fwnet_hwaddr *ha = (union fwnet_hwaddr *)hdr_buf.h_dest;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1301) 		__be64 guid = get_unaligned(&ha->uc.uniq_id);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1302) 		u8 generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1303) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1304) 		peer = fwnet_peer_find_by_guid(dev, be64_to_cpu(guid));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1305) 		if (!peer)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1306) 			goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1307) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1308) 		generation         = peer->generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1309) 		dest_node          = peer->node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1310) 		max_payload        = peer->max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1311) 		datagram_label_ptr = &peer->datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1312) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1313) 		ptask->fifo_addr   = fwnet_hwaddr_fifo(ha);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1314) 		ptask->generation  = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1315) 		ptask->dest_node   = dest_node;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1316) 		ptask->speed       = peer->speed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1317) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1318) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1319) 	ptask->hdr.w0 = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1320) 	ptask->hdr.w1 = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1321) 	ptask->skb = skb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1322) 	ptask->dev = dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1323) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1324) 	/* Does it all fit in one packet? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1325) 	if (dg_size <= max_payload) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1326) 		fwnet_make_uf_hdr(&ptask->hdr, ntohs(proto));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1327) 		ptask->outstanding_pkts = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1328) 		max_payload = dg_size + RFC2374_UNFRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1329) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1330) 		u16 datagram_label;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1331) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1332) 		max_payload -= RFC2374_FRAG_OVERHEAD;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1333) 		datagram_label = (*datagram_label_ptr)++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1334) 		fwnet_make_ff_hdr(&ptask->hdr, ntohs(proto), dg_size,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1335) 				  datagram_label);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1336) 		ptask->outstanding_pkts = DIV_ROUND_UP(dg_size, max_payload);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1337) 		max_payload += RFC2374_FRAG_HDR_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1338) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1339) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1340) 	if (++dev->queued_datagrams == FWNET_MAX_QUEUED_DATAGRAMS)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1341) 		netif_stop_queue(dev->netdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1342) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1343) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1344) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1345) 	ptask->max_payload = max_payload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1346) 	ptask->enqueued    = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1347) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1348) 	fwnet_send_packet(ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1349) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1350) 	return NETDEV_TX_OK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1351) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1352)  fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1353) 	spin_unlock_irqrestore(&dev->lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1354) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1355) 	if (ptask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1356) 		kmem_cache_free(fwnet_packet_task_cache, ptask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1358) 	if (skb != NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1359) 		dev_kfree_skb(skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1360) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1361) 	net->stats.tx_dropped++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1362) 	net->stats.tx_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1363) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1364) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1365) 	 * FIXME: According to a patch from 2003-02-26, "returning non-zero
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1366) 	 * causes serious problems" here, allegedly.  Before that patch,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1367) 	 * -ERRNO was returned which is not appropriate under Linux 2.6.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1368) 	 * Perhaps more needs to be done?  Stop the queue in serious
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1369) 	 * conditions and restart it elsewhere?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1370) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1371) 	return NETDEV_TX_OK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1372) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1373) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1374) static const struct ethtool_ops fwnet_ethtool_ops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1375) 	.get_link	= ethtool_op_get_link,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1376) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1377) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1378) static const struct net_device_ops fwnet_netdev_ops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1379) 	.ndo_open       = fwnet_open,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1380) 	.ndo_stop	= fwnet_stop,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1381) 	.ndo_start_xmit = fwnet_tx,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1382) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1383) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1384) static void fwnet_init_dev(struct net_device *net)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1385) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1386) 	net->header_ops		= &fwnet_header_ops;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1387) 	net->netdev_ops		= &fwnet_netdev_ops;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1388) 	net->watchdog_timeo	= 2 * HZ;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1389) 	net->flags		= IFF_BROADCAST | IFF_MULTICAST;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1390) 	net->features		= NETIF_F_HIGHDMA;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1391) 	net->addr_len		= FWNET_ALEN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1392) 	net->hard_header_len	= FWNET_HLEN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1393) 	net->type		= ARPHRD_IEEE1394;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1394) 	net->tx_queue_len	= FWNET_TX_QUEUE_LEN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1395) 	net->ethtool_ops	= &fwnet_ethtool_ops;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1396) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1397) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1398) /* caller must hold fwnet_device_mutex */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1399) static struct fwnet_device *fwnet_dev_find(struct fw_card *card)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1400) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1401) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1402) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1403) 	list_for_each_entry(dev, &fwnet_device_list, dev_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1404) 		if (dev->card == card)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1405) 			return dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1406) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1407) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1408) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1409) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1410) static int fwnet_add_peer(struct fwnet_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1411) 			  struct fw_unit *unit, struct fw_device *device)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1412) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1413) 	struct fwnet_peer *peer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1414) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1415) 	peer = kmalloc(sizeof(*peer), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1416) 	if (!peer)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1417) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1418) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1419) 	dev_set_drvdata(&unit->device, peer);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1420) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1421) 	peer->dev = dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1422) 	peer->guid = (u64)device->config_rom[3] << 32 | device->config_rom[4];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1423) 	INIT_LIST_HEAD(&peer->pd_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1424) 	peer->pdg_size = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1425) 	peer->datagram_label = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1426) 	peer->speed = device->max_speed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1427) 	peer->max_payload = fwnet_max_payload(device->max_rec, peer->speed);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1428) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1429) 	peer->generation = device->generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1430) 	smp_rmb();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1431) 	peer->node_id = device->node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1432) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1433) 	spin_lock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1434) 	list_add_tail(&peer->peer_link, &dev->peer_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1435) 	dev->peer_count++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1436) 	set_carrier_state(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1437) 	spin_unlock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1438) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1439) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1440) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1441) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1442) static int fwnet_probe(struct fw_unit *unit,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1443) 		       const struct ieee1394_device_id *id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1444) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1445) 	struct fw_device *device = fw_parent_device(unit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1446) 	struct fw_card *card = device->card;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1447) 	struct net_device *net;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1448) 	bool allocated_netdev = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1449) 	struct fwnet_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1450) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1451) 	union fwnet_hwaddr *ha;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1452) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1453) 	mutex_lock(&fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1454) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1455) 	dev = fwnet_dev_find(card);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1456) 	if (dev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1457) 		net = dev->netdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1458) 		goto have_dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1459) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1460) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1461) 	net = alloc_netdev(sizeof(*dev), "firewire%d", NET_NAME_UNKNOWN,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1462) 			   fwnet_init_dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1463) 	if (net == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1464) 		mutex_unlock(&fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1465) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1466) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1467) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1468) 	allocated_netdev = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1469) 	SET_NETDEV_DEV(net, card->device);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1470) 	dev = netdev_priv(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1471) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1472) 	spin_lock_init(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1473) 	dev->broadcast_state = FWNET_BROADCAST_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1474) 	dev->broadcast_rcv_context = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1475) 	dev->broadcast_xmt_max_payload = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1476) 	dev->broadcast_xmt_datagramlabel = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1477) 	dev->local_fifo = FWNET_NO_FIFO_ADDR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1478) 	dev->queued_datagrams = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1479) 	INIT_LIST_HEAD(&dev->peer_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1480) 	dev->card = card;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1481) 	dev->netdev = net;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1482) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1483) 	ret = fwnet_fifo_start(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1484) 	if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1485) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1486) 	dev->local_fifo = dev->handler.offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1487) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1488) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1489) 	 * default MTU: RFC 2734 cl. 4, RFC 3146 cl. 4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1490) 	 * maximum MTU: RFC 2734 cl. 4.2, fragment encapsulation header's
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1491) 	 *              maximum possible datagram_size + 1 = 0xfff + 1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1492) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1493) 	net->mtu = 1500U;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1494) 	net->min_mtu = ETH_MIN_MTU;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1495) 	net->max_mtu = 4096U;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1496) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1497) 	/* Set our hardware address while we're at it */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1498) 	ha = (union fwnet_hwaddr *)net->dev_addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1499) 	put_unaligned_be64(card->guid, &ha->uc.uniq_id);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1500) 	ha->uc.max_rec = dev->card->max_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1501) 	ha->uc.sspd = dev->card->link_speed;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1502) 	put_unaligned_be16(dev->local_fifo >> 32, &ha->uc.fifo_hi);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1503) 	put_unaligned_be32(dev->local_fifo & 0xffffffff, &ha->uc.fifo_lo);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1504) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1505) 	memset(net->broadcast, -1, net->addr_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1506) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1507) 	ret = register_netdev(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1508) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1509) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1510) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1511) 	list_add_tail(&dev->dev_link, &fwnet_device_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1512) 	dev_notice(&net->dev, "IP over IEEE 1394 on card %s\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1513) 		   dev_name(card->device));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1514)  have_dev:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1515) 	ret = fwnet_add_peer(dev, unit, device);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1516) 	if (ret && allocated_netdev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1517) 		unregister_netdev(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1518) 		list_del(&dev->dev_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1519)  out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1520) 		fwnet_fifo_stop(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1521) 		free_netdev(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1522) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1523) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1524) 	mutex_unlock(&fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1525) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1526) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1527) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1528) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1529) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1530)  * FIXME abort partially sent fragmented datagrams,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1531)  * discard partially received fragmented datagrams
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1532)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1533) static void fwnet_update(struct fw_unit *unit)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1534) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1535) 	struct fw_device *device = fw_parent_device(unit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1536) 	struct fwnet_peer *peer = dev_get_drvdata(&unit->device);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1537) 	int generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1538) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1539) 	generation = device->generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1540) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1541) 	spin_lock_irq(&peer->dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1542) 	peer->node_id    = device->node_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1543) 	peer->generation = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1544) 	spin_unlock_irq(&peer->dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1545) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1546) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1547) static void fwnet_remove_peer(struct fwnet_peer *peer, struct fwnet_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1548) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1549) 	struct fwnet_partial_datagram *pd, *pd_next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1550) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1551) 	spin_lock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1552) 	list_del(&peer->peer_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1553) 	dev->peer_count--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1554) 	set_carrier_state(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1555) 	spin_unlock_irq(&dev->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1556) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1557) 	list_for_each_entry_safe(pd, pd_next, &peer->pd_list, pd_link)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1558) 		fwnet_pd_delete(pd);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1559) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1560) 	kfree(peer);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1561) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1562) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1563) static void fwnet_remove(struct fw_unit *unit)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1564) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1565) 	struct fwnet_peer *peer = dev_get_drvdata(&unit->device);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1566) 	struct fwnet_device *dev = peer->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1567) 	struct net_device *net;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1568) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1569) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1570) 	mutex_lock(&fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1571) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1572) 	net = dev->netdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1573) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1574) 	fwnet_remove_peer(peer, dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1575) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1576) 	if (list_empty(&dev->peer_list)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1577) 		unregister_netdev(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1578) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1579) 		fwnet_fifo_stop(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1580) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1581) 		for (i = 0; dev->queued_datagrams && i < 5; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1582) 			ssleep(1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1583) 		WARN_ON(dev->queued_datagrams);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1584) 		list_del(&dev->dev_link);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1585) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1586) 		free_netdev(net);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1587) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1588) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1589) 	mutex_unlock(&fwnet_device_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1590) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1591) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1592) static const struct ieee1394_device_id fwnet_id_table[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1593) 	{
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1594) 		.match_flags  = IEEE1394_MATCH_SPECIFIER_ID |
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1595) 				IEEE1394_MATCH_VERSION,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1596) 		.specifier_id = IANA_SPECIFIER_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1597) 		.version      = RFC2734_SW_VERSION,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1598) 	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1599) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1600) 	{
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1601) 		.match_flags  = IEEE1394_MATCH_SPECIFIER_ID |
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1602) 				IEEE1394_MATCH_VERSION,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1603) 		.specifier_id = IANA_SPECIFIER_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1604) 		.version      = RFC3146_SW_VERSION,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1605) 	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1606) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1607) 	{ }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1608) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1609) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1610) static struct fw_driver fwnet_driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1611) 	.driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1612) 		.owner  = THIS_MODULE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1613) 		.name   = KBUILD_MODNAME,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1614) 		.bus    = &fw_bus_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1615) 	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1616) 	.probe    = fwnet_probe,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1617) 	.update   = fwnet_update,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1618) 	.remove   = fwnet_remove,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1619) 	.id_table = fwnet_id_table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1620) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1621) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1622) static const u32 rfc2374_unit_directory_data[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1623) 	0x00040000,	/* directory_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1624) 	0x1200005e,	/* unit_specifier_id: IANA	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1625) 	0x81000003,	/* textual descriptor offset	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1626) 	0x13000001,	/* unit_sw_version: RFC 2734	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1627) 	0x81000005,	/* textual descriptor offset	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1628) 	0x00030000,	/* descriptor_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1629) 	0x00000000,	/* text				*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1630) 	0x00000000,	/* minimal ASCII, en		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1631) 	0x49414e41,	/* I A N A			*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1632) 	0x00030000,	/* descriptor_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1633) 	0x00000000,	/* text				*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1634) 	0x00000000,	/* minimal ASCII, en		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1635) 	0x49507634,	/* I P v 4			*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1636) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1637) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1638) static struct fw_descriptor rfc2374_unit_directory = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1639) 	.length = ARRAY_SIZE(rfc2374_unit_directory_data),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1640) 	.key    = (CSR_DIRECTORY | CSR_UNIT) << 24,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1641) 	.data   = rfc2374_unit_directory_data
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1642) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1643) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1644) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1645) static const u32 rfc3146_unit_directory_data[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1646) 	0x00040000,	/* directory_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1647) 	0x1200005e,	/* unit_specifier_id: IANA	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1648) 	0x81000003,	/* textual descriptor offset	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1649) 	0x13000002,	/* unit_sw_version: RFC 3146	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1650) 	0x81000005,	/* textual descriptor offset	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1651) 	0x00030000,	/* descriptor_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1652) 	0x00000000,	/* text				*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1653) 	0x00000000,	/* minimal ASCII, en		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1654) 	0x49414e41,	/* I A N A			*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1655) 	0x00030000,	/* descriptor_length		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1656) 	0x00000000,	/* text				*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1657) 	0x00000000,	/* minimal ASCII, en		*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1658) 	0x49507636,	/* I P v 6			*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1659) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1660) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1661) static struct fw_descriptor rfc3146_unit_directory = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1662) 	.length = ARRAY_SIZE(rfc3146_unit_directory_data),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1663) 	.key    = (CSR_DIRECTORY | CSR_UNIT) << 24,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1664) 	.data   = rfc3146_unit_directory_data
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1665) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1666) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1667) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1668) static int __init fwnet_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1669) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1670) 	int err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1671) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1672) 	err = fw_core_add_descriptor(&rfc2374_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1673) 	if (err)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1674) 		return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1675) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1676) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1677) 	err = fw_core_add_descriptor(&rfc3146_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1678) 	if (err)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1679) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1680) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1681) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1682) 	fwnet_packet_task_cache = kmem_cache_create("packet_task",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1683) 			sizeof(struct fwnet_packet_task), 0, 0, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1684) 	if (!fwnet_packet_task_cache) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1685) 		err = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1686) 		goto out2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1687) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1688) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1689) 	err = driver_register(&fwnet_driver.driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1690) 	if (!err)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1691) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1692) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1693) 	kmem_cache_destroy(fwnet_packet_task_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1694) out2:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1695) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1696) 	fw_core_remove_descriptor(&rfc3146_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1697) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1698) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1699) 	fw_core_remove_descriptor(&rfc2374_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1700) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1701) 	return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1702) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1703) module_init(fwnet_init);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1704) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1705) static void __exit fwnet_cleanup(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1706) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1707) 	driver_unregister(&fwnet_driver.driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1708) 	kmem_cache_destroy(fwnet_packet_task_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1709) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1710) 	fw_core_remove_descriptor(&rfc3146_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1711) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1712) 	fw_core_remove_descriptor(&rfc2374_unit_directory);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1713) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1714) module_exit(fwnet_cleanup);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1715) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1716) MODULE_AUTHOR("Jay Fenlason <fenlason@redhat.com>");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1717) MODULE_DESCRIPTION("IP over IEEE1394 as per RFC 2734/3146");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1718) MODULE_LICENSE("GPL");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1719) MODULE_DEVICE_TABLE(ieee1394, fwnet_id_table);