mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 04:02:20 +00:00
net: Correct spelling in net/core
Correct spelling in net/core. As reported by codespell. Signed-off-by: Simon Horman <horms@kernel.org> Link: https://patch.msgid.link/20240822-net-spell-v1-13-3a98971ce2d2@kernel.org Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
parent
70d0bb45fa
commit
a8c924e987
@ -3707,7 +3707,7 @@ struct sk_buff *validate_xmit_skb_list(struct sk_buff *skb, struct net_device *d
|
||||
next = skb->next;
|
||||
skb_mark_not_on_list(skb);
|
||||
|
||||
/* in case skb wont be segmented, point to itself */
|
||||
/* in case skb won't be segmented, point to itself */
|
||||
skb->prev = skb;
|
||||
|
||||
skb = validate_xmit_skb(skb, dev, again);
|
||||
@ -11432,7 +11432,7 @@ void unregister_netdevice_many_notify(struct list_head *head,
|
||||
* @head: list of devices
|
||||
*
|
||||
* Note: As most callers use a stack allocated list_head,
|
||||
* we force a list_del() to make sure stack wont be corrupted later.
|
||||
* we force a list_del() to make sure stack won't be corrupted later.
|
||||
*/
|
||||
void unregister_netdevice_many(struct list_head *head)
|
||||
{
|
||||
@ -11490,7 +11490,7 @@ int __dev_change_net_namespace(struct net_device *dev, struct net *net,
|
||||
if (dev->features & NETIF_F_NETNS_LOCAL)
|
||||
goto out;
|
||||
|
||||
/* Ensure the device has been registrered */
|
||||
/* Ensure the device has been registered */
|
||||
if (dev->reg_state != NETREG_REGISTERED)
|
||||
goto out;
|
||||
|
||||
|
@ -262,7 +262,7 @@ static int __hw_addr_sync_multiple(struct netdev_hw_addr_list *to_list,
|
||||
}
|
||||
|
||||
/* This function only works where there is a strict 1-1 relationship
|
||||
* between source and destionation of they synch. If you ever need to
|
||||
* between source and destination of they synch. If you ever need to
|
||||
* sync addresses to more then 1 destination, you need to use
|
||||
* __hw_addr_sync_multiple().
|
||||
*/
|
||||
@ -299,8 +299,8 @@ void __hw_addr_unsync(struct netdev_hw_addr_list *to_list,
|
||||
EXPORT_SYMBOL(__hw_addr_unsync);
|
||||
|
||||
/**
|
||||
* __hw_addr_sync_dev - Synchonize device's multicast list
|
||||
* @list: address list to syncronize
|
||||
* __hw_addr_sync_dev - Synchronize device's multicast list
|
||||
* @list: address list to synchronize
|
||||
* @dev: device to sync
|
||||
* @sync: function to call if address should be added
|
||||
* @unsync: function to call if address should be removed
|
||||
|
@ -72,7 +72,7 @@ int fib_default_rule_add(struct fib_rules_ops *ops,
|
||||
r->suppress_prefixlen = -1;
|
||||
r->suppress_ifgroup = -1;
|
||||
|
||||
/* The lock is not required here, the list in unreacheable
|
||||
/* The lock is not required here, the list in unreachable
|
||||
* at the moment this function is called */
|
||||
list_add_tail(&r->list, &ops->rules_list);
|
||||
return 0;
|
||||
|
@ -374,7 +374,7 @@ static void gro_list_prepare(const struct list_head *head,
|
||||
skb_mac_header(skb),
|
||||
maclen);
|
||||
|
||||
/* in most common scenarions 'slow_gro' is 0
|
||||
/* in most common scenarios 'slow_gro' is 0
|
||||
* otherwise we are already on some slower paths
|
||||
* either skip all the infrequent tests altogether or
|
||||
* avoid trying too hard to skip each of them individually
|
||||
|
@ -162,7 +162,7 @@ static void poll_one_napi(struct napi_struct *napi)
|
||||
if (test_and_set_bit(NAPI_STATE_NPSVC, &napi->state))
|
||||
return;
|
||||
|
||||
/* We explicilty pass the polling call a budget of 0 to
|
||||
/* We explicitly pass the polling call a budget of 0 to
|
||||
* indicate that we are clearing the Tx path only.
|
||||
*/
|
||||
work = napi->poll(napi, 0);
|
||||
|
@ -69,7 +69,7 @@
|
||||
*
|
||||
* By design there should only be *one* "controlling" process. In practice
|
||||
* multiple write accesses gives unpredictable result. Understood by "write"
|
||||
* to /proc gives result code thats should be read be the "writer".
|
||||
* to /proc gives result code that should be read be the "writer".
|
||||
* For practical use this should be no problem.
|
||||
*
|
||||
* Note when adding devices to a specific CPU there good idea to also assign
|
||||
@ -2371,11 +2371,11 @@ static void get_ipsec_sa(struct pktgen_dev *pkt_dev, int flow)
|
||||
|
||||
if (pkt_dev->spi) {
|
||||
/* We need as quick as possible to find the right SA
|
||||
* Searching with minimum criteria to archieve this.
|
||||
* Searching with minimum criteria to achieve, this.
|
||||
*/
|
||||
x = xfrm_state_lookup_byspi(pn->net, htonl(pkt_dev->spi), AF_INET);
|
||||
} else {
|
||||
/* slow path: we dont already have xfrm_state */
|
||||
/* slow path: we don't already have xfrm_state */
|
||||
x = xfrm_stateonly_find(pn->net, DUMMY_MARK, 0,
|
||||
(xfrm_address_t *)&pkt_dev->cur_daddr,
|
||||
(xfrm_address_t *)&pkt_dev->cur_saddr,
|
||||
@ -3838,8 +3838,8 @@ static int pktgen_add_device(struct pktgen_thread *t, const char *ifname)
|
||||
pkt_dev->ipsmode = XFRM_MODE_TRANSPORT;
|
||||
pkt_dev->ipsproto = IPPROTO_ESP;
|
||||
|
||||
/* xfrm tunnel mode needs additional dst to extract outter
|
||||
* ip header protocol/ttl/id field, here creat a phony one.
|
||||
/* xfrm tunnel mode needs additional dst to extract outer
|
||||
* ip header protocol/ttl/id field, here create a phony one.
|
||||
* instead of looking for a valid rt, which definitely hurting
|
||||
* performance under such circumstance.
|
||||
*/
|
||||
|
@ -5163,7 +5163,7 @@ EXPORT_SYMBOL_GPL(skb_to_sgvec);
|
||||
* 3. sg_unmark_end
|
||||
* 4. skb_to_sgvec(payload2)
|
||||
*
|
||||
* When mapping mutilple payload conditionally, skb_to_sgvec_nomark
|
||||
* When mapping multiple payload conditionally, skb_to_sgvec_nomark
|
||||
* is more preferable.
|
||||
*/
|
||||
int skb_to_sgvec_nomark(struct sk_buff *skb, struct scatterlist *sg,
|
||||
@ -6021,7 +6021,7 @@ EXPORT_SYMBOL(skb_try_coalesce);
|
||||
* @skb: buffer to clean
|
||||
* @xnet: packet is crossing netns
|
||||
*
|
||||
* skb_scrub_packet can be used after encapsulating or decapsulting a packet
|
||||
* skb_scrub_packet can be used after encapsulating or decapsulating a packet
|
||||
* into/from a tunnel. Some information have to be cleared during these
|
||||
* operations.
|
||||
* skb_scrub_packet can also be used to clean a skb before injecting it in
|
||||
|
@ -2048,7 +2048,7 @@ static inline void sock_lock_init(struct sock *sk)
|
||||
|
||||
/*
|
||||
* Copy all fields from osk to nsk but nsk->sk_refcnt must not change yet,
|
||||
* even temporarly, because of RCU lookups. sk_node should also be left as is.
|
||||
* even temporarily, because of RCU lookups. sk_node should also be left as is.
|
||||
* We must not copy fields between sk_dontcopy_begin and sk_dontcopy_end
|
||||
*/
|
||||
static void sock_copy(struct sock *nsk, const struct sock *osk)
|
||||
@ -2538,7 +2538,7 @@ void skb_set_owner_w(struct sk_buff *skb, struct sock *sk)
|
||||
skb_set_hash_from_sk(skb, sk);
|
||||
/*
|
||||
* We used to take a refcount on sk, but following operation
|
||||
* is enough to guarantee sk_free() wont free this sock until
|
||||
* is enough to guarantee sk_free() won't free this sock until
|
||||
* all in-flight packets are completed
|
||||
*/
|
||||
refcount_add(skb->truesize, &sk->sk_wmem_alloc);
|
||||
@ -3697,7 +3697,7 @@ EXPORT_SYMBOL(sock_recv_errqueue);
|
||||
*
|
||||
* FIX: POSIX 1003.1g is very ambiguous here. It states that
|
||||
* asynchronous errors should be reported by getsockopt. We assume
|
||||
* this means if you specify SO_ERROR (otherwise whats the point of it).
|
||||
* this means if you specify SO_ERROR (otherwise what is the point of it).
|
||||
*/
|
||||
int sock_common_getsockopt(struct socket *sock, int level, int optname,
|
||||
char __user *optval, int __user *optlen)
|
||||
|
@ -1,6 +1,6 @@
|
||||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* Generic address resultion entity
|
||||
* Generic address resolution entity
|
||||
*
|
||||
* Authors:
|
||||
* net_random Alan Cox
|
||||
|
Loading…
Reference in New Issue
Block a user