VirtualBox

source: vbox/trunk/src/VBox/HostDrivers/VBoxNetFlt/linux/VBoxNetFlt-linux.c@ 21351

Last change on this file since 21351 was 21351, checked in by vboxsync, 16 years ago

HstDrv: Made vboxnetadp and vboxnetflt use IPRT from vboxdrv.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 37.8 KB
Line 
1/* $Id: VBoxNetFlt-linux.c 21351 2009-07-07 15:56:04Z vboxsync $ */
2/** @file
3 * VBoxNetFlt - Network Filter Driver (Host), Linux Specific Code.
4 */
5
6/*
7 * Copyright (C) 2006-2008 Sun Microsystems, Inc.
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 *
17 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa
18 * Clara, CA 95054 USA or visit http://www.sun.com if you need
19 * additional information or have any questions.
20 */
21
22/*******************************************************************************
23* Header Files *
24*******************************************************************************/
25#include "the-linux-kernel.h"
26#include "version-generated.h"
27#include <linux/netdevice.h>
28#include <linux/etherdevice.h>
29#include <linux/rtnetlink.h>
30#include <linux/miscdevice.h>
31#include <linux/ip.h>
32
33#define LOG_GROUP LOG_GROUP_NET_FLT_DRV
34#include <VBox/log.h>
35#include <VBox/err.h>
36#include <iprt/alloca.h>
37#include <iprt/assert.h>
38#include <iprt/spinlock.h>
39#include <iprt/semaphore.h>
40#include <iprt/initterm.h>
41#include <iprt/process.h>
42#include <iprt/mem.h>
43#include <iprt/log.h>
44#include <iprt/mp.h>
45#include <iprt/mem.h>
46#include <iprt/time.h>
47
48#define VBOXNETFLT_OS_SPECFIC 1
49#include "../VBoxNetFltInternal.h"
50
51#define VBOX_FLT_NB_TO_INST(pNB) ((PVBOXNETFLTINS)((uint8_t *)pNB - \
52 RT_OFFSETOF(VBOXNETFLTINS, u.s.Notifier)))
53#define VBOX_FLT_PT_TO_INST(pPT) ((PVBOXNETFLTINS)((uint8_t *)pPT - \
54 RT_OFFSETOF(VBOXNETFLTINS, u.s.PacketType)))
55#define VBOX_FLT_XT_TO_INST(pXT) ((PVBOXNETFLTINS)((uint8_t *)pXT - \
56 RT_OFFSETOF(VBOXNETFLTINS, u.s.XmitTask)))
57
58#define VBOX_GET_PCOUNT(pDev) (pDev->promiscuity)
59
60#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 22)
61# define VBOX_SKB_RESET_NETWORK_HDR(skb) skb_reset_network_header(skb)
62# define VBOX_SKB_RESET_MAC_HDR(skb) skb_reset_mac_header(skb)
63#else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 22) */
64# define VBOX_SKB_RESET_NETWORK_HDR(skb) skb->nh.raw = skb->data
65# define VBOX_SKB_RESET_MAC_HDR(skb) skb->mac.raw = skb->data
66#endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 22) */
67
68#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 19)
69# define VBOX_SKB_CHECKSUM_HELP(skb) skb_checksum_help(skb)
70#else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 19) */
71# define CHECKSUM_PARTIAL CHECKSUM_HW
72# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 10)
73# define VBOX_SKB_CHECKSUM_HELP(skb) skb_checksum_help(skb, 0)
74# else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 10) */
75# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 7)
76# define VBOX_SKB_CHECKSUM_HELP(skb) skb_checksum_help(&skb, 0)
77# else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 7) */
78# define VBOX_SKB_CHECKSUM_HELP(skb) (!skb_checksum_help(skb))
79# endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 7) */
80# endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 10) */
81#endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 19) */
82
83#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 18)
84# define VBOX_SKB_IS_GSO(skb) skb_is_gso(skb)
85 /* No features, very dumb device */
86# define VBOX_SKB_GSO_SEGMENT(skb) skb_gso_segment(skb, 0)
87#else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18) */
88# define VBOX_SKB_IS_GSO(skb) false
89# define VBOX_SKB_GSO_SEGMENT(skb) NULL
90#endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 18) */
91
92#ifndef NET_IP_ALIGN
93# define NET_IP_ALIGN 2
94#endif
95
96#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 12)
97unsigned dev_get_flags(const struct net_device *dev)
98{
99 unsigned flags;
100
101 flags = (dev->flags & ~(IFF_PROMISC |
102 IFF_ALLMULTI |
103 IFF_RUNNING)) |
104 (dev->gflags & (IFF_PROMISC |
105 IFF_ALLMULTI));
106
107 if (netif_running(dev) && netif_carrier_ok(dev))
108 flags |= IFF_RUNNING;
109
110 return flags;
111}
112#endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 12) */
113
114/*******************************************************************************
115* Internal Functions *
116*******************************************************************************/
117static int VBoxNetFltLinuxInit(void);
118static void VBoxNetFltLinuxUnload(void);
119
120
121/*******************************************************************************
122* Global Variables *
123*******************************************************************************/
124/**
125 * The (common) global data.
126 */
127static VBOXNETFLTGLOBALS g_VBoxNetFltGlobals;
128
129module_init(VBoxNetFltLinuxInit);
130module_exit(VBoxNetFltLinuxUnload);
131
132MODULE_AUTHOR("Sun Microsystems, Inc.");
133MODULE_DESCRIPTION("VirtualBox Network Filter Driver");
134MODULE_LICENSE("GPL");
135#ifdef MODULE_VERSION
136# define xstr(s) str(s)
137# define str(s) #s
138MODULE_VERSION(VBOX_VERSION_STRING " (" xstr(INTNETTRUNKIFPORT_VERSION) ")");
139#endif
140
141
142/**
143 * Initialize module.
144 *
145 * @returns appropriate status code.
146 */
147static int __init VBoxNetFltLinuxInit(void)
148{
149 int rc;
150 /*
151 * Initialize IPRT.
152 */
153 rc = RTR0Init(0);
154 if (RT_SUCCESS(rc))
155 {
156 Log(("VBoxNetFltLinuxInit\n"));
157
158 /*
159 * Initialize the globals and connect to the support driver.
160 *
161 * This will call back vboxNetFltOsOpenSupDrv (and maybe vboxNetFltOsCloseSupDrv)
162 * for establishing the connect to the support driver.
163 */
164 memset(&g_VBoxNetFltGlobals, 0, sizeof(g_VBoxNetFltGlobals));
165 rc = vboxNetFltInitGlobalsAndIdc(&g_VBoxNetFltGlobals);
166 if (RT_SUCCESS(rc))
167 {
168 LogRel(("VBoxNetFlt: Successfully started.\n"));
169 return 0;
170 }
171
172 LogRel(("VBoxNetFlt: failed to initialize device extension (rc=%d)\n", rc));
173 RTR0Term();
174 }
175 else
176 LogRel(("VBoxNetFlt: failed to initialize IPRT (rc=%d)\n", rc));
177
178 memset(&g_VBoxNetFltGlobals, 0, sizeof(g_VBoxNetFltGlobals));
179 return -RTErrConvertToErrno(rc);
180}
181
182
183/**
184 * Unload the module.
185 *
186 * @todo We have to prevent this if we're busy!
187 */
188static void __exit VBoxNetFltLinuxUnload(void)
189{
190 int rc;
191 Log(("VBoxNetFltLinuxUnload\n"));
192 Assert(vboxNetFltCanUnload(&g_VBoxNetFltGlobals));
193
194 /*
195 * Undo the work done during start (in reverse order).
196 */
197 rc = vboxNetFltTryDeleteIdcAndGlobals(&g_VBoxNetFltGlobals);
198 AssertRC(rc); NOREF(rc);
199
200 RTR0Term();
201
202 memset(&g_VBoxNetFltGlobals, 0, sizeof(g_VBoxNetFltGlobals));
203
204 Log(("VBoxNetFltLinuxUnload - done\n"));
205}
206
207
208/**
209 * Reads and retains the host interface handle.
210 *
211 * @returns The handle, NULL if detached.
212 * @param pThis
213 */
214DECLINLINE(struct net_device *) vboxNetFltLinuxRetainNetDev(PVBOXNETFLTINS pThis)
215{
216#if 0
217 RTSPINLOCKTMP Tmp = RTSPINLOCKTMP_INITIALIZER;
218 struct net_device *pDev = NULL;
219
220 Log(("vboxNetFltLinuxRetainNetDev\n"));
221 /*
222 * Be careful here to avoid problems racing the detached callback.
223 */
224 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
225 if (!ASMAtomicUoReadBool(&pThis->fDisconnectedFromHost))
226 {
227 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
228 if (pDev)
229 {
230 dev_hold(pDev);
231 Log(("vboxNetFltLinuxRetainNetDev: Device %p(%s) retained. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
232 }
233 }
234 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
235
236 Log(("vboxNetFltLinuxRetainNetDev - done\n"));
237 return pDev;
238#else
239 return (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
240#endif
241}
242
243
244/**
245 * Release the host interface handle previously retained
246 * by vboxNetFltLinuxRetainNetDev.
247 *
248 * @param pThis The instance.
249 * @param pDev The vboxNetFltLinuxRetainNetDev
250 * return value, NULL is fine.
251 */
252DECLINLINE(void) vboxNetFltLinuxReleaseNetDev(PVBOXNETFLTINS pThis, struct net_device *pDev)
253{
254#if 0
255 Log(("vboxNetFltLinuxReleaseNetDev\n"));
256 NOREF(pThis);
257 if (pDev)
258 {
259 dev_put(pDev);
260 Log(("vboxNetFltLinuxReleaseNetDev: Device %p(%s) released. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
261 }
262 Log(("vboxNetFltLinuxReleaseNetDev - done\n"));
263#endif
264}
265
266#define VBOXNETFLT_CB_TAG(skb) (0xA1C90000 | (skb->dev->ifindex & 0xFFFF))
267#define VBOXNETFLT_SKB_TAG(skb) (*(uint32_t*)&((skb)->cb[sizeof((skb)->cb)-sizeof(uint32_t)]))
268
269/**
270 * Checks whether this is an mbuf created by vboxNetFltLinuxMBufFromSG,
271 * i.e. a buffer which we're pushing and should be ignored by the filter callbacks.
272 *
273 * @returns true / false accordingly.
274 * @param pBuf The sk_buff.
275 */
276DECLINLINE(bool) vboxNetFltLinuxSkBufIsOur(struct sk_buff *pBuf)
277{
278 return VBOXNETFLT_SKB_TAG(pBuf) == VBOXNETFLT_CB_TAG(pBuf);
279}
280
281
282/**
283 * Internal worker that create a linux sk_buff for a
284 * (scatter/)gather list.
285 *
286 * @returns Pointer to the sk_buff.
287 * @param pThis The instance.
288 * @param pSG The (scatter/)gather list.
289 */
290static struct sk_buff *vboxNetFltLinuxSkBufFromSG(PVBOXNETFLTINS pThis, PINTNETSG pSG, bool fDstWire)
291{
292 struct sk_buff *pPkt;
293 struct net_device *pDev;
294 /*
295 * Because we're lazy, we will ASSUME that all SGs coming from INTNET
296 * will only contain one single segment.
297 */
298 if (pSG->cSegsUsed != 1 || pSG->cbTotal != pSG->aSegs[0].cb)
299 {
300 LogRel(("VBoxNetFlt: Dropped multi-segment(%d) packet coming from internal network.\n", pSG->cSegsUsed));
301 return NULL;
302 }
303 if (pSG->cbTotal == 0)
304 {
305 LogRel(("VBoxNetFlt: Dropped empty packet coming from internal network.\n"));
306 return NULL;
307 }
308
309 /*
310 * Allocate a packet and copy over the data.
311 *
312 */
313 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
314 pPkt = dev_alloc_skb(pSG->cbTotal + NET_IP_ALIGN);
315 if (pPkt)
316 {
317 pPkt->dev = pDev;
318 /* Align IP header on 16-byte boundary: 2 + 14 (ethernet hdr size). */
319 skb_reserve(pPkt, NET_IP_ALIGN);
320 skb_put(pPkt, pSG->cbTotal);
321 memcpy(pPkt->data, pSG->aSegs[0].pv, pSG->cbTotal);
322 /* Set protocol and packet_type fields. */
323 pPkt->protocol = eth_type_trans(pPkt, pDev);
324 pPkt->ip_summed = CHECKSUM_NONE;
325 if (fDstWire)
326 {
327 VBOX_SKB_RESET_NETWORK_HDR(pPkt);
328 /* Restore ethernet header back. */
329 skb_push(pPkt, ETH_HLEN);
330 VBOX_SKB_RESET_MAC_HDR(pPkt);
331 }
332 VBOXNETFLT_SKB_TAG(pPkt) = VBOXNETFLT_CB_TAG(pPkt);
333
334 return pPkt;
335 }
336 else
337 Log(("vboxNetFltLinuxSkBufFromSG: Failed to allocate sk_buff(%u).\n", pSG->cbTotal));
338 pSG->pvUserData = NULL;
339
340 return NULL;
341}
342
343
344/**
345 * Initializes a SG list from an sk_buff.
346 *
347 * @returns Number of segments.
348 * @param pThis The instance.
349 * @param pBuf The sk_buff.
350 * @param pSG The SG.
351 * @param pvFrame The frame pointer, optional.
352 * @param cSegs The number of segments allocated for the SG.
353 * This should match the number in the mbuf exactly!
354 * @param fSrc The source of the frame.
355 */
356DECLINLINE(void) vboxNetFltLinuxSkBufToSG(PVBOXNETFLTINS pThis, struct sk_buff *pBuf, PINTNETSG pSG, unsigned cSegs, uint32_t fSrc)
357{
358 int i;
359 NOREF(pThis);
360
361 Assert(!skb_shinfo(pBuf)->frag_list);
362 pSG->pvOwnerData = NULL;
363 pSG->pvUserData = NULL;
364 pSG->pvUserData2 = NULL;
365 pSG->cUsers = 1;
366 pSG->fFlags = INTNETSG_FLAGS_TEMP;
367 pSG->cSegsAlloc = cSegs;
368
369 if (fSrc & INTNETTRUNKDIR_WIRE)
370 {
371 /*
372 * The packet came from wire, ethernet header was removed by device driver.
373 * Restore it.
374 */
375 skb_push(pBuf, ETH_HLEN);
376 }
377 pSG->cbTotal = pBuf->len;
378#ifdef VBOXNETFLT_SG_SUPPORT
379 pSG->aSegs[0].cb = skb_headlen(pBuf);
380 pSG->aSegs[0].pv = pBuf->data;
381 pSG->aSegs[0].Phys = NIL_RTHCPHYS;
382
383 for (i = 0; i < skb_shinfo(pBuf)->nr_frags; i++)
384 {
385 skb_frag_t *pFrag = &skb_shinfo(pBuf)->frags[i];
386 pSG->aSegs[i+1].cb = pFrag->size;
387 pSG->aSegs[i+1].pv = kmap(pFrag->page);
388 printk("%p = kmap()\n", pSG->aSegs[i+1].pv);
389 pSG->aSegs[i+1].Phys = NIL_RTHCPHYS;
390 }
391 pSG->cSegsUsed = ++i;
392#else
393 pSG->aSegs[0].cb = pBuf->len;
394 pSG->aSegs[0].pv = pBuf->data;
395 pSG->aSegs[0].Phys = NIL_RTHCPHYS;
396 pSG->cSegsUsed = i = 1;
397#endif
398
399
400#ifdef PADD_RUNT_FRAMES_FROM_HOST
401 /*
402 * Add a trailer if the frame is too small.
403 *
404 * Since we're getting to the packet before it is framed, it has not
405 * yet been padded. The current solution is to add a segment pointing
406 * to a buffer containing all zeros and pray that works for all frames...
407 */
408 if (pSG->cbTotal < 60 && (fSrc & INTNETTRUNKDIR_HOST))
409 {
410 static uint8_t const s_abZero[128] = {0};
411
412 AssertReturnVoid(i < cSegs);
413
414 pSG->aSegs[i].Phys = NIL_RTHCPHYS;
415 pSG->aSegs[i].pv = (void *)&s_abZero[0];
416 pSG->aSegs[i].cb = 60 - pSG->cbTotal;
417 pSG->cbTotal = 60;
418 pSG->cSegsUsed++;
419 }
420#endif
421 Log4(("vboxNetFltLinuxSkBufToSG: allocated=%d, segments=%d frags=%d next=%p frag_list=%p pkt_type=%x fSrc=%x\n",
422 pSG->cSegsAlloc, pSG->cSegsUsed, skb_shinfo(pBuf)->nr_frags, pBuf->next, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type, fSrc));
423 for (i = 0; i < pSG->cSegsUsed; i++)
424 Log4(("vboxNetFltLinuxSkBufToSG: #%d: cb=%d pv=%p\n",
425 i, pSG->aSegs[i].cb, pSG->aSegs[i].pv));
426}
427
428/**
429 * Packet handler,
430 *
431 * @returns 0 or EJUSTRETURN.
432 * @param pThis The instance.
433 * @param pMBuf The mbuf.
434 * @param pvFrame The start of the frame, optional.
435 * @param fSrc Where the packet (allegedly) comes from, one INTNETTRUNKDIR_* value.
436 * @param eProtocol The protocol.
437 */
438#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 14)
439static int vboxNetFltLinuxPacketHandler(struct sk_buff *pBuf,
440 struct net_device *pSkbDev,
441 struct packet_type *pPacketType,
442 struct net_device *pOrigDev)
443#else
444static int vboxNetFltLinuxPacketHandler(struct sk_buff *pBuf,
445 struct net_device *pSkbDev,
446 struct packet_type *pPacketType)
447#endif
448{
449 PVBOXNETFLTINS pThis;
450 struct net_device *pDev;
451 LogFlow(("vboxNetFltLinuxPacketHandler: pBuf=%p pSkbDev=%p pPacketType=%p\n",
452 pBuf, pSkbDev, pPacketType));
453#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 18)
454 Log3(("vboxNetFltLinuxPacketHandler: skb len=%u data_len=%u truesize=%u next=%p nr_frags=%u gso_size=%u gso_seqs=%u gso_type=%x frag_list=%p pkt_type=%x\n",
455 pBuf->len, pBuf->data_len, pBuf->truesize, pBuf->next, skb_shinfo(pBuf)->nr_frags, skb_shinfo(pBuf)->gso_size, skb_shinfo(pBuf)->gso_segs, skb_shinfo(pBuf)->gso_type, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type));
456#else
457 Log3(("vboxNetFltLinuxPacketHandler: skb len=%u data_len=%u truesize=%u next=%p nr_frags=%u tso_size=%u tso_seqs=%u frag_list=%p pkt_type=%x\n",
458 pBuf->len, pBuf->data_len, pBuf->truesize, pBuf->next, skb_shinfo(pBuf)->nr_frags, skb_shinfo(pBuf)->tso_size, skb_shinfo(pBuf)->tso_segs, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type));
459#endif
460 /*
461 * Drop it immediately?
462 */
463 if (!pBuf)
464 return 0;
465
466 pThis = VBOX_FLT_PT_TO_INST(pPacketType);
467 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
468 if (pThis->u.s.pDev != pSkbDev)
469 {
470 Log(("vboxNetFltLinuxPacketHandler: Devices do not match, pThis may be wrong! pThis=%p\n", pThis));
471 return 0;
472 }
473
474 Log4(("vboxNetFltLinuxPacketHandler: pBuf->cb dump:\n%.*Rhxd\n", sizeof(pBuf->cb), pBuf->cb));
475 if (vboxNetFltLinuxSkBufIsOur(pBuf))
476 {
477 Log2(("vboxNetFltLinuxPacketHandler: got our own sk_buff, drop it.\n"));
478 dev_kfree_skb(pBuf);
479 return 0;
480 }
481
482#ifndef VBOXNETFLT_SG_SUPPORT
483 {
484 /*
485 * Get rid of fragmented packets, they cause too much trouble.
486 */
487 struct sk_buff *pCopy = skb_copy(pBuf, GFP_ATOMIC);
488 kfree_skb(pBuf);
489 if (!pCopy)
490 {
491 LogRel(("VBoxNetFlt: Failed to allocate packet buffer, dropping the packet.\n"));
492 return 0;
493 }
494 pBuf = pCopy;
495#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 18)
496 Log3(("vboxNetFltLinuxPacketHandler: skb copy len=%u data_len=%u truesize=%u next=%p nr_frags=%u gso_size=%u gso_seqs=%u gso_type=%x frag_list=%p pkt_type=%x\n",
497 pBuf->len, pBuf->data_len, pBuf->truesize, pBuf->next, skb_shinfo(pBuf)->nr_frags, skb_shinfo(pBuf)->gso_size, skb_shinfo(pBuf)->gso_segs, skb_shinfo(pBuf)->gso_type, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type));
498#else
499 Log3(("vboxNetFltLinuxPacketHandler: skb copy len=%u data_len=%u truesize=%u next=%p nr_frags=%u tso_size=%u tso_seqs=%u frag_list=%p pkt_type=%x\n",
500 pBuf->len, pBuf->data_len, pBuf->truesize, pBuf->next, skb_shinfo(pBuf)->nr_frags, skb_shinfo(pBuf)->tso_size, skb_shinfo(pBuf)->tso_segs, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type));
501#endif
502 }
503#endif
504
505 /* Add the packet to transmit queue and schedule the bottom half. */
506 skb_queue_tail(&pThis->u.s.XmitQueue, pBuf);
507 schedule_work(&pThis->u.s.XmitTask);
508 Log4(("vboxNetFltLinuxPacketHandler: scheduled work %p for sk_buff %p\n",
509 &pThis->u.s.XmitTask, pBuf));
510 /* It does not really matter what we return, it is ignored by the kernel. */
511 return 0;
512}
513
514static unsigned vboxNetFltLinuxSGSegments(PVBOXNETFLTINS pThis, struct sk_buff *pBuf)
515{
516#ifdef VBOXNETFLT_SG_SUPPORT
517 unsigned cSegs = 1 + skb_shinfo(pBuf)->nr_frags;
518#else
519 unsigned cSegs = 1;
520#endif
521#ifdef PADD_RUNT_FRAMES_FROM_HOST
522 /*
523 * Add a trailer if the frame is too small.
524 */
525 if (pBuf->len < 60)
526 cSegs++;
527#endif
528 return cSegs;
529}
530
531/* WARNING! This function should only be called after vboxNetFltLinuxSkBufToSG()! */
532static void vboxNetFltLinuxFreeSkBuff(struct sk_buff *pBuf, PINTNETSG pSG)
533{
534#ifdef VBOXNETFLT_SG_SUPPORT
535 int i;
536
537 for (i = 0; i < skb_shinfo(pBuf)->nr_frags; i++)
538 {
539 printk("kunmap(%p)\n", pSG->aSegs[i+1].pv);
540 kunmap(pSG->aSegs[i+1].pv);
541 }
542#endif
543
544 dev_kfree_skb(pBuf);
545}
546
547#ifndef LOG_ENABLED
548#define vboxNetFltDumpPacket(a, b, c, d)
549#else
550static void vboxNetFltDumpPacket(PINTNETSG pSG, bool fEgress, const char *pszWhere, int iIncrement)
551{
552 uint8_t *pInt, *pExt;
553 static int iPacketNo = 1;
554 iPacketNo += iIncrement;
555 if (fEgress)
556 {
557 pExt = pSG->aSegs[0].pv;
558 pInt = pExt + 6;
559 }
560 else
561 {
562 pInt = pSG->aSegs[0].pv;
563 pExt = pInt + 6;
564 }
565 Log(("VBoxNetFlt: (int)%02x:%02x:%02x:%02x:%02x:%02x"
566 " %s (%s)%02x:%02x:%02x:%02x:%02x:%02x (%u bytes) packet #%u\n",
567 pInt[0], pInt[1], pInt[2], pInt[3], pInt[4], pInt[5],
568 fEgress ? "-->" : "<--", pszWhere,
569 pExt[0], pExt[1], pExt[2], pExt[3], pExt[4], pExt[5],
570 pSG->cbTotal, iPacketNo));
571 Log3(("%.*Rhxd\n", pSG->aSegs[0].cb, pSG->aSegs[0].pv));
572}
573#endif
574
575static int vboxNetFltLinuxForwardSegment(PVBOXNETFLTINS pThis, struct sk_buff *pBuf, uint32_t fSrc)
576{
577 unsigned cSegs = vboxNetFltLinuxSGSegments(pThis, pBuf);
578 if (cSegs < MAX_SKB_FRAGS)
579 {
580 uint8_t *pTmp;
581 PINTNETSG pSG = (PINTNETSG)alloca(RT_OFFSETOF(INTNETSG, aSegs[cSegs]));
582 if (!pSG)
583 {
584 Log(("VBoxNetFlt: Failed to allocate SG buffer.\n"));
585 return VERR_NO_MEMORY;
586 }
587 vboxNetFltLinuxSkBufToSG(pThis, pBuf, pSG, cSegs, fSrc);
588
589 pTmp = pSG->aSegs[0].pv;
590 vboxNetFltDumpPacket(pSG, false, (fSrc & INTNETTRUNKDIR_HOST) ? "host" : "wire", 1);
591 pThis->pSwitchPort->pfnRecv(pThis->pSwitchPort, pSG, fSrc);
592 Log4(("VBoxNetFlt: Dropping the sk_buff.\n"));
593 vboxNetFltLinuxFreeSkBuff(pBuf, pSG);
594 }
595
596 return VINF_SUCCESS;
597}
598
599static void vboxNetFltLinuxForwardToIntNet(PVBOXNETFLTINS pThis, struct sk_buff *pBuf)
600{
601 uint32_t fSrc = pBuf->pkt_type == PACKET_OUTGOING ? INTNETTRUNKDIR_HOST : INTNETTRUNKDIR_WIRE;
602
603 if (VBOX_SKB_IS_GSO(pBuf))
604 {
605 /* Need to segment the packet */
606 struct sk_buff *pNext, *pSegment;
607#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 18)
608 Log3(("vboxNetFltLinuxForwardToIntNet: skb len=%u data_len=%u truesize=%u next=%p nr_frags=%u gso_size=%u gso_seqs=%u gso_type=%x frag_list=%p pkt_type=%x ip_summed=%d\n",
609 pBuf->len, pBuf->data_len, pBuf->truesize, pBuf->next, skb_shinfo(pBuf)->nr_frags, skb_shinfo(pBuf)->gso_size, skb_shinfo(pBuf)->gso_segs, skb_shinfo(pBuf)->gso_type, skb_shinfo(pBuf)->frag_list, pBuf->pkt_type, pBuf->ip_summed));
610#endif
611
612 pSegment = VBOX_SKB_GSO_SEGMENT(pBuf);
613 if (IS_ERR(pSegment))
614 {
615 dev_kfree_skb(pBuf);
616 LogRel(("VBoxNetFlt: Failed to segment a packet (%d).\n", PTR_ERR(pBuf)));
617 return;
618 }
619 for (; pSegment; pSegment = pNext)
620 {
621#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 18)
622 Log3(("vboxNetFltLinuxForwardToIntNet: segment len=%u data_len=%u truesize=%u next=%p nr_frags=%u gso_size=%u gso_seqs=%u gso_type=%x frag_list=%p pkt_type=%x\n",
623 pSegment->len, pSegment->data_len, pSegment->truesize, pSegment->next, skb_shinfo(pSegment)->nr_frags, skb_shinfo(pSegment)->gso_size, skb_shinfo(pSegment)->gso_segs, skb_shinfo(pSegment)->gso_type, skb_shinfo(pSegment)->frag_list, pSegment->pkt_type));
624#endif
625 pNext = pSegment->next;
626 pSegment->next = 0;
627 vboxNetFltLinuxForwardSegment(pThis, pSegment, fSrc);
628 }
629 dev_kfree_skb(pBuf);
630 }
631 else
632 {
633 if (pBuf->ip_summed == CHECKSUM_PARTIAL && pBuf->pkt_type == PACKET_OUTGOING)
634 {
635#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 18)
636 /*
637 * Try to work around the problem with CentOS 4.7 and 5.2 (2.6.9
638 * and 2.6.18 kernels), they pass wrong 'h' pointer down. We take IP
639 * header length from the header itself and reconstruct 'h' pointer
640 * to TCP (or whatever) header.
641 */
642 unsigned char *tmp = pBuf->h.raw;
643 if (pBuf->h.raw == pBuf->nh.raw && pBuf->protocol == htons(ETH_P_IP))
644 pBuf->h.raw = pBuf->nh.raw + pBuf->nh.iph->ihl * 4;
645#endif /* LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 18) */
646 if (VBOX_SKB_CHECKSUM_HELP(pBuf))
647 {
648 LogRel(("VBoxNetFlt: Failed to compute checksum, dropping the packet.\n"));
649 dev_kfree_skb(pBuf);
650 return;
651 }
652#if LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 18)
653 /* Restore the original (wrong) pointer. */
654 pBuf->h.raw = tmp;
655#endif /* LINUX_VERSION_CODE <= KERNEL_VERSION(2, 6, 18) */
656 }
657 vboxNetFltLinuxForwardSegment(pThis, pBuf, fSrc);
658 }
659 /*
660 * Create a (scatter/)gather list for the sk_buff and feed it to the internal network.
661 */
662}
663
664#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 20)
665static void vboxNetFltLinuxXmitTask(struct work_struct *pWork)
666#else
667static void vboxNetFltLinuxXmitTask(void *pWork)
668#endif
669{
670 struct sk_buff *pBuf;
671 bool fActive;
672 PVBOXNETFLTINS pThis;
673 RTSPINLOCKTMP Tmp = RTSPINLOCKTMP_INITIALIZER;
674
675 Log4(("vboxNetFltLinuxXmitTask: Got work %p.\n", pWork));
676 pThis = VBOX_FLT_XT_TO_INST(pWork);
677 /*
678 * Active? Retain the instance and increment the busy counter.
679 */
680 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
681 fActive = ASMAtomicUoReadBool(&pThis->fActive);
682 if (fActive)
683 vboxNetFltRetain(pThis, true /* fBusy */);
684 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
685 if (!fActive)
686 return;
687
688 while ((pBuf = skb_dequeue(&pThis->u.s.XmitQueue)) != 0)
689 vboxNetFltLinuxForwardToIntNet(pThis, pBuf);
690
691 vboxNetFltRelease(pThis, true /* fBusy */);
692}
693
694/**
695 * Internal worker for vboxNetFltOsInitInstance and vboxNetFltOsMaybeRediscovered.
696 *
697 * @returns VBox status code.
698 * @param pThis The instance.
699 * @param fRediscovery If set we're doing a rediscovery attempt, so, don't
700 * flood the release log.
701 */
702static int vboxNetFltLinuxAttachToInterface(PVBOXNETFLTINS pThis, struct net_device *pDev)
703{
704 struct packet_type *pt;
705 RTSPINLOCKTMP Tmp = RTSPINLOCKTMP_INITIALIZER;
706
707 LogFlow(("vboxNetFltLinuxAttachToInterface: pThis=%p (%s)\n", pThis, pThis->szName));
708
709 if (!pDev)
710 {
711 Log(("VBoxNetFlt: failed to find device '%s'\n", pThis->szName));
712 return VERR_INTNET_FLT_IF_NOT_FOUND;
713 }
714
715 dev_hold(pDev);
716 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
717 ASMAtomicUoWritePtr((void * volatile *)&pThis->u.s.pDev, pDev);
718 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
719
720 Log(("vboxNetFltLinuxAttachToInterface: Device %p(%s) retained. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
721 Log(("vboxNetFltLinuxAttachToInterface: Got pDev=%p pThis=%p pThis->u.s.pDev=%p\n", pDev, pThis, ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev)));
722 /*
723 * Get the mac address while we still have a valid ifnet reference.
724 */
725 memcpy(&pThis->u.s.Mac, pDev->dev_addr, sizeof(pThis->u.s.Mac));
726
727 pt = &pThis->u.s.PacketType;
728 pt->type = __constant_htons(ETH_P_ALL);
729 pt->dev = pDev;
730 pt->func = vboxNetFltLinuxPacketHandler;
731 dev_add_pack(pt);
732 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
733 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
734 if (pDev)
735 {
736 ASMAtomicUoWriteBool(&pThis->fDisconnectedFromHost, false);
737 ASMAtomicUoWriteBool(&pThis->u.s.fRegistered, true);
738 pDev = NULL; /* don't dereference it */
739 }
740 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
741 Log(("vboxNetFltLinuxAttachToInterface: this=%p: Packet handler installed.\n", pThis));
742
743 /* Release the interface on failure. */
744 if (pDev)
745 {
746 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
747 ASMAtomicUoWritePtr((void * volatile *)&pThis->u.s.pDev, NULL);
748 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
749 dev_put(pDev);
750 Log(("vboxNetFltLinuxAttachToInterface: Device %p(%s) released. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
751 }
752
753 LogRel(("VBoxNetFlt: attached to '%s' / %.*Rhxs\n", pThis->szName, sizeof(pThis->u.s.Mac), &pThis->u.s.Mac));
754 return VINF_SUCCESS;
755}
756
757
758static int vboxNetFltLinuxUnregisterDevice(PVBOXNETFLTINS pThis, struct net_device *pDev)
759{
760 RTSPINLOCKTMP Tmp = RTSPINLOCKTMP_INITIALIZER;
761
762 Assert(!pThis->fDisconnectedFromHost);
763 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
764 ASMAtomicWriteBool(&pThis->u.s.fRegistered, false);
765 ASMAtomicWriteBool(&pThis->fDisconnectedFromHost, true);
766 ASMAtomicUoWritePtr((void * volatile *)&pThis->u.s.pDev, NULL);
767 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
768
769 dev_remove_pack(&pThis->u.s.PacketType);
770 skb_queue_purge(&pThis->u.s.XmitQueue);
771 Log(("vboxNetFltLinuxUnregisterDevice: this=%p: Packet handler removed, xmit queue purged.\n", pThis));
772 Log(("vboxNetFltLinuxUnregisterDevice: Device %p(%s) released. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
773 dev_put(pDev);
774
775 return NOTIFY_OK;
776}
777
778static int vboxNetFltLinuxDeviceIsUp(PVBOXNETFLTINS pThis, struct net_device *pDev)
779{
780 /* Check if we are not suspended and promiscuous mode has not been set. */
781 if (ASMAtomicUoReadBool(&pThis->fActive) && !ASMAtomicUoReadBool(&pThis->u.s.fPromiscuousSet))
782 {
783 /* Note that there is no need for locking as the kernel got hold of the lock already. */
784 dev_set_promiscuity(pDev, 1);
785 ASMAtomicWriteBool(&pThis->u.s.fPromiscuousSet, true);
786 Log(("vboxNetFltLinuxDeviceIsUp: enabled promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
787 }
788 else
789 Log(("vboxNetFltLinuxDeviceIsUp: no need to enable promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
790 return NOTIFY_OK;
791}
792
793static int vboxNetFltLinuxDeviceGoingDown(PVBOXNETFLTINS pThis, struct net_device *pDev)
794{
795 /* Undo promiscuous mode if we has set it. */
796 if (ASMAtomicUoReadBool(&pThis->u.s.fPromiscuousSet))
797 {
798 /* Note that there is no need for locking as the kernel got hold of the lock already. */
799 dev_set_promiscuity(pDev, -1);
800 ASMAtomicWriteBool(&pThis->u.s.fPromiscuousSet, false);
801 Log(("vboxNetFltLinuxDeviceGoingDown: disabled promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
802 }
803 else
804 Log(("vboxNetFltLinuxDeviceGoingDown: no need to disable promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
805 return NOTIFY_OK;
806}
807
808static int vboxNetFltLinuxNotifierCallback(struct notifier_block *self, unsigned long ulEventType, void *ptr)
809
810{
811 int rc = NOTIFY_OK;
812#ifdef DEBUG
813 char *pszEvent = "<unknown>";
814#endif
815 struct net_device *pDev = (struct net_device *)ptr;
816 PVBOXNETFLTINS pThis = VBOX_FLT_NB_TO_INST(self);
817
818#ifdef DEBUG
819 switch (ulEventType)
820 {
821 case NETDEV_REGISTER: pszEvent = "NETDEV_REGISTER"; break;
822 case NETDEV_UNREGISTER: pszEvent = "NETDEV_UNREGISTER"; break;
823 case NETDEV_UP: pszEvent = "NETDEV_UP"; break;
824 case NETDEV_DOWN: pszEvent = "NETDEV_DOWN"; break;
825 case NETDEV_REBOOT: pszEvent = "NETDEV_REBOOT"; break;
826 case NETDEV_CHANGENAME: pszEvent = "NETDEV_CHANGENAME"; break;
827 case NETDEV_CHANGE: pszEvent = "NETDEV_CHANGE"; break;
828 case NETDEV_CHANGEMTU: pszEvent = "NETDEV_CHANGEMTU"; break;
829 case NETDEV_CHANGEADDR: pszEvent = "NETDEV_CHANGEADDR"; break;
830 case NETDEV_GOING_DOWN: pszEvent = "NETDEV_GOING_DOWN"; break;
831 }
832 Log(("VBoxNetFlt: got event %s(0x%lx) on %s, pDev=%p pThis=%p pThis->u.s.pDev=%p\n",
833 pszEvent, ulEventType, pDev->name, pDev, pThis, ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev)));
834#endif
835 if (ulEventType == NETDEV_REGISTER && !strcmp(pDev->name, pThis->szName))
836 {
837 vboxNetFltLinuxAttachToInterface(pThis, pDev);
838 }
839 else
840 {
841 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
842 if (pDev != ptr)
843 return NOTIFY_OK;
844 rc = NOTIFY_OK;
845 switch (ulEventType)
846 {
847 case NETDEV_UNREGISTER:
848 rc = vboxNetFltLinuxUnregisterDevice(pThis, pDev);
849 break;
850 case NETDEV_UP:
851 rc = vboxNetFltLinuxDeviceIsUp(pThis, pDev);
852 break;
853 case NETDEV_GOING_DOWN:
854 rc = vboxNetFltLinuxDeviceGoingDown(pThis, pDev);
855 break;
856 case NETDEV_CHANGENAME:
857 break;
858 }
859 }
860
861 return rc;
862}
863
864bool vboxNetFltOsMaybeRediscovered(PVBOXNETFLTINS pThis)
865{
866 return !ASMAtomicUoReadBool(&pThis->fDisconnectedFromHost);
867}
868
869int vboxNetFltPortOsXmit(PVBOXNETFLTINS pThis, PINTNETSG pSG, uint32_t fDst)
870{
871 struct net_device * pDev;
872 int err;
873 int rc = VINF_SUCCESS;
874
875 LogFlow(("vboxNetFltPortOsXmit: pThis=%p (%s)\n", pThis, pThis->szName));
876
877 pDev = vboxNetFltLinuxRetainNetDev(pThis);
878 if (pDev)
879 {
880 /*
881 * Create a sk_buff for the gather list and push it onto the wire.
882 */
883 if (fDst & INTNETTRUNKDIR_WIRE)
884 {
885 struct sk_buff *pBuf = vboxNetFltLinuxSkBufFromSG(pThis, pSG, true);
886 if (pBuf)
887 {
888 vboxNetFltDumpPacket(pSG, true, "wire", 1);
889 Log4(("vboxNetFltPortOsXmit: pBuf->cb dump:\n%.*Rhxd\n", sizeof(pBuf->cb), pBuf->cb));
890 Log4(("vboxNetFltPortOsXmit: dev_queue_xmit(%p)\n", pBuf));
891 err = dev_queue_xmit(pBuf);
892 if (err)
893 rc = RTErrConvertFromErrno(err);
894 }
895 else
896 rc = VERR_NO_MEMORY;
897 }
898
899 /*
900 * Create a sk_buff for the gather list and push it onto the host stack.
901 */
902 if (fDst & INTNETTRUNKDIR_HOST)
903 {
904 struct sk_buff *pBuf = vboxNetFltLinuxSkBufFromSG(pThis, pSG, false);
905 if (pBuf)
906 {
907 vboxNetFltDumpPacket(pSG, true, "host", (fDst & INTNETTRUNKDIR_WIRE) ? 0 : 1);
908 Log4(("vboxNetFltPortOsXmit: pBuf->cb dump:\n%.*Rhxd\n", sizeof(pBuf->cb), pBuf->cb));
909 Log4(("vboxNetFltPortOsXmit: netif_rx_ni(%p)\n", pBuf));
910 err = netif_rx_ni(pBuf);
911 if (err)
912 rc = RTErrConvertFromErrno(err);
913 }
914 else
915 rc = VERR_NO_MEMORY;
916 }
917
918 vboxNetFltLinuxReleaseNetDev(pThis, pDev);
919 }
920
921 return rc;
922}
923
924
925bool vboxNetFltPortOsIsPromiscuous(PVBOXNETFLTINS pThis)
926{
927 bool fRc = false;
928 struct net_device * pDev = vboxNetFltLinuxRetainNetDev(pThis);
929 if (pDev)
930 {
931 fRc = !!(pDev->promiscuity - (ASMAtomicUoReadBool(&pThis->u.s.fPromiscuousSet) & 1));
932 LogFlow(("vboxNetFltPortOsIsPromiscuous: returns %d, pDev->promiscuity=%d, fPromiscuousSet=%d\n",
933 fRc, pDev->promiscuity, pThis->u.s.fPromiscuousSet));
934 vboxNetFltLinuxReleaseNetDev(pThis, pDev);
935 }
936 return fRc;
937}
938
939
940void vboxNetFltPortOsGetMacAddress(PVBOXNETFLTINS pThis, PRTMAC pMac)
941{
942 *pMac = pThis->u.s.Mac;
943}
944
945
946bool vboxNetFltPortOsIsHostMac(PVBOXNETFLTINS pThis, PCRTMAC pMac)
947{
948 /* ASSUMES that the MAC address never changes. */
949 return pThis->u.s.Mac.au16[0] == pMac->au16[0]
950 && pThis->u.s.Mac.au16[1] == pMac->au16[1]
951 && pThis->u.s.Mac.au16[2] == pMac->au16[2];
952}
953
954
955void vboxNetFltPortOsSetActive(PVBOXNETFLTINS pThis, bool fActive)
956{
957 struct net_device * pDev;
958
959 LogFlow(("vboxNetFltPortOsSetActive: pThis=%p (%s), fActive=%s, fDisablePromiscuous=%s\n",
960 pThis, pThis->szName, fActive?"true":"false",
961 pThis->fDisablePromiscuous?"true":"false"));
962
963 if (pThis->fDisablePromiscuous)
964 return;
965
966 pDev = vboxNetFltLinuxRetainNetDev(pThis);
967 if (pDev)
968 {
969 /*
970 * This api is a bit weird, the best reference is the code.
971 *
972 * Also, we have a bit or race conditions wrt the maintance of
973 * host the interface promiscuity for vboxNetFltPortOsIsPromiscuous.
974 */
975#ifdef LOG_ENABLED
976 u_int16_t fIf;
977 unsigned const cPromiscBefore = VBOX_GET_PCOUNT(pDev);
978#endif
979 if (fActive)
980 {
981 Assert(!pThis->u.s.fPromiscuousSet);
982
983 rtnl_lock();
984 dev_set_promiscuity(pDev, 1);
985 rtnl_unlock();
986 pThis->u.s.fPromiscuousSet = true;
987 Log(("vboxNetFltPortOsSetActive: enabled promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
988 }
989 else
990 {
991 if (pThis->u.s.fPromiscuousSet)
992 {
993 rtnl_lock();
994 dev_set_promiscuity(pDev, -1);
995 rtnl_unlock();
996 Log(("vboxNetFltPortOsSetActive: disabled promiscuous mode on %s (%d)\n", pThis->szName, VBOX_GET_PCOUNT(pDev)));
997 }
998 pThis->u.s.fPromiscuousSet = false;
999
1000#ifdef LOG_ENABLED
1001 fIf = dev_get_flags(pDev);
1002 Log(("VBoxNetFlt: fIf=%#x; %d->%d\n", fIf, cPromiscBefore, VBOX_GET_PCOUNT(pDev)));
1003#endif
1004 }
1005
1006 vboxNetFltLinuxReleaseNetDev(pThis, pDev);
1007 }
1008}
1009
1010
1011int vboxNetFltOsDisconnectIt(PVBOXNETFLTINS pThis)
1012{
1013 /* Nothing to do here. */
1014 return VINF_SUCCESS;
1015}
1016
1017
1018int vboxNetFltOsConnectIt(PVBOXNETFLTINS pThis)
1019{
1020 /* Nothing to do here. */
1021 return VINF_SUCCESS;
1022}
1023
1024
1025void vboxNetFltOsDeleteInstance(PVBOXNETFLTINS pThis)
1026{
1027 struct net_device *pDev;
1028 bool fRegistered;
1029 RTSPINLOCKTMP Tmp = RTSPINLOCKTMP_INITIALIZER;
1030
1031 RTSpinlockAcquire(pThis->hSpinlock, &Tmp);
1032 pDev = (struct net_device *)ASMAtomicUoReadPtr((void * volatile *)&pThis->u.s.pDev);
1033 fRegistered = ASMAtomicUoReadBool(&pThis->u.s.fRegistered);
1034 RTSpinlockRelease(pThis->hSpinlock, &Tmp);
1035 if (fRegistered)
1036 {
1037 dev_remove_pack(&pThis->u.s.PacketType);
1038 skb_queue_purge(&pThis->u.s.XmitQueue);
1039 Log(("vboxNetFltOsDeleteInstance: this=%p: Packet handler removed, xmit queue purged.\n", pThis));
1040 Log(("vboxNetFltOsDeleteInstance: Device %p(%s) released. ref=%d\n", pDev, pDev->name, atomic_read(&pDev->refcnt)));
1041 dev_put(pDev);
1042 }
1043 Log(("vboxNetFltOsDeleteInstance: this=%p: Notifier removed.\n", pThis));
1044 unregister_netdevice_notifier(&pThis->u.s.Notifier);
1045 module_put(THIS_MODULE);
1046}
1047
1048
1049int vboxNetFltOsInitInstance(PVBOXNETFLTINS pThis, void *pvContext)
1050{
1051 int err;
1052 NOREF(pvContext);
1053
1054 pThis->u.s.Notifier.notifier_call = vboxNetFltLinuxNotifierCallback;
1055 err = register_netdevice_notifier(&pThis->u.s.Notifier);
1056 if (err)
1057 return VERR_INTNET_FLT_IF_FAILED;
1058 if (!pThis->u.s.fRegistered)
1059 {
1060 unregister_netdevice_notifier(&pThis->u.s.Notifier);
1061 LogRel(("VBoxNetFlt: failed to find %s.\n", pThis->szName));
1062 return VERR_INTNET_FLT_IF_NOT_FOUND;
1063 }
1064
1065 Log(("vboxNetFltOsInitInstance: this=%p: Notifier installed.\n", pThis));
1066 if ( pThis->fDisconnectedFromHost
1067 || !try_module_get(THIS_MODULE))
1068 return VERR_INTNET_FLT_IF_FAILED;
1069
1070 return VINF_SUCCESS;
1071}
1072
1073int vboxNetFltOsPreInitInstance(PVBOXNETFLTINS pThis)
1074{
1075 /*
1076 * Init the linux specific members.
1077 */
1078 pThis->u.s.pDev = NULL;
1079 pThis->u.s.fRegistered = false;
1080 pThis->u.s.fPromiscuousSet = false;
1081 memset(&pThis->u.s.PacketType, 0, sizeof(pThis->u.s.PacketType));
1082 skb_queue_head_init(&pThis->u.s.XmitQueue);
1083#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 20)
1084 INIT_WORK(&pThis->u.s.XmitTask, vboxNetFltLinuxXmitTask);
1085#else
1086 INIT_WORK(&pThis->u.s.XmitTask, vboxNetFltLinuxXmitTask, &pThis->u.s.XmitTask);
1087#endif
1088
1089 return VINF_SUCCESS;
1090}
1091
Note: See TracBrowser for help on using the repository browser.

© 2025 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette