adapt tun device to zerocopy (#57)

This commit is contained in:
Sijie.Sun
2024-04-25 23:25:37 +08:00
committed by GitHub
parent 3467890270
commit 57c9f11371
15 changed files with 405 additions and 150 deletions

4
Cargo.lock generated
View File

@@ -2518,9 +2518,9 @@ dependencies = [
[[package]] [[package]]
name = "libc" name = "libc"
version = "0.2.152" version = "0.2.153"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "13e3bf6590cbc649f4d1a3eefc9d5d6eb746f5200ffb04e5e142700b8faa56e7" checksum = "9c198f91728a82281a64e1f4f9eeb25d82cb32a5de251c6bd1b5154d63a8e7bd"
[[package]] [[package]]
name = "libloading" name = "libloading"

View File

@@ -46,7 +46,7 @@ chrono = "0.4.35"
gethostname = "0.4.3" gethostname = "0.4.3"
futures = "0.3" futures = { version = "0.3", features = ["bilock", "unstable"] }
tokio = { version = "1", features = ["full"] } tokio = { version = "1", features = ["full"] }
tokio-stream = "0.1" tokio-stream = "0.1"

View File

@@ -6,7 +6,7 @@ use tokio::process::Command;
use super::error::Error; use super::error::Error;
#[async_trait] #[async_trait]
pub trait IfConfiguerTrait { pub trait IfConfiguerTrait: Send + Sync {
async fn add_ipv4_route( async fn add_ipv4_route(
&self, &self,
name: &str, name: &str,

View File

@@ -118,6 +118,13 @@ and the vpn client is in network of 10.14.14.0/24"
#[arg(long, help = "default protocol to use when connecting to peers")] #[arg(long, help = "default protocol to use when connecting to peers")]
default_protocol: Option<String>, default_protocol: Option<String>,
#[arg(
long,
help = "use multi-thread runtime, default is single-thread",
default_value = "false"
)]
multi_thread: bool,
} }
impl From<Cli> for TomlConfigLoader { impl From<Cli> for TomlConfigLoader {
@@ -329,14 +336,8 @@ fn setup_panic_handler() {
})); }));
} }
#[tokio::main(flavor = "current_thread")]
#[tracing::instrument] #[tracing::instrument]
pub async fn main() { pub async fn async_main(cli: Cli) {
setup_panic_handler();
let cli = Cli::parse();
tracing::info!(cli = ?cli, "cli args parsed");
let cfg: TomlConfigLoader = cli.into(); let cfg: TomlConfigLoader = cli.into();
init_logger(&cfg); init_logger(&cfg);
@@ -427,3 +428,24 @@ pub async fn main() {
inst.wait().await; inst.wait().await;
} }
fn main() {
setup_panic_handler();
let cli = Cli::parse();
tracing::info!(cli = ?cli, "cli args parsed");
if cli.multi_thread {
tokio::runtime::Builder::new_multi_thread()
.enable_all()
.build()
.unwrap()
.block_on(async move { async_main(cli).await })
} else {
tokio::runtime::Builder::new_current_thread()
.enable_all()
.build()
.unwrap()
.block_on(async move { async_main(cli).await })
}
}

View File

@@ -1,5 +1,4 @@
use dashmap::DashSet; use std::sync::{Arc, Mutex};
use std::sync::Arc;
use tokio::task::JoinSet; use tokio::task::JoinSet;
use crate::common::global_ctx::ArcGlobalCtx; use crate::common::global_ctx::ArcGlobalCtx;
@@ -11,7 +10,7 @@ pub mod udp_proxy;
#[derive(Debug)] #[derive(Debug)]
struct CidrSet { struct CidrSet {
global_ctx: ArcGlobalCtx, global_ctx: ArcGlobalCtx,
cidr_set: Arc<DashSet<cidr::IpCidr>>, cidr_set: Arc<Mutex<Vec<cidr::IpCidr>>>,
tasks: JoinSet<()>, tasks: JoinSet<()>,
} }
@@ -19,7 +18,7 @@ impl CidrSet {
pub fn new(global_ctx: ArcGlobalCtx) -> Self { pub fn new(global_ctx: ArcGlobalCtx) -> Self {
let mut ret = Self { let mut ret = Self {
global_ctx, global_ctx,
cidr_set: Arc::new(DashSet::new()), cidr_set: Arc::new(Mutex::new(vec![])),
tasks: JoinSet::new(), tasks: JoinSet::new(),
}; };
ret.run_cidr_updater(); ret.run_cidr_updater();
@@ -35,9 +34,9 @@ impl CidrSet {
let cidrs = global_ctx.get_proxy_cidrs(); let cidrs = global_ctx.get_proxy_cidrs();
if cidrs != last_cidrs { if cidrs != last_cidrs {
last_cidrs = cidrs.clone(); last_cidrs = cidrs.clone();
cidr_set.clear(); cidr_set.lock().unwrap().clear();
for cidr in cidrs.iter() { for cidr in cidrs.iter() {
cidr_set.insert(cidr.clone()); cidr_set.lock().unwrap().push(cidr.clone());
} }
} }
tokio::time::sleep(std::time::Duration::from_secs(1)).await; tokio::time::sleep(std::time::Duration::from_secs(1)).await;
@@ -47,10 +46,16 @@ impl CidrSet {
pub fn contains_v4(&self, ip: std::net::Ipv4Addr) -> bool { pub fn contains_v4(&self, ip: std::net::Ipv4Addr) -> bool {
let ip = ip.into(); let ip = ip.into();
return self.cidr_set.iter().any(|cidr| cidr.contains(&ip)); let s = self.cidr_set.lock().unwrap();
for cidr in s.iter() {
if cidr.contains(&ip) {
return true;
}
}
false
} }
pub fn is_empty(&self) -> bool { pub fn is_empty(&self) -> bool {
return self.cidr_set.is_empty(); self.cidr_set.lock().unwrap().is_empty()
} }
} }

View File

@@ -1,13 +1,13 @@
use std::borrow::BorrowMut; use std::borrow::BorrowMut;
use std::net::Ipv4Addr; use std::net::Ipv4Addr;
use std::pin::Pin;
use std::sync::{Arc, Weak}; use std::sync::{Arc, Weak};
use anyhow::Context; use anyhow::Context;
use futures::{SinkExt, StreamExt}; use futures::{SinkExt, StreamExt};
use pnet::packet::ethernet::EthernetPacket;
use pnet::packet::ipv4::Ipv4Packet; use pnet::packet::ipv4::Ipv4Packet;
use bytes::BytesMut;
use tokio::{sync::Mutex, task::JoinSet}; use tokio::{sync::Mutex, task::JoinSet};
use tonic::transport::Server; use tonic::transport::Server;
@@ -30,11 +30,14 @@ use crate::rpc::vpn_portal_rpc_server::VpnPortalRpc;
use crate::rpc::{GetVpnPortalInfoRequest, GetVpnPortalInfoResponse, VpnPortalInfo}; use crate::rpc::{GetVpnPortalInfoRequest, GetVpnPortalInfoResponse, VpnPortalInfo};
use crate::tunnel::packet_def::ZCPacket; use crate::tunnel::packet_def::ZCPacket;
use crate::tunnel::{ZCPacketSink, ZCPacketStream};
use crate::vpn_portal::{self, VpnPortal}; use crate::vpn_portal::{self, VpnPortal};
use super::listeners::ListenerManager; use super::listeners::ListenerManager;
use super::virtual_nic; use super::virtual_nic;
use crate::common::ifcfg::IfConfiguerTrait;
#[derive(Clone)] #[derive(Clone)]
struct IpProxy { struct IpProxy {
tcp_proxy: Arc<TcpProxy>, tcp_proxy: Arc<TcpProxy>,
@@ -156,8 +159,8 @@ impl Instance {
self.conn_manager.clone() self.conn_manager.clone()
} }
async fn do_forward_nic_to_peers_ipv4(ret: BytesMut, mgr: &PeerManager) { async fn do_forward_nic_to_peers_ipv4(ret: ZCPacket, mgr: &PeerManager) {
if let Some(ipv4) = Ipv4Packet::new(&ret) { if let Some(ipv4) = Ipv4Packet::new(ret.payload()) {
if ipv4.get_version() != 4 { if ipv4.get_version() != 4 {
tracing::info!("[USER_PACKET] not ipv4 packet: {:?}", ipv4); tracing::info!("[USER_PACKET] not ipv4 packet: {:?}", ipv4);
return; return;
@@ -169,9 +172,7 @@ impl Instance {
); );
// TODO: use zero-copy // TODO: use zero-copy
let send_ret = mgr let send_ret = mgr.send_msg_ipv4(ret, dst_ipv4).await;
.send_msg_ipv4(ZCPacket::new_with_payload(ret.as_ref()), dst_ipv4)
.await;
if send_ret.is_err() { if send_ret.is_err() {
tracing::trace!(?send_ret, "[USER_PACKET] send_msg_ipv4 failed") tracing::trace!(?send_ret, "[USER_PACKET] send_msg_ipv4 failed")
} }
@@ -180,23 +181,23 @@ impl Instance {
} }
} }
async fn do_forward_nic_to_peers_ethernet(mut ret: BytesMut, mgr: &PeerManager) { // async fn do_forward_nic_to_peers_ethernet(mut ret: BytesMut, mgr: &PeerManager) {
if let Some(eth) = EthernetPacket::new(&ret) { // if let Some(eth) = EthernetPacket::new(&ret) {
log::warn!("begin to forward: {:?}, type: {}", eth, eth.get_ethertype()); // log::warn!("begin to forward: {:?}, type: {}", eth, eth.get_ethertype());
Self::do_forward_nic_to_peers_ipv4(ret.split_off(14), mgr).await; // Self::do_forward_nic_to_peers_ipv4(ret.split_off(14), mgr).await;
} else { // } else {
log::warn!("not ipv4 packet: {:?}", ret); // log::warn!("not ipv4 packet: {:?}", ret);
} // }
} // }
fn do_forward_nic_to_peers(&mut self) -> Result<(), Error> { fn do_forward_nic_to_peers(
&mut self,
mut stream: Pin<Box<dyn ZCPacketStream>>,
) -> Result<(), Error> {
// read from nic and write to corresponding tunnel // read from nic and write to corresponding tunnel
let nic = self.virtual_nic.as_ref().unwrap();
let nic = nic.clone();
let mgr = self.peer_manager.clone(); let mgr = self.peer_manager.clone();
self.tasks.spawn(async move { self.tasks.spawn(async move {
let mut stream = nic.pin_recv_stream();
while let Some(ret) = stream.next().await { while let Some(ret) = stream.next().await {
if ret.is_err() { if ret.is_err() {
log::error!("read from nic failed: {:?}", ret); log::error!("read from nic failed: {:?}", ret);
@@ -212,21 +213,17 @@ impl Instance {
fn do_forward_peers_to_nic( fn do_forward_peers_to_nic(
tasks: &mut JoinSet<()>, tasks: &mut JoinSet<()>,
nic: Arc<virtual_nic::VirtualNic>, mut sink: Pin<Box<dyn ZCPacketSink>>,
channel: Option<PacketRecvChanReceiver>, channel: Option<PacketRecvChanReceiver>,
) { ) {
tasks.spawn(async move { tasks.spawn(async move {
let mut send = nic.pin_send_stream();
let mut channel = channel.unwrap(); let mut channel = channel.unwrap();
while let Some(packet) = channel.recv().await { while let Some(packet) = channel.recv().await {
tracing::trace!( tracing::trace!(
"[USER_PACKET] forward packet from peers to nic. packet: {:?}", "[USER_PACKET] forward packet from peers to nic. packet: {:?}",
packet packet
); );
let mut b = BytesMut::new(); let ret = sink.send(packet).await;
b.extend_from_slice(packet.payload());
let ret = send.send(b.freeze()).await;
if ret.is_err() { if ret.is_err() {
panic!("do_forward_tunnel_to_nic"); panic!("do_forward_tunnel_to_nic");
} }
@@ -244,19 +241,19 @@ impl Instance {
} }
async fn prepare_tun_device(&mut self) -> Result<(), Error> { async fn prepare_tun_device(&mut self) -> Result<(), Error> {
let nic = virtual_nic::VirtualNic::new(self.get_global_ctx()) let mut nic = virtual_nic::VirtualNic::new(self.get_global_ctx());
.create_dev() let tunnel = nic.create_dev().await?;
.await?;
self.global_ctx self.global_ctx
.issue_event(GlobalCtxEvent::TunDeviceReady(nic.ifname().to_string())); .issue_event(GlobalCtxEvent::TunDeviceReady(nic.ifname().to_string()));
let (stream, sink) = tunnel.split();
self.virtual_nic = Some(Arc::new(nic)); self.virtual_nic = Some(Arc::new(nic));
self.do_forward_nic_to_peers().unwrap(); self.do_forward_nic_to_peers(stream).unwrap();
Self::do_forward_peers_to_nic( Self::do_forward_peers_to_nic(
self.tasks.borrow_mut(), self.tasks.borrow_mut(),
self.virtual_nic.as_ref().unwrap().clone(), sink,
self.peer_packet_receiver.take(), self.peer_packet_receiver.take(),
); );
@@ -438,6 +435,8 @@ impl Instance {
let global_ctx = self.global_ctx.clone(); let global_ctx = self.global_ctx.clone();
let net_ns = self.global_ctx.net_ns.clone(); let net_ns = self.global_ctx.net_ns.clone();
let nic = self.virtual_nic.as_ref().unwrap().clone(); let nic = self.virtual_nic.as_ref().unwrap().clone();
let ifcfg = nic.get_ifcfg();
let ifname = nic.ifname().to_owned();
self.tasks.spawn(async move { self.tasks.spawn(async move {
let mut cur_proxy_cidrs = vec![]; let mut cur_proxy_cidrs = vec![];
@@ -464,10 +463,9 @@ impl Instance {
} }
let _g = net_ns.guard(); let _g = net_ns.guard();
let ret = nic let ret = ifcfg
.get_ifcfg()
.remove_ipv4_route( .remove_ipv4_route(
nic.ifname(), ifname.as_str(),
cidr.first_address(), cidr.first_address(),
cidr.network_length(), cidr.network_length(),
) )
@@ -487,9 +485,12 @@ impl Instance {
continue; continue;
} }
let _g = net_ns.guard(); let _g = net_ns.guard();
let ret = nic let ret = ifcfg
.get_ifcfg() .add_ipv4_route(
.add_ipv4_route(nic.ifname(), cidr.first_address(), cidr.network_length()) ifname.as_str(),
cidr.first_address(),
cidr.network_length(),
)
.await; .await;
if ret.is_err() { if ret.is_err() {

View File

@@ -1,21 +1,207 @@
use std::{net::Ipv4Addr, pin::Pin}; use std::{
io,
net::Ipv4Addr,
pin::Pin,
task::{Context, Poll},
};
use crate::{ use crate::{
common::{ common::{
error::Result, error::Error,
global_ctx::ArcGlobalCtx, global_ctx::ArcGlobalCtx,
ifcfg::{IfConfiger, IfConfiguerTrait}, ifcfg::{IfConfiger, IfConfiguerTrait},
}, },
tunnels::{ tunnel::{
codec::BytesCodec, common::FramedTunnel, DatagramSink, DatagramStream, Tunnel, TunnelError, common::{FramedWriter, TunnelWrapper, ZCPacketToBytes},
packet_def::ZCPacket,
StreamItem, Tunnel, TunnelError,
}, },
}; };
use futures::{SinkExt, StreamExt}; use byteorder::WriteBytesExt as _;
use tokio_util::{bytes::Bytes, codec::Framed}; use futures::{lock::BiLock, ready, Stream};
use tun::Device; use pin_project_lite::pin_project;
use tokio::io::AsyncWrite;
use tokio_util::{bytes::Bytes, io::poll_read_buf};
use tun::{create_as_async, AsyncDevice, Configuration, Device as _, Layer};
use zerocopy::{NativeEndian, NetworkEndian};
use super::tun_codec::{TunPacket, TunPacketCodec}; pin_project! {
pub struct TunStream {
#[pin]
l: BiLock<AsyncDevice>,
cur_packet: Option<ZCPacket>,
}
}
impl TunStream {
pub fn new(l: BiLock<AsyncDevice>) -> Self {
Self {
l,
cur_packet: None,
}
}
}
impl Stream for TunStream {
type Item = StreamItem;
fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<StreamItem>> {
let self_mut = self.project();
let mut g = ready!(self_mut.l.poll_lock(cx));
if self_mut.cur_packet.is_none() {
*self_mut.cur_packet = Some(ZCPacket::new_with_reserved_payload(2048));
}
let cur_packet = self_mut.cur_packet.as_mut().unwrap();
match ready!(poll_read_buf(
g.as_pin_mut(),
cx,
&mut cur_packet.mut_inner()
)) {
Ok(0) => Poll::Ready(None),
Ok(_n) => Poll::Ready(Some(Ok(self_mut.cur_packet.take().unwrap()))),
Err(err) => {
println!("tun stream error: {:?}", err);
Poll::Ready(None)
}
}
}
}
#[derive(Debug, Clone, Copy, Default)]
enum PacketProtocol {
#[default]
IPv4,
IPv6,
Other(u8),
}
// Note: the protocol in the packet information header is platform dependent.
impl PacketProtocol {
#[cfg(any(target_os = "linux", target_os = "android"))]
fn into_pi_field(self) -> Result<u16, io::Error> {
use nix::libc;
match self {
PacketProtocol::IPv4 => Ok(libc::ETH_P_IP as u16),
PacketProtocol::IPv6 => Ok(libc::ETH_P_IPV6 as u16),
PacketProtocol::Other(_) => Err(io::Error::new(
io::ErrorKind::Other,
"neither an IPv4 nor IPv6 packet",
)),
}
}
#[cfg(any(target_os = "macos", target_os = "ios"))]
fn into_pi_field(self) -> Result<u16, io::Error> {
use nix::libc;
match self {
PacketProtocol::IPv4 => Ok(libc::PF_INET as u16),
PacketProtocol::IPv6 => Ok(libc::PF_INET6 as u16),
PacketProtocol::Other(_) => Err(io::Error::new(
io::ErrorKind::Other,
"neither an IPv4 nor IPv6 packet",
)),
}
}
#[cfg(target_os = "windows")]
fn into_pi_field(self) -> Result<u16, io::Error> {
unimplemented!()
}
}
/// Infer the protocol based on the first nibble in the packet buffer.
fn infer_proto(buf: &[u8]) -> PacketProtocol {
match buf[0] >> 4 {
4 => PacketProtocol::IPv4,
6 => PacketProtocol::IPv6,
p => PacketProtocol::Other(p),
}
}
struct TunZCPacketToBytes {
has_packet_info: bool,
}
impl TunZCPacketToBytes {
pub fn new(has_packet_info: bool) -> Self {
Self { has_packet_info }
}
pub fn fill_packet_info(&self, mut buf: &mut [u8]) -> Result<(), io::Error> {
// flags is always 0
buf.write_u16::<NativeEndian>(0)?;
// write the protocol as network byte order
buf.write_u16::<NetworkEndian>(infer_proto(&buf).into_pi_field()?)?;
Ok(())
}
}
impl ZCPacketToBytes for TunZCPacketToBytes {
fn into_bytes(&self, zc_packet: ZCPacket) -> Result<Bytes, TunnelError> {
let payload_offset = zc_packet.payload_offset();
let mut inner = zc_packet.inner();
// we have peer manager header, so payload offset must larger than 4
assert!(payload_offset >= 4);
let ret = if self.has_packet_info {
let mut inner = inner.split_off(payload_offset - 4);
self.fill_packet_info(&mut inner[0..4])?;
inner
} else {
inner.split_off(payload_offset)
};
tracing::debug!(?ret, ?payload_offset, "convert zc packet to tun packet");
Ok(ret.into())
}
}
pin_project! {
pub struct TunAsyncWrite {
#[pin]
l: BiLock<AsyncDevice>,
}
}
impl AsyncWrite for TunAsyncWrite {
fn poll_write(
self: Pin<&mut Self>,
cx: &mut Context<'_>,
buf: &[u8],
) -> Poll<Result<usize, io::Error>> {
let self_mut = self.project();
let mut g = ready!(self_mut.l.poll_lock(cx));
g.as_pin_mut().poll_write(cx, buf)
}
fn poll_flush(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> {
let self_mut = self.project();
let mut g = ready!(self_mut.l.poll_lock(cx));
g.as_pin_mut().poll_flush(cx)
}
fn poll_shutdown(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Result<(), io::Error>> {
let self_mut = self.project();
let mut g = ready!(self_mut.l.poll_lock(cx));
g.as_pin_mut().poll_shutdown(cx)
}
fn poll_write_vectored(
self: Pin<&mut Self>,
cx: &mut Context<'_>,
bufs: &[io::IoSlice<'_>],
) -> Poll<Result<usize, io::Error>> {
let self_mut = self.project();
let mut g = ready!(self_mut.l.poll_lock(cx));
g.as_pin_mut().poll_write_vectored(cx, bufs)
}
fn is_write_vectored(&self) -> bool {
true
}
}
pub struct VirtualNic { pub struct VirtualNic {
dev_name: String, dev_name: String,
@@ -24,7 +210,6 @@ pub struct VirtualNic {
global_ctx: ArcGlobalCtx, global_ctx: ArcGlobalCtx,
ifname: Option<String>, ifname: Option<String>,
tun: Option<Box<dyn Tunnel>>,
ifcfg: Box<dyn IfConfiguerTrait + Send + Sync + 'static>, ifcfg: Box<dyn IfConfiguerTrait + Send + Sync + 'static>,
} }
@@ -35,25 +220,24 @@ impl VirtualNic {
queue_num: 1, queue_num: 1,
global_ctx, global_ctx,
ifname: None, ifname: None,
tun: None,
ifcfg: Box::new(IfConfiger {}), ifcfg: Box::new(IfConfiger {}),
} }
} }
pub fn set_dev_name(mut self, dev_name: &str) -> Result<Self> { pub fn set_dev_name(mut self, dev_name: &str) -> Result<Self, Error> {
self.dev_name = dev_name.to_owned(); self.dev_name = dev_name.to_owned();
Ok(self) Ok(self)
} }
pub fn set_queue_num(mut self, queue_num: usize) -> Result<Self> { pub fn set_queue_num(mut self, queue_num: usize) -> Result<Self, Error> {
self.queue_num = queue_num; self.queue_num = queue_num;
Ok(self) Ok(self)
} }
async fn create_dev_ret_err(&mut self) -> Result<()> { async fn create_dev_ret_err(&mut self) -> Result<Box<dyn Tunnel>, Error> {
let mut config = tun::Configuration::default(); let mut config = Configuration::default();
let has_packet_info = cfg!(target_os = "macos"); let has_packet_info = cfg!(target_os = "macos");
config.layer(tun::Layer::L3); config.layer(Layer::L3);
#[cfg(target_os = "linux")] #[cfg(target_os = "linux")]
{ {
@@ -71,61 +255,42 @@ impl VirtualNic {
let dev = { let dev = {
let _g = self.global_ctx.net_ns.guard(); let _g = self.global_ctx.net_ns.guard();
tun::create_as_async(&config)? create_as_async(&config)?
}; };
let ifname = dev.get_ref().name()?; let ifname = dev.get_ref().name()?;
self.ifcfg.wait_interface_show(ifname.as_str()).await?; self.ifcfg.wait_interface_show(ifname.as_str()).await?;
let ft: Box<dyn Tunnel> = if has_packet_info { let (a, b) = BiLock::new(dev);
let framed = Framed::new(dev, TunPacketCodec::new(true, 2500));
let (sink, stream) = framed.split();
let new_stream = stream.map(|item| match item { let ft = TunnelWrapper::new(
Ok(item) => Ok(item.into_bytes_mut()), TunStream::new(a),
Err(err) => { FramedWriter::new_with_converter(
println!("tun stream error: {:?}", err); TunAsyncWrite { l: b },
Err(TunnelError::TunError(err.to_string())) TunZCPacketToBytes::new(has_packet_info),
} ),
}); None,
);
let new_sink = Box::pin(sink.with(|item: Bytes| async move {
if false {
return Err(TunnelError::TunError("tun sink error".to_owned()));
}
Ok(TunPacket::new(super::tun_codec::TunPacketBuffer::Bytes(
item,
)))
}));
Box::new(FramedTunnel::new(new_stream, new_sink, None))
} else {
let framed = Framed::new(dev, BytesCodec::new(2500));
let (sink, stream) = framed.split();
Box::new(FramedTunnel::new(stream, sink, None))
};
self.ifname = Some(ifname.to_owned()); self.ifname = Some(ifname.to_owned());
self.tun = Some(ft); Ok(Box::new(ft))
Ok(())
} }
pub async fn create_dev(mut self) -> Result<Self> { pub async fn create_dev(&mut self) -> Result<Box<dyn Tunnel>, Error> {
self.create_dev_ret_err().await?; self.create_dev_ret_err().await
Ok(self)
} }
pub fn ifname(&self) -> &str { pub fn ifname(&self) -> &str {
self.ifname.as_ref().unwrap().as_str() self.ifname.as_ref().unwrap().as_str()
} }
pub async fn link_up(&self) -> Result<()> { pub async fn link_up(&self) -> Result<(), Error> {
let _g = self.global_ctx.net_ns.guard(); let _g = self.global_ctx.net_ns.guard();
self.ifcfg.set_link_status(self.ifname(), true).await?; self.ifcfg.set_link_status(self.ifname(), true).await?;
Ok(()) Ok(())
} }
pub async fn add_route(&self, address: Ipv4Addr, cidr: u8) -> Result<()> { pub async fn add_route(&self, address: Ipv4Addr, cidr: u8) -> Result<(), Error> {
let _g = self.global_ctx.net_ns.guard(); let _g = self.global_ctx.net_ns.guard();
self.ifcfg self.ifcfg
.add_ipv4_route(self.ifname(), address, cidr) .add_ipv4_route(self.ifname(), address, cidr)
@@ -133,13 +298,13 @@ impl VirtualNic {
Ok(()) Ok(())
} }
pub async fn remove_ip(&self, ip: Option<Ipv4Addr>) -> Result<()> { pub async fn remove_ip(&self, ip: Option<Ipv4Addr>) -> Result<(), Error> {
let _g = self.global_ctx.net_ns.guard(); let _g = self.global_ctx.net_ns.guard();
self.ifcfg.remove_ip(self.ifname(), ip).await?; self.ifcfg.remove_ip(self.ifname(), ip).await?;
Ok(()) Ok(())
} }
pub async fn add_ip(&self, ip: Ipv4Addr, cidr: i32) -> Result<()> { pub async fn add_ip(&self, ip: Ipv4Addr, cidr: i32) -> Result<(), Error> {
let _g = self.global_ctx.net_ns.guard(); let _g = self.global_ctx.net_ns.guard();
self.ifcfg self.ifcfg
.add_ipv4_ip(self.ifname(), ip, cidr as u8) .add_ipv4_ip(self.ifname(), ip, cidr as u8)
@@ -147,16 +312,8 @@ impl VirtualNic {
Ok(()) Ok(())
} }
pub fn pin_recv_stream(&self) -> Pin<Box<dyn DatagramStream>> { pub fn get_ifcfg(&self) -> impl IfConfiguerTrait {
self.tun.as_ref().unwrap().pin_stream() IfConfiger {}
}
pub fn pin_send_stream(&self) -> Pin<Box<dyn DatagramSink>> {
self.tun.as_ref().unwrap().pin_sink()
}
pub fn get_ifcfg(&self) -> &dyn IfConfiguerTrait {
self.ifcfg.as_ref()
} }
} }
#[cfg(test)] #[cfg(test)]
@@ -166,7 +323,8 @@ mod tests {
use super::VirtualNic; use super::VirtualNic;
async fn run_test_helper() -> Result<VirtualNic, Error> { async fn run_test_helper() -> Result<VirtualNic, Error> {
let dev = VirtualNic::new(get_mock_global_ctx()).create_dev().await?; let mut dev = VirtualNic::new(get_mock_global_ctx());
let _tunnel = dev.create_dev().await?;
tokio::time::sleep(tokio::time::Duration::from_secs(1)).await; tokio::time::sleep(tokio::time::Duration::from_secs(1)).await;

View File

@@ -1,5 +1,6 @@
use std::sync::Arc; use std::sync::Arc;
use crossbeam::atomic::AtomicCell;
use dashmap::DashMap; use dashmap::DashMap;
use tokio::{ use tokio::{
@@ -38,6 +39,8 @@ pub struct Peer {
close_event_listener: JoinHandle<()>, close_event_listener: JoinHandle<()>,
shutdown_notifier: Arc<tokio::sync::Notify>, shutdown_notifier: Arc<tokio::sync::Notify>,
default_conn_id: AtomicCell<PeerConnId>,
} }
impl Peer { impl Peer {
@@ -99,6 +102,7 @@ impl Peer {
close_event_listener, close_event_listener,
shutdown_notifier, shutdown_notifier,
default_conn_id: AtomicCell::new(PeerConnId::default()),
} }
} }
@@ -112,8 +116,24 @@ impl Peer {
.insert(conn.get_conn_id(), Arc::new(Mutex::new(conn))); .insert(conn.get_conn_id(), Arc::new(Mutex::new(conn)));
} }
async fn select_conn(&self) -> Option<ArcPeerConn> {
let default_conn_id = self.default_conn_id.load();
if let Some(conn) = self.conns.get(&default_conn_id) {
return Some(conn.clone());
}
let conn = self.conns.iter().next();
if conn.is_none() {
return None;
}
let conn = conn.unwrap().clone();
self.default_conn_id.store(conn.lock().await.get_conn_id());
Some(conn)
}
pub async fn send_msg(&self, msg: ZCPacket) -> Result<(), Error> { pub async fn send_msg(&self, msg: ZCPacket) -> Result<(), Error> {
let Some(conn) = self.conns.iter().next() else { let Some(conn) = self.select_conn().await else {
return Err(Error::PeerNoConnectionError(self.peer_node_id)); return Err(Error::PeerNoConnectionError(self.peer_node_id));
}; };

View File

@@ -118,12 +118,6 @@ impl PeerMap {
pub async fn send_msg(&self, msg: ZCPacket, dst_peer_id: PeerId) -> Result<(), Error> { pub async fn send_msg(&self, msg: ZCPacket, dst_peer_id: PeerId) -> Result<(), Error> {
let Some(gateway_peer_id) = self.get_gateway_peer_id(dst_peer_id).await else { let Some(gateway_peer_id) = self.get_gateway_peer_id(dst_peer_id).await else {
tracing::trace!(
"no gateway for dst_peer_id: {}, peers: {:?}, my_peer_id: {}",
dst_peer_id,
self.peer_map.iter().map(|v| *v.key()).collect::<Vec<_>>(),
self.my_peer_id
);
return Err(Error::RouteError(Some(format!( return Err(Error::RouteError(Some(format!(
"peer map sengmsg no gateway for dst_peer_id: {}", "peer map sengmsg no gateway for dst_peer_id: {}",
dst_peer_id dst_peer_id

View File

@@ -1265,7 +1265,7 @@ impl Route for PeerRoute {
return Some(peer_id); return Some(peer_id);
} }
tracing::info!("no peer id for ipv4: {}", ipv4_addr); tracing::info!(?ipv4_addr, "no peer id for ipv4");
None None
} }
} }

View File

@@ -153,7 +153,7 @@ pub async fn basic_three_node_test(#[values("tcp", "udp", "wg")] proto: &str) {
wait_for_condition( wait_for_condition(
|| async { ping_test("net_c", "10.144.144.1").await }, || async { ping_test("net_c", "10.144.144.1").await },
Duration::from_secs(5), Duration::from_secs(5000),
) )
.await; .await;
} }

View File

@@ -167,16 +167,41 @@ where
} }
} }
pub trait ZCPacketToBytes {
fn into_bytes(&self, zc_packet: ZCPacket) -> Result<Bytes, TunnelError>;
}
pub struct TcpZCPacketToBytes;
impl ZCPacketToBytes for TcpZCPacketToBytes {
fn into_bytes(&self, mut item: ZCPacket) -> Result<Bytes, TunnelError> {
let tcp_len = PEER_MANAGER_HEADER_SIZE + item.payload_len();
let Some(header) = item.mut_tcp_tunnel_header() else {
return Err(TunnelError::InvalidPacket("packet too short".to_string()));
};
header.len.set(tcp_len.try_into().unwrap());
Ok(item.into_bytes(ZCPacketType::TCP))
}
}
pin_project! { pin_project! {
pub struct FramedWriter<W> { pub struct FramedWriter<W, C> {
#[pin] #[pin]
writer: W, writer: W,
sending_bufs: BufList<Bytes>, sending_bufs: BufList<Bytes>,
associate_data: Option<Box<dyn Any + Send + 'static>>, associate_data: Option<Box<dyn Any + Send + 'static>>,
converter: C,
} }
} }
impl<W> FramedWriter<W> { impl<W, C> FramedWriter<W, C> {
fn max_buffer_count(&self) -> usize {
64
}
}
impl<W> FramedWriter<W, TcpZCPacketToBytes> {
pub fn new(writer: W) -> Self { pub fn new(writer: W) -> Self {
Self::new_with_associate_data(writer, None) Self::new_with_associate_data(writer, None)
} }
@@ -188,18 +213,35 @@ impl<W> FramedWriter<W> {
FramedWriter { FramedWriter {
writer, writer,
sending_bufs: BufList::new(), sending_bufs: BufList::new(),
associate_data: associate_data, associate_data,
converter: TcpZCPacketToBytes {},
} }
} }
fn max_buffer_count(&self) -> usize {
64
}
} }
impl<W> Sink<SinkItem> for FramedWriter<W> impl<W, C: ZCPacketToBytes + Send + 'static> FramedWriter<W, C> {
pub fn new_with_converter(writer: W, converter: C) -> Self {
Self::new_with_converter_and_associate_data(writer, converter, None)
}
pub fn new_with_converter_and_associate_data(
writer: W,
converter: C,
associate_data: Option<Box<dyn Any + Send + 'static>>,
) -> Self {
FramedWriter {
writer,
sending_bufs: BufList::new(),
associate_data,
converter,
}
}
}
impl<W, C> Sink<SinkItem> for FramedWriter<W, C>
where where
W: AsyncWrite + Send + 'static, W: AsyncWrite + Send + 'static,
C: ZCPacketToBytes + Send + 'static,
{ {
type Error = TunnelError; type Error = TunnelError;
@@ -216,15 +258,9 @@ where
} }
} }
fn start_send(self: Pin<&mut Self>, mut item: ZCPacket) -> Result<(), Self::Error> { fn start_send(self: Pin<&mut Self>, item: ZCPacket) -> Result<(), Self::Error> {
let tcp_len = PEER_MANAGER_HEADER_SIZE + item.payload_len(); let pinned = self.project();
let Some(header) = item.mut_tcp_tunnel_header() else { pinned.sending_bufs.push(pinned.converter.into_bytes(item)?);
return Err(TunnelError::InvalidPacket("packet too short".to_string()));
};
header.len.set(tcp_len.try_into().unwrap());
let item = item.into_bytes(ZCPacketType::TCP);
self.project().sending_bufs.push(item);
Ok(()) Ok(())
} }

View File

@@ -52,6 +52,9 @@ pub enum TunnelError {
#[error("shutdown")] #[error("shutdown")]
Shutdown, Shutdown,
#[error("tunnel error: {0}")]
TunError(String),
} }
pub type StreamT = packet_def::ZCPacket; pub type StreamT = packet_def::ZCPacket;

View File

@@ -161,12 +161,25 @@ impl ZCPacket {
ret ret
} }
pub fn new_with_reserved_payload(cap: usize) -> Self {
let mut ret = Self::new_nic_packet();
ret.inner.reserve(cap);
let total_len = ret.packet_type.get_packet_offsets().payload_offset;
ret.inner.resize(total_len, 0);
ret
}
pub fn packet_type(&self) -> ZCPacketType { pub fn packet_type(&self) -> ZCPacketType {
self.packet_type self.packet_type
} }
pub fn payload_offset(&self) -> usize {
self.packet_type.get_packet_offsets().payload_offset
}
pub fn mut_payload(&mut self) -> &mut [u8] { pub fn mut_payload(&mut self) -> &mut [u8] {
&mut self.inner[self.packet_type.get_packet_offsets().payload_offset..] let offset = self.payload_offset();
&mut self.inner[offset..]
} }
pub fn mut_peer_manager_header(&mut self) -> Option<&mut PeerManagerHeader> { pub fn mut_peer_manager_header(&mut self) -> Option<&mut PeerManagerHeader> {
@@ -207,7 +220,7 @@ impl ZCPacket {
// ref versions // ref versions
pub fn payload(&self) -> &[u8] { pub fn payload(&self) -> &[u8] {
&self.inner[self.packet_type.get_packet_offsets().payload_offset..] &self.inner[self.payload_offset()..]
} }
pub fn peer_manager_header(&self) -> Option<&PeerManagerHeader> { pub fn peer_manager_header(&self) -> Option<&PeerManagerHeader> {
@@ -246,8 +259,7 @@ impl ZCPacket {
} }
pub fn payload_len(&self) -> usize { pub fn payload_len(&self) -> usize {
let payload_offset = self.packet_type.get_packet_offsets().payload_offset; self.inner.len() - self.payload_offset()
self.inner.len() - payload_offset
} }
pub fn buf_len(&self) -> usize { pub fn buf_len(&self) -> usize {
@@ -307,6 +319,10 @@ impl ZCPacket {
pub fn inner(self) -> BytesMut { pub fn inner(self) -> BytesMut {
self.inner self.inner
} }
pub fn mut_inner(&mut self) -> &mut BytesMut {
&mut self.inner
}
} }
#[cfg(test)] #[cfg(test)]

View File

@@ -176,8 +176,8 @@ impl RingSink {
return Err(TunnelError::Shutdown); return Err(TunnelError::Shutdown);
} }
log::trace!("id: {}, send buffer, buf: {:?}", self.tunnel.id(), &item); tracing::trace!(id=?self.tunnel.id(), ?item, "send buffer");
self.tunnel.ring.push(item).unwrap(); let _ = self.tunnel.ring.push(item);
self.tunnel.notify_new_item(); self.tunnel.notify_new_item();
Ok(()) Ok(())