mirror of
https://github.com/Bluemangoo/sekai-unpacker.git
synced 2026-05-06 20:44:47 +08:00
Compare commits
3 Commits
de7b63d366
...
cdf617b5c3
| Author | SHA1 | Date | |
|---|---|---|---|
| cdf617b5c3 | |||
| 46cf5cdf4b | |||
| adc3afc7d5 |
26
Cargo.lock
generated
26
Cargo.lock
generated
@ -126,6 +126,17 @@ dependencies = [
|
|||||||
"yaml_serde",
|
"yaml_serde",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "async-http-proxy"
|
||||||
|
version = "1.2.5"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "29faa5d4d308266048bd7505ba55484315a890102f9345b9ff4b87de64201592"
|
||||||
|
dependencies = [
|
||||||
|
"httparse",
|
||||||
|
"thiserror 1.0.69",
|
||||||
|
"tokio",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "async-lock"
|
name = "async-lock"
|
||||||
version = "3.4.1"
|
version = "3.4.1"
|
||||||
@ -444,6 +455,7 @@ name = "communicator"
|
|||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anyhow",
|
"anyhow",
|
||||||
|
"async-http-proxy",
|
||||||
"bytes",
|
"bytes",
|
||||||
"futures",
|
"futures",
|
||||||
"futures-util",
|
"futures-util",
|
||||||
@ -455,7 +467,9 @@ dependencies = [
|
|||||||
"serde_json",
|
"serde_json",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-rustls",
|
"tokio-rustls",
|
||||||
|
"tokio-socks",
|
||||||
"tokio-util",
|
"tokio-util",
|
||||||
|
"url",
|
||||||
"webpki-roots",
|
"webpki-roots",
|
||||||
]
|
]
|
||||||
|
|
||||||
@ -2715,6 +2729,18 @@ dependencies = [
|
|||||||
"tokio",
|
"tokio",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "tokio-socks"
|
||||||
|
version = "0.5.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "0d4770b8024672c1101b3f6733eab95b18007dbe0847a8afe341fcf79e06043f"
|
||||||
|
dependencies = [
|
||||||
|
"either",
|
||||||
|
"futures-util",
|
||||||
|
"thiserror 1.0.69",
|
||||||
|
"tokio",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "tokio-util"
|
name = "tokio-util"
|
||||||
version = "0.7.18"
|
version = "0.7.18"
|
||||||
|
|||||||
@ -34,4 +34,8 @@ lazy_static = "1.5.0"
|
|||||||
structopt = "0.3.26"
|
structopt = "0.3.26"
|
||||||
tokio-util = "0.7.18"
|
tokio-util = "0.7.18"
|
||||||
futures-util = "0.3.32"
|
futures-util = "0.3.32"
|
||||||
twox-hash = "2.1.2"
|
twox-hash = "2.1.2"
|
||||||
|
futures = "0.3.32"
|
||||||
|
tokio-socks = "0.5.2"
|
||||||
|
url = "2.5.8"
|
||||||
|
async-http-proxy = "1.2.5"
|
||||||
@ -183,7 +183,6 @@ async fn main() -> anyhow::Result<()> {
|
|||||||
let liveness_tx = liveness_tx.clone();
|
let liveness_tx = liveness_tx.clone();
|
||||||
join_set.spawn(async move {
|
join_set.spawn(async move {
|
||||||
let _guard = liveness_tx;
|
let _guard = liveness_tx;
|
||||||
let mut inner_set = JoinSet::new();
|
|
||||||
loop {
|
loop {
|
||||||
if cancel_token.is_cancelled() {
|
if cancel_token.is_cancelled() {
|
||||||
return;
|
return;
|
||||||
@ -205,7 +204,7 @@ async fn main() -> anyhow::Result<()> {
|
|||||||
let local_manifest = local_manifest.clone();
|
let local_manifest = local_manifest.clone();
|
||||||
let signal = signal.clone();
|
let signal = signal.clone();
|
||||||
|
|
||||||
inner_set.spawn(async move {
|
tokio::spawn(async move {
|
||||||
loop {
|
loop {
|
||||||
if client.get_client().await.is_err() {
|
if client.get_client().await.is_err() {
|
||||||
return;
|
return;
|
||||||
|
|||||||
@ -103,6 +103,23 @@ impl<T> SharedQueue<T> {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn clear(&self) {
|
||||||
|
let mut queue = self.inner.data.lock().unwrap();
|
||||||
|
|
||||||
|
let removed_count = queue.len();
|
||||||
|
|
||||||
|
if removed_count > 0 {
|
||||||
|
queue.clear();
|
||||||
|
|
||||||
|
let prev_pending = self.inner.pending.fetch_sub(removed_count, Ordering::SeqCst);
|
||||||
|
|
||||||
|
if prev_pending == removed_count {
|
||||||
|
self.inner.done_cond.notify_all();
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// 阻塞当前线程,直到所有在途任务(pending == 0)全部处理完
|
/// 阻塞当前线程,直到所有在途任务(pending == 0)全部处理完
|
||||||
pub fn wait_until_all_consumed(&self) {
|
pub fn wait_until_all_consumed(&self) {
|
||||||
let mut _queue_lock = self.inner.data.lock().unwrap();
|
let mut _queue_lock = self.inner.data.lock().unwrap();
|
||||||
|
|||||||
@ -55,6 +55,7 @@ pub async fn post_run(
|
|||||||
return Ok(None);
|
return Ok(None);
|
||||||
}
|
}
|
||||||
cnt.reset();
|
cnt.reset();
|
||||||
|
queue.clear();
|
||||||
queue.push_all(tasks);
|
queue.push_all(tasks);
|
||||||
Ok(Some(id))
|
Ok(Some(id))
|
||||||
}
|
}
|
||||||
@ -121,11 +122,6 @@ pub async fn run_main(
|
|||||||
match r {
|
match r {
|
||||||
Ok(Ok(())) => cnt.inc_success(),
|
Ok(Ok(())) => cnt.inc_success(),
|
||||||
Ok(Err(e)) => {
|
Ok(Err(e)) => {
|
||||||
if e.to_string()
|
|
||||||
.contains("Session did not reconnect within 15s")
|
|
||||||
{
|
|
||||||
return Err(anyhow!(e));
|
|
||||||
}
|
|
||||||
error!("{}", e);
|
error!("{}", e);
|
||||||
cnt.inc_failure()
|
cnt.inc_failure()
|
||||||
}
|
}
|
||||||
@ -163,12 +159,21 @@ pub async fn run_side(
|
|||||||
let n = p1.concurrent.unwrap_or(5);
|
let n = p1.concurrent.unwrap_or(5);
|
||||||
let semaphore = Arc::new(Semaphore::new(n));
|
let semaphore = Arc::new(Semaphore::new(n));
|
||||||
let mut join_set = JoinSet::new();
|
let mut join_set = JoinSet::new();
|
||||||
|
|
||||||
|
let cancel_token = CancellationToken::new();
|
||||||
while let Some(task) = queue.try_pop() {
|
while let Some(task) = queue.try_pop() {
|
||||||
|
if cancel_token.is_cancelled() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
let permit = semaphore.clone().acquire_owned().await?;
|
let permit = semaphore.clone().acquire_owned().await?;
|
||||||
|
if cancel_token.is_cancelled() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
let client = client.clone();
|
let client = client.clone();
|
||||||
let id = id.clone();
|
let id = id.clone();
|
||||||
let local_manifest = manifest.clone();
|
let local_manifest = manifest.clone();
|
||||||
let p1 = p1.clone();
|
let p1 = p1.clone();
|
||||||
|
let cancel_token = cancel_token.clone();
|
||||||
|
|
||||||
join_set.spawn(async move {
|
join_set.spawn(async move {
|
||||||
let guard = task;
|
let guard = task;
|
||||||
@ -185,6 +190,11 @@ pub async fn run_side(
|
|||||||
let mut retry_conn = client.get_client().await?;
|
let mut retry_conn = client.get_client().await?;
|
||||||
result = download(&mut retry_conn, &req, &p1).await;
|
result = download(&mut retry_conn, &req, &p1).await;
|
||||||
}
|
}
|
||||||
|
if let Err(e) = &result
|
||||||
|
&& e.downcast_ref::<h2::Error>().is_some()
|
||||||
|
{
|
||||||
|
cancel_token.cancel();
|
||||||
|
}
|
||||||
result?;
|
result?;
|
||||||
|
|
||||||
local_manifest
|
local_manifest
|
||||||
|
|||||||
@ -17,4 +17,7 @@ webpki-roots = { workspace = true }
|
|||||||
anyhow = { workspace = true }
|
anyhow = { workspace = true }
|
||||||
tokio-util = { workspace = true }
|
tokio-util = { workspace = true }
|
||||||
futures-util = { workspace = true }
|
futures-util = { workspace = true }
|
||||||
futures = "0.3.32"
|
futures = { workspace = true }
|
||||||
|
tokio-socks = { workspace = true }
|
||||||
|
url = {workspace = true}
|
||||||
|
async-http-proxy = {workspace = true, features = ["tokio", "runtime-tokio"]}
|
||||||
|
|||||||
@ -34,6 +34,7 @@ pub struct TcpClientTunnelConfig {
|
|||||||
pub host: Option<String>,
|
pub host: Option<String>,
|
||||||
pub url: String,
|
pub url: String,
|
||||||
pub token: String,
|
pub token: String,
|
||||||
|
pub proxy: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl TcpClientTunnelConfig {
|
impl TcpClientTunnelConfig {
|
||||||
@ -43,6 +44,7 @@ impl TcpClientTunnelConfig {
|
|||||||
host: self.host,
|
host: self.host,
|
||||||
url: self.url,
|
url: self.url,
|
||||||
token: self.token,
|
token: self.token,
|
||||||
|
proxy: self.proxy,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,10 +1,11 @@
|
|||||||
use anyhow::anyhow;
|
use anyhow::anyhow;
|
||||||
|
use async_http_proxy::http_connect_tokio;
|
||||||
use bytes::Bytes;
|
use bytes::Bytes;
|
||||||
use h2::{RecvStream, client, server};
|
use h2::{RecvStream, client, server};
|
||||||
use http::Request;
|
use http::Request;
|
||||||
use log::{debug, error, info};
|
use log::{debug, error, info};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::collections::{HashMap, HashSet};
|
use std::collections::HashMap;
|
||||||
use std::error::Error;
|
use std::error::Error;
|
||||||
use std::fs::File;
|
use std::fs::File;
|
||||||
use std::io::BufReader;
|
use std::io::BufReader;
|
||||||
@ -17,6 +18,8 @@ use tokio::time::{Duration, sleep, timeout};
|
|||||||
use tokio_rustls::rustls;
|
use tokio_rustls::rustls;
|
||||||
use tokio_rustls::rustls::pki_types::{CertificateDer, ServerName};
|
use tokio_rustls::rustls::pki_types::{CertificateDer, ServerName};
|
||||||
use tokio_rustls::{TlsAcceptor, TlsConnector};
|
use tokio_rustls::{TlsAcceptor, TlsConnector};
|
||||||
|
use tokio_socks::tcp::Socks5Stream;
|
||||||
|
use url::Url;
|
||||||
|
|
||||||
#[derive(Debug, Copy, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
#[derive(Debug, Copy, Clone, PartialEq, Eq, Serialize, Deserialize)]
|
||||||
pub enum Identity {
|
pub enum Identity {
|
||||||
@ -39,6 +42,7 @@ pub struct ClientTunnelConfig {
|
|||||||
pub url: String,
|
pub url: String,
|
||||||
pub token: String,
|
pub token: String,
|
||||||
pub identity: Identity,
|
pub identity: Identity,
|
||||||
|
pub proxy: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Default for ClientTunnelConfig {
|
impl Default for ClientTunnelConfig {
|
||||||
@ -48,6 +52,7 @@ impl Default for ClientTunnelConfig {
|
|||||||
url: "127.0.0.1:3333".to_string(),
|
url: "127.0.0.1:3333".to_string(),
|
||||||
token: "super_secret_magic_token".to_string(),
|
token: "super_secret_magic_token".to_string(),
|
||||||
identity: Identity::Client,
|
identity: Identity::Client,
|
||||||
|
proxy: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -95,6 +100,30 @@ pub enum TunnelEndpoint {
|
|||||||
Server(Arc<ServerManager>),
|
Server(Arc<ServerManager>),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
pub enum WeakTunnelEndpoint {
|
||||||
|
Client(std::sync::Weak<ClientManager>),
|
||||||
|
Server(std::sync::Weak<ServerManager>),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl TunnelEndpoint {
|
||||||
|
pub fn downgrade(&self) -> WeakTunnelEndpoint {
|
||||||
|
match self {
|
||||||
|
Self::Client(c) => WeakTunnelEndpoint::Client(Arc::downgrade(c)),
|
||||||
|
Self::Server(s) => WeakTunnelEndpoint::Server(Arc::downgrade(s)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl WeakTunnelEndpoint {
|
||||||
|
pub fn upgrade(&self) -> Option<TunnelEndpoint> {
|
||||||
|
match self {
|
||||||
|
Self::Client(c) => c.upgrade().map(TunnelEndpoint::Client),
|
||||||
|
Self::Server(s) => s.upgrade().map(TunnelEndpoint::Server),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub struct ClientManager {
|
pub struct ClientManager {
|
||||||
pub session_id: AtomicU64,
|
pub session_id: AtomicU64,
|
||||||
pub current_client: Mutex<Option<client::SendRequest<Bytes>>>,
|
pub current_client: Mutex<Option<client::SendRequest<Bytes>>>,
|
||||||
@ -231,9 +260,9 @@ enum ResumeResult {
|
|||||||
pub struct TunnelListener {
|
pub struct TunnelListener {
|
||||||
listener: TcpListener,
|
listener: TcpListener,
|
||||||
config: ServerTunnelConfig,
|
config: ServerTunnelConfig,
|
||||||
pending_plain_sessions: Mutex<HashSet<u64>>,
|
pending_plain_sessions: Mutex<HashMap<u64, std::time::Instant>>,
|
||||||
next_session_id: AtomicU64,
|
next_session_id: AtomicU64,
|
||||||
active_sessions: Mutex<HashMap<u64, TunnelEndpoint>>,
|
active_sessions: Mutex<HashMap<u64, WeakTunnelEndpoint>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl TunnelListener {
|
impl TunnelListener {
|
||||||
@ -243,7 +272,7 @@ impl TunnelListener {
|
|||||||
Ok(Self {
|
Ok(Self {
|
||||||
listener,
|
listener,
|
||||||
config,
|
config,
|
||||||
pending_plain_sessions: Mutex::new(HashSet::new()),
|
pending_plain_sessions: Mutex::new(HashMap::new()),
|
||||||
next_session_id: AtomicU64::new(1),
|
next_session_id: AtomicU64::new(1),
|
||||||
active_sessions: Mutex::new(HashMap::new()),
|
active_sessions: Mutex::new(HashMap::new()),
|
||||||
})
|
})
|
||||||
@ -264,7 +293,9 @@ impl TunnelListener {
|
|||||||
let mut stream = match self.try_resume_plain_session(stream, peer_addr).await? {
|
let mut stream = match self.try_resume_plain_session(stream, peer_addr).await? {
|
||||||
ResumeResult::NewSession(ep_raw, sid) => {
|
ResumeResult::NewSession(ep_raw, sid) => {
|
||||||
let ep = wrap_raw_endpoint(sid, ep_raw, None);
|
let ep = wrap_raw_endpoint(sid, ep_raw, None);
|
||||||
self.active_sessions.lock().await.insert(sid, ep.clone());
|
let mut sessions = self.active_sessions.lock().await;
|
||||||
|
sessions.retain(|_, weak_ep| weak_ep.upgrade().is_some());
|
||||||
|
sessions.insert(sid, ep.downgrade());
|
||||||
return Ok(ep);
|
return Ok(ep);
|
||||||
}
|
}
|
||||||
ResumeResult::ResumedExisting => continue,
|
ResumeResult::ResumedExisting => continue,
|
||||||
@ -290,7 +321,9 @@ impl TunnelListener {
|
|||||||
|
|
||||||
let sid = self.next_session_id.fetch_add(1, Ordering::Relaxed);
|
let sid = self.next_session_id.fetch_add(1, Ordering::Relaxed);
|
||||||
let ep = wrap_raw_endpoint(sid, ep_raw, None);
|
let ep = wrap_raw_endpoint(sid, ep_raw, None);
|
||||||
self.active_sessions.lock().await.insert(sid, ep.clone());
|
let mut sessions = self.active_sessions.lock().await;
|
||||||
|
sessions.retain(|_, weak_ep| weak_ep.upgrade().is_some());
|
||||||
|
sessions.insert(sid, ep.downgrade());
|
||||||
return Ok(ep);
|
return Ok(ep);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -320,7 +353,8 @@ impl TunnelListener {
|
|||||||
let session_id = self.next_session_id.fetch_add(1, Ordering::Relaxed);
|
let session_id = self.next_session_id.fetch_add(1, Ordering::Relaxed);
|
||||||
{
|
{
|
||||||
let mut pending = self.pending_plain_sessions.lock().await;
|
let mut pending = self.pending_plain_sessions.lock().await;
|
||||||
pending.insert(session_id);
|
pending.retain(|_, time| time.elapsed() < Duration::from_secs(30));
|
||||||
|
pending.insert(session_id, std::time::Instant::now());
|
||||||
}
|
}
|
||||||
|
|
||||||
tls_stream.write_all(TLS_BOOTSTRAP_MAGIC).await?;
|
tls_stream.write_all(TLS_BOOTSTRAP_MAGIC).await?;
|
||||||
@ -351,7 +385,7 @@ impl TunnelListener {
|
|||||||
|
|
||||||
let is_pending = {
|
let is_pending = {
|
||||||
let mut pending = self.pending_plain_sessions.lock().await;
|
let mut pending = self.pending_plain_sessions.lock().await;
|
||||||
pending.remove(&session_id)
|
pending.remove(&session_id).is_some()
|
||||||
};
|
};
|
||||||
|
|
||||||
if is_pending {
|
if is_pending {
|
||||||
@ -366,7 +400,11 @@ impl TunnelListener {
|
|||||||
let ep_raw = upgrade_to_h2_raw(stream, self.config.identity)
|
let ep_raw = upgrade_to_h2_raw(stream, self.config.identity)
|
||||||
.await
|
.await
|
||||||
.map_err(|e| anyhow::anyhow!("{}", e))?;
|
.map_err(|e| anyhow::anyhow!("{}", e))?;
|
||||||
update_endpoint(&ep, ep_raw).await;
|
update_endpoint(
|
||||||
|
&ep.upgrade().ok_or(anyhow!("Connection is cleared"))?,
|
||||||
|
ep_raw,
|
||||||
|
)
|
||||||
|
.await;
|
||||||
info!(
|
info!(
|
||||||
"[{}] Successfully resumed existing session {}",
|
"[{}] Successfully resumed existing session {}",
|
||||||
peer_addr, session_id
|
peer_addr, session_id
|
||||||
@ -542,6 +580,47 @@ pub async fn connect_tunnel(config: ClientTunnelConfig) -> Result<TunnelEndpoint
|
|||||||
Ok(wrap_raw_endpoint(sid, ep_raw, Some(config)))
|
Ok(wrap_raw_endpoint(sid, ep_raw, Some(config)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub async fn connect_with_auto_proxy(config: &ClientTunnelConfig) -> anyhow::Result<TcpStream> {
|
||||||
|
let Some(proxy) = &config.proxy else {
|
||||||
|
return Ok(TcpStream::connect(&config.url).await?);
|
||||||
|
};
|
||||||
|
|
||||||
|
let parsed_proxy = Url::parse(proxy)?;
|
||||||
|
|
||||||
|
match parsed_proxy.scheme() {
|
||||||
|
"socks5" => {
|
||||||
|
let host = parsed_proxy.host_str().unwrap();
|
||||||
|
let port = parsed_proxy.port().unwrap_or(1080);
|
||||||
|
let stream = Socks5Stream::connect((host, port), config.url.clone())
|
||||||
|
.await?
|
||||||
|
.into_inner();
|
||||||
|
Ok(stream)
|
||||||
|
}
|
||||||
|
"http" | "https" => {
|
||||||
|
let proxy_addr = format!(
|
||||||
|
"{}:{}",
|
||||||
|
parsed_proxy.host_str().unwrap(),
|
||||||
|
parsed_proxy.port().unwrap_or(80)
|
||||||
|
);
|
||||||
|
let mut stream = TcpStream::connect(proxy_addr).await?;
|
||||||
|
|
||||||
|
let target_url = Url::parse(&format!("tcp://{}", config.url))?;
|
||||||
|
http_connect_tokio(
|
||||||
|
&mut stream,
|
||||||
|
target_url.host_str().unwrap(),
|
||||||
|
target_url.port().unwrap_or(80),
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(stream)
|
||||||
|
}
|
||||||
|
_ => Err(anyhow::anyhow!(
|
||||||
|
"Unsupported proxy scheme: {}",
|
||||||
|
parsed_proxy.scheme()
|
||||||
|
)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
async fn do_client_reconnect(
|
async fn do_client_reconnect(
|
||||||
config: &ClientTunnelConfig,
|
config: &ClientTunnelConfig,
|
||||||
current_sid: &mut u64,
|
current_sid: &mut u64,
|
||||||
@ -563,7 +642,7 @@ async fn do_client_reconnect(
|
|||||||
*current_sid = sid;
|
*current_sid = sid;
|
||||||
resume_tunnel_client(config, sid).await
|
resume_tunnel_client(config, sid).await
|
||||||
} else {
|
} else {
|
||||||
let mut stream = TcpStream::connect(&config.url).await?;
|
let mut stream = connect_with_auto_proxy(config).await?;
|
||||||
perform_client_handshake(&mut stream, &config.token, config.identity).await?;
|
perform_client_handshake(&mut stream, &config.token, config.identity).await?;
|
||||||
let raw = upgrade_to_h2_raw(stream, config.identity).await?;
|
let raw = upgrade_to_h2_raw(stream, config.identity).await?;
|
||||||
*current_sid = 0;
|
*current_sid = 0;
|
||||||
@ -576,7 +655,7 @@ async fn bootstrap_tls_and_get_sid(
|
|||||||
host: &str,
|
host: &str,
|
||||||
) -> anyhow::Result<(u64, ())> {
|
) -> anyhow::Result<(u64, ())> {
|
||||||
let connector = build_client_tls_connector();
|
let connector = build_client_tls_connector();
|
||||||
let tcp = TcpStream::connect(&config.url).await?;
|
let tcp = connect_with_auto_proxy(config).await?;
|
||||||
let server_name = ServerName::try_from(host.to_string())
|
let server_name = ServerName::try_from(host.to_string())
|
||||||
.map_err(|_| anyhow!("Invalid TLS host: {}", host))?
|
.map_err(|_| anyhow!("Invalid TLS host: {}", host))?
|
||||||
.to_owned();
|
.to_owned();
|
||||||
@ -604,7 +683,7 @@ async fn resume_tunnel_client(
|
|||||||
config: &ClientTunnelConfig,
|
config: &ClientTunnelConfig,
|
||||||
session_id: u64,
|
session_id: u64,
|
||||||
) -> anyhow::Result<TunnelEndpointRaw> {
|
) -> anyhow::Result<TunnelEndpointRaw> {
|
||||||
let mut plain_stream = TcpStream::connect(&config.url).await?;
|
let mut plain_stream = connect_with_auto_proxy(config).await?;
|
||||||
plain_stream.write_all(RESUME_MAGIC).await?;
|
plain_stream.write_all(RESUME_MAGIC).await?;
|
||||||
plain_stream.write_all(&session_id.to_be_bytes()).await?;
|
plain_stream.write_all(&session_id.to_be_bytes()).await?;
|
||||||
upgrade_to_h2_raw(plain_stream, config.identity).await
|
upgrade_to_h2_raw(plain_stream, config.identity).await
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user