640 lines
18 KiB
C++
640 lines
18 KiB
C++
#include "./ngcft1.hpp"
|
|
|
|
#include "./flow_only.hpp"
|
|
#include "./cubic.hpp"
|
|
#include "./ledbat.hpp"
|
|
|
|
#include <solanaceae/util/utils.hpp>
|
|
|
|
#include <sodium.h>
|
|
|
|
#include <cstdint>
|
|
#include <iostream>
|
|
#include <set>
|
|
#include <algorithm>
|
|
#include <cassert>
|
|
#include <vector>
|
|
|
|
void NGCFT1::updateSendTransfer(float time_delta, uint32_t group_number, uint32_t peer_number, Group::Peer& peer, size_t idx, std::set<CCAI::SeqIDType>& timeouts_set, int64_t& can_packet_size) {
|
|
auto& tf_opt = peer.send_transfers.at(idx);
|
|
assert(tf_opt.has_value());
|
|
auto& tf = tf_opt.value();
|
|
|
|
tf.time_since_activity += time_delta;
|
|
|
|
switch (tf.state) {
|
|
using State = Group::Peer::SendTransfer::State;
|
|
case State::INIT_SENT:
|
|
if (tf.time_since_activity >= init_retry_timeout_after) {
|
|
if (tf.inits_sent >= 3) {
|
|
// delete, timed out 3 times
|
|
std::cerr << "NGCFT1 warning: ft init timed out, deleting\n";
|
|
dispatch(
|
|
NGCFT1_Event::send_done,
|
|
Events::NGCFT1_send_done{
|
|
group_number, peer_number,
|
|
static_cast<uint8_t>(idx),
|
|
}
|
|
);
|
|
tf_opt.reset();
|
|
} else {
|
|
// timed out, resend
|
|
std::cerr << "NGCFT1 warning: ft init timed out, resending\n";
|
|
//sendPKG_FT1_INIT(group_number, peer_number, tf.file_kind, tf.file_size, idx, tf.file_id.data(), tf.file_id.size());
|
|
_neep.send_ft1_init(group_number, peer_number, tf.file_kind, tf.file_size, idx, tf.file_id.data(), tf.file_id.size());
|
|
tf.inits_sent++;
|
|
tf.time_since_activity = 0.f;
|
|
}
|
|
}
|
|
//break;
|
|
return;
|
|
case State::FINISHING: // we still have unacked packets
|
|
tf.ssb.for_each(time_delta, [&](uint16_t id, const std::vector<uint8_t>& data, float& time_since_activity) {
|
|
if (can_packet_size >= data.size() && timeouts_set.count({idx, id})) {
|
|
_neep.send_ft1_data(group_number, peer_number, idx, id, data.data(), data.size());
|
|
peer.cca->onLoss({idx, id}, false);
|
|
time_since_activity = 0.f;
|
|
timeouts_set.erase({idx, id});
|
|
can_packet_size -= data.size();
|
|
}
|
|
});
|
|
if (tf.time_since_activity >= sending_give_up_after) {
|
|
// no ack after 30sec, close ft
|
|
// TODO: notify app
|
|
std::cerr << "NGCFT1 warning: sending ft finishing timed out, deleting\n";
|
|
|
|
// clean up cca
|
|
tf.ssb.for_each(time_delta, [&](uint16_t id, const std::vector<uint8_t>& data, float& time_since_activity) {
|
|
peer.cca->onLoss({idx, id}, true);
|
|
timeouts_set.erase({idx, id});
|
|
});
|
|
|
|
tf_opt.reset();
|
|
}
|
|
break;
|
|
case State::SENDING: {
|
|
// first handle overall timeout (could otherwise do resends directly before, which is useless)
|
|
// timeout increases with active transfers (otherwise we could starve them)
|
|
if (tf.time_since_activity >= (sending_give_up_after * peer.active_send_transfers)) {
|
|
// no ack after 30sec, close ft
|
|
std::cerr << "NGCFT1 warning: sending ft in progress timed out, deleting (ifc:" << peer.cca->inFlightCount() << ")\n";
|
|
dispatch(
|
|
NGCFT1_Event::send_done,
|
|
Events::NGCFT1_send_done{
|
|
group_number, peer_number,
|
|
static_cast<uint8_t>(idx),
|
|
}
|
|
);
|
|
|
|
// clean up cca
|
|
tf.ssb.for_each(time_delta, [&](uint16_t id, const std::vector<uint8_t>& data, float& time_since_activity) {
|
|
peer.cca->onLoss({idx, id}, true);
|
|
timeouts_set.erase({idx, id});
|
|
});
|
|
|
|
tf_opt.reset();
|
|
//continue; // dangerous control flow
|
|
return;
|
|
}
|
|
|
|
// do resends
|
|
tf.ssb.for_each(time_delta, [&](uint16_t id, const std::vector<uint8_t>& data, float& time_since_activity) {
|
|
if (can_packet_size >= data.size() && time_since_activity >= peer.cca->getCurrentDelay() && timeouts_set.count({idx, id})) {
|
|
// TODO: can fail
|
|
_neep.send_ft1_data(group_number, peer_number, idx, id, data.data(), data.size());
|
|
peer.cca->onLoss({idx, id}, false);
|
|
time_since_activity = 0.f;
|
|
timeouts_set.erase({idx, id});
|
|
can_packet_size -= data.size();
|
|
}
|
|
});
|
|
|
|
// if chunks in flight < window size (2)
|
|
while (can_packet_size > 0 && tf.file_size > 0) {
|
|
std::vector<uint8_t> new_data;
|
|
|
|
size_t chunk_size = std::min<size_t>({
|
|
peer.cca->MAXIMUM_SEGMENT_DATA_SIZE,
|
|
static_cast<size_t>(can_packet_size),
|
|
tf.file_size - tf.file_size_current
|
|
});
|
|
if (chunk_size == 0) {
|
|
tf.state = State::FINISHING;
|
|
break; // we done
|
|
}
|
|
|
|
new_data.resize(chunk_size);
|
|
|
|
assert(idx <= 0xffu);
|
|
// TODO: check return value
|
|
dispatch(
|
|
NGCFT1_Event::send_data,
|
|
Events::NGCFT1_send_data{
|
|
group_number, peer_number,
|
|
static_cast<uint8_t>(idx),
|
|
tf.file_size_current,
|
|
new_data.data(), new_data.size(),
|
|
}
|
|
);
|
|
|
|
uint16_t seq_id = tf.ssb.add(std::move(new_data));
|
|
const bool sent = _neep.send_ft1_data(group_number, peer_number, idx, seq_id, tf.ssb.entries.at(seq_id).data.data(), tf.ssb.entries.at(seq_id).data.size());
|
|
if (sent) {
|
|
peer.cca->onSent({idx, seq_id}, chunk_size);
|
|
} else {
|
|
std::cerr << "NGCFT1: failed to send packet (queue full?) --------------\n";
|
|
peer.cca->onLoss({idx, seq_id}, false); // HACK: fake congestion event
|
|
// TODO: onCongestion
|
|
can_packet_size = 0;
|
|
}
|
|
|
|
tf.file_size_current += chunk_size;
|
|
can_packet_size -= chunk_size;
|
|
}
|
|
}
|
|
break;
|
|
default: // invalid state, delete
|
|
std::cerr << "NGCFT1 error: ft in invalid state, deleting\n";
|
|
tf_opt.reset();
|
|
//continue;
|
|
return;
|
|
}
|
|
}
|
|
|
|
void NGCFT1::iteratePeer(float time_delta, uint32_t group_number, uint32_t peer_number, Group::Peer& peer) {
|
|
if (peer.cca) {
|
|
auto timeouts = peer.cca->getTimeouts();
|
|
std::set<CCAI::SeqIDType> timeouts_set{timeouts.cbegin(), timeouts.cend()};
|
|
|
|
int64_t can_packet_size {peer.cca->canSend(time_delta)}; // might get more space while iterating (time)
|
|
|
|
// get number current running transfers TODO: improve
|
|
peer.active_send_transfers = 0;
|
|
for (const auto& it : peer.send_transfers) {
|
|
if (it.has_value()) {
|
|
peer.active_send_transfers++;
|
|
}
|
|
}
|
|
|
|
// change iterate start position to not starve transfers in the back
|
|
size_t iterated_count = 0;
|
|
bool last_send_found = false;
|
|
for (size_t idx = peer.next_send_transfer_send_idx; iterated_count < peer.send_transfers.size(); idx++, iterated_count++) {
|
|
idx = idx % peer.send_transfers.size();
|
|
|
|
if (peer.send_transfers.at(idx).has_value()) {
|
|
if (!last_send_found && can_packet_size <= 0) {
|
|
peer.next_send_transfer_send_idx = idx;
|
|
last_send_found = true; // only set once
|
|
}
|
|
updateSendTransfer(time_delta, group_number, peer_number, peer, idx, timeouts_set, can_packet_size);
|
|
}
|
|
}
|
|
}
|
|
|
|
// TODO: receiving tranfers?
|
|
}
|
|
|
|
NGCFT1::NGCFT1(
|
|
ToxI& t,
|
|
ToxEventProviderI& tep,
|
|
NGCEXTEventProvider& neep
|
|
) : _t(t), _tep(tep), _neep(neep)
|
|
{
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_REQUEST);
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_INIT);
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_INIT_ACK);
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_DATA);
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_DATA_ACK);
|
|
_neep.subscribe(this, NGCEXT_Event::FT1_MESSAGE);
|
|
|
|
_tep.subscribe(this, Tox_Event_Type::TOX_EVENT_GROUP_PEER_EXIT);
|
|
}
|
|
|
|
float NGCFT1::iterate(float time_delta) {
|
|
bool transfer_in_progress {false};
|
|
for (auto& [group_number, group] : groups) {
|
|
for (auto& [peer_number, peer] : group.peers) {
|
|
iteratePeer(time_delta, group_number, peer_number, peer);
|
|
|
|
// find any active transfer
|
|
if (!transfer_in_progress) {
|
|
for (const auto& t : peer.send_transfers) {
|
|
if (t.has_value()) {
|
|
transfer_in_progress = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (!transfer_in_progress) {
|
|
for (const auto& t : peer.recv_transfers) {
|
|
if (t.has_value()) {
|
|
transfer_in_progress = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (transfer_in_progress) {
|
|
// ~15ms for up to 1mb/s
|
|
// ~5ms for up to 4mb/s
|
|
return 0.005f; // 5ms
|
|
} else {
|
|
return 1.f; // once a sec might be too little
|
|
}
|
|
}
|
|
|
|
void NGCFT1::NGC_FT1_send_request_private(
|
|
uint32_t group_number, uint32_t peer_number,
|
|
uint32_t file_kind,
|
|
const uint8_t* file_id, size_t file_id_size
|
|
) {
|
|
// TODO: error check
|
|
_neep.send_ft1_request(group_number, peer_number, file_kind, file_id, file_id_size);
|
|
}
|
|
|
|
bool NGCFT1::NGC_FT1_send_init_private(
|
|
uint32_t group_number, uint32_t peer_number,
|
|
uint32_t file_kind,
|
|
const uint8_t* file_id, size_t file_id_size,
|
|
size_t file_size,
|
|
uint8_t* transfer_id
|
|
) {
|
|
if (std::get<0>(_t.toxGroupPeerGetConnectionStatus(group_number, peer_number)).value_or(TOX_CONNECTION_NONE) == TOX_CONNECTION_NONE) {
|
|
std::cerr << "NGCFT1 error: cant init ft, peer offline\n";
|
|
return false;
|
|
}
|
|
|
|
auto& peer = groups[group_number].peers[peer_number];
|
|
|
|
// allocate transfer_id
|
|
size_t idx = peer.next_send_transfer_idx;
|
|
peer.next_send_transfer_idx = (peer.next_send_transfer_idx + 1) % 256;
|
|
{ // TODO: extract
|
|
size_t i = idx;
|
|
bool found = false;
|
|
do {
|
|
if (!peer.send_transfers[i].has_value()) {
|
|
// free slot
|
|
idx = i;
|
|
found = true;
|
|
break;
|
|
}
|
|
|
|
i = (i + 1) % 256;
|
|
} while (i != idx);
|
|
|
|
if (!found) {
|
|
std::cerr << "NGCFT1 error: cant init ft, no free transfer slot\n";
|
|
return false;
|
|
}
|
|
}
|
|
|
|
// TODO: check return value
|
|
_neep.send_ft1_init(group_number, peer_number, file_kind, file_size, idx, file_id, file_id_size);
|
|
|
|
peer.send_transfers[idx] = Group::Peer::SendTransfer{
|
|
file_kind,
|
|
std::vector(file_id, file_id+file_id_size),
|
|
Group::Peer::SendTransfer::State::INIT_SENT,
|
|
1,
|
|
0.f,
|
|
file_size,
|
|
0,
|
|
{}, // ssb
|
|
};
|
|
|
|
if (transfer_id != nullptr) {
|
|
*transfer_id = idx;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool NGCFT1::NGC_FT1_send_message_public(
|
|
uint32_t group_number,
|
|
uint32_t& message_id,
|
|
uint32_t file_kind,
|
|
const uint8_t* file_id, size_t file_id_size
|
|
) {
|
|
// create msg_id
|
|
message_id = randombytes_random();
|
|
|
|
// TODO: check return value
|
|
return _neep.send_all_ft1_message(group_number, message_id, file_kind, file_id, file_id_size);
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_request& e) {
|
|
//#if !NDEBUG
|
|
std::cout << "NGCFT1: FT1_REQUEST fk:" << e.file_kind << " [" << bin2hex(e.file_id) << "]\n";
|
|
//#endif
|
|
|
|
// .... just rethrow??
|
|
// TODO: dont
|
|
return dispatch(
|
|
NGCFT1_Event::recv_request,
|
|
Events::NGCFT1_recv_request{
|
|
e.group_number, e.peer_number,
|
|
static_cast<NGCFT1_file_kind>(e.file_kind),
|
|
e.file_id.data(), e.file_id.size()
|
|
}
|
|
);
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_init& e) {
|
|
//#if !NDEBUG
|
|
std::cout << "NGCFT1: FT1_INIT fk:" << e.file_kind << " fs:" << e.file_size << " tid:" << int(e.transfer_id) << " [" << bin2hex(e.file_id) << "]\n";
|
|
//#endif
|
|
|
|
bool accept = false;
|
|
dispatch(
|
|
NGCFT1_Event::recv_init,
|
|
Events::NGCFT1_recv_init{
|
|
e.group_number, e.peer_number,
|
|
static_cast<NGCFT1_file_kind>(e.file_kind),
|
|
e.file_id.data(), e.file_id.size(),
|
|
e.transfer_id,
|
|
e.file_size,
|
|
accept
|
|
}
|
|
);
|
|
|
|
if (!accept) {
|
|
std::cout << "NGCFT1: rejected init\n";
|
|
return true; // return true?
|
|
}
|
|
|
|
_neep.send_ft1_init_ack(e.group_number, e.peer_number, e.transfer_id);
|
|
|
|
std::cout << "NGCFT1: accepted init\n";
|
|
|
|
auto& peer = groups[e.group_number].peers[e.peer_number];
|
|
if (peer.recv_transfers[e.transfer_id].has_value()) {
|
|
std::cerr << "NGCFT1 warning: overwriting existing recv_transfer " << int(e.transfer_id) << "\n";
|
|
}
|
|
|
|
peer.recv_transfers[e.transfer_id] = Group::Peer::RecvTransfer{
|
|
e.file_kind,
|
|
e.file_id,
|
|
Group::Peer::RecvTransfer::State::INITED,
|
|
e.file_size,
|
|
0u,
|
|
{} // rsb
|
|
};
|
|
|
|
return true;
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_init_ack& e) {
|
|
//#if !NDEBUG
|
|
std::cout << "NGCFT1: FT1_INIT_ACK mds:" << e.max_lossy_data_size << "\n";
|
|
//#endif
|
|
|
|
// we now should start sending data
|
|
|
|
if (!groups.count(e.group_number)) {
|
|
std::cerr << "NGCFT1 warning: init_ack for unknown group\n";
|
|
return true;
|
|
}
|
|
|
|
Group::Peer& peer = groups[e.group_number].peers[e.peer_number];
|
|
if (!peer.send_transfers[e.transfer_id].has_value()) {
|
|
std::cerr << "NGCFT1 warning: init_ack for unknown transfer\n";
|
|
return true;
|
|
}
|
|
|
|
Group::Peer::SendTransfer& transfer = peer.send_transfers[e.transfer_id].value();
|
|
|
|
using State = Group::Peer::SendTransfer::State;
|
|
if (transfer.state != State::INIT_SENT) {
|
|
std::cerr << "NGCFT1 error: init_ack but not in INIT_SENT state\n";
|
|
return true;
|
|
}
|
|
|
|
// negotiated packet_data_size
|
|
const auto negotiated_packet_data_size = std::min<uint32_t>(e.max_lossy_data_size, _t.toxGroupMaxCustomLossyPacketLength()-4);
|
|
// TODO: reset cca with new pkg size
|
|
if (!peer.cca) {
|
|
// make random max of [1020-1220]
|
|
const uint32_t random_max_data_size = (1024-4) + _rng()%201;
|
|
const uint32_t randomized_negotiated_packet_data_size = std::min(negotiated_packet_data_size, random_max_data_size);
|
|
|
|
peer.max_packet_data_size = randomized_negotiated_packet_data_size;
|
|
|
|
std::cerr << "NGCFT1: creating cca with max:" << peer.max_packet_data_size << "\n";
|
|
|
|
peer.cca = std::make_unique<CUBIC>(peer.max_packet_data_size);
|
|
//peer.cca = std::make_unique<LEDBAT>(peer.max_packet_data_size);
|
|
//peer.cca = std::make_unique<FlowOnly>(peer.max_packet_data_size);
|
|
//peer.cca->max_byterate_allowed = 1.f *1024*1024;
|
|
}
|
|
|
|
// iterate will now call NGC_FT1_send_data_cb
|
|
transfer.state = State::SENDING;
|
|
transfer.time_since_activity = 0.f;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_data& e) {
|
|
#if !NDEBUG
|
|
//std::cout << "NGCFT1: FT1_DATA\n";
|
|
#endif
|
|
|
|
if (e.data.empty()) {
|
|
std::cerr << "NGCFT1 error: data of size 0!\n";
|
|
return true;
|
|
}
|
|
|
|
if (!groups.count(e.group_number)) {
|
|
std::cerr << "NGCFT1 warning: data for unknown group\n";
|
|
return true;
|
|
}
|
|
|
|
Group::Peer& peer = groups[e.group_number].peers[e.peer_number];
|
|
if (!peer.recv_transfers[e.transfer_id].has_value()) {
|
|
std::cerr << "NGCFT1 warning: data for unknown transfer\n";
|
|
return true;
|
|
}
|
|
|
|
auto& transfer = peer.recv_transfers[e.transfer_id].value();
|
|
|
|
// do reassembly, ignore dups
|
|
transfer.rsb.add(e.sequence_id, std::vector<uint8_t>(e.data)); // TODO: ugly explicit copy for what should just be a move
|
|
|
|
// loop for chunks without holes
|
|
while (transfer.rsb.canPop()) {
|
|
auto data = transfer.rsb.pop();
|
|
|
|
// TODO: check return value
|
|
dispatch(
|
|
NGCFT1_Event::recv_data,
|
|
Events::NGCFT1_recv_data{
|
|
e.group_number, e.peer_number,
|
|
e.transfer_id,
|
|
transfer.file_size_current,
|
|
data.data(), data.size()
|
|
}
|
|
);
|
|
|
|
transfer.file_size_current += data.size();
|
|
}
|
|
|
|
// send acks
|
|
// reverse, last seq is most recent
|
|
std::vector<uint16_t> ack_seq_ids(transfer.rsb.ack_seq_ids.crbegin(), transfer.rsb.ack_seq_ids.crend());
|
|
// TODO: check if this caps at max acks
|
|
if (!ack_seq_ids.empty()) {
|
|
// TODO: check return value
|
|
_neep.send_ft1_data_ack(e.group_number, e.peer_number, e.transfer_id, ack_seq_ids.data(), ack_seq_ids.size());
|
|
}
|
|
|
|
|
|
if (transfer.file_size_current == transfer.file_size) {
|
|
// TODO: set all data received, and clean up
|
|
//transfer.state = Group::Peer::RecvTransfer::State::RECV;
|
|
dispatch(
|
|
NGCFT1_Event::recv_done,
|
|
Events::NGCFT1_recv_done{
|
|
e.group_number, e.peer_number,
|
|
e.transfer_id
|
|
}
|
|
);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_data_ack& e) {
|
|
#if !NDEBUG
|
|
//std::cout << "NGCFT1: FT1_DATA_ACK\n";
|
|
#endif
|
|
|
|
if (!groups.count(e.group_number)) {
|
|
std::cerr << "NGCFT1 warning: data_ack for unknown group\n";
|
|
return true;
|
|
}
|
|
|
|
Group::Peer& peer = groups[e.group_number].peers[e.peer_number];
|
|
if (!peer.send_transfers[e.transfer_id].has_value()) {
|
|
// we delete directly, packets might still be in flight (in practice they are when ce)
|
|
//std::cerr << "NGCFT1 warning: data_ack for unknown transfer\n";
|
|
return true;
|
|
}
|
|
|
|
Group::Peer::SendTransfer& transfer = peer.send_transfers[e.transfer_id].value();
|
|
|
|
using State = Group::Peer::SendTransfer::State;
|
|
if (transfer.state != State::SENDING && transfer.state != State::FINISHING) {
|
|
std::cerr << "NGCFT1 error: data_ack but not in SENDING or FINISHING state (" << int(transfer.state) << ")\n";
|
|
return true;
|
|
}
|
|
|
|
transfer.time_since_activity = 0.f;
|
|
|
|
{
|
|
std::vector<CCAI::SeqIDType> seqs;
|
|
for (const auto it : e.sequence_ids) {
|
|
// TODO: improve this o.o
|
|
seqs.push_back({e.transfer_id, it});
|
|
transfer.ssb.erase(it);
|
|
}
|
|
peer.cca->onAck(std::move(seqs));
|
|
}
|
|
|
|
// delete if all packets acked
|
|
if (transfer.file_size == transfer.file_size_current && transfer.ssb.size() == 0) {
|
|
std::cout << "NGCFT1: " << int(e.transfer_id) << " done. wnd:" << peer.cca->getWindow() << "\n";
|
|
dispatch(
|
|
NGCFT1_Event::send_done,
|
|
Events::NGCFT1_send_done{
|
|
e.group_number, e.peer_number,
|
|
e.transfer_id,
|
|
}
|
|
);
|
|
// TODO: check for FINISHING state
|
|
peer.send_transfers[e.transfer_id].reset();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool NGCFT1::onEvent(const Events::NGCEXT_ft1_message& e) {
|
|
std::cout << "NGCFT1: FT1_MESSAGE mid:" << e.message_id << " fk:" << e.file_kind << " [" << bin2hex(e.file_id) << "]\n";
|
|
|
|
// .... just rethrow??
|
|
// TODO: dont
|
|
return dispatch(
|
|
NGCFT1_Event::recv_message,
|
|
Events::NGCFT1_recv_message{
|
|
e.group_number, e.peer_number,
|
|
e.message_id,
|
|
static_cast<NGCFT1_file_kind>(e.file_kind),
|
|
e.file_id.data(), e.file_id.size()
|
|
}
|
|
);
|
|
}
|
|
|
|
bool NGCFT1::onToxEvent(const Tox_Event_Group_Peer_Exit* e) {
|
|
const auto group_number = tox_event_group_peer_exit_get_group_number(e);
|
|
const auto peer_number = tox_event_group_peer_exit_get_peer_id(e);
|
|
|
|
// peer disconnected, end all transfers
|
|
|
|
if (!groups.count(group_number)) {
|
|
return false;
|
|
}
|
|
|
|
auto& group = groups.at(group_number);
|
|
|
|
if (!group.peers.count(peer_number)) {
|
|
return false;
|
|
}
|
|
|
|
auto& peer = group.peers.at(peer_number);
|
|
|
|
for (size_t i = 0; i < peer.send_transfers.size(); i++) {
|
|
auto& it_opt = peer.send_transfers.at(i);
|
|
if (!it_opt.has_value()) {
|
|
continue;
|
|
}
|
|
|
|
std::cout << "NGCFT1: sending " << int(i) << " canceled bc peer offline\n";
|
|
dispatch(
|
|
NGCFT1_Event::send_done,
|
|
Events::NGCFT1_send_done{
|
|
group_number, peer_number,
|
|
static_cast<uint8_t>(i),
|
|
}
|
|
);
|
|
|
|
it_opt.reset();
|
|
}
|
|
|
|
for (size_t i = 0; i < peer.recv_transfers.size(); i++) {
|
|
auto& it_opt = peer.recv_transfers.at(i);
|
|
if (!it_opt.has_value()) {
|
|
continue;
|
|
}
|
|
|
|
std::cout << "NGCFT1: receiving " << int(i) << " canceled bc peer offline\n";
|
|
dispatch(
|
|
NGCFT1_Event::recv_done,
|
|
Events::NGCFT1_recv_done{
|
|
group_number, peer_number,
|
|
static_cast<uint8_t>(i),
|
|
}
|
|
);
|
|
|
|
it_opt.reset();
|
|
}
|
|
|
|
// reset cca
|
|
peer.cca.reset(); // dont actually reallocate
|
|
|
|
return false;
|
|
}
|
|
|