|
use dora_coordinator::{ControlEvent, Event}; |
|
use dora_core::{ |
|
descriptor::Descriptor, |
|
topics::{ |
|
ControlRequest, ControlRequestReply, DataflowId, DORA_COORDINATOR_PORT_CONTROL_DEFAULT, |
|
DORA_COORDINATOR_PORT_DEFAULT, |
|
}, |
|
}; |
|
use dora_tracing::set_up_tracing; |
|
use eyre::{bail, Context}; |
|
|
|
use std::{ |
|
collections::BTreeSet, |
|
net::{IpAddr, Ipv4Addr, SocketAddr}, |
|
path::Path, |
|
time::Duration, |
|
}; |
|
use tokio::{ |
|
sync::{ |
|
mpsc::{self, Sender}, |
|
oneshot, |
|
}, |
|
task::JoinSet, |
|
}; |
|
use tokio_stream::wrappers::ReceiverStream; |
|
use uuid::Uuid; |
|
|
|
#[tokio::main] |
|
async fn main() -> eyre::Result<()> { |
|
set_up_tracing("multiple-daemon-runner").wrap_err("failed to set up tracing subscriber")?; |
|
|
|
let root = Path::new(env!("CARGO_MANIFEST_DIR")); |
|
std::env::set_current_dir(root.join(file!()).parent().unwrap()) |
|
.wrap_err("failed to set working dir")?; |
|
|
|
let dataflow = Path::new("dataflow.yml"); |
|
build_dataflow(dataflow).await?; |
|
|
|
let (coordinator_events_tx, coordinator_events_rx) = mpsc::channel(1); |
|
let coordinator_bind = SocketAddr::new( |
|
IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), |
|
DORA_COORDINATOR_PORT_DEFAULT, |
|
); |
|
let coordinator_control_bind = SocketAddr::new( |
|
IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), |
|
DORA_COORDINATOR_PORT_CONTROL_DEFAULT, |
|
); |
|
let (coordinator_port, coordinator) = dora_coordinator::start( |
|
coordinator_bind, |
|
coordinator_control_bind, |
|
ReceiverStream::new(coordinator_events_rx), |
|
) |
|
.await?; |
|
let coordinator_addr = SocketAddr::new(Ipv4Addr::LOCALHOST.into(), coordinator_port); |
|
let daemon_a = run_daemon(coordinator_addr.to_string(), "A"); |
|
let daemon_b = run_daemon(coordinator_addr.to_string(), "B"); |
|
|
|
tracing::info!("Spawning coordinator and daemons"); |
|
let mut tasks = JoinSet::new(); |
|
tasks.spawn(coordinator); |
|
tasks.spawn(daemon_a); |
|
tasks.spawn(daemon_b); |
|
|
|
tracing::info!("waiting until daemons are connected to coordinator"); |
|
let mut retries = 0; |
|
loop { |
|
let connected_machines = connected_machines(&coordinator_events_tx).await?; |
|
if connected_machines.contains("A") && connected_machines.contains("B") { |
|
break; |
|
} else if retries > 20 { |
|
bail!("daemon not connected after {retries} retries"); |
|
} else { |
|
std::thread::sleep(Duration::from_millis(500)); |
|
retries += 1 |
|
} |
|
} |
|
|
|
tracing::info!("starting dataflow"); |
|
let uuid = start_dataflow(dataflow, &coordinator_events_tx).await?; |
|
tracing::info!("started dataflow under ID `{uuid}`"); |
|
|
|
let running = running_dataflows(&coordinator_events_tx).await?; |
|
if !running.iter().map(|d| d.uuid).any(|id| id == uuid) { |
|
bail!("dataflow `{uuid}` is not running"); |
|
} |
|
|
|
tracing::info!("waiting for dataflow `{uuid}` to finish"); |
|
let mut retries = 0; |
|
loop { |
|
let running = running_dataflows(&coordinator_events_tx).await?; |
|
if running.is_empty() { |
|
break; |
|
} else if retries > 100 { |
|
bail!("dataflow not finished after {retries} retries"); |
|
} else { |
|
tracing::debug!("not done yet"); |
|
std::thread::sleep(Duration::from_millis(500)); |
|
retries += 1 |
|
} |
|
} |
|
tracing::info!("dataflow `{uuid}` finished, destroying coordinator"); |
|
destroy(&coordinator_events_tx).await?; |
|
|
|
tracing::info!("joining tasks"); |
|
while let Some(res) = tasks.join_next().await { |
|
res.unwrap()?; |
|
} |
|
|
|
tracing::info!("done"); |
|
Ok(()) |
|
} |
|
|
|
async fn start_dataflow( |
|
dataflow: &Path, |
|
coordinator_events_tx: &Sender<Event>, |
|
) -> eyre::Result<Uuid> { |
|
let dataflow_descriptor = Descriptor::read(dataflow) |
|
.await |
|
.wrap_err("failed to read yaml dataflow")?; |
|
let working_dir = dataflow |
|
.canonicalize() |
|
.context("failed to canonicalize dataflow path")? |
|
.parent() |
|
.ok_or_else(|| eyre::eyre!("dataflow path has no parent dir"))? |
|
.to_owned(); |
|
dataflow_descriptor |
|
.check(&working_dir) |
|
.wrap_err("could not validate yaml")?; |
|
|
|
let (reply_sender, reply) = oneshot::channel(); |
|
coordinator_events_tx |
|
.send(Event::Control(ControlEvent::IncomingRequest { |
|
request: ControlRequest::Start { |
|
dataflow: dataflow_descriptor, |
|
local_working_dir: working_dir, |
|
name: None, |
|
}, |
|
reply_sender, |
|
})) |
|
.await?; |
|
let result = reply.await??; |
|
let uuid = match result { |
|
ControlRequestReply::DataflowStarted { uuid } => uuid, |
|
ControlRequestReply::Error(err) => bail!("{err}"), |
|
other => bail!("unexpected start dataflow reply: {other:?}"), |
|
}; |
|
Ok(uuid) |
|
} |
|
|
|
async fn connected_machines( |
|
coordinator_events_tx: &Sender<Event>, |
|
) -> eyre::Result<BTreeSet<String>> { |
|
let (reply_sender, reply) = oneshot::channel(); |
|
coordinator_events_tx |
|
.send(Event::Control(ControlEvent::IncomingRequest { |
|
request: ControlRequest::ConnectedMachines, |
|
reply_sender, |
|
})) |
|
.await?; |
|
let result = reply.await??; |
|
let machines = match result { |
|
ControlRequestReply::ConnectedMachines(machines) => machines, |
|
ControlRequestReply::Error(err) => bail!("{err}"), |
|
other => bail!("unexpected start dataflow reply: {other:?}"), |
|
}; |
|
Ok(machines) |
|
} |
|
|
|
async fn running_dataflows(coordinator_events_tx: &Sender<Event>) -> eyre::Result<Vec<DataflowId>> { |
|
let (reply_sender, reply) = oneshot::channel(); |
|
coordinator_events_tx |
|
.send(Event::Control(ControlEvent::IncomingRequest { |
|
request: ControlRequest::List, |
|
reply_sender, |
|
})) |
|
.await?; |
|
let result = reply.await??; |
|
let dataflows = match result { |
|
ControlRequestReply::DataflowList { dataflows } => dataflows, |
|
ControlRequestReply::Error(err) => bail!("{err}"), |
|
other => bail!("unexpected start dataflow reply: {other:?}"), |
|
}; |
|
Ok(dataflows) |
|
} |
|
|
|
async fn destroy(coordinator_events_tx: &Sender<Event>) -> eyre::Result<()> { |
|
let (reply_sender, reply) = oneshot::channel(); |
|
coordinator_events_tx |
|
.send(Event::Control(ControlEvent::IncomingRequest { |
|
request: ControlRequest::Destroy, |
|
reply_sender, |
|
})) |
|
.await?; |
|
let result = reply.await??; |
|
match result { |
|
ControlRequestReply::DestroyOk => Ok(()), |
|
ControlRequestReply::Error(err) => bail!("{err}"), |
|
other => bail!("unexpected start dataflow reply: {other:?}"), |
|
} |
|
} |
|
|
|
async fn build_dataflow(dataflow: &Path) -> eyre::Result<()> { |
|
let cargo = std::env::var("CARGO").unwrap(); |
|
let mut cmd = tokio::process::Command::new(&cargo); |
|
cmd.arg("run"); |
|
cmd.arg("--package").arg("dora-cli"); |
|
cmd.arg("--").arg("build").arg(dataflow); |
|
if !cmd.status().await?.success() { |
|
bail!("failed to build dataflow"); |
|
}; |
|
Ok(()) |
|
} |
|
|
|
async fn run_daemon(coordinator: String, machine_id: &str) -> eyre::Result<()> { |
|
let cargo = std::env::var("CARGO").unwrap(); |
|
let mut cmd = tokio::process::Command::new(&cargo); |
|
cmd.arg("run"); |
|
cmd.arg("--package").arg("dora-cli"); |
|
cmd.arg("--") |
|
.arg("daemon") |
|
.arg("--machine-id") |
|
.arg(machine_id) |
|
.arg("--coordinator-addr") |
|
.arg(coordinator); |
|
if !cmd.status().await?.success() { |
|
bail!("failed to run dataflow"); |
|
}; |
|
Ok(()) |
|
} |
|
|