cistine's picture
Upload 490 files
b98ffbb verified
raw
history blame
7.55 kB
use dora_coordinator::{ControlEvent, Event};
use dora_core::{
descriptor::Descriptor,
topics::{
ControlRequest, ControlRequestReply, DataflowId, DORA_COORDINATOR_PORT_CONTROL_DEFAULT,
DORA_COORDINATOR_PORT_DEFAULT,
},
};
use dora_tracing::set_up_tracing;
use eyre::{bail, Context};
use std::{
collections::BTreeSet,
net::{IpAddr, Ipv4Addr, SocketAddr},
path::Path,
time::Duration,
};
use tokio::{
sync::{
mpsc::{self, Sender},
oneshot,
},
task::JoinSet,
};
use tokio_stream::wrappers::ReceiverStream;
use uuid::Uuid;
#[tokio::main]
async fn main() -> eyre::Result<()> {
set_up_tracing("multiple-daemon-runner").wrap_err("failed to set up tracing subscriber")?;
let root = Path::new(env!("CARGO_MANIFEST_DIR"));
std::env::set_current_dir(root.join(file!()).parent().unwrap())
.wrap_err("failed to set working dir")?;
let dataflow = Path::new("dataflow.yml");
build_dataflow(dataflow).await?;
let (coordinator_events_tx, coordinator_events_rx) = mpsc::channel(1);
let coordinator_bind = SocketAddr::new(
IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)),
DORA_COORDINATOR_PORT_DEFAULT,
);
let coordinator_control_bind = SocketAddr::new(
IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)),
DORA_COORDINATOR_PORT_CONTROL_DEFAULT,
);
let (coordinator_port, coordinator) = dora_coordinator::start(
coordinator_bind,
coordinator_control_bind,
ReceiverStream::new(coordinator_events_rx),
)
.await?;
let coordinator_addr = SocketAddr::new(Ipv4Addr::LOCALHOST.into(), coordinator_port);
let daemon_a = run_daemon(coordinator_addr.to_string(), "A");
let daemon_b = run_daemon(coordinator_addr.to_string(), "B");
tracing::info!("Spawning coordinator and daemons");
let mut tasks = JoinSet::new();
tasks.spawn(coordinator);
tasks.spawn(daemon_a);
tasks.spawn(daemon_b);
tracing::info!("waiting until daemons are connected to coordinator");
let mut retries = 0;
loop {
let connected_machines = connected_machines(&coordinator_events_tx).await?;
if connected_machines.contains("A") && connected_machines.contains("B") {
break;
} else if retries > 20 {
bail!("daemon not connected after {retries} retries");
} else {
std::thread::sleep(Duration::from_millis(500));
retries += 1
}
}
tracing::info!("starting dataflow");
let uuid = start_dataflow(dataflow, &coordinator_events_tx).await?;
tracing::info!("started dataflow under ID `{uuid}`");
let running = running_dataflows(&coordinator_events_tx).await?;
if !running.iter().map(|d| d.uuid).any(|id| id == uuid) {
bail!("dataflow `{uuid}` is not running");
}
tracing::info!("waiting for dataflow `{uuid}` to finish");
let mut retries = 0;
loop {
let running = running_dataflows(&coordinator_events_tx).await?;
if running.is_empty() {
break;
} else if retries > 100 {
bail!("dataflow not finished after {retries} retries");
} else {
tracing::debug!("not done yet");
std::thread::sleep(Duration::from_millis(500));
retries += 1
}
}
tracing::info!("dataflow `{uuid}` finished, destroying coordinator");
destroy(&coordinator_events_tx).await?;
tracing::info!("joining tasks");
while let Some(res) = tasks.join_next().await {
res.unwrap()?;
}
tracing::info!("done");
Ok(())
}
async fn start_dataflow(
dataflow: &Path,
coordinator_events_tx: &Sender<Event>,
) -> eyre::Result<Uuid> {
let dataflow_descriptor = Descriptor::read(dataflow)
.await
.wrap_err("failed to read yaml dataflow")?;
let working_dir = dataflow
.canonicalize()
.context("failed to canonicalize dataflow path")?
.parent()
.ok_or_else(|| eyre::eyre!("dataflow path has no parent dir"))?
.to_owned();
dataflow_descriptor
.check(&working_dir)
.wrap_err("could not validate yaml")?;
let (reply_sender, reply) = oneshot::channel();
coordinator_events_tx
.send(Event::Control(ControlEvent::IncomingRequest {
request: ControlRequest::Start {
dataflow: dataflow_descriptor,
local_working_dir: working_dir,
name: None,
},
reply_sender,
}))
.await?;
let result = reply.await??;
let uuid = match result {
ControlRequestReply::DataflowStarted { uuid } => uuid,
ControlRequestReply::Error(err) => bail!("{err}"),
other => bail!("unexpected start dataflow reply: {other:?}"),
};
Ok(uuid)
}
async fn connected_machines(
coordinator_events_tx: &Sender<Event>,
) -> eyre::Result<BTreeSet<String>> {
let (reply_sender, reply) = oneshot::channel();
coordinator_events_tx
.send(Event::Control(ControlEvent::IncomingRequest {
request: ControlRequest::ConnectedMachines,
reply_sender,
}))
.await?;
let result = reply.await??;
let machines = match result {
ControlRequestReply::ConnectedMachines(machines) => machines,
ControlRequestReply::Error(err) => bail!("{err}"),
other => bail!("unexpected start dataflow reply: {other:?}"),
};
Ok(machines)
}
async fn running_dataflows(coordinator_events_tx: &Sender<Event>) -> eyre::Result<Vec<DataflowId>> {
let (reply_sender, reply) = oneshot::channel();
coordinator_events_tx
.send(Event::Control(ControlEvent::IncomingRequest {
request: ControlRequest::List,
reply_sender,
}))
.await?;
let result = reply.await??;
let dataflows = match result {
ControlRequestReply::DataflowList { dataflows } => dataflows,
ControlRequestReply::Error(err) => bail!("{err}"),
other => bail!("unexpected start dataflow reply: {other:?}"),
};
Ok(dataflows)
}
async fn destroy(coordinator_events_tx: &Sender<Event>) -> eyre::Result<()> {
let (reply_sender, reply) = oneshot::channel();
coordinator_events_tx
.send(Event::Control(ControlEvent::IncomingRequest {
request: ControlRequest::Destroy,
reply_sender,
}))
.await?;
let result = reply.await??;
match result {
ControlRequestReply::DestroyOk => Ok(()),
ControlRequestReply::Error(err) => bail!("{err}"),
other => bail!("unexpected start dataflow reply: {other:?}"),
}
}
async fn build_dataflow(dataflow: &Path) -> eyre::Result<()> {
let cargo = std::env::var("CARGO").unwrap();
let mut cmd = tokio::process::Command::new(&cargo);
cmd.arg("run");
cmd.arg("--package").arg("dora-cli");
cmd.arg("--").arg("build").arg(dataflow);
if !cmd.status().await?.success() {
bail!("failed to build dataflow");
};
Ok(())
}
async fn run_daemon(coordinator: String, machine_id: &str) -> eyre::Result<()> {
let cargo = std::env::var("CARGO").unwrap();
let mut cmd = tokio::process::Command::new(&cargo);
cmd.arg("run");
cmd.arg("--package").arg("dora-cli");
cmd.arg("--")
.arg("daemon")
.arg("--machine-id")
.arg(machine_id)
.arg("--coordinator-addr")
.arg(coordinator);
if !cmd.status().await?.success() {
bail!("failed to run dataflow");
};
Ok(())
}