Add connected controllers hashmap for controller-ws
This commit is contained in:
parent
6459804e1f
commit
6536ff0792
17 changed files with 245 additions and 260 deletions
BIN
Cargo.lock
generated
BIN
Cargo.lock
generated
Binary file not shown.
|
@ -20,6 +20,6 @@ serde = "1.0"
|
||||||
serde_json = "1.0"
|
serde_json = "1.0"
|
||||||
serde_derive = "1.0"
|
serde_derive = "1.0"
|
||||||
|
|
||||||
sqlx = { version = "0.7", features = ["sqlite", "runtime-async-std", "macros", "chrono"] }
|
sqlx = { version = "0.7", features = ["sqlite", "runtime-tokio", "macros", "chrono"] }
|
||||||
|
|
||||||
futures = "0.3"
|
futures = "0.3"
|
||||||
|
|
|
@ -2,11 +2,11 @@ use std::str;
|
||||||
|
|
||||||
use crate::relay_loop::run_relay_loop;
|
use crate::relay_loop::run_relay_loop;
|
||||||
use crate::settings::Settings;
|
use crate::settings::Settings;
|
||||||
|
use emgauwa_lib::db;
|
||||||
use emgauwa_lib::db::{DbController, DbRelay};
|
use emgauwa_lib::db::{DbController, DbRelay};
|
||||||
use emgauwa_lib::handlers::v1::ws::controllers::ControllerWsAction;
|
use emgauwa_lib::handlers::v1::ws::controllers::ControllerWsAction;
|
||||||
use emgauwa_lib::models::convert_db_list;
|
use emgauwa_lib::models::{Controller, FromDbModel};
|
||||||
use emgauwa_lib::types::ControllerUid;
|
use emgauwa_lib::types::ControllerUid;
|
||||||
use emgauwa_lib::{db, models};
|
|
||||||
use futures::channel::mpsc;
|
use futures::channel::mpsc;
|
||||||
use futures::{future, pin_mut, SinkExt, StreamExt};
|
use futures::{future, pin_mut, SinkExt, StreamExt};
|
||||||
use sqlx::pool::PoolConnection;
|
use sqlx::pool::PoolConnection;
|
||||||
|
@ -64,37 +64,23 @@ async fn main() {
|
||||||
futures::executor::block_on(create_this_controller(&mut conn, &settings))
|
futures::executor::block_on(create_this_controller(&mut conn, &settings))
|
||||||
});
|
});
|
||||||
|
|
||||||
let db_relays: Vec<DbRelay> = settings
|
for relay in &settings.relays {
|
||||||
.relays
|
if DbRelay::get_by_controller_and_num(&mut conn, &db_controller, relay.number.unwrap())
|
||||||
.iter()
|
|
||||||
.map(|relay| {
|
|
||||||
futures::executor::block_on(async {
|
|
||||||
match DbRelay::get_by_controller_and_num(
|
|
||||||
&mut conn,
|
|
||||||
&db_controller,
|
|
||||||
relay.number.unwrap(),
|
|
||||||
)
|
|
||||||
.await
|
.await
|
||||||
.expect("Failed to get relay from database")
|
.expect("Failed to get relay from database")
|
||||||
|
.is_none()
|
||||||
{
|
{
|
||||||
None => create_this_relay(&mut conn, &db_controller, relay).await,
|
create_this_relay(&mut conn, &db_controller, relay).await;
|
||||||
Some(relay) => relay,
|
}
|
||||||
}
|
}
|
||||||
})
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
|
|
||||||
let db_controller = db_controller
|
let db_controller = db_controller
|
||||||
.update(&mut conn, &db_controller.name, db_relays.len() as i64)
|
.update(&mut conn, &db_controller.name, settings.relays.len() as i64)
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
let relays = convert_db_list(&mut conn, db_relays).expect("Failed to convert relays");
|
let this = Controller::from_db_model(&mut conn, db_controller)
|
||||||
|
.expect("Failed to convert database models");
|
||||||
let this = models::Controller {
|
|
||||||
controller: db_controller,
|
|
||||||
relays,
|
|
||||||
};
|
|
||||||
|
|
||||||
let url = format!(
|
let url = format!(
|
||||||
"ws://{}:{}/api/v1/ws/controllers",
|
"ws://{}:{}/api/v1/ws/controllers",
|
||||||
|
|
|
@ -1,12 +1,15 @@
|
||||||
use actix_cors::Cors;
|
use actix_cors::Cors;
|
||||||
|
use std::collections::HashMap;
|
||||||
use std::net::TcpListener;
|
use std::net::TcpListener;
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
use std::sync::{Arc, Mutex};
|
||||||
|
|
||||||
use crate::utils::drop_privileges;
|
use crate::utils::drop_privileges;
|
||||||
use actix_web::middleware::TrailingSlash;
|
use actix_web::middleware::TrailingSlash;
|
||||||
use actix_web::{middleware, web, App, HttpServer};
|
use actix_web::{middleware, web, App, HttpServer};
|
||||||
use emgauwa_lib::db::DbController;
|
use emgauwa_lib::db::DbController;
|
||||||
use emgauwa_lib::handlers;
|
use emgauwa_lib::handlers;
|
||||||
|
use emgauwa_lib::types::ConnectedControllersType;
|
||||||
use log::{trace, LevelFilter};
|
use log::{trace, LevelFilter};
|
||||||
use simple_logger::SimpleLogger;
|
use simple_logger::SimpleLogger;
|
||||||
|
|
||||||
|
@ -41,6 +44,8 @@ async fn main() -> std::io::Result<()> {
|
||||||
.expect("Error setting all controllers inactive");
|
.expect("Error setting all controllers inactive");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let connected_controllers: ConnectedControllersType = Arc::new(Mutex::new(HashMap::new()));
|
||||||
|
|
||||||
log::info!("Starting server on {}:{}", settings.host, settings.port);
|
log::info!("Starting server on {}:{}", settings.host, settings.port);
|
||||||
HttpServer::new(move || {
|
HttpServer::new(move || {
|
||||||
let cors = Cors::default().allow_any_method().allow_any_header();
|
let cors = Cors::default().allow_any_method().allow_any_header();
|
||||||
|
@ -59,6 +64,8 @@ async fn main() -> std::io::Result<()> {
|
||||||
.wrap(middleware::NormalizePath::new(TrailingSlash::Trim))
|
.wrap(middleware::NormalizePath::new(TrailingSlash::Trim))
|
||||||
.app_data(web::JsonConfig::default().error_handler(handlers::json_error_handler))
|
.app_data(web::JsonConfig::default().error_handler(handlers::json_error_handler))
|
||||||
.app_data(web::Data::new(pool.clone()))
|
.app_data(web::Data::new(pool.clone()))
|
||||||
|
.app_data(web::Data::new(connected_controllers.clone()))
|
||||||
|
.service(handlers::v1::controllers::index)
|
||||||
.service(handlers::v1::relays::index)
|
.service(handlers::v1::relays::index)
|
||||||
.service(handlers::v1::schedules::index)
|
.service(handlers::v1::schedules::index)
|
||||||
.service(handlers::v1::schedules::tagged)
|
.service(handlers::v1::schedules::tagged)
|
||||||
|
@ -67,7 +74,7 @@ async fn main() -> std::io::Result<()> {
|
||||||
.service(handlers::v1::schedules::add_list)
|
.service(handlers::v1::schedules::add_list)
|
||||||
.service(handlers::v1::schedules::update)
|
.service(handlers::v1::schedules::update)
|
||||||
.service(handlers::v1::schedules::delete)
|
.service(handlers::v1::schedules::delete)
|
||||||
.service(handlers::v1::ws::controllers::index)
|
.service(handlers::v1::ws::ws_controllers)
|
||||||
})
|
})
|
||||||
.listen(listener)?
|
.listen(listener)?
|
||||||
.run()
|
.run()
|
||||||
|
|
|
@ -18,7 +18,7 @@ config = "0.13"
|
||||||
|
|
||||||
chrono = { version = "0.4", features = ["serde"] }
|
chrono = { version = "0.4", features = ["serde"] }
|
||||||
|
|
||||||
sqlx = { version = "0.7", features = ["sqlite", "runtime-async-std", "macros", "chrono"] }
|
sqlx = { version = "0.7", features = ["sqlite", "runtime-tokio", "macros", "chrono"] }
|
||||||
libsqlite3-sys = { version = "*", features = ["bundled"] }
|
libsqlite3-sys = { version = "*", features = ["bundled"] }
|
||||||
|
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
|
|
|
@ -1 +1,5 @@
|
||||||
|
use std::time::Duration;
|
||||||
|
|
||||||
pub const DEFAULT_PORT: u16 = 4419;
|
pub const DEFAULT_PORT: u16 = 4419;
|
||||||
|
pub const HEARTBEAT_INTERVAL: Duration = Duration::from_secs(5);
|
||||||
|
pub const HEARTBEAT_TIMEOUT: Duration = Duration::from_secs(15);
|
||||||
|
|
|
@ -5,7 +5,7 @@ use sqlx::pool::PoolConnection;
|
||||||
use sqlx::Sqlite;
|
use sqlx::Sqlite;
|
||||||
|
|
||||||
use crate::db::errors::DatabaseError;
|
use crate::db::errors::DatabaseError;
|
||||||
use crate::db::DbTag;
|
use crate::db::{DbRelay, DbTag};
|
||||||
use crate::types::ControllerUid;
|
use crate::types::ControllerUid;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
@ -144,6 +144,20 @@ impl DbController {
|
||||||
.ok_or(DatabaseError::UpdateGetError)
|
.ok_or(DatabaseError::UpdateGetError)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub async fn get_relays(
|
||||||
|
&self,
|
||||||
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
|
) -> Result<Vec<DbRelay>, DatabaseError> {
|
||||||
|
sqlx::query_as!(
|
||||||
|
DbRelay,
|
||||||
|
"SELECT * FROM relays WHERE controller_id = ?",
|
||||||
|
self.id
|
||||||
|
)
|
||||||
|
.fetch_all(conn.deref_mut())
|
||||||
|
.await
|
||||||
|
.map_err(DatabaseError::from)
|
||||||
|
}
|
||||||
|
|
||||||
pub async fn all_inactive(conn: &mut PoolConnection<Sqlite>) -> Result<(), DatabaseError> {
|
pub async fn all_inactive(conn: &mut PoolConnection<Sqlite>) -> Result<(), DatabaseError> {
|
||||||
sqlx::query!("UPDATE controllers SET active = 0")
|
sqlx::query!("UPDATE controllers SET active = 0")
|
||||||
.execute(conn.deref_mut())
|
.execute(conn.deref_mut())
|
||||||
|
|
|
@ -20,20 +20,20 @@ pub struct DbRelay {
|
||||||
|
|
||||||
impl DbRelay {
|
impl DbRelay {
|
||||||
pub async fn get_all(conn: &mut PoolConnection<Sqlite>) -> Result<Vec<DbRelay>, DatabaseError> {
|
pub async fn get_all(conn: &mut PoolConnection<Sqlite>) -> Result<Vec<DbRelay>, DatabaseError> {
|
||||||
Ok(sqlx::query_as!(DbRelay, "SELECT * FROM relays")
|
sqlx::query_as!(DbRelay, "SELECT * FROM relays")
|
||||||
.fetch_all(conn.deref_mut())
|
.fetch_all(conn.deref_mut())
|
||||||
.await?)
|
.await
|
||||||
|
.map_err(DatabaseError::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn get(
|
pub async fn get(
|
||||||
conn: &mut PoolConnection<Sqlite>,
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
id: i64,
|
id: i64,
|
||||||
) -> Result<Option<DbRelay>, DatabaseError> {
|
) -> Result<Option<DbRelay>, DatabaseError> {
|
||||||
Ok(
|
|
||||||
sqlx::query_as!(DbRelay, "SELECT * FROM relays WHERE id = ?", id)
|
sqlx::query_as!(DbRelay, "SELECT * FROM relays WHERE id = ?", id)
|
||||||
.fetch_optional(conn.deref_mut())
|
.fetch_optional(conn.deref_mut())
|
||||||
.await?,
|
.await
|
||||||
)
|
.map_err(DatabaseError::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn get_by_controller_and_num(
|
pub async fn get_by_controller_and_num(
|
||||||
|
@ -41,14 +41,15 @@ impl DbRelay {
|
||||||
controller: &DbController,
|
controller: &DbController,
|
||||||
number: i64,
|
number: i64,
|
||||||
) -> Result<Option<DbRelay>, DatabaseError> {
|
) -> Result<Option<DbRelay>, DatabaseError> {
|
||||||
Ok(sqlx::query_as!(
|
sqlx::query_as!(
|
||||||
DbRelay,
|
DbRelay,
|
||||||
"SELECT * FROM relays WHERE controller_id = ? AND number = ?",
|
"SELECT * FROM relays WHERE controller_id = ? AND number = ?",
|
||||||
controller.id,
|
controller.id,
|
||||||
number
|
number
|
||||||
)
|
)
|
||||||
.fetch_optional(conn.deref_mut())
|
.fetch_optional(conn.deref_mut())
|
||||||
.await?)
|
.await
|
||||||
|
.map_err(DatabaseError::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn get_by_controller_and_num_or_create(
|
pub async fn get_by_controller_and_num_or_create(
|
||||||
|
@ -67,9 +68,10 @@ impl DbRelay {
|
||||||
conn: &mut PoolConnection<Sqlite>,
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
tag: &DbTag,
|
tag: &DbTag,
|
||||||
) -> Result<Vec<DbRelay>, DatabaseError> {
|
) -> Result<Vec<DbRelay>, DatabaseError> {
|
||||||
Ok(sqlx::query_as!(DbRelay, "SELECT schedule.* FROM relays AS schedule INNER JOIN junction_tag ON junction_tag.schedule_id = schedule.id WHERE junction_tag.tag_id = ?", tag.id)
|
sqlx::query_as!(DbRelay, "SELECT schedule.* FROM relays AS schedule INNER JOIN junction_tag ON junction_tag.schedule_id = schedule.id WHERE junction_tag.tag_id = ?", tag.id)
|
||||||
.fetch_all(conn.deref_mut())
|
.fetch_all(conn.deref_mut())
|
||||||
.await?)
|
.await
|
||||||
|
.map_err(DatabaseError::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn create(
|
pub async fn create(
|
||||||
|
@ -135,9 +137,10 @@ impl DbRelay {
|
||||||
&self,
|
&self,
|
||||||
conn: &mut PoolConnection<Sqlite>,
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
) -> Result<Vec<String>, DatabaseError> {
|
) -> Result<Vec<String>, DatabaseError> {
|
||||||
Ok(sqlx::query_scalar!("SELECT tag FROM tags INNER JOIN junction_tag ON junction_tag.tag_id = tags.id WHERE junction_tag.relay_id = ?", self.id)
|
sqlx::query_scalar!("SELECT tag FROM tags INNER JOIN junction_tag ON junction_tag.tag_id = tags.id WHERE junction_tag.relay_id = ?", self.id)
|
||||||
.fetch_all(conn.deref_mut())
|
.fetch_all(conn.deref_mut())
|
||||||
.await?)
|
.await
|
||||||
|
.map_err(DatabaseError::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn set_tags(
|
pub async fn set_tags(
|
||||||
|
|
26
emgauwa-lib/src/handlers/v1/controllers.rs
Normal file
26
emgauwa-lib/src/handlers/v1/controllers.rs
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
use actix_web::{get, web, HttpResponse};
|
||||||
|
|
||||||
|
use sqlx::{Pool, Sqlite};
|
||||||
|
|
||||||
|
use crate::db::DbController;
|
||||||
|
|
||||||
|
use crate::handlers::errors::ApiError;
|
||||||
|
use crate::models::{convert_db_list, Controller};
|
||||||
|
use crate::types::ConnectedControllersType;
|
||||||
|
|
||||||
|
#[get("/api/v1/controllers")]
|
||||||
|
pub async fn index(
|
||||||
|
pool: web::Data<Pool<Sqlite>>,
|
||||||
|
connected_controllers: web::Data<ConnectedControllersType>,
|
||||||
|
) -> Result<HttpResponse, ApiError> {
|
||||||
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
|
||||||
|
let db_controllers = DbController::get_all(&mut pool_conn).await?;
|
||||||
|
|
||||||
|
let controllers: Vec<Controller> = convert_db_list(&mut pool_conn, db_controllers)?;
|
||||||
|
|
||||||
|
let data = connected_controllers.lock().unwrap();
|
||||||
|
println!("{:?}", *data);
|
||||||
|
|
||||||
|
Ok(HttpResponse::Ok().json(controllers))
|
||||||
|
}
|
|
@ -1,3 +1,4 @@
|
||||||
|
pub mod controllers;
|
||||||
pub mod relays;
|
pub mod relays;
|
||||||
pub mod schedules;
|
pub mod schedules;
|
||||||
pub mod ws;
|
pub mod ws;
|
||||||
|
|
|
@ -24,138 +24,3 @@ pub async fn index(pool: web::Data<Pool<Sqlite>>) -> Result<HttpResponse, ApiErr
|
||||||
|
|
||||||
Ok(HttpResponse::Ok().json(relays))
|
Ok(HttpResponse::Ok().json(relays))
|
||||||
}
|
}
|
||||||
|
|
||||||
//#[get("/api/v1/tags/tag/{tag}")]
|
|
||||||
//pub async fn tagged(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// path: web::Path<(String,)>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let (tag,) = path.into_inner();
|
|
||||||
// let tag_db = Tag::get_by_tag(&mut pool_conn, &tag).await?;
|
|
||||||
//
|
|
||||||
// let relays = Relay::get_by_tag(&mut pool_conn, &tag_db).await?;
|
|
||||||
//
|
|
||||||
// let return_relays: Vec<ReturnRelay> =
|
|
||||||
// relays.into_iter().map(|s| ReturnRelay::from_relay(s, &mut pool_conn)).collect();
|
|
||||||
//
|
|
||||||
// Ok(HttpResponse::Ok().json(return_relays))
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//#[get("/api/v1/tags/{relay_id}")]
|
|
||||||
//pub async fn show(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// path: web::Path<(String,)>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let (relay_uid,) = path.into_inner();
|
|
||||||
// let uid = RelayUid::try_from(relay_uid.as_str()).or(Err(ApiError::BadUid))?;
|
|
||||||
//
|
|
||||||
// let relay = Relay::get_by_uid(&mut pool_conn, &uid).await?;
|
|
||||||
//
|
|
||||||
// let return_relay = ReturnRelay::from_relay(relay, &mut pool_conn);
|
|
||||||
// Ok(HttpResponse::Ok().json(return_relay))
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//#[post("/api/v1/tags")]
|
|
||||||
//pub async fn add(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// data: web::Json<RequestRelay>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let new_relay = Relay::create(&mut pool_conn, &data.name, &data.periods).await?;
|
|
||||||
//
|
|
||||||
// new_relay
|
|
||||||
// .set_tags(&mut pool_conn, data.tags.as_slice())
|
|
||||||
// .await?;
|
|
||||||
//
|
|
||||||
// let return_relay = ReturnRelay::from_relay(new_relay, &mut pool_conn);
|
|
||||||
// Ok(HttpResponse::Created().json(return_relay))
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//async fn add_list_single(
|
|
||||||
// conn: &mut PoolConnection<Sqlite>,
|
|
||||||
// request_relay: &RequestRelay,
|
|
||||||
//) -> Result<Relay, DatabaseError> {
|
|
||||||
// let new_relay =
|
|
||||||
// Relay::create(conn, &request_relay.name, &request_relay.periods).await?;
|
|
||||||
//
|
|
||||||
// new_relay
|
|
||||||
// .set_tags(conn, request_relay.tags.as_slice())
|
|
||||||
// .await?;
|
|
||||||
//
|
|
||||||
// Ok(new_relay)
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//#[post("/api/v1/tags/list")]
|
|
||||||
//pub async fn add_list(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// data: web::Json<Vec<RequestRelay>>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let result: Vec<Result<Relay, DatabaseError>> = data
|
|
||||||
// .as_slice()
|
|
||||||
// .iter()
|
|
||||||
// .map(|request_relay| {
|
|
||||||
// futures::executor::block_on(add_list_single(&mut pool_conn, request_relay))
|
|
||||||
// })
|
|
||||||
// .collect();
|
|
||||||
//
|
|
||||||
// let mut return_relays: Vec<ReturnRelay> = Vec::new();
|
|
||||||
// for relay in result {
|
|
||||||
// match relay {
|
|
||||||
// Ok(relay) => return_relays.push(ReturnRelay::from_relay(relay, &mut pool_conn)),
|
|
||||||
// Err(e) => return Ok(HttpResponse::from(e)),
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// Ok(HttpResponse::Created().json(return_relays))
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//#[put("/api/v1/tags/{relay_id}")]
|
|
||||||
//pub async fn update(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// path: web::Path<(String,)>,
|
|
||||||
// data: web::Json<RequestRelay>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let (relay_uid,) = path.into_inner();
|
|
||||||
// let uid = RelayUid::try_from(relay_uid.as_str()).or(Err(ApiError::BadUid))?;
|
|
||||||
//
|
|
||||||
// let relay = Relay::get_by_uid(&mut pool_conn, &uid).await?;
|
|
||||||
//
|
|
||||||
// let relay = relay
|
|
||||||
// .update(&mut pool_conn, data.name.as_str(), &data.periods)
|
|
||||||
// .await?;
|
|
||||||
//
|
|
||||||
// relay
|
|
||||||
// .set_tags(&mut pool_conn, data.tags.as_slice())
|
|
||||||
// .await?;
|
|
||||||
//
|
|
||||||
// let return_relay = ReturnRelay::from_relay(relay, &mut pool_conn);
|
|
||||||
// Ok(HttpResponse::Ok().json(return_relay))
|
|
||||||
//}
|
|
||||||
//
|
|
||||||
//#[delete("/api/v1/tags/{relay_id}")]
|
|
||||||
//pub async fn delete(
|
|
||||||
// pool: web::Data<Pool<Sqlite>>,
|
|
||||||
// path: web::Path<(String,)>,
|
|
||||||
//) -> Result<HttpResponse, ApiError> {
|
|
||||||
// let mut pool_conn = pool.acquire().await?;
|
|
||||||
//
|
|
||||||
// let (relay_uid,) = path.into_inner();
|
|
||||||
// let uid = RelayUid::try_from(relay_uid.as_str()).or(Err(ApiError::BadUid))?;
|
|
||||||
//
|
|
||||||
// match uid {
|
|
||||||
// RelayUid::Off => Err(ApiError::ProtectedRelay),
|
|
||||||
// RelayUid::On => Err(ApiError::ProtectedRelay),
|
|
||||||
// RelayUid::Any(_) => {
|
|
||||||
// Relay::delete_by_uid(&mut pool_conn, uid).await?;
|
|
||||||
// Ok(HttpResponse::Ok().json("relay got deleted"))
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
//}
|
|
||||||
|
|
|
@ -104,16 +104,10 @@ pub async fn add_list(
|
||||||
let mut pool_conn = pool.acquire().await?;
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
|
||||||
let mut db_schedules: Vec<DbSchedule> = Vec::new();
|
let mut db_schedules: Vec<DbSchedule> = Vec::new();
|
||||||
data.iter().try_for_each(|s| {
|
for s in data.iter() {
|
||||||
let new_s = futures::executor::block_on(add_list_single(&mut pool_conn, s));
|
let new_s = futures::executor::block_on(add_list_single(&mut pool_conn, s))?;
|
||||||
match new_s {
|
|
||||||
Ok(new_s) => {
|
|
||||||
db_schedules.push(new_s);
|
db_schedules.push(new_s);
|
||||||
Ok(())
|
|
||||||
}
|
}
|
||||||
Err(e) => Err(e),
|
|
||||||
}
|
|
||||||
})?;
|
|
||||||
|
|
||||||
let schedules: Vec<Schedule> = convert_db_list(&mut pool_conn, db_schedules)?;
|
let schedules: Vec<Schedule> = convert_db_list(&mut pool_conn, db_schedules)?;
|
||||||
Ok(HttpResponse::Created().json(schedules))
|
Ok(HttpResponse::Created().json(schedules))
|
||||||
|
|
|
@ -1,14 +1,15 @@
|
||||||
|
use crate::constants::{HEARTBEAT_INTERVAL, HEARTBEAT_TIMEOUT};
|
||||||
use crate::db::errors::DatabaseError;
|
use crate::db::errors::DatabaseError;
|
||||||
use crate::db::{DbController, DbRelay};
|
use crate::db::{DbController, DbRelay};
|
||||||
use crate::handlers::errors::ApiError;
|
use crate::models::{Controller, FromDbModel};
|
||||||
use crate::models::Controller;
|
use crate::types::{ConnectedControllersType, ControllerUid};
|
||||||
use actix::{Actor, StreamHandler};
|
use actix::{Actor, ActorContext, AsyncContext, StreamHandler};
|
||||||
use actix_web::{get, web, HttpRequest, HttpResponse};
|
|
||||||
use actix_web_actors::ws;
|
use actix_web_actors::ws;
|
||||||
use actix_web_actors::ws::ProtocolError;
|
use actix_web_actors::ws::ProtocolError;
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
use sqlx::pool::PoolConnection;
|
use sqlx::pool::PoolConnection;
|
||||||
use sqlx::{Pool, Sqlite};
|
use sqlx::{Pool, Sqlite};
|
||||||
|
use std::time::Instant;
|
||||||
use ws::Message;
|
use ws::Message;
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
@ -16,48 +17,35 @@ pub enum ControllerWsAction {
|
||||||
Register(Controller),
|
Register(Controller),
|
||||||
}
|
}
|
||||||
|
|
||||||
struct ControllerWs {
|
pub struct ControllerWs {
|
||||||
pub pool: Pool<Sqlite>,
|
pub pool: Pool<Sqlite>,
|
||||||
pub controller: Option<DbController>,
|
pub controller_uid: Option<ControllerUid>,
|
||||||
|
pub connected_controllers: ConnectedControllersType,
|
||||||
|
pub hb: Instant,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Actor for ControllerWs {
|
impl Actor for ControllerWs {
|
||||||
type Context = ws::WebsocketContext<Self>;
|
type Context = ws::WebsocketContext<Self>;
|
||||||
|
|
||||||
|
fn started(&mut self, ctx: &mut Self::Context) {
|
||||||
|
self.hb(ctx);
|
||||||
|
}
|
||||||
|
|
||||||
fn stopped(&mut self, _ctx: &mut Self::Context) {
|
fn stopped(&mut self, _ctx: &mut Self::Context) {
|
||||||
if let Some(controller) = &self.controller {
|
if let Some(controller_uid) = &self.controller_uid {
|
||||||
let mut pool_conn = futures::executor::block_on(self.pool.acquire()).unwrap();
|
|
||||||
futures::executor::block_on(controller.update_active(&mut pool_conn, false)).unwrap();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl StreamHandler<Result<Message, ProtocolError>> for ControllerWs {
|
|
||||||
fn handle(&mut self, msg: Result<Message, ProtocolError>, ctx: &mut Self::Context) {
|
|
||||||
let mut pool_conn = futures::executor::block_on(self.pool.acquire()).unwrap();
|
let mut pool_conn = futures::executor::block_on(self.pool.acquire()).unwrap();
|
||||||
|
|
||||||
match msg {
|
let mut data = self.connected_controllers.lock().unwrap();
|
||||||
Ok(Message::Ping(msg)) => ctx.pong(&msg),
|
if let Some(controller) = data.remove(controller_uid) {
|
||||||
Ok(Message::Text(text)) => {
|
futures::executor::block_on(controller.c.update_active(&mut pool_conn, false))
|
||||||
let action: ControllerWsAction = serde_json::from_str(&text).unwrap();
|
.unwrap();
|
||||||
let action_res =
|
|
||||||
futures::executor::block_on(self.handle_action(&mut pool_conn, action));
|
|
||||||
if let Err(e) = action_res {
|
|
||||||
log::error!("Error handling action: {:?}", e);
|
|
||||||
ctx.text(serde_json::to_string(&e).unwrap());
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
_ => {}
|
|
||||||
}
|
|
||||||
|
|
||||||
//let schedules = futures::executor::block_on(DbSchedule::get_all(&mut pool_conn)).unwrap();
|
|
||||||
//let schedules_json = serde_json::to_string(&schedules).unwrap();
|
|
||||||
//ctx.text(schedules_json);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ControllerWs {
|
impl ControllerWs {
|
||||||
pub async fn handle_action(
|
pub fn handle_action(
|
||||||
&mut self,
|
&mut self,
|
||||||
conn: &mut PoolConnection<Sqlite>,
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
action: ControllerWsAction,
|
action: ControllerWsAction,
|
||||||
|
@ -65,49 +53,94 @@ impl ControllerWs {
|
||||||
match action {
|
match action {
|
||||||
ControllerWsAction::Register(controller) => {
|
ControllerWsAction::Register(controller) => {
|
||||||
log::info!("Registering controller: {:?}", controller);
|
log::info!("Registering controller: {:?}", controller);
|
||||||
let c = &controller.controller;
|
let c = &controller.c;
|
||||||
let controller_db =
|
let controller_db = futures::executor::block_on(
|
||||||
DbController::get_by_uid_or_create(conn, &c.uid, &c.name, c.relay_count)
|
DbController::get_by_uid_or_create(conn, &c.uid, &c.name, c.relay_count),
|
||||||
.await?;
|
)?;
|
||||||
controller_db.update_active(conn, true).await?;
|
futures::executor::block_on(controller_db.update_active(conn, true))?;
|
||||||
|
|
||||||
println!("Controller: {:?}", controller_db);
|
|
||||||
|
|
||||||
for relay in &controller.relays {
|
for relay in &controller.relays {
|
||||||
let r = &relay.relay;
|
let r = &relay.r;
|
||||||
let relay_db = DbRelay::get_by_controller_and_num_or_create(
|
futures::executor::block_on(DbRelay::get_by_controller_and_num_or_create(
|
||||||
conn,
|
conn,
|
||||||
&controller_db,
|
&controller_db,
|
||||||
r.number,
|
r.number,
|
||||||
&r.name,
|
&r.name,
|
||||||
)
|
))?;
|
||||||
.await?;
|
|
||||||
println!("Controller relay: {:?}", relay_db);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
self.controller = Some(controller_db);
|
let controller = Controller::from_db_model(conn, controller_db)?;
|
||||||
|
|
||||||
|
let controller_uid = &controller.c.uid;
|
||||||
|
self.controller_uid = Some(controller_uid.clone());
|
||||||
|
|
||||||
|
let mut data = self.connected_controllers.lock().unwrap();
|
||||||
|
data.insert(controller_uid.clone(), controller);
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// helper method that sends ping to client every 5 seconds (HEARTBEAT_INTERVAL).
|
||||||
|
///
|
||||||
|
/// also this method checks heartbeats from client
|
||||||
|
fn hb(&self, ctx: &mut ws::WebsocketContext<Self>) {
|
||||||
|
ctx.run_interval(HEARTBEAT_INTERVAL, |act, ctx| {
|
||||||
|
// check client heartbeats
|
||||||
|
if Instant::now().duration_since(act.hb) > HEARTBEAT_TIMEOUT {
|
||||||
|
log::warn!("Websocket Controller heartbeat failed, disconnecting!");
|
||||||
|
ctx.stop();
|
||||||
|
// don't try to send a ping
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
#[get("/api/v1/ws/controllers")]
|
ctx.ping(&[]);
|
||||||
pub async fn index(
|
});
|
||||||
pool: web::Data<Pool<Sqlite>>,
|
}
|
||||||
req: HttpRequest,
|
}
|
||||||
stream: web::Payload,
|
|
||||||
) -> Result<HttpResponse, ApiError> {
|
impl StreamHandler<Result<Message, ProtocolError>> for ControllerWs {
|
||||||
let resp = ws::start(
|
fn handle(&mut self, msg: Result<Message, ProtocolError>, ctx: &mut Self::Context) {
|
||||||
ControllerWs {
|
let mut pool_conn = futures::executor::block_on(self.pool.acquire()).unwrap();
|
||||||
pool: pool.get_ref().clone(),
|
|
||||||
controller: None,
|
let msg = match msg {
|
||||||
},
|
Err(_) => {
|
||||||
&req,
|
ctx.stop();
|
||||||
stream,
|
return;
|
||||||
)
|
}
|
||||||
.map_err(|_| ApiError::InternalError(String::from("error starting websocket")));
|
Ok(msg) => msg,
|
||||||
println!("{:?}", resp);
|
};
|
||||||
resp
|
|
||||||
|
match msg {
|
||||||
|
Message::Ping(msg) => {
|
||||||
|
self.hb = Instant::now();
|
||||||
|
ctx.pong(&msg)
|
||||||
|
}
|
||||||
|
Message::Pong(_) => {
|
||||||
|
self.hb = Instant::now();
|
||||||
|
}
|
||||||
|
Message::Text(text) => {
|
||||||
|
let action: ControllerWsAction = serde_json::from_str(&text).unwrap();
|
||||||
|
let action_res = self.handle_action(&mut pool_conn, action);
|
||||||
|
if let Err(e) = action_res {
|
||||||
|
log::error!("Error handling action: {:?}", e);
|
||||||
|
ctx.text(serde_json::to_string(&e).unwrap());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Message::Binary(_) => log::warn!("Received unexpected binary in controller ws"),
|
||||||
|
Message::Close(reason) => {
|
||||||
|
ctx.close(reason);
|
||||||
|
ctx.stop();
|
||||||
|
}
|
||||||
|
Message::Continuation(_) => {
|
||||||
|
ctx.stop();
|
||||||
|
}
|
||||||
|
Message::Nop => (),
|
||||||
|
}
|
||||||
|
|
||||||
|
//let schedules = futures::executor::block_on(DbSchedule::get_all(&mut pool_conn)).unwrap();
|
||||||
|
//let schedules_json = serde_json::to_string(&schedules).unwrap();
|
||||||
|
//ctx.text(schedules_json);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1 +1,31 @@
|
||||||
|
use crate::handlers::errors::ApiError;
|
||||||
|
use crate::handlers::v1::ws::controllers::ControllerWs;
|
||||||
|
use crate::types::ConnectedControllersType;
|
||||||
|
use actix_web::{get, web, HttpRequest, HttpResponse};
|
||||||
|
use actix_web_actors::ws;
|
||||||
|
use sqlx::{Pool, Sqlite};
|
||||||
|
use std::time::Instant;
|
||||||
|
|
||||||
pub mod controllers;
|
pub mod controllers;
|
||||||
|
|
||||||
|
#[get("/api/v1/ws/controllers")]
|
||||||
|
pub async fn ws_controllers(
|
||||||
|
pool: web::Data<Pool<Sqlite>>,
|
||||||
|
connected_controllers: web::Data<ConnectedControllersType>,
|
||||||
|
req: HttpRequest,
|
||||||
|
stream: web::Payload,
|
||||||
|
) -> Result<HttpResponse, ApiError> {
|
||||||
|
let resp = ws::start(
|
||||||
|
ControllerWs {
|
||||||
|
pool: pool.get_ref().clone(),
|
||||||
|
controller_uid: None,
|
||||||
|
connected_controllers: connected_controllers.get_ref().clone(),
|
||||||
|
hb: Instant::now(),
|
||||||
|
},
|
||||||
|
&req,
|
||||||
|
stream,
|
||||||
|
)
|
||||||
|
.map_err(|_| ApiError::InternalError(String::from("error starting websocket")));
|
||||||
|
println!("{:?}", resp);
|
||||||
|
resp
|
||||||
|
}
|
||||||
|
|
|
@ -1,6 +1,5 @@
|
||||||
use crate::db;
|
|
||||||
use crate::db::errors::DatabaseError;
|
use crate::db::errors::DatabaseError;
|
||||||
use crate::db::{DbRelay, DbSchedule};
|
use crate::db::{DbController, DbRelay, DbSchedule};
|
||||||
use crate::types::ControllerUid;
|
use crate::types::ControllerUid;
|
||||||
use futures::executor;
|
use futures::executor;
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
|
@ -21,15 +20,15 @@ pub trait FromDbModel {
|
||||||
#[derive(Serialize, Deserialize, Debug)]
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
pub struct Schedule {
|
pub struct Schedule {
|
||||||
#[serde(flatten)]
|
#[serde(flatten)]
|
||||||
pub schedule: DbSchedule,
|
pub s: DbSchedule,
|
||||||
pub tags: Vec<String>,
|
pub tags: Vec<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Serialize, Deserialize, Debug)]
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
pub struct Relay {
|
pub struct Relay {
|
||||||
#[serde(flatten)]
|
#[serde(flatten)]
|
||||||
pub relay: DbRelay,
|
pub r: DbRelay,
|
||||||
pub controller: db::DbController,
|
pub controller: DbController,
|
||||||
pub controller_id: ControllerUid,
|
pub controller_id: ControllerUid,
|
||||||
pub tags: Vec<String>,
|
pub tags: Vec<String>,
|
||||||
}
|
}
|
||||||
|
@ -37,7 +36,7 @@ pub struct Relay {
|
||||||
#[derive(Serialize, Deserialize, Debug)]
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
pub struct Controller {
|
pub struct Controller {
|
||||||
#[serde(flatten)]
|
#[serde(flatten)]
|
||||||
pub controller: db::DbController,
|
pub c: DbController,
|
||||||
pub relays: Vec<Relay>,
|
pub relays: Vec<Relay>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -51,7 +50,7 @@ impl FromDbModel for Schedule {
|
||||||
let schedule = db_model.clone();
|
let schedule = db_model.clone();
|
||||||
let tags = executor::block_on(schedule.get_tags(conn))?;
|
let tags = executor::block_on(schedule.get_tags(conn))?;
|
||||||
|
|
||||||
Ok(Schedule { schedule, tags })
|
Ok(Schedule { s: schedule, tags })
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -68,7 +67,7 @@ impl FromDbModel for Relay {
|
||||||
let tags = executor::block_on(relay.get_tags(conn))?;
|
let tags = executor::block_on(relay.get_tags(conn))?;
|
||||||
|
|
||||||
Ok(Relay {
|
Ok(Relay {
|
||||||
relay,
|
r: relay,
|
||||||
controller,
|
controller,
|
||||||
controller_id,
|
controller_id,
|
||||||
tags,
|
tags,
|
||||||
|
@ -76,6 +75,22 @@ impl FromDbModel for Relay {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl FromDbModel for Controller {
|
||||||
|
type DbModel = DbController;
|
||||||
|
|
||||||
|
fn from_db_model(
|
||||||
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
|
db_model: Self::DbModel,
|
||||||
|
) -> Result<Self, DatabaseError> {
|
||||||
|
let relays_db = executor::block_on(db_model.get_relays(conn))?;
|
||||||
|
let relays = convert_db_list(conn, relays_db)?;
|
||||||
|
Ok(Controller {
|
||||||
|
c: db_model,
|
||||||
|
relays,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub fn convert_db_list<T: FromDbModel>(
|
pub fn convert_db_list<T: FromDbModel>(
|
||||||
conn: &mut PoolConnection<Sqlite>,
|
conn: &mut PoolConnection<Sqlite>,
|
||||||
db_models: Vec<T::DbModel>,
|
db_models: Vec<T::DbModel>,
|
||||||
|
|
|
@ -7,7 +7,7 @@ use sqlx::{Decode, Encode, Sqlite, Type};
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
|
||||||
pub struct ControllerUid(Uuid);
|
pub struct ControllerUid(Uuid);
|
||||||
|
|
||||||
impl Default for ControllerUid {
|
impl Default for ControllerUid {
|
||||||
|
|
|
@ -1,5 +1,12 @@
|
||||||
mod controller_uid;
|
mod controller_uid;
|
||||||
mod schedule_uid;
|
mod schedule_uid;
|
||||||
|
|
||||||
|
use crate::models::Controller;
|
||||||
|
|
||||||
pub use controller_uid::ControllerUid;
|
pub use controller_uid::ControllerUid;
|
||||||
pub use schedule_uid::ScheduleUid;
|
pub use schedule_uid::ScheduleUid;
|
||||||
|
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::sync::{Arc, Mutex};
|
||||||
|
|
||||||
|
pub type ConnectedControllersType = Arc<Mutex<HashMap<ControllerUid, Controller>>>;
|
||||||
|
|
Loading…
Reference in a new issue