Improve controller endpoint and model
This commit is contained in:
parent
2f51ebf91e
commit
6400b7745c
5 changed files with 103 additions and 7 deletions
|
@ -1,10 +1,18 @@
|
||||||
use actix_web::{get, web, HttpResponse};
|
use actix_web::{delete, get, put, web, HttpResponse};
|
||||||
|
use emgauwa_lib::db::errors::DatabaseError;
|
||||||
use emgauwa_lib::db::DbController;
|
use emgauwa_lib::db::DbController;
|
||||||
use emgauwa_lib::models::{convert_db_list, Controller};
|
use emgauwa_lib::models::{convert_db_list, Controller, FromDbModel};
|
||||||
|
use emgauwa_lib::types::ControllerUid;
|
||||||
|
use serde_derive::{Deserialize, Serialize};
|
||||||
use sqlx::{Pool, Sqlite};
|
use sqlx::{Pool, Sqlite};
|
||||||
|
|
||||||
use crate::handlers::errors::ApiError;
|
use crate::handlers::errors::ApiError;
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
pub struct RequestController {
|
||||||
|
name: String,
|
||||||
|
}
|
||||||
|
|
||||||
#[get("/api/v1/controllers")]
|
#[get("/api/v1/controllers")]
|
||||||
pub async fn index(pool: web::Data<Pool<Sqlite>>) -> Result<HttpResponse, ApiError> {
|
pub async fn index(pool: web::Data<Pool<Sqlite>>) -> Result<HttpResponse, ApiError> {
|
||||||
let mut pool_conn = pool.acquire().await?;
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
@ -15,3 +23,58 @@ pub async fn index(pool: web::Data<Pool<Sqlite>>) -> Result<HttpResponse, ApiErr
|
||||||
|
|
||||||
Ok(HttpResponse::Ok().json(controllers))
|
Ok(HttpResponse::Ok().json(controllers))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[get("/api/v1/controllers/{controller_id}")]
|
||||||
|
pub async fn show(
|
||||||
|
pool: web::Data<Pool<Sqlite>>,
|
||||||
|
path: web::Path<(String,)>,
|
||||||
|
) -> Result<HttpResponse, ApiError> {
|
||||||
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
|
||||||
|
let (controller_uid,) = path.into_inner();
|
||||||
|
let uid = ControllerUid::try_from(controller_uid.as_str()).or(Err(ApiError::BadUid))?;
|
||||||
|
|
||||||
|
let controller = DbController::get_by_uid(&mut pool_conn, &uid)
|
||||||
|
.await?
|
||||||
|
.ok_or(DatabaseError::NotFound)?;
|
||||||
|
|
||||||
|
let return_controller = Controller::from_db_model(&mut pool_conn, controller)?;
|
||||||
|
Ok(HttpResponse::Ok().json(return_controller))
|
||||||
|
}
|
||||||
|
|
||||||
|
#[put("/api/v1/controllers/{controller_id}")]
|
||||||
|
pub async fn update(
|
||||||
|
pool: web::Data<Pool<Sqlite>>,
|
||||||
|
path: web::Path<(String,)>,
|
||||||
|
data: web::Json<RequestController>,
|
||||||
|
) -> Result<HttpResponse, ApiError> {
|
||||||
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
|
||||||
|
let (controller_uid,) = path.into_inner();
|
||||||
|
let uid = ControllerUid::try_from(controller_uid.as_str()).or(Err(ApiError::BadUid))?;
|
||||||
|
|
||||||
|
let controller = DbController::get_by_uid(&mut pool_conn, &uid)
|
||||||
|
.await?
|
||||||
|
.ok_or(DatabaseError::NotFound)?;
|
||||||
|
|
||||||
|
let controller = controller
|
||||||
|
.update(&mut pool_conn, data.name.as_str(), controller.relay_count)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
let return_controller = Controller::from_db_model(&mut pool_conn, controller)?;
|
||||||
|
Ok(HttpResponse::Ok().json(return_controller))
|
||||||
|
}
|
||||||
|
|
||||||
|
#[delete("/api/v1/controllers/{controller_id}")]
|
||||||
|
pub async fn delete(
|
||||||
|
pool: web::Data<Pool<Sqlite>>,
|
||||||
|
path: web::Path<(String,)>,
|
||||||
|
) -> Result<HttpResponse, ApiError> {
|
||||||
|
let mut pool_conn = pool.acquire().await?;
|
||||||
|
|
||||||
|
let (controller_uid,) = path.into_inner();
|
||||||
|
let uid = ControllerUid::try_from(controller_uid.as_str()).or(Err(ApiError::BadUid))?;
|
||||||
|
|
||||||
|
DbController::delete_by_uid(&mut pool_conn, uid).await?;
|
||||||
|
Ok(HttpResponse::Ok().json("controller got deleted"))
|
||||||
|
}
|
||||||
|
|
|
@ -57,6 +57,9 @@ async fn main() -> std::io::Result<()> {
|
||||||
.app_data(web::Data::new(pool.clone()))
|
.app_data(web::Data::new(pool.clone()))
|
||||||
.app_data(web::Data::new(connected_controllers.clone()))
|
.app_data(web::Data::new(connected_controllers.clone()))
|
||||||
.service(handlers::v1::controllers::index)
|
.service(handlers::v1::controllers::index)
|
||||||
|
.service(handlers::v1::controllers::show)
|
||||||
|
.service(handlers::v1::controllers::update)
|
||||||
|
.service(handlers::v1::controllers::delete)
|
||||||
.service(handlers::v1::relays::index)
|
.service(handlers::v1::relays::index)
|
||||||
.service(handlers::v1::schedules::index)
|
.service(handlers::v1::schedules::index)
|
||||||
.service(handlers::v1::schedules::tagged)
|
.service(handlers::v1::schedules::tagged)
|
||||||
|
|
|
@ -4,8 +4,9 @@ use sqlx::pool::PoolConnection;
|
||||||
use sqlx::Sqlite;
|
use sqlx::Sqlite;
|
||||||
|
|
||||||
use crate::db::errors::DatabaseError;
|
use crate::db::errors::DatabaseError;
|
||||||
use crate::db::{DbController, DbRelay, DbSchedule};
|
use crate::db::{DbController, DbJunctionRelaySchedule, DbRelay, DbSchedule};
|
||||||
use crate::types::ControllerUid;
|
use crate::types::{ControllerUid, Weekday};
|
||||||
|
use crate::utils;
|
||||||
|
|
||||||
pub trait FromDbModel {
|
pub trait FromDbModel {
|
||||||
type DbModel: Clone;
|
type DbModel: Clone;
|
||||||
|
@ -40,6 +41,8 @@ pub struct Relay {
|
||||||
pub r: DbRelay,
|
pub r: DbRelay,
|
||||||
pub controller: DbController,
|
pub controller: DbController,
|
||||||
pub controller_id: ControllerUid,
|
pub controller_id: ControllerUid,
|
||||||
|
pub schedules: Vec<DbSchedule>,
|
||||||
|
pub active_schedule: DbSchedule,
|
||||||
pub tags: Vec<String>,
|
pub tags: Vec<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -96,10 +99,23 @@ impl FromDbModel for Relay {
|
||||||
let tags = executor::block_on(db_model.get_tags(conn))?;
|
let tags = executor::block_on(db_model.get_tags(conn))?;
|
||||||
let controller_id = cache.uid.clone();
|
let controller_id = cache.uid.clone();
|
||||||
|
|
||||||
|
let schedules =
|
||||||
|
executor::block_on(DbJunctionRelaySchedule::get_schedules(conn, &db_model))?;
|
||||||
|
|
||||||
|
let weekday = utils::get_weekday();
|
||||||
|
let active_schedule = executor::block_on(DbJunctionRelaySchedule::get_schedule(
|
||||||
|
conn,
|
||||||
|
&db_model,
|
||||||
|
weekday as Weekday,
|
||||||
|
))?
|
||||||
|
.ok_or(DatabaseError::NotFound)?;
|
||||||
|
|
||||||
Ok(Relay {
|
Ok(Relay {
|
||||||
r: db_model,
|
r: db_model,
|
||||||
controller: cache,
|
controller: cache,
|
||||||
controller_id,
|
controller_id,
|
||||||
|
schedules,
|
||||||
|
active_schedule,
|
||||||
tags,
|
tags,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
|
@ -67,9 +67,12 @@ impl<'r> Decode<'r, Sqlite> for ControllerUid {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl From<&str> for ControllerUid {
|
impl TryFrom<&str> for ControllerUid {
|
||||||
fn from(value: &str) -> Self {
|
type Error = uuid::Error;
|
||||||
Self(Uuid::from_str(value).unwrap())
|
|
||||||
|
fn try_from(value: &str) -> Result<Self, Self::Error> {
|
||||||
|
let uuid = Uuid::from_str(value)?;
|
||||||
|
Ok(Self(uuid))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,8 +1,11 @@
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
|
||||||
|
use chrono::Datelike;
|
||||||
use log::LevelFilter;
|
use log::LevelFilter;
|
||||||
use simple_logger::SimpleLogger;
|
use simple_logger::SimpleLogger;
|
||||||
|
|
||||||
|
use crate::types::Weekday;
|
||||||
|
|
||||||
pub fn load_settings<T>(config_name: &str, env_prefix: &str) -> T
|
pub fn load_settings<T>(config_name: &str, env_prefix: &str) -> T
|
||||||
where
|
where
|
||||||
for<'de> T: serde::Deserialize<'de>,
|
for<'de> T: serde::Deserialize<'de>,
|
||||||
|
@ -31,3 +34,11 @@ pub fn init_logging(level: &str) {
|
||||||
.init()
|
.init()
|
||||||
.expect("Error initializing logger.");
|
.expect("Error initializing logger.");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn get_weekday() -> Weekday {
|
||||||
|
(chrono::offset::Local::now()
|
||||||
|
.date_naive()
|
||||||
|
.weekday()
|
||||||
|
.number_from_monday()
|
||||||
|
- 1) as Weekday
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in a new issue