crates/server: enhance auth middleware and error responses
Signed-off-by: NotAShelf <raf@notashelf.dev> Change-Id: I48a780779d884c4a7730347f920b91216a6a6964
This commit is contained in:
parent
000fb8994f
commit
92153bf9aa
8 changed files with 272 additions and 99 deletions
|
|
@ -9,8 +9,8 @@ use sha2::{Digest, Sha256};
|
||||||
|
|
||||||
use crate::state::AppState;
|
use crate::state::AppState;
|
||||||
|
|
||||||
/// Extract and validate an API key from the Authorization header.
|
/// Extract and validate an API key from the Authorization header or session cookie.
|
||||||
/// Keys use the format: `Bearer fc_xxxx`.
|
/// Keys use the format: `Bearer fc_xxxx`. Session cookies use `fc_session=<id>`.
|
||||||
/// Write endpoints (POST/PUT/DELETE/PATCH) require a valid key.
|
/// Write endpoints (POST/PUT/DELETE/PATCH) require a valid key.
|
||||||
/// Read endpoints (GET/HEAD/OPTIONS) try to extract optionally (for dashboard admin UI).
|
/// Read endpoints (GET/HEAD/OPTIONS) try to extract optionally (for dashboard admin UI).
|
||||||
pub async fn require_api_key(
|
pub async fn require_api_key(
|
||||||
|
|
@ -33,42 +33,44 @@ pub async fn require_api_key(
|
||||||
.as_deref()
|
.as_deref()
|
||||||
.and_then(|h| h.strip_prefix("Bearer "));
|
.and_then(|h| h.strip_prefix("Bearer "));
|
||||||
|
|
||||||
match token {
|
// Try Bearer token first
|
||||||
Some(token) => {
|
if let Some(token) = token {
|
||||||
let mut hasher = Sha256::new();
|
let mut hasher = Sha256::new();
|
||||||
hasher.update(token.as_bytes());
|
hasher.update(token.as_bytes());
|
||||||
let key_hash = hex::encode(hasher.finalize());
|
let key_hash = hex::encode(hasher.finalize());
|
||||||
|
|
||||||
match fc_common::repo::api_keys::get_by_hash(&state.pool, &key_hash).await {
|
if let Ok(Some(api_key)) =
|
||||||
Ok(Some(api_key)) => {
|
fc_common::repo::api_keys::get_by_hash(&state.pool, &key_hash).await
|
||||||
// Touch last_used_at (fire and forget)
|
{
|
||||||
let pool = state.pool.clone();
|
let pool = state.pool.clone();
|
||||||
let key_id = api_key.id;
|
let key_id = api_key.id;
|
||||||
tokio::spawn(async move {
|
tokio::spawn(async move {
|
||||||
let _ = fc_common::repo::api_keys::touch_last_used(&pool, key_id).await;
|
let _ = fc_common::repo::api_keys::touch_last_used(&pool, key_id).await;
|
||||||
});
|
});
|
||||||
|
|
||||||
request.extensions_mut().insert(api_key);
|
request.extensions_mut().insert(api_key);
|
||||||
Ok(next.run(request).await)
|
return Ok(next.run(request).await);
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
if is_read {
|
|
||||||
// Invalid token on read is still allowed, just no ApiKey in extensions
|
|
||||||
Ok(next.run(request).await)
|
|
||||||
} else {
|
|
||||||
Err(StatusCode::UNAUTHORIZED)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
if is_read {
|
|
||||||
Ok(next.run(request).await)
|
|
||||||
} else {
|
|
||||||
Err(StatusCode::UNAUTHORIZED)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Fall back to session cookie (so dashboard JS fetches work)
|
||||||
|
if let Some(cookie_header) = request
|
||||||
|
.headers()
|
||||||
|
.get("cookie")
|
||||||
|
.and_then(|v| v.to_str().ok())
|
||||||
|
&& let Some(session_id) = parse_cookie(cookie_header, "fc_session")
|
||||||
|
&& let Some(session) = state.sessions.get(&session_id)
|
||||||
|
&& session.created_at.elapsed() < std::time::Duration::from_secs(24 * 60 * 60) {
|
||||||
|
request.extensions_mut().insert(session.api_key.clone());
|
||||||
|
return Ok(next.run(request).await);
|
||||||
|
}
|
||||||
|
|
||||||
|
// No valid auth found
|
||||||
|
if is_read {
|
||||||
|
Ok(next.run(request).await)
|
||||||
|
} else {
|
||||||
|
Err(StatusCode::UNAUTHORIZED)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Extractor that requires an authenticated admin user.
|
/// Extractor that requires an authenticated admin user.
|
||||||
|
|
@ -129,9 +131,8 @@ pub async fn extract_session(
|
||||||
.headers()
|
.headers()
|
||||||
.get("cookie")
|
.get("cookie")
|
||||||
.and_then(|v| v.to_str().ok())
|
.and_then(|v| v.to_str().ok())
|
||||||
{
|
&& let Some(session_id) = parse_cookie(cookie_header, "fc_session")
|
||||||
if let Some(session_id) = parse_cookie(cookie_header, "fc_session") {
|
&& let Some(session) = state.sessions.get(&session_id) {
|
||||||
if let Some(session) = state.sessions.get(&session_id) {
|
|
||||||
// Check session expiry (24 hours)
|
// Check session expiry (24 hours)
|
||||||
if session.created_at.elapsed() < std::time::Duration::from_secs(24 * 60 * 60) {
|
if session.created_at.elapsed() < std::time::Duration::from_secs(24 * 60 * 60) {
|
||||||
request.extensions_mut().insert(session.api_key.clone());
|
request.extensions_mut().insert(session.api_key.clone());
|
||||||
|
|
@ -141,12 +142,10 @@ pub async fn extract_session(
|
||||||
state.sessions.remove(&session_id);
|
state.sessions.remove(&session_id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
|
||||||
next.run(request).await
|
next.run(request).await
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_cookie<'a>(header: &'a str, name: &str) -> Option<String> {
|
fn parse_cookie(header: &str, name: &str) -> Option<String> {
|
||||||
header
|
header
|
||||||
.split(';')
|
.split(';')
|
||||||
.filter_map(|pair| {
|
.filter_map(|pair| {
|
||||||
|
|
|
||||||
|
|
@ -22,18 +22,60 @@ impl IntoResponse for ApiError {
|
||||||
CiError::Timeout(msg) => (StatusCode::REQUEST_TIMEOUT, "TIMEOUT", msg.clone()),
|
CiError::Timeout(msg) => (StatusCode::REQUEST_TIMEOUT, "TIMEOUT", msg.clone()),
|
||||||
CiError::Unauthorized(msg) => (StatusCode::UNAUTHORIZED, "UNAUTHORIZED", msg.clone()),
|
CiError::Unauthorized(msg) => (StatusCode::UNAUTHORIZED, "UNAUTHORIZED", msg.clone()),
|
||||||
CiError::Forbidden(msg) => (StatusCode::FORBIDDEN, "FORBIDDEN", msg.clone()),
|
CiError::Forbidden(msg) => (StatusCode::FORBIDDEN, "FORBIDDEN", msg.clone()),
|
||||||
CiError::Database(_) => (
|
CiError::NixEval(msg) => (
|
||||||
StatusCode::INTERNAL_SERVER_ERROR,
|
StatusCode::UNPROCESSABLE_ENTITY,
|
||||||
"DATABASE_ERROR",
|
"NIX_EVAL_ERROR",
|
||||||
"Internal database error".to_string(),
|
msg.clone(),
|
||||||
),
|
),
|
||||||
_ => (
|
CiError::Build(msg) => (StatusCode::UNPROCESSABLE_ENTITY, "BUILD_ERROR", msg.clone()),
|
||||||
|
CiError::Config(msg) => (
|
||||||
StatusCode::INTERNAL_SERVER_ERROR,
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
"INTERNAL_ERROR",
|
"CONFIG_ERROR",
|
||||||
"Internal server error".to_string(),
|
msg.clone(),
|
||||||
),
|
),
|
||||||
|
CiError::Database(e) => {
|
||||||
|
tracing::error!(error = %e, "Database error in API handler");
|
||||||
|
(
|
||||||
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
|
"DATABASE_ERROR",
|
||||||
|
"Internal database error".to_string(),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
CiError::Git(e) => {
|
||||||
|
tracing::error!(error = %e, "Git error in API handler");
|
||||||
|
(
|
||||||
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
|
"GIT_ERROR",
|
||||||
|
format!("Git operation failed: {e}"),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
CiError::Serialization(e) => {
|
||||||
|
tracing::error!(error = %e, "Serialization error in API handler");
|
||||||
|
(
|
||||||
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
|
"SERIALIZATION_ERROR",
|
||||||
|
format!("Data serialization error: {e}"),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
CiError::Io(e) => {
|
||||||
|
tracing::error!(error = %e, "IO error in API handler");
|
||||||
|
(
|
||||||
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
|
"IO_ERROR",
|
||||||
|
format!("IO error: {e}"),
|
||||||
|
)
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
if status.is_server_error() {
|
||||||
|
tracing::warn!(
|
||||||
|
status = %status,
|
||||||
|
code = code,
|
||||||
|
"API error response: {}",
|
||||||
|
message
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
let body = axum::Json(json!({ "error": message, "error_code": code }));
|
let body = axum::Json(json!({ "error": message, "error_code": code }));
|
||||||
(status, body).into_response()
|
(status, body).into_response()
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -45,16 +45,19 @@ async fn shutdown_signal() {
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() -> anyhow::Result<()> {
|
async fn main() -> anyhow::Result<()> {
|
||||||
tracing_subscriber::fmt::init();
|
let config = Config::load()?;
|
||||||
|
fc_common::init_tracing(&config.tracing);
|
||||||
|
|
||||||
let cli = Cli::parse();
|
let cli = Cli::parse();
|
||||||
let config = Config::load()?;
|
|
||||||
|
|
||||||
let host = cli.host.unwrap_or(config.server.host.clone());
|
let host = cli.host.unwrap_or(config.server.host.clone());
|
||||||
let port = cli.port.unwrap_or(config.server.port);
|
let port = cli.port.unwrap_or(config.server.port);
|
||||||
|
|
||||||
let db = Database::new(config.database.clone()).await?;
|
let db = Database::new(config.database.clone()).await?;
|
||||||
|
|
||||||
|
// Bootstrap declarative projects, jobsets, and API keys from config
|
||||||
|
fc_common::bootstrap::run(db.pool(), &config.declarative).await?;
|
||||||
|
|
||||||
let state = AppState {
|
let state = AppState {
|
||||||
pool: db.pool().clone(),
|
pool: db.pool().clone(),
|
||||||
config: config.clone(),
|
config: config.clone(),
|
||||||
|
|
|
||||||
|
|
@ -6,9 +6,7 @@ use axum::{
|
||||||
response::{IntoResponse, Response},
|
response::{IntoResponse, Response},
|
||||||
routing::{get, post},
|
routing::{get, post},
|
||||||
};
|
};
|
||||||
use fc_common::{
|
use fc_common::{Build, BuildProduct, BuildStep, PaginatedResponse, PaginationParams};
|
||||||
Build, BuildProduct, BuildStatus, BuildStep, CreateBuild, PaginatedResponse, PaginationParams,
|
|
||||||
};
|
|
||||||
use serde::Deserialize;
|
use serde::Deserialize;
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
|
||||||
|
|
@ -155,44 +153,17 @@ async fn restart_build(
|
||||||
Path(id): Path<Uuid>,
|
Path(id): Path<Uuid>,
|
||||||
) -> Result<Json<Build>, ApiError> {
|
) -> Result<Json<Build>, ApiError> {
|
||||||
check_role(&extensions, &["restart-jobs"])?;
|
check_role(&extensions, &["restart-jobs"])?;
|
||||||
let original = fc_common::repo::builds::get(&state.pool, id)
|
let build = fc_common::repo::builds::restart(&state.pool, id)
|
||||||
.await
|
.await
|
||||||
.map_err(ApiError)?;
|
.map_err(ApiError)?;
|
||||||
|
|
||||||
// Can only restart completed or failed builds
|
|
||||||
if original.status != BuildStatus::Failed
|
|
||||||
&& original.status != BuildStatus::Completed
|
|
||||||
&& original.status != BuildStatus::Cancelled
|
|
||||||
{
|
|
||||||
return Err(ApiError(fc_common::CiError::Validation(
|
|
||||||
"Can only restart failed, completed, or cancelled builds".to_string(),
|
|
||||||
)));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create a new build with the same parameters
|
|
||||||
let new_build = fc_common::repo::builds::create(
|
|
||||||
&state.pool,
|
|
||||||
CreateBuild {
|
|
||||||
evaluation_id: original.evaluation_id,
|
|
||||||
job_name: original.job_name.clone(),
|
|
||||||
drv_path: original.drv_path.clone(),
|
|
||||||
system: original.system.clone(),
|
|
||||||
outputs: original.outputs.clone(),
|
|
||||||
is_aggregate: Some(original.is_aggregate),
|
|
||||||
constituents: original.constituents.clone(),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.map_err(ApiError)?;
|
|
||||||
|
|
||||||
tracing::info!(
|
tracing::info!(
|
||||||
original_id = %id,
|
build_id = %id,
|
||||||
new_id = %new_build.id,
|
job = %build.job_name,
|
||||||
job = %original.job_name,
|
|
||||||
"Build restarted"
|
"Build restarted"
|
||||||
);
|
);
|
||||||
|
|
||||||
Ok(Json(new_build))
|
Ok(Json(build))
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn bump_build(
|
async fn bump_build(
|
||||||
|
|
|
||||||
|
|
@ -170,9 +170,9 @@ async fn sign_narinfo(narinfo: &str, key_file: &std::path::Path) -> String {
|
||||||
.output()
|
.output()
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
if let Ok(o) = re_output {
|
if let Ok(o) = re_output
|
||||||
if let Ok(parsed) = serde_json::from_slice::<serde_json::Value>(&o.stdout) {
|
&& let Ok(parsed) = serde_json::from_slice::<serde_json::Value>(&o.stdout)
|
||||||
if let Some(sigs) = parsed
|
&& let Some(sigs) = parsed
|
||||||
.as_array()
|
.as_array()
|
||||||
.and_then(|a| a.first())
|
.and_then(|a| a.first())
|
||||||
.and_then(|e| e.get("signatures"))
|
.and_then(|e| e.get("signatures"))
|
||||||
|
|
@ -187,8 +187,6 @@ async fn sign_narinfo(narinfo: &str, key_file: &std::path::Path) -> String {
|
||||||
return format!("{narinfo}{}\n", sig_lines.join("\n"));
|
return format!("{narinfo}{}\n", sig_lines.join("\n"));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
|
||||||
narinfo.to_string()
|
narinfo.to_string()
|
||||||
}
|
}
|
||||||
_ => narinfo.to_string(),
|
_ => narinfo.to_string(),
|
||||||
|
|
|
||||||
|
|
@ -316,6 +316,14 @@ struct AdminTemplate {
|
||||||
auth_name: String,
|
auth_name: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Template)]
|
||||||
|
#[template(path = "project_setup.html")]
|
||||||
|
#[allow(dead_code)]
|
||||||
|
struct ProjectSetupTemplate {
|
||||||
|
is_admin: bool,
|
||||||
|
auth_name: String,
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Template)]
|
#[derive(Template)]
|
||||||
#[template(path = "login.html")]
|
#[template(path = "login.html")]
|
||||||
struct LoginTemplate {
|
struct LoginTemplate {
|
||||||
|
|
@ -353,14 +361,13 @@ async fn home(State(state): State<AppState>, extensions: Extensions) -> Html<Str
|
||||||
fc_common::repo::evaluations::list_filtered(&state.pool, Some(js.id), None, 1, 0)
|
fc_common::repo::evaluations::list_filtered(&state.pool, Some(js.id), None, 1, 0)
|
||||||
.await
|
.await
|
||||||
.unwrap_or_default();
|
.unwrap_or_default();
|
||||||
if let Some(e) = js_evals.into_iter().next() {
|
if let Some(e) = js_evals.into_iter().next()
|
||||||
if last_eval
|
&& last_eval
|
||||||
.as_ref()
|
.as_ref()
|
||||||
.map_or(true, |le| e.evaluation_time > le.evaluation_time)
|
.is_none_or(|le| e.evaluation_time > le.evaluation_time)
|
||||||
{
|
{
|
||||||
last_eval = Some(e);
|
last_eval = Some(e);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
}
|
||||||
let (status, class, time) = match &last_eval {
|
let (status, class, time) = match &last_eval {
|
||||||
Some(e) => {
|
Some(e) => {
|
||||||
|
|
@ -884,6 +891,19 @@ async fn admin_page(State(state): State<AppState>, extensions: Extensions) -> Ht
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// --- Setup Wizard ---
|
||||||
|
|
||||||
|
async fn project_setup_page(extensions: Extensions) -> Html<String> {
|
||||||
|
let tmpl = ProjectSetupTemplate {
|
||||||
|
is_admin: is_admin(&extensions),
|
||||||
|
auth_name: auth_name(&extensions),
|
||||||
|
};
|
||||||
|
Html(
|
||||||
|
tmpl.render()
|
||||||
|
.unwrap_or_else(|e| format!("Template error: {e}")),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
// --- Login / Logout ---
|
// --- Login / Logout ---
|
||||||
|
|
||||||
async fn login_page() -> Html<String> {
|
async fn login_page() -> Html<String> {
|
||||||
|
|
@ -950,7 +970,28 @@ async fn login_action(State(state): State<AppState>, Form(form): Form<LoginForm>
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn logout_action() -> Response {
|
async fn logout_action(State(state): State<AppState>, request: axum::extract::Request) -> Response {
|
||||||
|
// Remove server-side session
|
||||||
|
if let Some(cookie_header) = request
|
||||||
|
.headers()
|
||||||
|
.get("cookie")
|
||||||
|
.and_then(|v| v.to_str().ok())
|
||||||
|
&& let Some(session_id) = cookie_header
|
||||||
|
.split(';')
|
||||||
|
.filter_map(|pair| {
|
||||||
|
let pair = pair.trim();
|
||||||
|
let (k, v) = pair.split_once('=')?;
|
||||||
|
if k.trim() == "fc_session" {
|
||||||
|
Some(v.trim().to_string())
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.next()
|
||||||
|
{
|
||||||
|
state.sessions.remove(&session_id);
|
||||||
|
}
|
||||||
|
|
||||||
let cookie = "fc_session=; HttpOnly; SameSite=Strict; Path=/; Max-Age=0";
|
let cookie = "fc_session=; HttpOnly; SameSite=Strict; Path=/; Max-Age=0";
|
||||||
(
|
(
|
||||||
[(axum::http::header::SET_COOKIE, cookie.to_string())],
|
[(axum::http::header::SET_COOKIE, cookie.to_string())],
|
||||||
|
|
@ -966,6 +1007,7 @@ pub fn router(state: AppState) -> Router<AppState> {
|
||||||
.route("/logout", axum::routing::post(logout_action))
|
.route("/logout", axum::routing::post(logout_action))
|
||||||
.route("/", get(home))
|
.route("/", get(home))
|
||||||
.route("/projects", get(projects_page))
|
.route("/projects", get(projects_page))
|
||||||
|
.route("/projects/new", get(project_setup_page))
|
||||||
.route("/project/{id}", get(project_page))
|
.route("/project/{id}", get(project_page))
|
||||||
.route("/jobset/{id}", get(jobset_page))
|
.route("/jobset/{id}", get(jobset_page))
|
||||||
.route("/evaluations", get(evaluations_page))
|
.route("/evaluations", get(evaluations_page))
|
||||||
|
|
|
||||||
|
|
@ -5,7 +5,8 @@ use axum::{
|
||||||
routing::{get, post},
|
routing::{get, post},
|
||||||
};
|
};
|
||||||
use fc_common::{CreateEvaluation, Evaluation, PaginatedResponse, PaginationParams, Validate};
|
use fc_common::{CreateEvaluation, Evaluation, PaginatedResponse, PaginationParams, Validate};
|
||||||
use serde::Deserialize;
|
use serde::{Deserialize, Serialize};
|
||||||
|
use std::collections::HashMap;
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
|
||||||
use crate::auth_middleware::RequireRoles;
|
use crate::auth_middleware::RequireRoles;
|
||||||
|
|
@ -85,9 +86,127 @@ async fn trigger_evaluation(
|
||||||
Ok(Json(evaluation))
|
Ok(Json(evaluation))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Deserialize)]
|
||||||
|
struct CompareParams {
|
||||||
|
to: Uuid,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize)]
|
||||||
|
struct EvalComparison {
|
||||||
|
from_id: Uuid,
|
||||||
|
to_id: Uuid,
|
||||||
|
new_jobs: Vec<JobDiff>,
|
||||||
|
removed_jobs: Vec<JobDiff>,
|
||||||
|
changed_jobs: Vec<JobChange>,
|
||||||
|
unchanged_count: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize)]
|
||||||
|
struct JobDiff {
|
||||||
|
job_name: String,
|
||||||
|
system: Option<String>,
|
||||||
|
drv_path: String,
|
||||||
|
status: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize)]
|
||||||
|
struct JobChange {
|
||||||
|
job_name: String,
|
||||||
|
system: Option<String>,
|
||||||
|
old_drv: String,
|
||||||
|
new_drv: String,
|
||||||
|
old_status: String,
|
||||||
|
new_status: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn compare_evaluations(
|
||||||
|
State(state): State<AppState>,
|
||||||
|
Path(id): Path<Uuid>,
|
||||||
|
Query(params): Query<CompareParams>,
|
||||||
|
) -> Result<Json<EvalComparison>, ApiError> {
|
||||||
|
// Verify both evaluations exist
|
||||||
|
let _from_eval = fc_common::repo::evaluations::get(&state.pool, id)
|
||||||
|
.await
|
||||||
|
.map_err(ApiError)?;
|
||||||
|
let _to_eval = fc_common::repo::evaluations::get(&state.pool, params.to)
|
||||||
|
.await
|
||||||
|
.map_err(ApiError)?;
|
||||||
|
|
||||||
|
let from_builds = fc_common::repo::builds::list_for_evaluation(&state.pool, id)
|
||||||
|
.await
|
||||||
|
.map_err(ApiError)?;
|
||||||
|
let to_builds = fc_common::repo::builds::list_for_evaluation(&state.pool, params.to)
|
||||||
|
.await
|
||||||
|
.map_err(ApiError)?;
|
||||||
|
|
||||||
|
let from_map: HashMap<&str, &fc_common::Build> = from_builds
|
||||||
|
.iter()
|
||||||
|
.map(|b| (b.job_name.as_str(), b))
|
||||||
|
.collect();
|
||||||
|
let to_map: HashMap<&str, &fc_common::Build> =
|
||||||
|
to_builds.iter().map(|b| (b.job_name.as_str(), b)).collect();
|
||||||
|
|
||||||
|
let mut new_jobs = Vec::new();
|
||||||
|
let mut removed_jobs = Vec::new();
|
||||||
|
let mut changed_jobs = Vec::new();
|
||||||
|
let mut unchanged_count = 0;
|
||||||
|
|
||||||
|
// Jobs in `to` but not in `from` are new
|
||||||
|
for (name, build) in &to_map {
|
||||||
|
if !from_map.contains_key(name) {
|
||||||
|
new_jobs.push(JobDiff {
|
||||||
|
job_name: name.to_string(),
|
||||||
|
system: build.system.clone(),
|
||||||
|
drv_path: build.drv_path.clone(),
|
||||||
|
status: format!("{:?}", build.status),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Jobs in `from` but not in `to` are removed
|
||||||
|
for (name, build) in &from_map {
|
||||||
|
if !to_map.contains_key(name) {
|
||||||
|
removed_jobs.push(JobDiff {
|
||||||
|
job_name: name.to_string(),
|
||||||
|
system: build.system.clone(),
|
||||||
|
drv_path: build.drv_path.clone(),
|
||||||
|
status: format!("{:?}", build.status),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Jobs in both: compare derivation paths
|
||||||
|
for (name, from_build) in &from_map {
|
||||||
|
if let Some(to_build) = to_map.get(name) {
|
||||||
|
if from_build.drv_path != to_build.drv_path {
|
||||||
|
changed_jobs.push(JobChange {
|
||||||
|
job_name: name.to_string(),
|
||||||
|
system: to_build.system.clone(),
|
||||||
|
old_drv: from_build.drv_path.clone(),
|
||||||
|
new_drv: to_build.drv_path.clone(),
|
||||||
|
old_status: format!("{:?}", from_build.status),
|
||||||
|
new_status: format!("{:?}", to_build.status),
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
unchanged_count += 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(Json(EvalComparison {
|
||||||
|
from_id: id,
|
||||||
|
to_id: params.to,
|
||||||
|
new_jobs,
|
||||||
|
removed_jobs,
|
||||||
|
changed_jobs,
|
||||||
|
unchanged_count,
|
||||||
|
}))
|
||||||
|
}
|
||||||
|
|
||||||
pub fn router() -> Router<AppState> {
|
pub fn router() -> Router<AppState> {
|
||||||
Router::new()
|
Router::new()
|
||||||
.route("/evaluations", get(list_evaluations))
|
.route("/evaluations", get(list_evaluations))
|
||||||
.route("/evaluations/{id}", get(get_evaluation))
|
.route("/evaluations/{id}", get(get_evaluation))
|
||||||
|
.route("/evaluations/{id}/compare", get(compare_evaluations))
|
||||||
.route("/evaluations/trigger", post(trigger_evaluation))
|
.route("/evaluations/trigger", post(trigger_evaluation))
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -96,13 +96,12 @@ async fn stream_build_log(
|
||||||
consecutive_empty += 1;
|
consecutive_empty += 1;
|
||||||
if consecutive_empty > 5 {
|
if consecutive_empty > 5 {
|
||||||
// Check build status
|
// Check build status
|
||||||
if let Ok(b) = fc_common::repo::builds::get(&pool, build_id).await {
|
if let Ok(b) = fc_common::repo::builds::get(&pool, build_id).await
|
||||||
if b.status != fc_common::models::BuildStatus::Running
|
&& b.status != fc_common::models::BuildStatus::Running
|
||||||
&& b.status != fc_common::models::BuildStatus::Pending {
|
&& b.status != fc_common::models::BuildStatus::Pending {
|
||||||
yield Ok(Event::default().event("done").data("Build completed"));
|
yield Ok(Event::default().event("done").data("Build completed"));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
consecutive_empty = 0;
|
consecutive_empty = 0;
|
||||||
}
|
}
|
||||||
tokio::time::sleep(std::time::Duration::from_millis(500)).await;
|
tokio::time::sleep(std::time::Duration::from_millis(500)).await;
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue