feat: add websocket support for real-time scrape job monitoring

This commit is contained in:
2026-01-29 19:31:04 -06:00
parent d861888e5e
commit cfe098d193
26 changed files with 1118 additions and 173 deletions
Generated
+34 -2
View File
@@ -182,6 +182,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8b52af3cb4058c895d37317bb27508dccc8e5f2d39454016b297bf4a400597b8" checksum = "8b52af3cb4058c895d37317bb27508dccc8e5f2d39454016b297bf4a400597b8"
dependencies = [ dependencies = [
"axum-core", "axum-core",
"base64 0.22.1",
"bytes", "bytes",
"form_urlencoded", "form_urlencoded",
"futures-util", "futures-util",
@@ -200,8 +201,10 @@ dependencies = [
"serde_json", "serde_json",
"serde_path_to_error", "serde_path_to_error",
"serde_urlencoded", "serde_urlencoded",
"sha1",
"sync_wrapper 1.0.2", "sync_wrapper 1.0.2",
"tokio", "tokio",
"tokio-tungstenite 0.28.0",
"tower", "tower",
"tower-layer", "tower-layer",
"tower-service", "tower-service",
@@ -2916,7 +2919,7 @@ dependencies = [
"static_assertions", "static_assertions",
"time", "time",
"tokio", "tokio",
"tokio-tungstenite", "tokio-tungstenite 0.21.0",
"tracing", "tracing",
"typemap_rev", "typemap_rev",
"typesize", "typesize",
@@ -3604,10 +3607,22 @@ dependencies = [
"rustls-pki-types", "rustls-pki-types",
"tokio", "tokio",
"tokio-rustls 0.25.0", "tokio-rustls 0.25.0",
"tungstenite", "tungstenite 0.21.0",
"webpki-roots 0.26.11", "webpki-roots 0.26.11",
] ]
[[package]]
name = "tokio-tungstenite"
version = "0.28.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d25a406cddcc431a75d3d9afc6a7c0f7428d4891dd973e4d54c56b46127bf857"
dependencies = [
"futures-util",
"log",
"tokio",
"tungstenite 0.28.0",
]
[[package]] [[package]]
name = "tokio-util" name = "tokio-util"
version = "0.7.16" version = "0.7.16"
@@ -3845,6 +3860,23 @@ dependencies = [
"utf-8", "utf-8",
] ]
[[package]]
name = "tungstenite"
version = "0.28.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8628dcc84e5a09eb3d8423d6cb682965dea9133204e8fb3efee74c2a0c259442"
dependencies = [
"bytes",
"data-encoding",
"http 1.3.1",
"httparse",
"log",
"rand 0.9.2",
"sha1",
"thiserror 2.0.16",
"utf-8",
]
[[package]] [[package]]
name = "typemap_rev" name = "typemap_rev"
version = "0.3.0" version = "0.3.0"
+1 -1
View File
@@ -11,7 +11,7 @@ embed-assets = ["dep:rust-embed", "dep:mime_guess"]
[dependencies] [dependencies]
anyhow = "1.0.99" anyhow = "1.0.99"
async-trait = "0.1" async-trait = "0.1"
axum = "0.8.4" axum = { version = "0.8.4", features = ["ws"] }
bitflags = { version = "2.9.4", features = ["serde"] } bitflags = { version = "2.9.4", features = ["serde"] }
chrono = { version = "0.4.42", features = ["serde"] } chrono = { version = "0.4.42", features = ["serde"] }
compile-time = "0.2.0" compile-time = "0.2.0"
@@ -0,0 +1,7 @@
-- Add queued_at column to track when a job last entered the "ready to pick up" state.
-- For fresh jobs this equals execute_at; for retried jobs it is updated to NOW().
ALTER TABLE scrape_jobs
ADD COLUMN queued_at TIMESTAMPTZ NOT NULL DEFAULT NOW();
-- Backfill existing rows: set queued_at = execute_at (best approximation)
UPDATE scrape_jobs SET queued_at = execute_at;
+1
View File
@@ -126,6 +126,7 @@ impl App {
self.banner_api.clone(), self.banner_api.clone(),
self.app_state.reference_cache.clone(), self.app_state.reference_cache.clone(),
self.app_state.service_statuses.clone(), self.app_state.service_statuses.clone(),
self.app_state.scrape_job_tx.clone(),
)); ));
self.service_manager self.service_manager
.register_service(ServiceName::Scraper.as_str(), scraper_service); .register_service(ServiceName::Scraper.as_str(), scraper_service);
+35
View File
@@ -176,6 +176,20 @@ pub enum TargetType {
SingleCrn, SingleCrn,
} }
/// Computed status for a scrape job, derived from existing fields.
#[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize)]
#[serde(rename_all = "camelCase")]
pub enum ScrapeJobStatus {
Processing,
StaleLock,
Exhausted,
Scheduled,
Pending,
}
/// How long a lock can be held before it is considered stale (mirrors `scrape_jobs::LOCK_EXPIRY`).
const LOCK_EXPIRY_SECS: i64 = 10 * 60;
/// Represents a queryable job from the database. /// Represents a queryable job from the database.
#[allow(dead_code)] #[allow(dead_code)]
#[derive(sqlx::FromRow, Debug, Clone)] #[derive(sqlx::FromRow, Debug, Clone)]
@@ -191,6 +205,27 @@ pub struct ScrapeJob {
pub retry_count: i32, pub retry_count: i32,
/// Maximum number of retry attempts allowed (non-negative, enforced by CHECK constraint) /// Maximum number of retry attempts allowed (non-negative, enforced by CHECK constraint)
pub max_retries: i32, pub max_retries: i32,
/// When the job last entered the "ready to pick up" state.
/// Set to NOW() on creation; updated to NOW() on retry.
pub queued_at: DateTime<Utc>,
}
impl ScrapeJob {
/// Compute the current status of this job from its fields.
pub fn status(&self) -> ScrapeJobStatus {
let now = Utc::now();
match self.locked_at {
Some(locked) if (now - locked).num_seconds() < LOCK_EXPIRY_SECS => {
ScrapeJobStatus::Processing
}
Some(_) => ScrapeJobStatus::StaleLock,
None if self.retry_count >= self.max_retries && self.max_retries > 0 => {
ScrapeJobStatus::Exhausted
}
None if self.execute_at > now => ScrapeJobStatus::Scheduled,
None => ScrapeJobStatus::Pending,
}
}
} }
/// A user authenticated via Discord OAuth. /// A user authenticated via Discord OAuth.
+22 -18
View File
@@ -13,7 +13,9 @@ use std::collections::HashSet;
/// # Returns /// # Returns
/// The number of jobs that were unlocked. /// The number of jobs that were unlocked.
pub async fn force_unlock_all(db_pool: &PgPool) -> Result<u64> { pub async fn force_unlock_all(db_pool: &PgPool) -> Result<u64> {
let result = sqlx::query("UPDATE scrape_jobs SET locked_at = NULL WHERE locked_at IS NOT NULL") let result = sqlx::query(
"UPDATE scrape_jobs SET locked_at = NULL, queued_at = NOW() WHERE locked_at IS NOT NULL",
)
.execute(db_pool) .execute(db_pool)
.await?; .await?;
Ok(result.rows_affected()) Ok(result.rows_affected())
@@ -97,10 +99,11 @@ pub async fn unlock_job(job_id: i32, db_pool: &PgPool) -> Result<()> {
Ok(()) Ok(())
} }
/// Atomically unlock a job and increment its retry count. /// Atomically unlock a job, increment its retry count, and reset `queued_at`.
/// ///
/// Returns whether the job still has retries remaining. This is determined /// Returns the new `queued_at` timestamp if retries remain, or `None` if
/// atomically in the database to avoid race conditions between workers. /// the job has exhausted its retries. This is determined atomically in the
/// database to avoid race conditions between workers.
/// ///
/// # Arguments /// # Arguments
/// * `job_id` - The database ID of the job /// * `job_id` - The database ID of the job
@@ -108,25 +111,25 @@ pub async fn unlock_job(job_id: i32, db_pool: &PgPool) -> Result<()> {
/// * `db_pool` - PostgreSQL connection pool /// * `db_pool` - PostgreSQL connection pool
/// ///
/// # Returns /// # Returns
/// * `Ok(true)` if the job was unlocked and retries remain /// * `Ok(Some(queued_at))` if the job was unlocked and retries remain
/// * `Ok(false)` if the job has exhausted its retries /// * `Ok(None)` if the job has exhausted its retries
pub async fn unlock_and_increment_retry( pub async fn unlock_and_increment_retry(
job_id: i32, job_id: i32,
max_retries: i32, max_retries: i32,
db_pool: &PgPool, db_pool: &PgPool,
) -> Result<bool> { ) -> Result<Option<chrono::DateTime<chrono::Utc>>> {
let result = sqlx::query_scalar::<_, Option<i32>>( let result = sqlx::query_scalar::<_, Option<chrono::DateTime<chrono::Utc>>>(
"UPDATE scrape_jobs "UPDATE scrape_jobs
SET locked_at = NULL, retry_count = retry_count + 1 SET locked_at = NULL, retry_count = retry_count + 1, queued_at = NOW()
WHERE id = $1 WHERE id = $1
RETURNING CASE WHEN retry_count <= $2 THEN retry_count ELSE NULL END", RETURNING CASE WHEN retry_count <= $2 THEN queued_at ELSE NULL END",
) )
.bind(job_id) .bind(job_id)
.bind(max_retries) .bind(max_retries)
.fetch_one(db_pool) .fetch_one(db_pool)
.await?; .await?;
Ok(result.is_some()) Ok(result)
} }
/// Find existing job payloads matching the given target type and candidates. /// Find existing job payloads matching the given target type and candidates.
@@ -173,9 +176,9 @@ pub async fn find_existing_job_payloads(
pub async fn batch_insert_jobs( pub async fn batch_insert_jobs(
jobs: &[(serde_json::Value, TargetType, ScrapePriority)], jobs: &[(serde_json::Value, TargetType, ScrapePriority)],
db_pool: &PgPool, db_pool: &PgPool,
) -> Result<()> { ) -> Result<Vec<ScrapeJob>> {
if jobs.is_empty() { if jobs.is_empty() {
return Ok(()); return Ok(Vec::new());
} }
let mut target_types: Vec<String> = Vec::with_capacity(jobs.len()); let mut target_types: Vec<String> = Vec::with_capacity(jobs.len());
@@ -188,19 +191,20 @@ pub async fn batch_insert_jobs(
priorities.push(format!("{priority:?}")); priorities.push(format!("{priority:?}"));
} }
sqlx::query( let inserted = sqlx::query_as::<_, ScrapeJob>(
r#" r#"
INSERT INTO scrape_jobs (target_type, target_payload, priority, execute_at) INSERT INTO scrape_jobs (target_type, target_payload, priority, execute_at, queued_at)
SELECT v.target_type::target_type, v.payload, v.priority::scrape_priority, NOW() SELECT v.target_type::target_type, v.payload, v.priority::scrape_priority, NOW(), NOW()
FROM UNNEST($1::text[], $2::jsonb[], $3::text[]) FROM UNNEST($1::text[], $2::jsonb[], $3::text[])
AS v(target_type, payload, priority) AS v(target_type, payload, priority)
RETURNING *
"#, "#,
) )
.bind(&target_types) .bind(&target_types)
.bind(&payloads) .bind(&payloads)
.bind(&priorities) .bind(&priorities)
.execute(db_pool) .fetch_all(db_pool)
.await?; .await?;
Ok(()) Ok(inserted)
} }
+11 -1
View File
@@ -7,6 +7,7 @@ use crate::data::scrape_jobs;
use crate::services::Service; use crate::services::Service;
use crate::state::ReferenceCache; use crate::state::ReferenceCache;
use crate::status::{ServiceStatus, ServiceStatusRegistry}; use crate::status::{ServiceStatus, ServiceStatusRegistry};
use crate::web::ws::ScrapeJobEvent;
use sqlx::PgPool; use sqlx::PgPool;
use std::sync::Arc; use std::sync::Arc;
use tokio::sync::{RwLock, broadcast}; use tokio::sync::{RwLock, broadcast};
@@ -25,6 +26,7 @@ pub struct ScraperService {
banner_api: Arc<BannerApi>, banner_api: Arc<BannerApi>,
reference_cache: Arc<RwLock<ReferenceCache>>, reference_cache: Arc<RwLock<ReferenceCache>>,
service_statuses: ServiceStatusRegistry, service_statuses: ServiceStatusRegistry,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
scheduler_handle: Option<JoinHandle<()>>, scheduler_handle: Option<JoinHandle<()>>,
worker_handles: Vec<JoinHandle<()>>, worker_handles: Vec<JoinHandle<()>>,
shutdown_tx: Option<broadcast::Sender<()>>, shutdown_tx: Option<broadcast::Sender<()>>,
@@ -37,12 +39,14 @@ impl ScraperService {
banner_api: Arc<BannerApi>, banner_api: Arc<BannerApi>,
reference_cache: Arc<RwLock<ReferenceCache>>, reference_cache: Arc<RwLock<ReferenceCache>>,
service_statuses: ServiceStatusRegistry, service_statuses: ServiceStatusRegistry,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
) -> Self { ) -> Self {
Self { Self {
db_pool, db_pool,
banner_api, banner_api,
reference_cache, reference_cache,
service_statuses, service_statuses,
job_events_tx,
scheduler_handle: None, scheduler_handle: None,
worker_handles: Vec::new(), worker_handles: Vec::new(),
shutdown_tx: None, shutdown_tx: None,
@@ -71,6 +75,7 @@ impl ScraperService {
self.db_pool.clone(), self.db_pool.clone(),
self.banner_api.clone(), self.banner_api.clone(),
self.reference_cache.clone(), self.reference_cache.clone(),
self.job_events_tx.clone(),
); );
let shutdown_rx = shutdown_tx.subscribe(); let shutdown_rx = shutdown_tx.subscribe();
let scheduler_handle = tokio::spawn(async move { let scheduler_handle = tokio::spawn(async move {
@@ -81,7 +86,12 @@ impl ScraperService {
let worker_count = 4; // This could be configurable let worker_count = 4; // This could be configurable
for i in 0..worker_count { for i in 0..worker_count {
let worker = Worker::new(i, self.db_pool.clone(), self.banner_api.clone()); let worker = Worker::new(
i,
self.db_pool.clone(),
self.banner_api.clone(),
self.job_events_tx.clone(),
);
let shutdown_rx = shutdown_tx.subscribe(); let shutdown_rx = shutdown_tx.subscribe();
let worker_handle = tokio::spawn(async move { let worker_handle = tokio::spawn(async move {
worker.run(shutdown_rx).await; worker.run(shutdown_rx).await;
+21 -3
View File
@@ -5,6 +5,7 @@ use crate::error::Result;
use crate::rmp::RmpClient; use crate::rmp::RmpClient;
use crate::scraper::jobs::subject::SubjectJob; use crate::scraper::jobs::subject::SubjectJob;
use crate::state::ReferenceCache; use crate::state::ReferenceCache;
use crate::web::ws::{ScrapeJobDto, ScrapeJobEvent};
use serde_json::json; use serde_json::json;
use sqlx::PgPool; use sqlx::PgPool;
use std::sync::Arc; use std::sync::Arc;
@@ -25,6 +26,7 @@ pub struct Scheduler {
db_pool: PgPool, db_pool: PgPool,
banner_api: Arc<BannerApi>, banner_api: Arc<BannerApi>,
reference_cache: Arc<RwLock<ReferenceCache>>, reference_cache: Arc<RwLock<ReferenceCache>>,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
} }
impl Scheduler { impl Scheduler {
@@ -32,11 +34,13 @@ impl Scheduler {
db_pool: PgPool, db_pool: PgPool,
banner_api: Arc<BannerApi>, banner_api: Arc<BannerApi>,
reference_cache: Arc<RwLock<ReferenceCache>>, reference_cache: Arc<RwLock<ReferenceCache>>,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
) -> Self { ) -> Self {
Self { Self {
db_pool, db_pool,
banner_api, banner_api,
reference_cache, reference_cache,
job_events_tx,
} }
} }
@@ -74,6 +78,7 @@ impl Scheduler {
let banner_api = self.banner_api.clone(); let banner_api = self.banner_api.clone();
let cancel_token = cancel_token.clone(); let cancel_token = cancel_token.clone();
let reference_cache = self.reference_cache.clone(); let reference_cache = self.reference_cache.clone();
let job_events_tx = self.job_events_tx.clone();
async move { async move {
tokio::select! { tokio::select! {
@@ -99,7 +104,7 @@ impl Scheduler {
tokio::join!(rmp_fut, ref_fut); tokio::join!(rmp_fut, ref_fut);
if let Err(e) = Self::schedule_jobs_impl(&db_pool, &banner_api).await { if let Err(e) = Self::schedule_jobs_impl(&db_pool, &banner_api, Some(&job_events_tx)).await {
error!(error = ?e, "Failed to schedule jobs"); error!(error = ?e, "Failed to schedule jobs");
} }
} => {} } => {}
@@ -150,7 +155,11 @@ impl Scheduler {
/// ///
/// This is a static method (not &self) to allow it to be called from spawned tasks. /// This is a static method (not &self) to allow it to be called from spawned tasks.
#[tracing::instrument(skip_all, fields(term))] #[tracing::instrument(skip_all, fields(term))]
async fn schedule_jobs_impl(db_pool: &PgPool, banner_api: &BannerApi) -> Result<()> { async fn schedule_jobs_impl(
db_pool: &PgPool,
banner_api: &BannerApi,
job_events_tx: Option<&broadcast::Sender<ScrapeJobEvent>>,
) -> Result<()> {
// For now, we will implement a simple baseline scheduling strategy: // For now, we will implement a simple baseline scheduling strategy:
// 1. Get a list of all subjects from the Banner API. // 1. Get a list of all subjects from the Banner API.
// 2. Query existing jobs for all subjects in a single query. // 2. Query existing jobs for all subjects in a single query.
@@ -213,7 +222,16 @@ impl Scheduler {
.map(|(payload, _)| (payload, TargetType::Subject, ScrapePriority::Low)) .map(|(payload, _)| (payload, TargetType::Subject, ScrapePriority::Low))
.collect(); .collect();
scrape_jobs::batch_insert_jobs(&jobs, db_pool).await?; let inserted = scrape_jobs::batch_insert_jobs(&jobs, db_pool).await?;
if let Some(tx) = job_events_tx {
inserted.iter().for_each(|job| {
debug!(job_id = job.id, "Emitting JobCreated event");
let _ = tx.send(ScrapeJobEvent::JobCreated {
job: ScrapeJobDto::from(job),
});
});
}
} }
debug!("Job scheduling complete"); debug!("Job scheduling complete");
+49 -5
View File
@@ -1,8 +1,10 @@
use crate::banner::{BannerApi, BannerApiError}; use crate::banner::{BannerApi, BannerApiError};
use crate::data::models::ScrapeJob; use crate::data::models::{ScrapeJob, ScrapeJobStatus};
use crate::data::scrape_jobs; use crate::data::scrape_jobs;
use crate::error::Result; use crate::error::Result;
use crate::scraper::jobs::{JobError, JobType}; use crate::scraper::jobs::{JobError, JobType};
use crate::web::ws::ScrapeJobEvent;
use chrono::Utc;
use sqlx::PgPool; use sqlx::PgPool;
use std::sync::Arc; use std::sync::Arc;
use std::time::Duration; use std::time::Duration;
@@ -21,14 +23,21 @@ pub struct Worker {
id: usize, // For logging purposes id: usize, // For logging purposes
db_pool: PgPool, db_pool: PgPool,
banner_api: Arc<BannerApi>, banner_api: Arc<BannerApi>,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
} }
impl Worker { impl Worker {
pub fn new(id: usize, db_pool: PgPool, banner_api: Arc<BannerApi>) -> Self { pub fn new(
id: usize,
db_pool: PgPool,
banner_api: Arc<BannerApi>,
job_events_tx: broadcast::Sender<ScrapeJobEvent>,
) -> Self {
Self { Self {
id, id,
db_pool, db_pool,
banner_api, banner_api,
job_events_tx,
} }
} }
@@ -65,6 +74,15 @@ impl Worker {
let max_retries = job.max_retries; let max_retries = job.max_retries;
let start = std::time::Instant::now(); let start = std::time::Instant::now();
// Emit JobLocked event
let locked_at = Utc::now().to_rfc3339();
debug!(job_id, "Emitting JobLocked event");
let _ = self.job_events_tx.send(ScrapeJobEvent::JobLocked {
id: job_id,
locked_at,
status: ScrapeJobStatus::Processing,
});
// Process the job, racing against shutdown signal and timeout // Process the job, racing against shutdown signal and timeout
let process_result = tokio::select! { let process_result = tokio::select! {
_ = shutdown_rx.recv() => { _ = shutdown_rx.recv() => {
@@ -143,7 +161,11 @@ impl Worker {
scrape_jobs::unlock_job(job_id, &self.db_pool).await scrape_jobs::unlock_job(job_id, &self.db_pool).await
} }
async fn unlock_and_increment_retry(&self, job_id: i32, max_retries: i32) -> Result<bool> { async fn unlock_and_increment_retry(
&self,
job_id: i32,
max_retries: i32,
) -> Result<Option<chrono::DateTime<chrono::Utc>>> {
scrape_jobs::unlock_and_increment_retry(job_id, max_retries, &self.db_pool).await scrape_jobs::unlock_and_increment_retry(job_id, max_retries, &self.db_pool).await
} }
@@ -188,6 +210,10 @@ impl Worker {
if let Err(e) = self.delete_job(job_id).await { if let Err(e) = self.delete_job(job_id).await {
error!(worker_id = self.id, job_id, error = ?e, "Failed to delete completed job"); error!(worker_id = self.id, job_id, error = ?e, "Failed to delete completed job");
} }
debug!(job_id, "Emitting JobCompleted event");
let _ = self
.job_events_tx
.send(ScrapeJobEvent::JobCompleted { id: job_id });
} }
Err(JobError::Recoverable(e)) => { Err(JobError::Recoverable(e)) => {
self.handle_recoverable_error(job_id, retry_count, max_retries, e, duration) self.handle_recoverable_error(job_id, retry_count, max_retries, e, duration)
@@ -204,6 +230,10 @@ impl Worker {
if let Err(e) = self.delete_job(job_id).await { if let Err(e) = self.delete_job(job_id).await {
error!(worker_id = self.id, job_id, error = ?e, "Failed to delete corrupted job"); error!(worker_id = self.id, job_id, error = ?e, "Failed to delete corrupted job");
} }
debug!(job_id, "Emitting JobDeleted event");
let _ = self
.job_events_tx
.send(ScrapeJobEvent::JobDeleted { id: job_id });
} }
} }
} }
@@ -246,7 +276,7 @@ impl Worker {
// Atomically unlock and increment retry count, checking if retry is allowed // Atomically unlock and increment retry count, checking if retry is allowed
match self.unlock_and_increment_retry(job_id, max_retries).await { match self.unlock_and_increment_retry(job_id, max_retries).await {
Ok(can_retry) if can_retry => { Ok(Some(queued_at)) => {
debug!( debug!(
worker_id = self.id, worker_id = self.id,
job_id, job_id,
@@ -254,8 +284,15 @@ impl Worker {
remaining_retries = remaining_retries, remaining_retries = remaining_retries,
"Job unlocked for retry" "Job unlocked for retry"
); );
debug!(job_id, "Emitting JobRetried event");
let _ = self.job_events_tx.send(ScrapeJobEvent::JobRetried {
id: job_id,
retry_count: next_attempt,
queued_at: queued_at.to_rfc3339(),
status: ScrapeJobStatus::Pending,
});
} }
Ok(_) => { Ok(None) => {
// Max retries exceeded (detected atomically) // Max retries exceeded (detected atomically)
error!( error!(
worker_id = self.id, worker_id = self.id,
@@ -269,6 +306,13 @@ impl Worker {
if let Err(e) = self.delete_job(job_id).await { if let Err(e) = self.delete_job(job_id).await {
error!(worker_id = self.id, job_id, error = ?e, "Failed to delete failed job"); error!(worker_id = self.id, job_id, error = ?e, "Failed to delete failed job");
} }
debug!(job_id, "Emitting JobExhausted and JobDeleted events");
let _ = self
.job_events_tx
.send(ScrapeJobEvent::JobExhausted { id: job_id });
let _ = self
.job_events_tx
.send(ScrapeJobEvent::JobDeleted { id: job_id });
} }
Err(e) => { Err(e) => {
error!(worker_id = self.id, job_id, error = ?e, "Failed to unlock and increment retry count"); error!(worker_id = self.id, job_id, error = ?e, "Failed to unlock and increment retry count");
+10 -1
View File
@@ -5,11 +5,12 @@ use crate::banner::Course;
use crate::data::models::ReferenceData; use crate::data::models::ReferenceData;
use crate::status::ServiceStatusRegistry; use crate::status::ServiceStatusRegistry;
use crate::web::session_cache::{OAuthStateStore, SessionCache}; use crate::web::session_cache::{OAuthStateStore, SessionCache};
use crate::web::ws::ScrapeJobEvent;
use anyhow::Result; use anyhow::Result;
use sqlx::PgPool; use sqlx::PgPool;
use std::collections::HashMap; use std::collections::HashMap;
use std::sync::Arc; use std::sync::Arc;
use tokio::sync::RwLock; use tokio::sync::{RwLock, broadcast};
/// In-memory cache for reference data (code→description lookups). /// In-memory cache for reference data (code→description lookups).
/// ///
@@ -75,10 +76,12 @@ pub struct AppState {
pub reference_cache: Arc<RwLock<ReferenceCache>>, pub reference_cache: Arc<RwLock<ReferenceCache>>,
pub session_cache: SessionCache, pub session_cache: SessionCache,
pub oauth_state_store: OAuthStateStore, pub oauth_state_store: OAuthStateStore,
pub scrape_job_tx: broadcast::Sender<ScrapeJobEvent>,
} }
impl AppState { impl AppState {
pub fn new(banner_api: Arc<BannerApi>, db_pool: PgPool) -> Self { pub fn new(banner_api: Arc<BannerApi>, db_pool: PgPool) -> Self {
let (scrape_job_tx, _) = broadcast::channel(64);
Self { Self {
session_cache: SessionCache::new(db_pool.clone()), session_cache: SessionCache::new(db_pool.clone()),
oauth_state_store: OAuthStateStore::new(), oauth_state_store: OAuthStateStore::new(),
@@ -86,9 +89,15 @@ impl AppState {
db_pool, db_pool,
service_statuses: ServiceStatusRegistry::new(), service_statuses: ServiceStatusRegistry::new(),
reference_cache: Arc::new(RwLock::new(ReferenceCache::new())), reference_cache: Arc::new(RwLock::new(ReferenceCache::new())),
scrape_job_tx,
} }
} }
/// Subscribe to scrape job lifecycle events.
pub fn scrape_job_events(&self) -> broadcast::Receiver<ScrapeJobEvent> {
self.scrape_job_tx.subscribe()
}
/// Initialize the reference cache from the database. /// Initialize the reference cache from the database.
pub async fn load_reference_cache(&self) -> Result<()> { pub async fn load_reference_cache(&self) -> Result<()> {
let entries = crate::data::reference::get_all(&self.db_pool).await?; let entries = crate::data::reference::get_all(&self.db_pool).await?;
+2
View File
@@ -163,6 +163,8 @@ pub async fn list_scrape_jobs(
"lockedAt": j.locked_at.map(|t| t.to_rfc3339()), "lockedAt": j.locked_at.map(|t| t.to_rfc3339()),
"retryCount": j.retry_count, "retryCount": j.retry_count,
"maxRetries": j.max_retries, "maxRetries": j.max_retries,
"queuedAt": j.queued_at.to_rfc3339(),
"status": j.status(),
}) })
}) })
.collect(); .collect();
+1
View File
@@ -9,5 +9,6 @@ pub mod encoding;
pub mod extractors; pub mod extractors;
pub mod routes; pub mod routes;
pub mod session_cache; pub mod session_cache;
pub mod ws;
pub use routes::*; pub use routes::*;
+2
View File
@@ -11,6 +11,7 @@ use axum::{
use crate::web::admin; use crate::web::admin;
use crate::web::auth::{self, AuthConfig}; use crate::web::auth::{self, AuthConfig};
use crate::web::ws;
#[cfg(feature = "embed-assets")] #[cfg(feature = "embed-assets")]
use axum::{ use axum::{
http::{HeaderMap, StatusCode, Uri}, http::{HeaderMap, StatusCode, Uri},
@@ -63,6 +64,7 @@ pub fn create_router(app_state: AppState, auth_config: AuthConfig) -> Router {
put(admin::set_user_admin), put(admin::set_user_admin),
) )
.route("/admin/scrape-jobs", get(admin::list_scrape_jobs)) .route("/admin/scrape-jobs", get(admin::list_scrape_jobs))
.route("/admin/scrape-jobs/ws", get(ws::scrape_jobs_ws))
.route("/admin/audit-log", get(admin::list_audit_log)) .route("/admin/audit-log", get(admin::list_audit_log))
.with_state(app_state); .with_state(app_state);
+205
View File
@@ -0,0 +1,205 @@
//! WebSocket event types and handler for real-time scrape job updates.
use axum::{
extract::{
State,
ws::{Message, WebSocket, WebSocketUpgrade},
},
response::IntoResponse,
};
use futures::{SinkExt, StreamExt};
use serde::Serialize;
use sqlx::PgPool;
use tokio::sync::broadcast;
use tracing::debug;
use crate::data::models::{ScrapeJob, ScrapeJobStatus};
use crate::state::AppState;
use crate::web::extractors::AdminUser;
/// A serializable DTO for `ScrapeJob` with computed `status`.
#[derive(Debug, Clone, Serialize)]
#[serde(rename_all = "camelCase")]
pub struct ScrapeJobDto {
pub id: i32,
pub target_type: String,
pub target_payload: serde_json::Value,
pub priority: String,
pub execute_at: String,
pub created_at: String,
pub locked_at: Option<String>,
pub retry_count: i32,
pub max_retries: i32,
pub queued_at: String,
pub status: ScrapeJobStatus,
}
impl From<&ScrapeJob> for ScrapeJobDto {
fn from(job: &ScrapeJob) -> Self {
Self {
id: job.id,
target_type: format!("{:?}", job.target_type),
target_payload: job.target_payload.clone(),
priority: format!("{:?}", job.priority),
execute_at: job.execute_at.to_rfc3339(),
created_at: job.created_at.to_rfc3339(),
locked_at: job.locked_at.map(|t| t.to_rfc3339()),
retry_count: job.retry_count,
max_retries: job.max_retries,
queued_at: job.queued_at.to_rfc3339(),
status: job.status(),
}
}
}
/// Events broadcast when scrape job state changes.
#[derive(Debug, Clone, Serialize)]
#[serde(tag = "type", rename_all = "camelCase")]
pub enum ScrapeJobEvent {
Init {
jobs: Vec<ScrapeJobDto>,
},
JobCreated {
job: ScrapeJobDto,
},
JobLocked {
id: i32,
locked_at: String,
status: ScrapeJobStatus,
},
JobCompleted {
id: i32,
},
JobRetried {
id: i32,
retry_count: i32,
queued_at: String,
status: ScrapeJobStatus,
},
JobExhausted {
id: i32,
},
JobDeleted {
id: i32,
},
}
/// Fetch current scrape jobs from the DB and build an `Init` event.
async fn build_init_event(db_pool: &PgPool) -> Result<ScrapeJobEvent, sqlx::Error> {
let rows = sqlx::query_as::<_, ScrapeJob>(
"SELECT * FROM scrape_jobs ORDER BY priority DESC, execute_at ASC LIMIT 100",
)
.fetch_all(db_pool)
.await?;
let jobs = rows.iter().map(ScrapeJobDto::from).collect();
Ok(ScrapeJobEvent::Init { jobs })
}
/// WebSocket endpoint for real-time scrape job updates.
///
/// Auth is checked via `AdminUser` before the upgrade occurs — if rejected,
/// a 401/403 is returned and the upgrade never happens.
pub async fn scrape_jobs_ws(
ws: WebSocketUpgrade,
AdminUser(_user): AdminUser,
State(state): State<AppState>,
) -> impl IntoResponse {
ws.on_upgrade(|socket| handle_scrape_jobs_ws(socket, state))
}
/// Serialize an event and send it over the WebSocket sink.
/// Returns `true` if the message was sent, `false` if the client disconnected.
async fn send_event(
sink: &mut futures::stream::SplitSink<WebSocket, Message>,
event: &ScrapeJobEvent,
) -> bool {
let Ok(json) = serde_json::to_string(event) else {
return true; // serialization failed, but connection is still alive
};
sink.send(Message::Text(json.into())).await.is_ok()
}
async fn handle_scrape_jobs_ws(socket: WebSocket, state: AppState) {
debug!("scrape-jobs WebSocket connected");
let (mut sink, mut stream) = socket.split();
// Send initial state
let init_event = match build_init_event(&state.db_pool).await {
Ok(event) => event,
Err(e) => {
debug!(error = %e, "failed to build init event, closing WebSocket");
return;
}
};
if !send_event(&mut sink, &init_event).await {
debug!("client disconnected during init send");
return;
}
// Subscribe to broadcast events
let mut rx = state.scrape_job_events();
loop {
tokio::select! {
result = rx.recv() => {
match result {
Ok(ref event) => {
if !send_event(&mut sink, event).await {
debug!("client disconnected during event send");
break;
}
}
Err(broadcast::error::RecvError::Lagged(n)) => {
debug!(missed = n, "broadcast lagged, resyncing");
match build_init_event(&state.db_pool).await {
Ok(ref event) => {
if !send_event(&mut sink, event).await {
debug!("client disconnected during resync send");
break;
}
}
Err(e) => {
debug!(error = %e, "failed to build resync init event");
}
}
}
Err(broadcast::error::RecvError::Closed) => {
debug!("broadcast channel closed");
break;
}
}
}
msg = stream.next() => {
match msg {
Some(Ok(Message::Text(text))) => {
if let Ok(parsed) = serde_json::from_str::<serde_json::Value>(&text)
&& parsed.get("type").and_then(|t| t.as_str()) == Some("resync")
{
debug!("client requested resync");
match build_init_event(&state.db_pool).await {
Ok(ref event) => {
if !send_event(&mut sink, event).await {
debug!("client disconnected during resync send");
break;
}
}
Err(e) => {
debug!(error = %e, "failed to build resync init event");
}
}
}
}
Some(Ok(Message::Close(_))) | None => {
debug!("client disconnected");
break;
}
_ => {}
}
}
}
}
debug!("scrape-jobs WebSocket disconnected");
}
+9 -6
View File
@@ -217,10 +217,13 @@ async fn unlock_and_increment_retry_has_retries_remaining(pool: PgPool) {
) )
.await; .await;
let has_retries = scrape_jobs::unlock_and_increment_retry(id, 3, &pool) let result = scrape_jobs::unlock_and_increment_retry(id, 3, &pool)
.await .await
.unwrap(); .unwrap();
assert!(has_retries, "should have retries remaining (0→1, max=3)"); assert!(
result.is_some(),
"should have retries remaining (0→1, max=3)"
);
// Verify state in DB // Verify state in DB
let (retry_count, locked_at): (i32, Option<chrono::DateTime<chrono::Utc>>) = let (retry_count, locked_at): (i32, Option<chrono::DateTime<chrono::Utc>>) =
@@ -246,11 +249,11 @@ async fn unlock_and_increment_retry_exhausted(pool: PgPool) {
) )
.await; .await;
let has_retries = scrape_jobs::unlock_and_increment_retry(id, 3, &pool) let result = scrape_jobs::unlock_and_increment_retry(id, 3, &pool)
.await .await
.unwrap(); .unwrap();
assert!( assert!(
!has_retries, result.is_none(),
"should NOT have retries remaining (3→4, max=3)" "should NOT have retries remaining (3→4, max=3)"
); );
@@ -276,11 +279,11 @@ async fn unlock_and_increment_retry_already_exceeded(pool: PgPool) {
) )
.await; .await;
let has_retries = scrape_jobs::unlock_and_increment_retry(id, 3, &pool) let result = scrape_jobs::unlock_and_increment_retry(id, 3, &pool)
.await .await
.unwrap(); .unwrap();
assert!( assert!(
!has_retries, result.is_none(),
"should NOT have retries remaining (5→6, max=3)" "should NOT have retries remaining (5→6, max=3)"
); );
+2
View File
@@ -53,6 +53,8 @@ export interface ScrapeJob {
lockedAt: string | null; lockedAt: string | null;
retryCount: number; retryCount: number;
maxRetries: number; maxRetries: number;
queuedAt: string;
status: "processing" | "staleLock" | "exhausted" | "scheduled" | "pending";
} }
export interface ScrapeJobsResponse { export interface ScrapeJobsResponse {
@@ -0,0 +1,56 @@
<script lang="ts">
import { page } from "$app/state";
import { TriangleAlert, RotateCcw } from "@lucide/svelte";
interface Props {
/** Heading shown in the error card */
title?: string;
/** The error value from svelte:boundary */
error: unknown;
/** Reset callback from svelte:boundary */
reset: () => void;
}
let { title = "Something went wrong", error, reset }: Props = $props();
let errorName = $derived(error instanceof Error ? error.constructor.name : "Error");
let errorMessage = $derived(error instanceof Error ? error.message : String(error));
let errorStack = $derived(error instanceof Error ? error.stack : null);
</script>
<div class="flex items-center justify-center py-16 px-4">
<div class="w-full max-w-lg rounded-lg border border-status-red/25 bg-status-red/5 overflow-hidden text-sm">
<div class="px-4 py-2.5 border-b border-status-red/15 flex items-center justify-between gap-4">
<div class="flex items-center gap-2 text-status-red">
<TriangleAlert size={16} strokeWidth={2.25} />
<span class="font-semibold">{title}</span>
</div>
<span class="text-xs text-muted-foreground font-mono">{page.url.pathname}</span>
</div>
<div class="px-4 py-3 border-b border-status-red/15">
<span class="text-xs text-muted-foreground/70 font-mono">{errorName}</span>
<pre class="mt-1 text-xs text-foreground/80 overflow-auto whitespace-pre-wrap break-words">{errorMessage}</pre>
</div>
{#if errorStack}
<details class="border-b border-status-red/15">
<summary class="px-4 py-2 text-xs text-muted-foreground/70 cursor-pointer hover:text-muted-foreground select-none">
Stack trace
</summary>
<pre class="px-4 py-3 text-xs text-muted-foreground/60 overflow-auto whitespace-pre-wrap break-words max-h-48">{errorStack}</pre>
</details>
{/if}
<div class="px-4 py-2.5 flex items-center justify-end gap-3">
<span class="text-xs text-muted-foreground/60">Retries this section, not the full page</span>
<button
class="shrink-0 cursor-pointer inline-flex items-center gap-1.5 rounded-md bg-status-red px-3 py-1.5 text-sm font-medium text-white hover:brightness-110 transition-all"
onclick={reset}
>
<RotateCcw size={14} strokeWidth={2.25} />
Try again
</button>
</div>
</div>
</div>
+3 -3
View File
@@ -48,7 +48,7 @@ function inTransition(_node: HTMLElement): TransitionConfig {
function outTransition(_node: HTMLElement): TransitionConfig { function outTransition(_node: HTMLElement): TransitionConfig {
const dir = navigationStore.direction; const dir = navigationStore.direction;
const base = "position: absolute; top: 0; left: 0; width: 100%"; const base = "position: absolute; top: 0; left: 0; width: 100%; height: 100%";
if (dir === "fade") { if (dir === "fade") {
return { return {
duration: DURATION, duration: DURATION,
@@ -67,9 +67,9 @@ function outTransition(_node: HTMLElement): TransitionConfig {
} }
</script> </script>
<div class="relative overflow-hidden"> <div class="relative flex flex-1 flex-col overflow-hidden">
{#key key} {#key key}
<div in:inTransition out:outTransition class="w-full"> <div in:inTransition out:outTransition class="flex flex-1 flex-col">
{@render children()} {@render children()}
</div> </div>
{/key} {/key}
@@ -39,7 +39,7 @@ export function createSvelteTable<TData extends RowData>(options: TableOptions<T
// eslint-disable-next-line @typescript-eslint/no-explicit-any // eslint-disable-next-line @typescript-eslint/no-explicit-any
onStateChange: (updater: any) => { onStateChange: (updater: any) => {
if (updater instanceof Function) state = updater(state); if (updater instanceof Function) state = updater(state);
else state = mergeObjects(state, updater as Partial<TableState>); else state = { ...state, ...(updater as Partial<TableState>) };
options.onStateChange?.(updater); options.onStateChange?.(updater);
}, },
+3 -1
View File
@@ -7,7 +7,9 @@ export function formatRelativeDate(date: string | Date): string {
} }
/** Returns a full absolute datetime string for tooltip display, e.g. "Jan 29, 2026, 3:45:12 PM". */ /** Returns a full absolute datetime string for tooltip display, e.g. "Jan 29, 2026, 3:45:12 PM". */
export function formatAbsoluteDate(date: string | Date): string { export function formatAbsoluteDate(date: string | Date | null | undefined): string {
if (date == null) return "—";
const d = typeof date === "string" ? new Date(date) : date; const d = typeof date === "string" ? new Date(date) : date;
if (Number.isNaN(d.getTime())) return "—";
return format(d, "MMM d, yyyy, h:mm:ss a"); return format(d, "MMM d, yyyy, h:mm:ss a");
} }
+40 -31
View File
@@ -14,7 +14,7 @@ interface RelativeTimeResult {
} }
/** /**
* Compute a compact relative time string and the interval until it next changes. * Format a duration in milliseconds as a compact human-readable string.
* *
* Format tiers: * Format tiers:
* - < 60s: seconds only ("45s") * - < 60s: seconds only ("45s")
@@ -22,6 +22,33 @@ interface RelativeTimeResult {
* - < 24h: hours + minutes ("1h 23m") * - < 24h: hours + minutes ("1h 23m")
* - >= 24h: days only ("3d") * - >= 24h: days only ("3d")
*/ */
export function formatDuration(ms: number): string {
const totalSeconds = Math.floor(Math.abs(ms) / 1000);
if (totalSeconds < 60) return `${totalSeconds}s`;
const totalMinutes = Math.floor(totalSeconds / 60);
if (totalMinutes < 60) {
const secs = totalSeconds % 60;
return `${totalMinutes}m ${secs}s`;
}
const totalHours = Math.floor(totalMinutes / 60);
if (totalHours < 24) {
const mins = totalMinutes % 60;
return `${totalHours}h ${mins}m`;
}
const days = Math.floor(totalHours / 24);
return `${days}d`;
}
/**
* Compute a compact relative time string and the interval until it next changes.
*
* Uses {@link formatDuration} for the text, plus computes the optimal refresh
* interval so callers can schedule the next update efficiently.
*/
export function relativeTime(date: Date, ref: Date): RelativeTimeResult { export function relativeTime(date: Date, ref: Date): RelativeTimeResult {
const diffMs = ref.getTime() - date.getTime(); const diffMs = ref.getTime() - date.getTime();
const totalSeconds = Math.floor(diffMs / 1000); const totalSeconds = Math.floor(diffMs / 1000);
@@ -30,40 +57,22 @@ export function relativeTime(date: Date, ref: Date): RelativeTimeResult {
return { text: "now", nextUpdateMs: 1000 - (diffMs % 1000) || 1000 }; return { text: "now", nextUpdateMs: 1000 - (diffMs % 1000) || 1000 };
} }
if (totalSeconds < 60) { const text = formatDuration(diffMs);
const remainder = 1000 - (diffMs % 1000);
return {
text: `${totalSeconds}s`,
nextUpdateMs: remainder || 1000,
};
}
// Compute optimal next-update interval based on the current tier
const totalMinutes = Math.floor(totalSeconds / 60); const totalMinutes = Math.floor(totalSeconds / 60);
if (totalMinutes < 60) {
const secs = totalSeconds % 60;
const remainder = 1000 - (diffMs % 1000);
return {
text: `${totalMinutes}m ${secs}s`,
nextUpdateMs: remainder || 1000,
};
}
const totalHours = Math.floor(totalMinutes / 60); const totalHours = Math.floor(totalMinutes / 60);
if (totalHours < 24) {
const mins = totalMinutes % 60; let nextUpdateMs: number;
if (totalHours >= 24) {
const msIntoCurrentDay = diffMs % 86_400_000;
nextUpdateMs = 86_400_000 - msIntoCurrentDay || 86_400_000;
} else if (totalMinutes >= 60) {
const msIntoCurrentMinute = diffMs % 60_000; const msIntoCurrentMinute = diffMs % 60_000;
const msUntilNextMinute = 60_000 - msIntoCurrentMinute; nextUpdateMs = 60_000 - msIntoCurrentMinute || 60_000;
return { } else {
text: `${totalHours}h ${mins}m`, nextUpdateMs = 1000 - (diffMs % 1000) || 1000;
nextUpdateMs: msUntilNextMinute || 60_000,
};
} }
const days = Math.floor(totalHours / 24); return { text, nextUpdateMs };
const msIntoCurrentDay = diffMs % 86_400_000;
const msUntilNextDay = 86_400_000 - msIntoCurrentDay;
return {
text: `${days}d`,
nextUpdateMs: msUntilNextDay || 86_400_000,
};
} }
+210
View File
@@ -0,0 +1,210 @@
import type { ScrapeJob } from "$lib/api";
export type ScrapeJobStatus = "processing" | "staleLock" | "exhausted" | "scheduled" | "pending";
export type ScrapeJobEvent =
| { type: "init"; jobs: ScrapeJob[] }
| { type: "jobCreated"; job: ScrapeJob }
| { type: "jobLocked"; id: number; lockedAt: string; status: ScrapeJobStatus }
| { type: "jobCompleted"; id: number }
| {
type: "jobRetried";
id: number;
retryCount: number;
queuedAt: string;
status: ScrapeJobStatus;
}
| { type: "jobExhausted"; id: number }
| { type: "jobDeleted"; id: number };
export type ConnectionState = "connected" | "reconnecting" | "disconnected";
const PRIORITY_ORDER: Record<string, number> = {
critical: 0,
high: 1,
medium: 2,
low: 3,
};
const MAX_RECONNECT_DELAY = 30_000;
const MAX_RECONNECT_ATTEMPTS = 10;
function sortJobs(jobs: Iterable<ScrapeJob>): ScrapeJob[] {
return Array.from(jobs).sort((a, b) => {
const pa = PRIORITY_ORDER[a.priority.toLowerCase()] ?? 2;
const pb = PRIORITY_ORDER[b.priority.toLowerCase()] ?? 2;
if (pa !== pb) return pa - pb;
return new Date(a.executeAt).getTime() - new Date(b.executeAt).getTime();
});
}
export class ScrapeJobsStore {
private ws: WebSocket | null = null;
private jobs = new Map<number, ScrapeJob>();
private _connectionState: ConnectionState = "disconnected";
private _initialized = false;
private onUpdate: () => void;
private reconnectAttempts = 0;
private reconnectTimer: ReturnType<typeof setTimeout> | null = null;
private intentionalClose = false;
/** Cached sorted array, invalidated on data mutations. */
private cachedJobs: ScrapeJob[] = [];
private cacheDirty = false;
constructor(onUpdate: () => void) {
this.onUpdate = onUpdate;
}
getJobs(): ScrapeJob[] {
if (this.cacheDirty) {
this.cachedJobs = sortJobs(this.jobs.values());
this.cacheDirty = false;
}
return this.cachedJobs;
}
getConnectionState(): ConnectionState {
return this._connectionState;
}
isInitialized(): boolean {
return this._initialized;
}
connect(): void {
this.intentionalClose = false;
const protocol = window.location.protocol === "https:" ? "wss:" : "ws:";
const url = `${protocol}//${window.location.host}/api/admin/scrape-jobs/ws`;
try {
this.ws = new WebSocket(url);
} catch {
this.scheduleReconnect();
return;
}
this.ws.onopen = () => {
this._connectionState = "connected";
this.reconnectAttempts = 0;
this.onUpdate();
};
this.ws.onmessage = (event) => {
try {
const parsed = JSON.parse(event.data as string) as ScrapeJobEvent;
this.handleEvent(parsed);
} catch {
// Ignore malformed messages
}
};
this.ws.onclose = () => {
this.ws = null;
if (!this.intentionalClose) {
this.scheduleReconnect();
}
};
this.ws.onerror = () => {
// onclose will fire after onerror, so reconnect is handled there
};
}
handleEvent(event: ScrapeJobEvent): void {
switch (event.type) {
case "init":
this.jobs.clear();
for (const job of event.jobs) {
this.jobs.set(job.id, job);
}
this._initialized = true;
break;
case "jobCreated":
this.jobs.set(event.job.id, event.job);
break;
case "jobLocked": {
const job = this.jobs.get(event.id);
if (job) {
this.jobs.set(event.id, { ...job, lockedAt: event.lockedAt, status: event.status });
}
break;
}
case "jobCompleted":
this.jobs.delete(event.id);
break;
case "jobRetried": {
const job = this.jobs.get(event.id);
if (job) {
this.jobs.set(event.id, {
...job,
retryCount: event.retryCount,
queuedAt: event.queuedAt,
status: event.status,
lockedAt: null,
});
}
break;
}
case "jobExhausted": {
const job = this.jobs.get(event.id);
if (job) {
this.jobs.set(event.id, { ...job, status: "exhausted" });
}
break;
}
case "jobDeleted":
this.jobs.delete(event.id);
break;
}
this.cacheDirty = true;
this.onUpdate();
}
disconnect(): void {
this.intentionalClose = true;
if (this.reconnectTimer !== null) {
clearTimeout(this.reconnectTimer);
this.reconnectTimer = null;
}
if (this.ws) {
this.ws.close();
this.ws = null;
}
this._connectionState = "disconnected";
this.onUpdate();
}
resync(): void {
if (this.ws && this.ws.readyState === WebSocket.OPEN) {
this.ws.send(JSON.stringify({ type: "resync" }));
}
}
/** Attempt to reconnect after being disconnected. Resets attempt counter. */
retry(): void {
this.reconnectAttempts = 0;
this._connectionState = "reconnecting";
this.onUpdate();
this.connect();
}
private scheduleReconnect(): void {
if (this.reconnectAttempts >= MAX_RECONNECT_ATTEMPTS) {
this._connectionState = "disconnected";
this.onUpdate();
return;
}
this._connectionState = "reconnecting";
this.onUpdate();
const delay = Math.min(1000 * 2 ** this.reconnectAttempts, MAX_RECONNECT_DELAY);
this.reconnectAttempts++;
this.reconnectTimer = setTimeout(() => {
this.reconnectTimer = null;
this.connect();
}, delay);
}
}
+30 -1
View File
@@ -3,6 +3,7 @@ import { goto } from "$app/navigation";
import { page } from "$app/state"; import { page } from "$app/state";
import { authStore } from "$lib/auth.svelte"; import { authStore } from "$lib/auth.svelte";
import PageTransition from "$lib/components/PageTransition.svelte"; import PageTransition from "$lib/components/PageTransition.svelte";
import ErrorBoundaryFallback from "$lib/components/ErrorBoundaryFallback.svelte";
import { import {
ClipboardList, ClipboardList,
FileText, FileText,
@@ -12,10 +13,32 @@ import {
User, User,
Users, Users,
} from "@lucide/svelte"; } from "@lucide/svelte";
import { onMount } from "svelte"; import { onMount, tick } from "svelte";
let { children } = $props(); let { children } = $props();
// Track boundary reset function so navigation can auto-clear errors
let boundaryReset = $state<(() => void) | null>(null);
let errorPathname = $state<string | null>(null);
function onBoundaryError(e: unknown, reset: () => void) {
console.error("[page boundary]", e);
boundaryReset = reset;
errorPathname = page.url.pathname;
}
// Auto-reset the boundary only when the user navigates away from the errored page
$effect(() => {
const currentPath = page.url.pathname;
if (boundaryReset && errorPathname && currentPath !== errorPathname) {
const reset = boundaryReset;
boundaryReset = null;
errorPathname = null;
tick().then(() => reset());
}
});
onMount(async () => { onMount(async () => {
if (authStore.isLoading) { if (authStore.isLoading) {
await authStore.init(); await authStore.init();
@@ -115,9 +138,15 @@ function isActive(href: string): boolean {
<!-- Content --> <!-- Content -->
<main class="flex-1 min-w-0"> <main class="flex-1 min-w-0">
<svelte:boundary onerror={onBoundaryError}>
<PageTransition key={page.url.pathname} axis="vertical"> <PageTransition key={page.url.pathname} axis="vertical">
{@render children()} {@render children()}
</PageTransition> </PageTransition>
{#snippet failed(error, reset)}
<ErrorBoundaryFallback title="Page error" {error} {reset} />
{/snippet}
</svelte:boundary>
</main> </main>
</div> </div>
</div> </div>
+346 -90
View File
@@ -1,9 +1,9 @@
<script lang="ts"> <script lang="ts">
import { type ScrapeJob, type ScrapeJobsResponse, client } from "$lib/api"; import { type ScrapeJob, client } from "$lib/api";
import SimpleTooltip from "$lib/components/SimpleTooltip.svelte";
import { FlexRender, createSvelteTable } from "$lib/components/ui/data-table/index.js"; import { FlexRender, createSvelteTable } from "$lib/components/ui/data-table/index.js";
import { formatAbsoluteDate, formatRelativeDate } from "$lib/date"; import { formatAbsoluteDate } from "$lib/date";
import { ArrowDown, ArrowUp, ArrowUpDown } from "@lucide/svelte"; import { formatDuration } from "$lib/time";
import { ArrowDown, ArrowUp, ArrowUpDown, TriangleAlert } from "@lucide/svelte";
import { import {
type ColumnDef, type ColumnDef,
type SortingState, type SortingState,
@@ -12,26 +12,116 @@ import {
getSortedRowModel, getSortedRowModel,
} from "@tanstack/table-core"; } from "@tanstack/table-core";
import { onMount } from "svelte"; import { onMount } from "svelte";
import { type ConnectionState, ScrapeJobsStore } from "$lib/ws";
let data = $state<ScrapeJobsResponse | null>(null); let jobs = $state<ScrapeJob[]>([]);
let connectionState = $state<ConnectionState>("disconnected");
let initialized = $state(false);
let error = $state<string | null>(null); let error = $state<string | null>(null);
let sorting: SortingState = $state([]); let sorting: SortingState = $state([]);
let tick = $state(0);
let subjectMap = $state(new Map<string, string>());
onMount(async () => { let store: ScrapeJobsStore | undefined;
try {
data = await client.getAdminScrapeJobs(); // Shared tooltip state — single tooltip for all timing cells via event delegation
} catch (e) { let tooltipText = $state<string | null>(null);
error = e instanceof Error ? e.message : "Failed to load scrape jobs"; let tooltipX = $state(0);
let tooltipY = $state(0);
function showTooltip(event: MouseEvent) {
const target = (event.target as HTMLElement).closest<HTMLElement>("[data-timing-tooltip]");
if (!target) return;
tooltipText = target.dataset.timingTooltip ?? null;
tooltipX = event.clientX;
tooltipY = event.clientY;
} }
function moveTooltip(event: MouseEvent) {
if (tooltipText === null) return;
const target = (event.target as HTMLElement).closest<HTMLElement>("[data-timing-tooltip]");
if (!target) {
tooltipText = null;
return;
}
tooltipText = target.dataset.timingTooltip ?? null;
tooltipX = event.clientX;
tooltipY = event.clientY;
}
function hideTooltip() {
tooltipText = null;
}
onMount(() => {
// Tick every second for live time displays
const tickInterval = setInterval(() => {
tick++;
}, 1000);
// Load subject reference data
client
.getReference("subject")
.then((entries) => {
const map = new Map<string, string>();
for (const entry of entries) {
map.set(entry.code, entry.description);
}
subjectMap = map;
})
.catch(() => {
// Subject lookup is best-effort
});
// Initialize WebSocket store
store = new ScrapeJobsStore(() => {
if (!store) return;
connectionState = store.getConnectionState();
initialized = store.isInitialized();
// getJobs() returns a cached array when unchanged, so only reassign
// when the reference differs to avoid triggering reactive table rebuilds.
const next = store.getJobs();
if (next !== jobs) jobs = next;
});
store.connect();
return () => {
clearInterval(tickInterval);
store?.disconnect();
};
}); });
function handleSortingChange(updater: Updater<SortingState>) { function handleSortingChange(updater: Updater<SortingState>) {
sorting = typeof updater === "function" ? updater(sorting) : updater; sorting = typeof updater === "function" ? updater(sorting) : updater;
} }
// --- Helper functions ---
function formatJobDetails(job: ScrapeJob, subjects: Map<string, string>): string {
const payload = job.targetPayload as Record<string, unknown>;
switch (job.targetType) {
case "Subject": {
const code = payload.subject as string;
const desc = subjects.get(code);
return desc ? `${code} \u2014 ${desc}` : code;
}
case "CrnList": {
const crns = payload.crns as string[];
return `${crns.length} CRNs`;
}
case "SingleCrn":
return `CRN ${payload.crn as string}`;
case "CourseRange":
return `${payload.subject as string} ${payload.low as number}\u2013${payload.high as number}`;
default:
return JSON.stringify(payload);
}
}
function priorityColor(priority: string): string { function priorityColor(priority: string): string {
const p = priority.toLowerCase(); const p = priority.toLowerCase();
if (p === "urgent") return "text-red-500"; if (p === "urgent" || p === "critical") return "text-red-500";
if (p === "high") return "text-orange-500";
if (p === "low") return "text-muted-foreground"; if (p === "low") return "text-muted-foreground";
return "text-foreground"; return "text-foreground";
} }
@@ -42,6 +132,43 @@ function retryColor(retryCount: number, maxRetries: number): string {
return "text-muted-foreground"; return "text-muted-foreground";
} }
function statusColor(status: string): { text: string; dot: string } {
switch (status) {
case "processing":
return { text: "text-blue-500", dot: "bg-blue-500" };
case "pending":
return { text: "text-green-500", dot: "bg-green-500" };
case "scheduled":
return { text: "text-muted-foreground", dot: "bg-muted-foreground" };
case "staleLock":
return { text: "text-red-500", dot: "bg-red-500" };
case "exhausted":
return { text: "text-red-500", dot: "bg-red-500" };
default:
return { text: "text-muted-foreground", dot: "bg-muted-foreground" };
}
}
function formatStatusLabel(status: string): string {
// Convert camelCase to separate words, capitalize first letter
return status.replace(/([a-z])([A-Z])/g, "$1 $2").replace(/^\w/, (c) => c.toUpperCase());
}
function lockDurationColor(ms: number): string {
const minutes = ms / 60_000;
if (minutes >= 8) return "text-red-500";
if (minutes >= 5) return "text-amber-500";
return "text-foreground";
}
function overdueDurationColor(ms: number): string {
const minutes = ms / 60_000;
if (minutes >= 5) return "text-red-500";
return "text-amber-500";
}
// --- Table columns ---
const columns: ColumnDef<ScrapeJob, unknown>[] = [ const columns: ColumnDef<ScrapeJob, unknown>[] = [
{ {
id: "id", id: "id",
@@ -49,53 +176,69 @@ const columns: ColumnDef<ScrapeJob, unknown>[] = [
header: "ID", header: "ID",
enableSorting: false, enableSorting: false,
}, },
{
id: "status",
accessorKey: "status",
header: "Status",
enableSorting: true,
sortingFn: (rowA, rowB) => {
const order: Record<string, number> = {
processing: 0,
staleLock: 1,
pending: 2,
scheduled: 3,
exhausted: 4,
};
const a = order[rowA.original.status] ?? 3;
const b = order[rowB.original.status] ?? 3;
return a - b;
},
},
{ {
id: "targetType", id: "targetType",
accessorKey: "targetType", accessorKey: "targetType",
header: "Type", header: "Type",
enableSorting: false, enableSorting: false,
}, },
{
id: "details",
accessorFn: () => "",
header: "Details",
enableSorting: false,
},
{ {
id: "priority", id: "priority",
accessorKey: "priority", accessorKey: "priority",
header: "Priority", header: "Priority",
enableSorting: true, enableSorting: true,
sortingFn: (rowA, rowB) => { sortingFn: (rowA, rowB) => {
const order: Record<string, number> = { urgent: 0, high: 1, normal: 2, low: 3 }; const order: Record<string, number> = {
critical: 0,
urgent: 0,
high: 1,
normal: 2,
medium: 2,
low: 3,
};
const a = order[String(rowA.original.priority).toLowerCase()] ?? 2; const a = order[String(rowA.original.priority).toLowerCase()] ?? 2;
const b = order[String(rowB.original.priority).toLowerCase()] ?? 2; const b = order[String(rowB.original.priority).toLowerCase()] ?? 2;
return a - b; return a - b;
}, },
}, },
{ {
id: "executeAt", id: "timing",
accessorKey: "executeAt", accessorFn: (row) => {
header: "Execute At", if (row.lockedAt) return Date.now() - new Date(row.lockedAt).getTime();
enableSorting: true, return Date.now() - new Date(row.queuedAt).getTime();
}, },
{ header: "Timing",
id: "createdAt",
accessorKey: "createdAt",
header: "Created At",
enableSorting: false,
},
{
id: "retries",
accessorFn: (row) => row.retryCount,
header: "Retries",
enableSorting: false,
},
{
id: "status",
accessorFn: (row) => (row.lockedAt ? "Locked" : "Pending"),
header: "Status",
enableSorting: true, enableSorting: true,
}, },
]; ];
const table = createSvelteTable({ const table = createSvelteTable({
get data() { get data() {
return data?.jobs ?? []; return jobs;
}, },
getRowId: (row) => String(row.id), getRowId: (row) => String(row.id),
columns, columns,
@@ -111,29 +254,138 @@ const table = createSvelteTable({
}); });
const skeletonWidths: Record<string, string> = { const skeletonWidths: Record<string, string> = {
id: "w-8", id: "w-6",
targetType: "w-16",
priority: "w-16",
executeAt: "w-28",
createdAt: "w-28",
retries: "w-12",
status: "w-20", status: "w-20",
targetType: "w-16",
details: "w-32",
priority: "w-16",
timing: "w-32",
}; };
// Unified timing display: shows the most relevant duration for the job's current state.
// Uses _tick dependency so Svelte re-evaluates every second.
function getTimingDisplay(
job: ScrapeJob,
_tick: number
): { text: string; colorClass: string; icon: "warning" | "none"; tooltip: string } {
const now = Date.now();
const queuedTime = new Date(job.queuedAt).getTime();
const executeTime = new Date(job.executeAt).getTime();
if (job.status === "processing" || job.status === "staleLock") {
const lockedTime = job.lockedAt ? new Date(job.lockedAt).getTime() : now;
const processingMs = now - lockedTime;
const waitedMs = lockedTime - queuedTime;
const prefix = job.status === "staleLock" ? "stale" : "processing";
const colorClass =
job.status === "staleLock" ? "text-red-500" : lockDurationColor(processingMs);
const tooltipLines = [
`Queued: ${formatAbsoluteDate(job.queuedAt)}`,
`Waited: ${formatDuration(Math.max(0, waitedMs))}`,
`Locked: ${formatAbsoluteDate(job.lockedAt)}`,
`${job.status === "staleLock" ? "Stale for" : "Processing"}: ${formatDuration(processingMs)}`,
];
return {
text: `${prefix} ${formatDuration(processingMs)}`,
colorClass,
icon: job.status === "staleLock" ? "warning" : "none",
tooltip: tooltipLines.join("\n"),
};
}
if (job.status === "exhausted") {
const tooltipLines = [
`Queued: ${formatAbsoluteDate(job.queuedAt)}`,
`Retries: ${job.retryCount}/${job.maxRetries} exhausted`,
];
return {
text: "exhausted",
colorClass: "text-red-500",
icon: "warning",
tooltip: tooltipLines.join("\n"),
};
}
// Scheduled (future execute_at)
const executeAtDiff = now - executeTime;
if (job.status === "scheduled" || executeAtDiff < 0) {
const tooltipLines = [
`Queued: ${formatAbsoluteDate(job.queuedAt)}`,
`Executes: ${formatAbsoluteDate(job.executeAt)}`,
];
return {
text: `in ${formatDuration(Math.abs(executeAtDiff))}`,
colorClass: "text-muted-foreground",
icon: "none",
tooltip: tooltipLines.join("\n"),
};
}
// Pending (overdue — execute_at is in the past, waiting to be picked up)
const waitingMs = now - queuedTime;
const tooltipLines = [
`Queued: ${formatAbsoluteDate(job.queuedAt)}`,
`Waiting: ${formatDuration(waitingMs)}`,
];
return {
text: `waiting ${formatDuration(waitingMs)}`,
colorClass: overdueDurationColor(waitingMs),
icon: "warning",
tooltip: tooltipLines.join("\n"),
};
}
</script> </script>
<h1 class="mb-4 text-lg font-semibold text-foreground">Scrape Jobs</h1> <div class="flex items-center justify-between mb-4">
<h1 class="text-lg font-semibold text-foreground">Scrape Jobs</h1>
<div class="flex items-center gap-2 text-sm">
{#if connectionState === "connected"}
<span class="inline-flex items-center gap-1.5">
<span class="size-2 shrink-0 rounded-full bg-green-500"></span>
<span class="text-green-500">Live</span>
</span>
{:else if connectionState === "reconnecting"}
<span class="inline-flex items-center gap-1.5">
<span class="size-2 shrink-0 rounded-full bg-amber-500"></span>
<span class="text-amber-500">Reconnecting...</span>
</span>
{:else}
<span class="inline-flex items-center gap-2">
<span class="inline-flex items-center gap-1.5">
<span class="size-2 shrink-0 rounded-full bg-red-500"></span>
<span class="text-red-500">Disconnected</span>
</span>
<button
class="rounded-md bg-muted px-2 py-0.5 text-xs font-medium text-foreground hover:bg-muted/80 transition-colors"
onclick={() => store?.retry()}
>
Retry
</button>
</span>
{/if}
</div>
</div>
{#if error} {#if error}
<p class="text-destructive">{error}</p> <p class="text-destructive">{error}</p>
{:else} {:else}
<div class="bg-card border-border overflow-hidden rounded-lg border"> <div class="bg-card border-border overflow-hidden rounded-lg border">
<table class="w-full border-collapse text-sm"> <!-- svelte-ignore a11y_no_noninteractive_element_interactions -->
<table
class="w-full border-collapse text-xs"
onmouseenter={showTooltip}
onmousemove={moveTooltip}
onmouseleave={hideTooltip}
>
<thead> <thead>
{#each table.getHeaderGroups() as headerGroup} {#each table.getHeaderGroups() as headerGroup}
<tr class="border-b border-border text-left text-muted-foreground"> <tr class="border-b border-border text-left text-muted-foreground">
{#each headerGroup.headers as header} {#each headerGroup.headers as header}
<th <th
class="px-4 py-3 font-medium" class="px-3 py-2.5 font-medium whitespace-nowrap"
class:cursor-pointer={header.column.getCanSort()} class:cursor-pointer={header.column.getCanSort()}
class:select-none={header.column.getCanSort()} class:select-none={header.column.getCanSort()}
onclick={header.column.getToggleSortingHandler()} onclick={header.column.getToggleSortingHandler()}
@@ -169,21 +421,21 @@ const skeletonWidths: Record<string, string> = {
</tr> </tr>
{/each} {/each}
</thead> </thead>
{#if !data} {#if !initialized}
<tbody> <tbody>
{#each Array(5) as _} {#each Array(5) as _}
<tr class="border-b border-border"> <tr class="border-b border-border">
{#each columns as col} {#each columns as col}
<td class="px-4 py-3"> <td class="px-3 py-2.5">
<div <div
class="h-4 rounded bg-muted animate-pulse {skeletonWidths[col.id ?? ''] ?? 'w-20'}" class="h-3.5 rounded bg-muted animate-pulse {skeletonWidths[col.id ?? ''] ?? 'w-20'}"
></div> ></div>
</td> </td>
{/each} {/each}
</tr> </tr>
{/each} {/each}
</tbody> </tbody>
{:else if data.jobs.length === 0} {:else if jobs.length === 0}
<tbody> <tbody>
<tr> <tr>
<td colspan={columns.length} class="py-12 text-center text-muted-foreground"> <td colspan={columns.length} class="py-12 text-center text-muted-foreground">
@@ -193,67 +445,62 @@ const skeletonWidths: Record<string, string> = {
</tbody> </tbody>
{:else} {:else}
<tbody> <tbody>
{#each table.getRowModel().rows as row} {#each table.getRowModel().rows as row (row.id)}
{@const job = row.original} {@const job = row.original}
<tr class="border-b border-border last:border-b-0 hover:bg-muted/50 transition-colors"> {@const sc = statusColor(job.status)}
{@const timingDisplay = getTimingDisplay(job, tick)}
<tr
class="border-b border-border last:border-b-0 hover:bg-muted/50 transition-colors"
>
{#each row.getVisibleCells() as cell (cell.id)} {#each row.getVisibleCells() as cell (cell.id)}
{@const colId = cell.column.id} {@const colId = cell.column.id}
{#if colId === "id"} {#if colId === "id"}
<td class="px-4 py-3 tabular-nums text-muted-foreground">{job.id}</td> <td class="px-3 py-2.5 tabular-nums text-muted-foreground/70 w-12">{job.id}</td>
{:else if colId === "status"}
<td class="px-3 py-2.5 whitespace-nowrap">
<span class="inline-flex items-center gap-1.5">
<span class="size-1.5 shrink-0 rounded-full {sc.dot}"></span>
<span class="flex flex-col leading-tight">
<span class={sc.text}>{formatStatusLabel(job.status)}</span>
{#if job.maxRetries > 0}
<span class="text-[10px] {retryColor(job.retryCount, job.maxRetries)}">
{job.retryCount}/{job.maxRetries} retries
</span>
{/if}
</span>
</span>
</td>
{:else if colId === "targetType"} {:else if colId === "targetType"}
<td class="px-4 py-3"> <td class="px-3 py-2.5 whitespace-nowrap">
<span <span
class="inline-flex items-center rounded-md bg-muted/60 px-2 py-0.5 font-mono text-xs text-muted-foreground" class="inline-flex items-center rounded-md bg-muted/60 px-1.5 py-0.5 font-mono text-[11px] text-muted-foreground"
> >
{job.targetType} {job.targetType}
</span> </span>
</td> </td>
{:else if colId === "details"}
<td class="px-3 py-2.5 max-w-48 truncate text-muted-foreground" title={formatJobDetails(job, subjectMap)}>
{formatJobDetails(job, subjectMap)}
</td>
{:else if colId === "priority"} {:else if colId === "priority"}
<td class="px-4 py-3"> <td class="px-3 py-2.5 whitespace-nowrap">
<span class="font-medium capitalize {priorityColor(job.priority)}"> <span class="font-medium capitalize {priorityColor(job.priority)}">
{job.priority} {job.priority}
</span> </span>
</td> </td>
{:else if colId === "executeAt"} {:else if colId === "timing"}
<td class="px-4 py-3"> <td class="px-3 py-2.5 whitespace-nowrap">
<SimpleTooltip text={formatAbsoluteDate(job.executeAt)} passthrough> <span
<span class="text-muted-foreground"> class="inline-flex items-center gap-1.5 tabular-nums text-foreground"
{formatRelativeDate(job.executeAt)} data-timing-tooltip={timingDisplay.tooltip}
</span>
</SimpleTooltip>
</td>
{:else if colId === "createdAt"}
<td class="px-4 py-3">
<SimpleTooltip text={formatAbsoluteDate(job.createdAt)} passthrough>
<span class="text-muted-foreground">
{formatRelativeDate(job.createdAt)}
</span>
</SimpleTooltip>
</td>
{:else if colId === "retries"}
<td class="px-4 py-3">
<span class="tabular-nums {retryColor(job.retryCount, job.maxRetries)}">
{job.retryCount}/{job.maxRetries}
</span>
</td>
{:else if colId === "status"}
<td class="px-4 py-3">
{#if job.lockedAt}
<SimpleTooltip
text="Locked since {formatAbsoluteDate(job.lockedAt)}"
passthrough
> >
<span class="inline-flex items-center gap-1.5"> <span class="size-3.5 shrink-0 inline-flex items-center justify-center {timingDisplay.colorClass}">
<span class="size-2 shrink-0 rounded-full bg-amber-500"></span> {#if timingDisplay.icon === "warning"}
<span class="text-amber-500">Locked</span> <TriangleAlert class="size-3.5" />
</span>
</SimpleTooltip>
{:else}
<span class="inline-flex items-center gap-1.5">
<span class="size-2 shrink-0 rounded-full bg-green-500"></span>
<span class="text-green-500">Pending</span>
</span>
{/if} {/if}
</span>
{timingDisplay.text}
</span>
</td> </td>
{/if} {/if}
{/each} {/each}
@@ -263,4 +510,13 @@ const skeletonWidths: Record<string, string> = {
{/if} {/if}
</table> </table>
</div> </div>
{#if tooltipText !== null}
<div
class="pointer-events-none fixed z-50 bg-card text-card-foreground text-xs border border-border rounded-md px-2.5 py-1.5 shadow-sm whitespace-pre-line max-w-max text-left"
style="left: {tooltipX + 12}px; top: {tooltipY + 12}px;"
>
{tooltipText}
</div>
{/if}
{/if} {/if}
+7
View File
@@ -8,6 +8,7 @@ import { useOverlayScrollbars } from "$lib/composables/useOverlayScrollbars.svel
import { initNavigation } from "$lib/stores/navigation.svelte"; import { initNavigation } from "$lib/stores/navigation.svelte";
import { themeStore } from "$lib/stores/theme.svelte"; import { themeStore } from "$lib/stores/theme.svelte";
import { Tooltip } from "bits-ui"; import { Tooltip } from "bits-ui";
import ErrorBoundaryFallback from "$lib/components/ErrorBoundaryFallback.svelte";
import { onMount } from "svelte"; import { onMount } from "svelte";
let { children } = $props(); let { children } = $props();
@@ -40,8 +41,14 @@ onMount(() => {
<div class="flex min-h-screen flex-col"> <div class="flex min-h-screen flex-col">
<NavBar /> <NavBar />
<svelte:boundary onerror={(e) => console.error("[root boundary]", e)}>
<PageTransition key={transitionKey}> <PageTransition key={transitionKey}>
{@render children()} {@render children()}
</PageTransition> </PageTransition>
{#snippet failed(error, reset)}
<ErrorBoundaryFallback {error} {reset} />
{/snippet}
</svelte:boundary>
</div> </div>
</Tooltip.Provider> </Tooltip.Provider>
+1
View File
@@ -38,6 +38,7 @@ export default defineConfig({
target: "http://localhost:8080", target: "http://localhost:8080",
changeOrigin: true, changeOrigin: true,
secure: false, secure: false,
ws: true,
}, },
}, },
}, },