Use tokio::sync::mutex for Crawls

This commit is contained in:
Tyler Hallada 2023-07-19 23:02:25 -04:00
parent d17f909312
commit e6a37703be
3 changed files with 13 additions and 14 deletions

View File

@ -133,9 +133,7 @@ pub async fn post(
let receiver = crawl_scheduler.schedule(feed.feed_id).await;
{
let mut crawls = crawls.lock().map_err(|_| {
AddFeedError::CreateFeedError(add_feed.url.clone(), Error::InternalServerError)
})?;
let mut crawls = crawls.lock().await;
crawls.insert(feed.feed_id, receiver);
}
@ -164,7 +162,7 @@ pub async fn stream(
State(crawls): State<Crawls>,
) -> Result<impl IntoResponse> {
let receiver = {
let mut crawls = crawls.lock().expect("crawls lock poisoned");
let mut crawls = crawls.lock().await;
crawls.remove(&id.as_uuid())
}
.ok_or_else(|| Error::NotFound("feed stream", id.as_uuid()))?;

View File

@ -2,7 +2,7 @@ use std::{
collections::HashMap,
net::SocketAddr,
path::Path,
sync::{Arc, Mutex},
sync::Arc,
};
use anyhow::Result;
@ -17,6 +17,7 @@ use notify::Watcher;
use reqwest::Client;
use sqlx::postgres::PgPoolOptions;
use tokio::sync::watch::channel;
use tokio::sync::Mutex;
use tower::ServiceBuilder;
use tower_http::{services::ServeDir, trace::TraceLayer};
use tower_livereload::LiveReloadLayer;

View File

@ -1,13 +1,13 @@
use std::collections::HashMap;
use std::sync::{Arc, Mutex};
use std::sync::Arc;
use tokio::sync::{broadcast, watch};
use tokio::sync::{broadcast, watch, Mutex};
use axum::extract::FromRef;
use bytes::Bytes;
use reqwest::Client;
use sqlx::PgPool;
use uuid::Uuid;
use reqwest::Client;
use crate::actors::crawl_scheduler::{CrawlSchedulerHandle, CrawlSchedulerHandleMessage};
use crate::config::Config;