resolved error handling todos
This commit is contained in:
parent
6f4093302a
commit
477a94a374
@ -3,7 +3,6 @@ Simple planet like planet venus but in rust and maintained.
|
|||||||
** todo
|
** todo
|
||||||
Also see todos in the source files
|
Also see todos in the source files
|
||||||
|
|
||||||
*** TODO error handling everywhere
|
|
||||||
*** use a nice lib to process the config file
|
*** use a nice lib to process the config file
|
||||||
- should check whether dirs exists and are writeable
|
- should check whether dirs exists and are writeable
|
||||||
- should check whether feed urls can be parsed
|
- should check whether feed urls can be parsed
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
use anyhow::bail;
|
||||||
use anyhow::Result;
|
use anyhow::Result;
|
||||||
use feed_rs::model::Entry;
|
use feed_rs::model::Entry;
|
||||||
use feed_rs::model::Feed;
|
use feed_rs::model::Feed;
|
||||||
@ -29,26 +30,32 @@ impl FeedStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn slugify_url(url: &Url) -> String {
|
fn slugify_url(url: &Url) -> Result<String> {
|
||||||
let domain = url.domain().unwrap(); // todo don't hide error
|
let Some(domain) = url.domain() else {
|
||||||
|
bail!("Url has no domain: '{url}'.")
|
||||||
|
};
|
||||||
let query = url.query().unwrap_or("");
|
let query = url.query().unwrap_or("");
|
||||||
slug::slugify(format!("{domain}{}{query}", url.path()))
|
Ok(slug::slugify(format!("{domain}{}{query}", url.path())))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn generic_path(&self, url: &Url, ext: &str) -> String {
|
fn generic_path(&self, url: &Url, ext: &str) -> Result<String> {
|
||||||
format!("{}/{}{ext}", self.dir.display(), Self::slugify_url(url))
|
Ok(format!(
|
||||||
|
"{}/{}{ext}",
|
||||||
|
self.dir.display(),
|
||||||
|
Self::slugify_url(url)?
|
||||||
|
))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn feed_path(&self, url: &Url) -> String {
|
fn feed_path(&self, url: &Url) -> Result<String> {
|
||||||
self.generic_path(url, "")
|
self.generic_path(url, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fetchdata_path(&self, url: &Url) -> String {
|
fn fetchdata_path(&self, url: &Url) -> Result<String> {
|
||||||
self.generic_path(url, ".toml")
|
self.generic_path(url, ".toml")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn load_fetchdata(&self, url: &Url) -> Result<FetchData> {
|
pub fn load_fetchdata(&self, url: &Url) -> Result<FetchData> {
|
||||||
let path = self.fetchdata_path(url);
|
let path = self.fetchdata_path(url)?;
|
||||||
if !fs::exists(path.clone())? {
|
if !fs::exists(path.clone())? {
|
||||||
return Ok(FetchData::default());
|
return Ok(FetchData::default());
|
||||||
}
|
}
|
||||||
@ -56,7 +63,7 @@ impl FeedStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn has_changed(&self, url: &Url, new_feed: &Feed) -> Result<bool> {
|
fn has_changed(&self, url: &Url, new_feed: &Feed) -> Result<bool> {
|
||||||
let Some(old_feed) = self.load_feed(url, false) else {
|
let Some(old_feed) = self.load_feed(url, false)? else {
|
||||||
return Ok(true);
|
return Ok(true);
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -90,12 +97,7 @@ impl FeedStore {
|
|||||||
last_modified: hv(headers, "last_modified"),
|
last_modified: hv(headers, "last_modified"),
|
||||||
};
|
};
|
||||||
|
|
||||||
let body = response
|
let body = response.body_mut().with_config().read_to_vec()?;
|
||||||
.body_mut()
|
|
||||||
.with_config()
|
|
||||||
// .limit(MAX_BODY_SIZE)
|
|
||||||
.read_to_vec()
|
|
||||||
.unwrap();
|
|
||||||
let feed = match feed_rs::parser::parse(body.as_slice()) {
|
let feed = match feed_rs::parser::parse(body.as_slice()) {
|
||||||
Ok(f) => f,
|
Ok(f) => f,
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
@ -109,39 +111,44 @@ impl FeedStore {
|
|||||||
debug!("Storing feed for {url}.");
|
debug!("Storing feed for {url}.");
|
||||||
// todo don't serialize to string but to writer
|
// todo don't serialize to string but to writer
|
||||||
Self::write(
|
Self::write(
|
||||||
self.generic_path(url, ".ron"),
|
self.generic_path(url, ".ron")?,
|
||||||
to_string_pretty(&feed, PrettyConfig::default())?,
|
to_string_pretty(&feed, PrettyConfig::default())?,
|
||||||
)?;
|
)?;
|
||||||
Self::write(self.feed_path(url), body)?;
|
Self::write(self.feed_path(url)?, body)?;
|
||||||
Self::write(
|
Self::write(self.fetchdata_path(url)?, toml::to_string(&fetchdata)?)?;
|
||||||
self.fetchdata_path(url),
|
|
||||||
toml::to_string(&fetchdata).unwrap(),
|
|
||||||
)?;
|
|
||||||
Ok(true)
|
Ok(true)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn load_feed(&self, url: &Url, sanitize: bool) -> Option<Feed> {
|
fn load_feed(&self, url: &Url, sanitize: bool) -> Result<Option<Feed>> {
|
||||||
let parser = feed_rs::parser::Builder::new()
|
let parser = feed_rs::parser::Builder::new()
|
||||||
.sanitize_content(sanitize)
|
.sanitize_content(sanitize)
|
||||||
.build();
|
.build();
|
||||||
|
|
||||||
let path = self.feed_path(url);
|
let path = self.feed_path(url)?;
|
||||||
if !fs::exists(path.clone()).unwrap() {
|
if !fs::exists(path.clone())? {
|
||||||
return None;
|
return Ok(None);
|
||||||
}
|
}
|
||||||
let file = fs::File::open(path).unwrap();
|
let file = fs::File::open(path)?;
|
||||||
Some(parser.parse(BufReader::new(file)).unwrap())
|
Ok(Some(parser.parse(BufReader::new(file))?))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn collect(&self, feed_configs: &Vec<super::FeedConfig>) -> Vec<Entry> {
|
pub fn collect(&self, feed_configs: &Vec<super::FeedConfig>) -> Vec<Entry> {
|
||||||
let mut entries = vec![];
|
let mut entries = vec![];
|
||||||
|
|
||||||
for feed_config in feed_configs {
|
for feed_config in feed_configs {
|
||||||
let url = Url::parse(&feed_config.url).unwrap();
|
let mut feed = match (|| {
|
||||||
let Some(mut feed) = self.load_feed(&url, true) else {
|
let url = Url::parse(&feed_config.url)?;
|
||||||
// todo error handling!
|
self.load_feed(&url, true)
|
||||||
warn!("Problem parsing feed file for feed {}", feed_config.url);
|
})() {
|
||||||
continue;
|
Err(e) => {
|
||||||
|
warn!(
|
||||||
|
"Problem parsing feed file for feed {}: {e:?}",
|
||||||
|
feed_config.url
|
||||||
|
);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
Ok(None) => continue,
|
||||||
|
Ok(Some(f)) => f,
|
||||||
};
|
};
|
||||||
|
|
||||||
entries.append(&mut feed.entries);
|
entries.append(&mut feed.entries);
|
||||||
|
@ -14,8 +14,13 @@ pub struct Fetcher {
|
|||||||
|
|
||||||
impl Fetcher {
|
impl Fetcher {
|
||||||
pub fn new(bot_name: &str, from: &str) -> Fetcher {
|
pub fn new(bot_name: &str, from: &str) -> Fetcher {
|
||||||
// TODO Get URL from a better place, e.g. Cargo.toml?
|
let ua_name = format!(
|
||||||
let ua_name = format!("{bot_name}/{} https://TODO", env!("CARGO_PKG_VERSION"));
|
"{bot_name}/{} https://{} software: {}",
|
||||||
|
env!("CARGO_PKG_VERSION"),
|
||||||
|
env!("CARGO_PKG_HOMEPAGE"),
|
||||||
|
env!("CARGO_PKG_NAME")
|
||||||
|
);
|
||||||
|
info!("useragent: {ua_name}");
|
||||||
let agent = Agent::config_builder()
|
let agent = Agent::config_builder()
|
||||||
.http_status_as_error(false)
|
.http_status_as_error(false)
|
||||||
.user_agent(ua_name)
|
.user_agent(ua_name)
|
||||||
@ -49,7 +54,7 @@ impl Fetcher {
|
|||||||
let result = builder.call();
|
let result = builder.call();
|
||||||
let duration = start_instant.elapsed();
|
let duration = start_instant.elapsed();
|
||||||
|
|
||||||
let response = result?; // todo log and return false
|
let response = result?;
|
||||||
debug!(
|
debug!(
|
||||||
"fetched with status {} in {} ms: {url}",
|
"fetched with status {} in {} ms: {url}",
|
||||||
response.status(),
|
response.status(),
|
||||||
|
Loading…
x
Reference in New Issue
Block a user