Refactor
This commit is contained in:
parent
e9f6051f60
commit
b8a1c5461b
3 changed files with 286 additions and 108 deletions
241
src/main.rs
241
src/main.rs
|
@ -1,127 +1,180 @@
|
|||
use anyhow::{format_err, Result};
|
||||
use chrono::{DateTime, Local, NaiveDate, TimeZone, Utc};
|
||||
use anyhow::{bail, format_err, Result};
|
||||
use chrono::{DateTime, Local, LocalResult, NaiveDate, TimeZone, Utc};
|
||||
use html2md::parse_html;
|
||||
use log::{debug, trace};
|
||||
use megalodon::{
|
||||
entities::status::Status, generator, megalodon::GetLocalTimelineInputOptions,
|
||||
response::Response,
|
||||
entities::Status, generator, megalodon::GetLocalTimelineInputOptions, response::Response,
|
||||
Megalodon,
|
||||
};
|
||||
use tokio_stream::{iter, StreamExt};
|
||||
|
||||
use std::env;
|
||||
|
||||
#[derive(Debug)]
|
||||
struct Post {
|
||||
id: String,
|
||||
content: String,
|
||||
created_at: DateTime<Utc>,
|
||||
struct Range {
|
||||
start: DateTime<Local>,
|
||||
end: DateTime<Local>,
|
||||
}
|
||||
|
||||
// TODO implement try_from that looks for descendants and adds them
|
||||
impl From<&Status> for Post {
|
||||
fn from(status: &Status) -> Self {
|
||||
let Status {
|
||||
id,
|
||||
created_at,
|
||||
content,
|
||||
..
|
||||
} = status;
|
||||
let id = id.clone();
|
||||
let created_at = created_at.clone();
|
||||
let content = parse_html(&content);
|
||||
Post {
|
||||
id,
|
||||
created_at,
|
||||
content,
|
||||
}
|
||||
}
|
||||
#[derive(Debug)]
|
||||
struct Page<'a> {
|
||||
oldest_id: Option<String>,
|
||||
oldest: Option<&'a DateTime<Utc>>,
|
||||
newest: Option<&'a DateTime<Utc>>,
|
||||
}
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() -> Result<()> {
|
||||
env::set_var("RUST_LOG", format!("{}=debug", module_path!()));
|
||||
env_logger::init();
|
||||
|
||||
// TODO add clap and argument for date
|
||||
let start = Local
|
||||
.from_local_datetime(
|
||||
&NaiveDate::from_ymd_opt(2023, 7, 1)
|
||||
.ok_or_else(|| format_err!("Invalid date!"))?
|
||||
.and_hms_opt(0, 0, 0)
|
||||
.expect("Failed to construct time!"),
|
||||
)
|
||||
.unwrap();
|
||||
let day = try_create_range("2023-07-01")?;
|
||||
|
||||
let end = Local
|
||||
.from_local_datetime(
|
||||
&NaiveDate::from_ymd_opt(2023, 7, 1)
|
||||
.ok_or_else(|| format_err!("Invallid date!"))?
|
||||
.and_hms_opt(23, 59, 59)
|
||||
.expect("Failed to construct time!"),
|
||||
)
|
||||
.unwrap();
|
||||
debug!("Date {}", day.end.format("%Y-%m-%d"));
|
||||
|
||||
println!("Date {:#?}", start);
|
||||
let client = create_client()?;
|
||||
|
||||
let url = env::var("MASTODON_URL")?;
|
||||
let token = env::var("MASTODON_ACCESS_TOKEN")?;
|
||||
let client = generator(megalodon::SNS::Mastodon, url, Some(token), None);
|
||||
let mut max_id: Option<String> = None;
|
||||
let mut last_id_on_page: Option<String> = None;
|
||||
debug!("Fetching posts");
|
||||
loop {
|
||||
let Response { json, .. } = if let Some(max_id) = max_id.as_ref() {
|
||||
client
|
||||
.get_local_timeline(Some(&GetLocalTimelineInputOptions {
|
||||
max_id: Some(max_id.clone()),
|
||||
..GetLocalTimelineInputOptions::default()
|
||||
}))
|
||||
.await?
|
||||
} else {
|
||||
client.get_local_timeline(None).await?
|
||||
let json = fetch_page(&client, &last_id_on_page).await?;
|
||||
let page = Page {
|
||||
newest: json.first().map(|s| &s.created_at),
|
||||
oldest_id: json.last().map(|s| s.id.clone()),
|
||||
oldest: json.last().map(|s| &s.created_at),
|
||||
};
|
||||
trace!("Page bounds {:?}", page);
|
||||
|
||||
if let Some(last) = json.last() {
|
||||
if last.created_at > start {
|
||||
max_id.replace(last.id.clone());
|
||||
continue;
|
||||
}
|
||||
if last_id_on_page.is_some() && page_start_older_than(&page, &day) {
|
||||
break;
|
||||
}
|
||||
println!(
|
||||
"{}",
|
||||
json.iter()
|
||||
.filter(|json| start <= json.created_at && json.created_at <= end)
|
||||
.map(Post::from)
|
||||
.map(|post| {
|
||||
format!(
|
||||
"{} ({})
|
||||
|
||||
if let Some(oldest_id) = page_newer_than(&page, &day) {
|
||||
last_id_on_page.replace(oldest_id);
|
||||
continue;
|
||||
}
|
||||
|
||||
let json = json
|
||||
.clone()
|
||||
.into_iter()
|
||||
.filter(|json| day.start <= json.created_at && json.created_at <= day.end)
|
||||
.collect::<Vec<Status>>();
|
||||
trace!("Filtered to {} post(s)", json.len());
|
||||
|
||||
let mut stream = iter(json);
|
||||
|
||||
while let Some(status) = stream.next().await {
|
||||
println!(
|
||||
"{}
|
||||
> {}",
|
||||
post.created_at.format("%H:%M"),
|
||||
post.id,
|
||||
post.content
|
||||
)
|
||||
})
|
||||
.collect::<Vec<String>>()
|
||||
.join("\n\n")
|
||||
);
|
||||
let context = client
|
||||
.get_status_context(String::from("110638913257555200"), None)
|
||||
.await?;
|
||||
println!(
|
||||
"{}",
|
||||
context
|
||||
.json
|
||||
status.created_at.with_timezone(&Local).format("%H:%M"),
|
||||
parse_html(&status.content)
|
||||
);
|
||||
let Response { json, .. } = client.get_status_context(status.id, None).await?;
|
||||
let thread = json
|
||||
.descendants
|
||||
.iter()
|
||||
.map(Post::from)
|
||||
.map(|post| {
|
||||
.into_iter()
|
||||
.map(|status| {
|
||||
format!(
|
||||
">
|
||||
> {} ({})
|
||||
> {}
|
||||
>> {}",
|
||||
post.created_at.format("%H:%M"),
|
||||
post.id,
|
||||
post.content
|
||||
status.created_at.with_timezone(&Local).format("%H:%M"),
|
||||
parse_html(&status.content)
|
||||
)
|
||||
})
|
||||
.collect::<Vec<String>>()
|
||||
.join("\n")
|
||||
);
|
||||
break;
|
||||
.join("\n");
|
||||
println!("{}", thread);
|
||||
}
|
||||
|
||||
if page_end_older_than(&page, &day) {
|
||||
debug!("No more posts in range.");
|
||||
break;
|
||||
}
|
||||
|
||||
if let Some(id) = page.oldest_id {
|
||||
last_id_on_page.replace(id.clone());
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn create_client() -> Result<Box<dyn Megalodon + Send + Sync>> {
|
||||
let url = env::var("MASTODON_URL")?;
|
||||
let token = env::var("MASTODON_ACCESS_TOKEN")?;
|
||||
Ok(generator(megalodon::SNS::Mastodon, url, Some(token), None))
|
||||
}
|
||||
|
||||
async fn fetch_page(
|
||||
client: &Box<dyn Megalodon + Send + Sync>,
|
||||
last_id_on_page: &Option<String>,
|
||||
) -> Result<Vec<Status>> {
|
||||
let Response { json, .. } = if let Some(max_id) = last_id_on_page.as_ref() {
|
||||
debug!("Fetching next page");
|
||||
client
|
||||
.get_local_timeline(Some(&GetLocalTimelineInputOptions {
|
||||
max_id: Some(max_id.clone()),
|
||||
..GetLocalTimelineInputOptions::default()
|
||||
}))
|
||||
.await?
|
||||
} else {
|
||||
debug!("Fetching first page");
|
||||
client.get_local_timeline(None).await?
|
||||
};
|
||||
Ok(json)
|
||||
}
|
||||
|
||||
fn try_create_range<S: AsRef<str>>(date: S) -> Result<Range> {
|
||||
Ok(Range {
|
||||
start: create_day_bound(&date, 0, 0, 0)?,
|
||||
end: create_day_bound(date, 23, 59, 59)?,
|
||||
})
|
||||
}
|
||||
|
||||
fn create_day_bound<S: AsRef<str>>(
|
||||
day: S,
|
||||
hour: u32,
|
||||
minute: u32,
|
||||
second: u32,
|
||||
) -> Result<DateTime<Local>> {
|
||||
let ts: Vec<&str> = day.as_ref().split("-").collect();
|
||||
if ts.len() != 3 {
|
||||
bail!("Invalid date format! {}", day.as_ref())
|
||||
}
|
||||
let (year, month, day) = if let [year, month, day, ..] = &ts[..] {
|
||||
(year, month, day)
|
||||
} else {
|
||||
bail!("Invalid date format! {}", day.as_ref())
|
||||
};
|
||||
let b = Local.from_local_datetime(
|
||||
&NaiveDate::from_ymd_opt(year.parse()?, month.parse()?, day.parse()?)
|
||||
.ok_or_else(|| format_err!("Invalid date!"))?
|
||||
.and_hms_opt(hour, minute, second)
|
||||
.ok_or_else(|| format_err!("Invalid time!"))?,
|
||||
);
|
||||
if let LocalResult::Single(b) = b {
|
||||
Ok(b)
|
||||
} else {
|
||||
bail!("Cannot construct day boundary!")
|
||||
}
|
||||
}
|
||||
|
||||
fn page_newer_than(page: &Page, range: &Range) -> Option<String> {
|
||||
page.oldest
|
||||
.filter(|oldest| *oldest > &range.end)
|
||||
.and_then(|_| page.oldest_id.clone())
|
||||
}
|
||||
|
||||
fn page_end_older_than(page: &Page, range: &Range) -> bool {
|
||||
status_older_than(&page.oldest, &range.start)
|
||||
}
|
||||
|
||||
fn page_start_older_than(page: &Page, range: &Range) -> bool {
|
||||
status_older_than(&page.newest, &range.start)
|
||||
}
|
||||
|
||||
fn status_older_than(status: &Option<&DateTime<Utc>>, dt: &DateTime<Local>) -> bool {
|
||||
status.map(|status| status < dt).unwrap_or_default()
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue