refactor: fix clippy warnings

This commit is contained in:
neri 2022-11-04 11:37:10 +01:00
parent 7664ba3ec9
commit 756d4b67a0
7 changed files with 41 additions and 41 deletions

View File

@ -1,6 +1,7 @@
use std::env;
use std::path::PathBuf;
use std::str::FromStr;
use time::ext::NumericalDuration;
use time::Duration;
use tokio::fs;
@ -25,7 +26,7 @@ pub struct NoAuthLimits {
pub large_file_size: u64,
}
pub async fn get_config() -> Config {
pub async fn from_env() -> Config {
let max_file_size = env::var("UPLOAD_MAX_BYTES")
.ok()
.and_then(|variable| variable.parse().ok())
@ -68,15 +69,15 @@ pub async fn get_config() -> Config {
fn get_no_auth_limits() -> Option<NoAuthLimits> {
match (
env::var("AUTH_PASSWORD").ok(),
env_number("NO_AUTH_MAX_TIME"),
env_number("NO_AUTH_LARGE_FILE_MAX_TIME"),
env_number::<i64>("NO_AUTH_MAX_TIME"),
env_number::<i64>("NO_AUTH_LARGE_FILE_MAX_TIME"),
env_number("NO_AUTH_LARGE_FILE_SIZE"),
) {
(Some(auth_password), Some(max_time), Some(large_file_max_time), Some(large_file_size)) => {
Some(NoAuthLimits {
auth_password,
max_time: (max_time as i64).seconds(),
large_file_max_time: (large_file_max_time as i64).seconds(),
max_time: max_time.seconds(),
large_file_max_time: large_file_max_time.seconds(),
large_file_size,
})
}
@ -87,6 +88,6 @@ fn get_no_auth_limits() -> Option<NoAuthLimits> {
}
}
fn env_number(variable: &str) -> Option<u64> {
env::var(variable).ok().and_then(|n| n.parse::<u64>().ok())
fn env_number<T: FromStr>(variable: &str) -> Option<T> {
env::var(variable).ok().and_then(|n| n.parse::<T>().ok())
}

View File

@ -2,7 +2,7 @@ use sqlx::postgres::{PgPool, PgPoolOptions};
use std::env;
use time::ext::NumericalStdDuration;
pub async fn setup_db() -> PgPool {
pub async fn setup() -> PgPool {
let conn_url = &get_db_url();
log::info!("Using Connection string {}", conn_url);

View File

@ -42,8 +42,8 @@ pub async fn download(
let mime = Mime::from_str(&content_type).unwrap_or(APPLICATION_OCTET_STREAM);
let mut response = match get_view_type(&req, &mime, &path, delete).await {
ViewType::Raw => build_file_response(false, &file_name, path, mime, &req).await,
ViewType::Download => build_file_response(true, &file_name, path, mime, &req).await,
ViewType::Raw => build_file_response(false, &file_name, path, mime, &req),
ViewType::Download => build_file_response(true, &file_name, path, mime, &req),
ViewType::Html => build_text_response(&path).await,
}?;
@ -141,7 +141,7 @@ async fn build_text_response(path: &Path) -> Result<HttpResponse, Error> {
.body(html))
}
async fn build_file_response(
fn build_file_response(
download: bool,
file_name: &str,
path: PathBuf,
@ -196,7 +196,7 @@ fn get_disposition_params(filename: &str) -> Vec<DispositionParam> {
charset: Charset::Ext(String::from("UTF-8")),
language_tag: None,
value: filename.to_owned().into_bytes(),
}))
}));
}
parameters
}
@ -204,7 +204,8 @@ fn get_disposition_params(filename: &str) -> Vec<DispositionParam> {
fn insert_cache_headers(response: &mut HttpResponse, valid_till: OffsetDateTime) {
if response.status().is_success() {
let valid_duration = valid_till - OffsetDateTime::now_utc();
let valid_cache_seconds = valid_duration.whole_seconds().clamp(0, u32::MAX as i64) as u32;
let valid_cache_seconds =
valid_duration.whole_seconds().clamp(0, i64::from(u32::MAX)) as u32;
response.headers_mut().insert(
CACHE_CONTROL,
CacheControl(vec![

View File

@ -37,8 +37,8 @@ async fn not_found() -> Result<HttpResponse, Error> {
async fn main() -> std::io::Result<()> {
env_logger::Builder::from_env(Env::default().default_filter_or("info,sqlx=warn")).init();
let pool: PgPool = db::setup_db().await;
let config = config::get_config().await;
let pool: PgPool = db::setup().await;
let config = config::from_env().await;
let (sender, receiver) = channel(8);
log::info!("omnomnom");

View File

@ -15,16 +15,16 @@ lazy_static! {
fn load_mime_aliases() -> HashMap<Mime, Mime> {
tree_magic_db::aliases()
.lines()
.flat_map(|line| line.split_once(' '))
.flat_map(|(alias, mime)| Some((Mime::from_str(alias).ok()?, Mime::from_str(mime).ok()?)))
.filter_map(|line| line.split_once(' '))
.filter_map(|(alias, mime)| Some((Mime::from_str(alias).ok()?, Mime::from_str(mime).ok()?)))
.collect()
}
fn load_mime_parent_relations() -> Vec<(Mime, Mime)> {
tree_magic_db::subclasses()
.lines()
.flat_map(|line| line.split_once(' '))
.flat_map(|(child, parent)| {
.filter_map(|line| line.split_once(' '))
.filter_map(|(child, parent)| {
Some((Mime::from_str(child).ok()?, Mime::from_str(parent).ok()?))
})
.collect()
@ -34,11 +34,9 @@ fn load_mime_extensions() -> HashMap<Mime, &'static str> {
include_str!("../mime.types")
.lines()
.filter(|line| !line.is_empty() && !line.starts_with('#'))
.map(|line| line.split_whitespace())
.flat_map(|mut elements| Some((Mime::from_str(elements.next()?).ok()?, elements.next()?)))
.flat_map(|(mime, extension)| {
Some((ALIASES.get(&mime).unwrap_or(&mime).clone(), extension))
})
.map(str::split_whitespace)
.filter_map(|mut elements| Some((Mime::from_str(elements.next()?).ok()?, elements.next()?)))
.map(|(mime, extension)| (ALIASES.get(&mime).cloned().unwrap_or(mime), extension))
.collect()
}

View File

@ -32,16 +32,15 @@ pub(crate) async fn parse_multipart(
let mut password = None;
let mut size = 0;
while let Ok(Some(field)) = payload.try_next().await {
let name = get_field_name(&field)?;
let name = name.as_str();
match name {
while let Ok(Some(mut field)) = payload.try_next().await {
let name = get_field_name(&field)?.to_owned();
match name.as_str() {
"keep_for" => {
keep_for_seconds = Some(parse_string(name, field).await?);
keep_for_seconds = Some(parse_string(&name, &mut field).await?);
}
"file" => {
let (mime, uploaded_name) = get_file_metadata(&field);
if uploaded_name == None || uploaded_name.as_deref() == Some("") {
if uploaded_name.is_none() || uploaded_name.as_deref() == Some("") {
continue;
}
original_name = uploaded_name;
@ -62,10 +61,10 @@ pub(crate) async fn parse_multipart(
content_type = Some(get_content_type(&first_bytes));
}
"delete_on_download" => {
delete_on_download = parse_string(name, field).await? != "false";
delete_on_download = parse_string(&name, &mut field).await? != "false";
}
"password" => {
password = Some(parse_string(name, field).await?);
password = Some(parse_string(&name, &mut field).await?);
}
_ => {}
};
@ -77,8 +76,7 @@ pub(crate) async fn parse_multipart(
.map(|k| k.parse())
.transpose()
.map_err(|e| error::ErrorBadRequest(format!("field keep_for is not a number: {e}")))?
.map(Duration::seconds)
.unwrap_or(DEFAULT_UPLOAD_DURATION);
.map_or(DEFAULT_UPLOAD_DURATION, Duration::seconds);
let valid_till = OffsetDateTime::now_utc() + keep_for;
let upload_config = UploadConfig {
@ -88,7 +86,7 @@ pub(crate) async fn parse_multipart(
delete_on_download,
};
check_requirements(&upload_config, size, password, &keep_for, config)?;
check_requirements(&upload_config, size, &password, &keep_for, config)?;
Ok(upload_config)
}
@ -96,7 +94,7 @@ pub(crate) async fn parse_multipart(
fn check_requirements(
upload_config: &UploadConfig,
size: u64,
password: Option<String>,
password: &Option<String>,
keep_for: &Duration,
config: &config::Config,
) -> Result<(), error::Error> {
@ -127,21 +125,23 @@ fn check_requirements(
Ok(())
}
fn get_field_name(field: &Field) -> Result<String, error::Error> {
fn get_field_name(field: &Field) -> Result<&str, error::Error> {
Ok(field
.content_disposition()
.get_name()
.map(|s| s.to_owned())
.ok_or(error::ParseError::Incomplete)?)
}
async fn parse_string(name: &str, field: actix_multipart::Field) -> Result<String, error::Error> {
async fn parse_string(
name: &str,
field: &mut actix_multipart::Field,
) -> Result<String, error::Error> {
let data = read_content(field).await?;
String::from_utf8(data)
.map_err(|_| error::ErrorBadRequest(format!("could not parse field {name} as utf-8")))
}
async fn read_content(mut field: actix_multipart::Field) -> Result<Vec<u8>, error::Error> {
async fn read_content(field: &mut actix_multipart::Field) -> Result<Vec<u8>, error::Error> {
let mut data = Vec::new();
while let Some(chunk) = field.try_next().await.map_err(error::ErrorBadRequest)? {
data.extend(chunk);

View File

@ -52,7 +52,7 @@ fn build_index_html(config: &Config) -> String {
.replace("{max_size_snippet}", MAX_SIZE_SNIPPET_HTML.trim_end())
.replace("{max_size}", &render_file_size(max_file_size));
} else {
html = html.replace("{max_size_snippet}", "")
html = html.replace("{max_size_snippet}", "");
};
html
}