Compare commits
10 Commits
cccd3cbdc9
...
2d7bd0ae48
Author | SHA1 | Date | |
---|---|---|---|
2d7bd0ae48 | |||
fd942bb416 | |||
fd7f551726 | |||
155e09b159 | |||
d00a7ef8dc | |||
d7f056f77e | |||
aeb66e878c | |||
dc7f07476f | |||
870f457f1b | |||
a41dcab889 |
2
.containerignore
Normal file
2
.containerignore
Normal file
@ -0,0 +1,2 @@
|
|||||||
|
/target
|
||||||
|
Containerfile
|
12
.gitea/workflows/container.yaml
Normal file
12
.gitea/workflows/container.yaml
Normal file
@ -0,0 +1,12 @@
|
|||||||
|
name: Build Multiarch Container Image
|
||||||
|
on: [push]
|
||||||
|
jobs:
|
||||||
|
call-reusable-workflow:
|
||||||
|
uses: container/multiarch-build-workflow/.gitea/workflows/build.yaml@main
|
||||||
|
with:
|
||||||
|
repository: ${{ gitea.repository }}
|
||||||
|
ref_name: ${{ gitea.ref_name }}
|
||||||
|
sha: ${{ gitea.sha }}
|
||||||
|
registry_url: ${{ secrets.REGISTRY_URL }}
|
||||||
|
registry_user: ${{ secrets.REGISTRY_USER }}
|
||||||
|
registry_pw: ${{ secrets.REGISTRY_PW }}
|
55
Cargo.lock
generated
55
Cargo.lock
generated
@ -837,6 +837,29 @@ dependencies = [
|
|||||||
"vcpkg",
|
"vcpkg",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "little-hesinde"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"calibre-db",
|
||||||
|
"clap",
|
||||||
|
"once_cell",
|
||||||
|
"poem",
|
||||||
|
"quick-xml",
|
||||||
|
"rust-embed",
|
||||||
|
"serde",
|
||||||
|
"serde_json",
|
||||||
|
"serde_with",
|
||||||
|
"tera",
|
||||||
|
"thiserror",
|
||||||
|
"time",
|
||||||
|
"tokio",
|
||||||
|
"tokio-util",
|
||||||
|
"tracing",
|
||||||
|
"tracing-subscriber",
|
||||||
|
"uuid",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lock_api"
|
name = "lock_api"
|
||||||
version = "0.4.12"
|
version = "0.4.12"
|
||||||
@ -1354,28 +1377,6 @@ version = "0.1.23"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "d626bb9dae77e28219937af045c257c28bfd3f69333c512553507f5f9798cb76"
|
checksum = "d626bb9dae77e28219937af045c257c28bfd3f69333c512553507f5f9798cb76"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "rusty-library"
|
|
||||||
version = "0.1.0"
|
|
||||||
dependencies = [
|
|
||||||
"calibre-db",
|
|
||||||
"clap",
|
|
||||||
"once_cell",
|
|
||||||
"poem",
|
|
||||||
"quick-xml",
|
|
||||||
"rust-embed",
|
|
||||||
"serde",
|
|
||||||
"serde_json",
|
|
||||||
"serde_with",
|
|
||||||
"tera",
|
|
||||||
"thiserror",
|
|
||||||
"time",
|
|
||||||
"tokio",
|
|
||||||
"tracing",
|
|
||||||
"tracing-subscriber",
|
|
||||||
"uuid",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "ryu"
|
name = "ryu"
|
||||||
version = "1.0.17"
|
version = "1.0.17"
|
||||||
@ -1510,6 +1511,15 @@ dependencies = [
|
|||||||
"lazy_static",
|
"lazy_static",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "signal-hook-registry"
|
||||||
|
version = "1.4.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "a9e9e0b4211b72e7b8b6e85c807d36c212bdb33ea8587f7569562a84df5465b1"
|
||||||
|
dependencies = [
|
||||||
|
"libc",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "siphasher"
|
name = "siphasher"
|
||||||
version = "0.3.11"
|
version = "0.3.11"
|
||||||
@ -1678,6 +1688,7 @@ dependencies = [
|
|||||||
"mio",
|
"mio",
|
||||||
"num_cpus",
|
"num_cpus",
|
||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
|
"signal-hook-registry",
|
||||||
"socket2",
|
"socket2",
|
||||||
"tokio-macros",
|
"tokio-macros",
|
||||||
"windows-sys 0.48.0",
|
"windows-sys 0.48.0",
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
[workspace]
|
[workspace]
|
||||||
resolver = "2"
|
resolver = "2"
|
||||||
members = [
|
members = [
|
||||||
"calibre-db", "rusty-library",
|
"calibre-db", "little-hesinde",
|
||||||
]
|
]
|
||||||
|
|
||||||
[workspace.dependencies]
|
[workspace.dependencies]
|
||||||
|
22
Containerfile
Normal file
22
Containerfile
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
FROM docker.io/rust:1-alpine3.19 AS builder
|
||||||
|
|
||||||
|
RUN apk --no-cache add musl-dev
|
||||||
|
|
||||||
|
ENV CARGO_CARGO_NEW_VCS="none"
|
||||||
|
ENV CARGO_BUILD_RUSTFLAGS="-C target-feature=+crt-static"
|
||||||
|
|
||||||
|
WORKDIR /work
|
||||||
|
|
||||||
|
COPY . .
|
||||||
|
|
||||||
|
RUN cargo build --release --target=$(arch)-unknown-linux-musl
|
||||||
|
RUN cp "./target/$(arch)-unknown-linux-musl/release/little-hesinde" /app
|
||||||
|
|
||||||
|
|
||||||
|
FROM scratch
|
||||||
|
|
||||||
|
COPY --from=builder /app /app
|
||||||
|
CMD ["/app", "--", "/library"]
|
||||||
|
|
||||||
|
VOLUME ["/library"]
|
||||||
|
EXPOSE 3000
|
78
README.md
Normal file
78
README.md
Normal file
@ -0,0 +1,78 @@
|
|||||||
|
# Little Hesinde
|
||||||
|
|
||||||
|
I have been a long time user of [cops](https://github.com/seblucas/cops) and its
|
||||||
|
[fork](https://github.com/mikespub-org/seblucas-cops) in order to access my
|
||||||
|
ebook collection in a nice way from a browser. My biggest use always has been
|
||||||
|
the [ODPS](https://en.wikipedia.org/wiki/Open_Publication_Distribution_System)
|
||||||
|
feed for getting all of that from within [KOReader](https://koreader.rocks/).
|
||||||
|
|
||||||
|
It so happened that exactly when I was looking for a small side project I also
|
||||||
|
had trouble with a new cops version (not through their fault, only because I
|
||||||
|
insist on writing my own containers). _How hard can it be_ I thought and went on
|
||||||
|
hacking something together. The result does at most one tenth of what cops can
|
||||||
|
do but luckily enough it is the part I need for myself.
|
||||||
|
|
||||||
|
# Building
|
||||||
|
|
||||||
|
## Nix
|
||||||
|
|
||||||
|
A [nix](https://nixos.org/download/) environment with enabled
|
||||||
|
[nix-commands](https://nixos.wiki/wiki/Flakes) in order to use `nix develop` and
|
||||||
|
`nix build`.
|
||||||
|
|
||||||
|
A statically linked binary for linux systems (using
|
||||||
|
[musl](https://musl.libc.org/)) can be compiled by running `nix build` (run
|
||||||
|
`nix flake show` to get a list of available targets).
|
||||||
|
|
||||||
|
Otherwise run `nix develop` to be dropped into a shell with everything installed
|
||||||
|
and configured. From there all the usual `cargo` commands are accessible.
|
||||||
|
|
||||||
|
## Classic
|
||||||
|
|
||||||
|
A recent [rust](https://www.rust-lang.org/learn/get-started) installation is all
|
||||||
|
that is needed.
|
||||||
|
|
||||||
|
From there on `cargo run` and `cargo build` and so on can be used.
|
||||||
|
|
||||||
|
# Configuration
|
||||||
|
|
||||||
|
The binary takes exactly one argument, the path to the calibre library folder.
|
||||||
|
|
||||||
|
The listening port is hardcoded to `3000` for now, as is the listening on all
|
||||||
|
interfaces.
|
||||||
|
|
||||||
|
# Usage
|
||||||
|
|
||||||
|
Run the binary with the calibre library as an argument and open
|
||||||
|
http://localhost:3000 (or wherever it should be accessible).
|
||||||
|
http://localhost:3000/opds is the entry point for the OPDS feed.
|
||||||
|
|
||||||
|
# FAQ
|
||||||
|
|
||||||
|
## No authentication?
|
||||||
|
|
||||||
|
Not planned, put a reverse proxy in front of it that handles access.
|
||||||
|
|
||||||
|
## No search?
|
||||||
|
|
||||||
|
On my todo list once I feel like I need it.
|
||||||
|
|
||||||
|
## Why are the OPDS entries not paginated?
|
||||||
|
|
||||||
|
My hardware (a Kobo Aura One from ~2016) with KOReader works perfectly fine with
|
||||||
|
parsing the 1MB book feed from own library. Once that changes I might get over
|
||||||
|
my laziness and implement it.
|
||||||
|
|
||||||
|
## Aren't these database access patterns inefficient?
|
||||||
|
|
||||||
|
They most probably are but my elaborate testing setup (my own calibre library)
|
||||||
|
works fine with it.
|
||||||
|
|
||||||
|
## Why rust?
|
||||||
|
|
||||||
|
I like the language and wanted to try the
|
||||||
|
[poem](https://github.com/poem-web/poem) framework.
|
||||||
|
|
||||||
|
## Is it webscale?
|
||||||
|
|
||||||
|
Go away.
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Bundle all functions together.
|
||||||
|
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
|
|
||||||
use r2d2::Pool;
|
use r2d2::Pool;
|
||||||
@ -7,12 +9,17 @@ use crate::data::{
|
|||||||
author::Author, book::Book, error::DataStoreError, pagination::SortOrder, series::Series,
|
author::Author, book::Book, error::DataStoreError, pagination::SortOrder, series::Series,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/// Top level calibre functions, bundling all sub functions in one place and providing secure access to
|
||||||
|
/// the database.
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub struct Calibre {
|
pub struct Calibre {
|
||||||
pool: Pool<SqliteConnectionManager>,
|
pool: Pool<SqliteConnectionManager>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Calibre {
|
impl Calibre {
|
||||||
|
/// Open a connection to the calibre database.
|
||||||
|
///
|
||||||
|
/// Fail if the database file can not be opened or not be found.
|
||||||
pub fn load(path: &Path) -> Result<Self, DataStoreError> {
|
pub fn load(path: &Path) -> Result<Self, DataStoreError> {
|
||||||
let manager = SqliteConnectionManager::file(path);
|
let manager = SqliteConnectionManager::file(path);
|
||||||
let pool = r2d2::Pool::new(manager)?;
|
let pool = r2d2::Pool::new(manager)?;
|
||||||
@ -20,6 +27,8 @@ impl Calibre {
|
|||||||
Ok(Self { pool })
|
Ok(Self { pool })
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch book data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn books(
|
pub fn books(
|
||||||
&self,
|
&self,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -30,6 +39,8 @@ impl Calibre {
|
|||||||
Book::multiple(&conn, limit, cursor, sort_order)
|
Book::multiple(&conn, limit, cursor, sort_order)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch author data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn authors(
|
pub fn authors(
|
||||||
&self,
|
&self,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -40,6 +51,8 @@ impl Calibre {
|
|||||||
Author::multiple(&conn, limit, cursor, sort_order)
|
Author::multiple(&conn, limit, cursor, sort_order)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch books for an author specified by `author_id`, paginate the books by starting at `cursor`,
|
||||||
|
/// fetching up to an amount of `limit` and ordering by `sort_order`.
|
||||||
pub fn author_books(
|
pub fn author_books(
|
||||||
&self,
|
&self,
|
||||||
author_id: u64,
|
author_id: u64,
|
||||||
@ -51,21 +64,26 @@ impl Calibre {
|
|||||||
Book::author_books(&conn, author_id, limit, cursor, sort_order)
|
Book::author_books(&conn, author_id, limit, cursor, sort_order)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get recent books up to a limit of `limit`.
|
||||||
pub fn recent_books(&self, limit: u64) -> Result<Vec<Book>, DataStoreError> {
|
pub fn recent_books(&self, limit: u64) -> Result<Vec<Book>, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Book::recents(&conn, limit)
|
Book::recents(&conn, limit)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get a single book, specified `id`.
|
||||||
pub fn scalar_book(&self, id: u64) -> Result<Book, DataStoreError> {
|
pub fn scalar_book(&self, id: u64) -> Result<Book, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Book::scalar_book(&conn, id)
|
Book::scalar_book(&conn, id)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get the author to a book with id `id`.
|
||||||
pub fn book_author(&self, id: u64) -> Result<Author, DataStoreError> {
|
pub fn book_author(&self, id: u64) -> Result<Author, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Author::book_author(&conn, id)
|
Author::book_author(&conn, id)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch series data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn series(
|
pub fn series(
|
||||||
&self,
|
&self,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -76,51 +94,61 @@ impl Calibre {
|
|||||||
Series::multiple(&conn, limit, cursor, sort_order)
|
Series::multiple(&conn, limit, cursor, sort_order)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get the series a book with id `id` is in, as well as the book's position within the series.
|
||||||
pub fn book_series(&self, id: u64) -> Result<Option<(Series, f64)>, DataStoreError> {
|
pub fn book_series(&self, id: u64) -> Result<Option<(Series, f64)>, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Series::book_series(&conn, id)
|
Series::book_series(&conn, id)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get all books belonging to the series with id `id`.
|
||||||
pub fn series_books(&self, id: u64) -> Result<Vec<Book>, DataStoreError> {
|
pub fn series_books(&self, id: u64) -> Result<Vec<Book>, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Book::series_books(&conn, id)
|
Book::series_books(&conn, id)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more authors before the specified cursor.
|
||||||
pub fn has_previous_authors(&self, author_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_previous_authors(&self, author_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Author::has_previous_authors(&conn, author_sort)
|
Author::has_previous_authors(&conn, author_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more authors after the specified cursor.
|
||||||
pub fn has_more_authors(&self, author_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_authors(&self, author_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Author::has_more_authors(&conn, author_sort)
|
Author::has_more_authors(&conn, author_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more books before the specified cursor.
|
||||||
pub fn has_previous_books(&self, book_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_previous_books(&self, book_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Book::has_previous_books(&conn, book_sort)
|
Book::has_previous_books(&conn, book_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more books after the specified cursor.
|
||||||
pub fn has_more_books(&self, book_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_books(&self, book_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Book::has_more_books(&conn, book_sort)
|
Book::has_more_books(&conn, book_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more series before the specified cursor.
|
||||||
pub fn has_previous_series(&self, series_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_previous_series(&self, series_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Series::has_previous_series(&conn, series_sort)
|
Series::has_previous_series(&conn, series_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more series after the specified cursor.
|
||||||
pub fn has_more_series(&self, series_sort: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_series(&self, series_sort: &str) -> Result<bool, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Series::has_more_series(&conn, series_sort)
|
Series::has_more_series(&conn, series_sort)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch a single author with id `id`.
|
||||||
pub fn scalar_author(&self, id: u64) -> Result<Author, DataStoreError> {
|
pub fn scalar_author(&self, id: u64) -> Result<Author, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Author::scalar_author(&conn, id)
|
Author::scalar_author(&conn, id)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch a single series with id `id`.
|
||||||
pub fn scalar_series(&self, id: u64) -> Result<Series, DataStoreError> {
|
pub fn scalar_series(&self, id: u64) -> Result<Series, DataStoreError> {
|
||||||
let conn = self.pool.get()?;
|
let conn = self.pool.get()?;
|
||||||
Series::scalar_series(&conn, id)
|
Series::scalar_series(&conn, id)
|
||||||
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Author data.
|
||||||
|
|
||||||
use rusqlite::{named_params, Connection, Row};
|
use rusqlite::{named_params, Connection, Row};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
@ -6,10 +8,14 @@ use super::{
|
|||||||
pagination::{Pagination, SortOrder},
|
pagination::{Pagination, SortOrder},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/// Author in calibre.
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct Author {
|
pub struct Author {
|
||||||
|
/// Id in database.
|
||||||
pub id: u64,
|
pub id: u64,
|
||||||
|
/// Full name.
|
||||||
pub name: String,
|
pub name: String,
|
||||||
|
/// Full name for sorting.
|
||||||
pub sort: String,
|
pub sort: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -22,6 +28,8 @@ impl Author {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch author data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn multiple(
|
pub fn multiple(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -37,6 +45,7 @@ impl Author {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get the author to a book with id `id`.
|
||||||
pub fn book_author(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
pub fn book_author(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
||||||
let mut stmt = conn.prepare(
|
let mut stmt = conn.prepare(
|
||||||
"SELECT authors.id, authors.name, authors.sort FROM authors \
|
"SELECT authors.id, authors.name, authors.sort FROM authors \
|
||||||
@ -47,12 +56,14 @@ impl Author {
|
|||||||
Ok(stmt.query_row(params, Self::from_row)?)
|
Ok(stmt.query_row(params, Self::from_row)?)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch a single author with id `id`.
|
||||||
pub fn scalar_author(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
pub fn scalar_author(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
||||||
let mut stmt = conn.prepare("SELECT id, name, sort FROM authors WHERE id = (:id)")?;
|
let mut stmt = conn.prepare("SELECT id, name, sort FROM authors WHERE id = (:id)")?;
|
||||||
let params = named_params! { ":id": id };
|
let params = named_params! { ":id": id };
|
||||||
Ok(stmt.query_row(params, Self::from_row)?)
|
Ok(stmt.query_row(params, Self::from_row)?)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more authors before the specified cursor.
|
||||||
pub fn has_previous_authors(
|
pub fn has_previous_authors(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
sort_name: &str,
|
sort_name: &str,
|
||||||
@ -60,6 +71,7 @@ impl Author {
|
|||||||
Pagination::has_prev_or_more(conn, "authors", sort_name, &SortOrder::DESC)
|
Pagination::has_prev_or_more(conn, "authors", sort_name, &SortOrder::DESC)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more authors after the specified cursor.
|
||||||
pub fn has_more_authors(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_authors(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
||||||
Pagination::has_prev_or_more(conn, "authors", sort_name, &SortOrder::ASC)
|
Pagination::has_prev_or_more(conn, "authors", sort_name, &SortOrder::ASC)
|
||||||
}
|
}
|
||||||
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Book data.
|
||||||
|
|
||||||
use rusqlite::{named_params, Connection, Row};
|
use rusqlite::{named_params, Connection, Row};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
@ -7,14 +9,22 @@ use super::{
|
|||||||
pagination::{Pagination, SortOrder},
|
pagination::{Pagination, SortOrder},
|
||||||
};
|
};
|
||||||
|
|
||||||
#[derive(Debug, Serialize)]
|
/// Book in calibre.
|
||||||
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct Book {
|
pub struct Book {
|
||||||
|
/// Id in database.
|
||||||
pub id: u64,
|
pub id: u64,
|
||||||
|
/// Book title.
|
||||||
pub title: String,
|
pub title: String,
|
||||||
|
/// Book title for sorting.
|
||||||
pub sort: String,
|
pub sort: String,
|
||||||
|
/// Folder of the book within the calibre library.
|
||||||
pub path: String,
|
pub path: String,
|
||||||
|
/// Uuid of the book.
|
||||||
pub uuid: String,
|
pub uuid: String,
|
||||||
|
/// When was the book last modified.
|
||||||
pub last_modified: OffsetDateTime,
|
pub last_modified: OffsetDateTime,
|
||||||
|
/// Optional description.
|
||||||
pub description: Option<String>,
|
pub description: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -31,6 +41,8 @@ impl Book {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch book data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn multiple(
|
pub fn multiple(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -47,6 +59,8 @@ impl Book {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch books for an author specified by `author_id`, paginate the books by starting at `cursor`,
|
||||||
|
/// fetching up to an amount of `limit` and ordering by `sort_order`.
|
||||||
pub fn author_books(
|
pub fn author_books(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
author_id: u64,
|
author_id: u64,
|
||||||
@ -66,6 +80,7 @@ impl Book {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get all books belonging to the series with id `id`.
|
||||||
pub fn series_books(conn: &Connection, id: u64) -> Result<Vec<Book>, DataStoreError> {
|
pub fn series_books(conn: &Connection, id: u64) -> Result<Vec<Book>, DataStoreError> {
|
||||||
let mut stmt = conn.prepare(
|
let mut stmt = conn.prepare(
|
||||||
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text FROM series \
|
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text FROM series \
|
||||||
@ -80,6 +95,7 @@ impl Book {
|
|||||||
Ok(iter.filter_map(Result::ok).collect())
|
Ok(iter.filter_map(Result::ok).collect())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get recent books up to a limit of `limit`.
|
||||||
pub fn recents(conn: &Connection, limit: u64) -> Result<Vec<Self>, DataStoreError> {
|
pub fn recents(conn: &Connection, limit: u64) -> Result<Vec<Self>, DataStoreError> {
|
||||||
let mut stmt = conn.prepare(
|
let mut stmt = conn.prepare(
|
||||||
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text \
|
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text \
|
||||||
@ -90,6 +106,7 @@ impl Book {
|
|||||||
Ok(iter.filter_map(Result::ok).collect())
|
Ok(iter.filter_map(Result::ok).collect())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get a single book, specified `id`.
|
||||||
pub fn scalar_book(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
pub fn scalar_book(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
||||||
let mut stmt = conn.prepare(
|
let mut stmt = conn.prepare(
|
||||||
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text \
|
"SELECT books.id, books.title, books.sort, books.path, books.uuid, books.last_modified, comments.text \
|
||||||
@ -99,10 +116,12 @@ impl Book {
|
|||||||
Ok(stmt.query_row(params, Self::from_row)?)
|
Ok(stmt.query_row(params, Self::from_row)?)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more books before the specified cursor.
|
||||||
pub fn has_previous_books(conn: &Connection, sort_title: &str) -> Result<bool, DataStoreError> {
|
pub fn has_previous_books(conn: &Connection, sort_title: &str) -> Result<bool, DataStoreError> {
|
||||||
Pagination::has_prev_or_more(conn, "books", sort_title, &SortOrder::DESC)
|
Pagination::has_prev_or_more(conn, "books", sort_title, &SortOrder::DESC)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more books after the specified cursor.
|
||||||
pub fn has_more_books(conn: &Connection, sort_title: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_books(conn: &Connection, sort_title: &str) -> Result<bool, DataStoreError> {
|
||||||
Pagination::has_prev_or_more(conn, "books", sort_title, &SortOrder::ASC)
|
Pagination::has_prev_or_more(conn, "books", sort_title, &SortOrder::ASC)
|
||||||
}
|
}
|
||||||
|
@ -1,19 +1,28 @@
|
|||||||
|
//! Error handling for calibre database access.
|
||||||
|
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
use time::error::Parse;
|
use time::error::Parse;
|
||||||
|
|
||||||
|
/// Errors from accessing the calibre database.
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
#[error("data store error")]
|
#[error("data store error")]
|
||||||
pub enum DataStoreError {
|
pub enum DataStoreError {
|
||||||
|
/// Found no entries for the query.
|
||||||
#[error("no results")]
|
#[error("no results")]
|
||||||
NoResults(rusqlite::Error),
|
NoResults(rusqlite::Error),
|
||||||
|
/// Error with SQLite.
|
||||||
#[error("sqlite error")]
|
#[error("sqlite error")]
|
||||||
SqliteError(rusqlite::Error),
|
SqliteError(rusqlite::Error),
|
||||||
|
/// Error connecting to the database.
|
||||||
#[error("connection error")]
|
#[error("connection error")]
|
||||||
ConnectionError(#[from] r2d2::Error),
|
ConnectionError(#[from] r2d2::Error),
|
||||||
|
/// Error wparsing a datetime from the database.
|
||||||
#[error("failed to parse datetime")]
|
#[error("failed to parse datetime")]
|
||||||
DateTimeError(#[from] Parse),
|
DateTimeError(#[from] Parse),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert an SQLite error into a proper NoResults one if the query
|
||||||
|
/// returned no rows, return others as is.
|
||||||
impl From<rusqlite::Error> for DataStoreError {
|
impl From<rusqlite::Error> for DataStoreError {
|
||||||
fn from(error: rusqlite::Error) -> Self {
|
fn from(error: rusqlite::Error) -> Self {
|
||||||
match error {
|
match error {
|
||||||
|
@ -1,22 +1,33 @@
|
|||||||
|
//! Cursor pagination handling.
|
||||||
|
|
||||||
use rusqlite::{named_params, Connection, Row, ToSql};
|
use rusqlite::{named_params, Connection, Row, ToSql};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use super::error::DataStoreError;
|
use super::error::DataStoreError;
|
||||||
|
|
||||||
|
/// How to sort query results. Signifying whether we are paginating forwards or backwards.
|
||||||
#[derive(Debug, Copy, Clone, PartialEq, Deserialize, Serialize)]
|
#[derive(Debug, Copy, Clone, PartialEq, Deserialize, Serialize)]
|
||||||
pub enum SortOrder {
|
pub enum SortOrder {
|
||||||
|
/// Forwards
|
||||||
ASC,
|
ASC,
|
||||||
|
/// Backwards
|
||||||
DESC,
|
DESC,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Pagination data.
|
||||||
pub struct Pagination<'a> {
|
pub struct Pagination<'a> {
|
||||||
|
/// Sort by this column.
|
||||||
pub sort_col: &'a str,
|
pub sort_col: &'a str,
|
||||||
|
/// Limit returned results.
|
||||||
pub limit: u64,
|
pub limit: u64,
|
||||||
|
/// Where to start paginating.
|
||||||
pub cursor: Option<&'a str>,
|
pub cursor: Option<&'a str>,
|
||||||
|
/// Paginating forwards or backwards.
|
||||||
pub sort_order: SortOrder,
|
pub sort_order: SortOrder,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'a> Pagination<'a> {
|
impl<'a> Pagination<'a> {
|
||||||
|
/// Create a new pagination.
|
||||||
pub fn new(
|
pub fn new(
|
||||||
sort_col: &'a str,
|
sort_col: &'a str,
|
||||||
cursor: Option<&'a str>,
|
cursor: Option<&'a str>,
|
||||||
@ -40,6 +51,7 @@ impl<'a> Pagination<'a> {
|
|||||||
.to_string()
|
.to_string()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more items forwards or backwards from `cursor` (direction specified by `sort_order`).
|
||||||
pub fn has_prev_or_more(
|
pub fn has_prev_or_more(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
table: &str,
|
table: &str,
|
||||||
@ -57,6 +69,7 @@ impl<'a> Pagination<'a> {
|
|||||||
Ok(count > 0)
|
Ok(count > 0)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Paginate a statement.
|
||||||
pub fn paginate<T, F>(
|
pub fn paginate<T, F>(
|
||||||
&self,
|
&self,
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
@ -77,7 +90,8 @@ impl<'a> Pagination<'a> {
|
|||||||
};
|
};
|
||||||
|
|
||||||
let sort_col = self.sort_col;
|
let sort_col = self.sort_col;
|
||||||
// otherwise paginated statements with join will fail
|
// otherwise paginated statements with join will fails, not happy with this but fine for
|
||||||
|
// now
|
||||||
let sort_col_wrapped = if let Some(index) = sort_col.find('.') {
|
let sort_col_wrapped = if let Some(index) = sort_col.find('.') {
|
||||||
let right_part = &sort_col[index..];
|
let right_part = &sort_col[index..];
|
||||||
"t".to_owned() + right_part
|
"t".to_owned() + right_part
|
||||||
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Series data.
|
||||||
|
|
||||||
use rusqlite::{named_params, Connection, Row};
|
use rusqlite::{named_params, Connection, Row};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
@ -6,10 +8,14 @@ use super::{
|
|||||||
pagination::{Pagination, SortOrder},
|
pagination::{Pagination, SortOrder},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/// Series in calibre.
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct Series {
|
pub struct Series {
|
||||||
|
/// Id in database.
|
||||||
pub id: u64,
|
pub id: u64,
|
||||||
|
/// Series name.
|
||||||
pub name: String,
|
pub name: String,
|
||||||
|
/// Series name for sorting.
|
||||||
pub sort: String,
|
pub sort: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -22,6 +28,8 @@ impl Series {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch series data from calibre, starting at `cursor`, fetching up to an amount of `limit` and
|
||||||
|
/// ordering by `sort_order`.
|
||||||
pub fn multiple(
|
pub fn multiple(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
limit: u64,
|
limit: u64,
|
||||||
@ -37,12 +45,14 @@ impl Series {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Fetch a single series with id `id`.
|
||||||
pub fn scalar_series(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
pub fn scalar_series(conn: &Connection, id: u64) -> Result<Self, DataStoreError> {
|
||||||
let mut stmt = conn.prepare("SELECT id, name, sort FROM series WHERE id = (:id)")?;
|
let mut stmt = conn.prepare("SELECT id, name, sort FROM series WHERE id = (:id)")?;
|
||||||
let params = named_params! { ":id": id };
|
let params = named_params! { ":id": id };
|
||||||
Ok(stmt.query_row(params, Self::from_row)?)
|
Ok(stmt.query_row(params, Self::from_row)?)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get the series a book with id `id` is in, as well as the book's position within the series.
|
||||||
pub fn book_series(
|
pub fn book_series(
|
||||||
conn: &Connection,
|
conn: &Connection,
|
||||||
book_id: u64,
|
book_id: u64,
|
||||||
@ -68,10 +78,12 @@ impl Series {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more series before the specified cursor.
|
||||||
pub fn has_previous_series(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
pub fn has_previous_series(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
||||||
Pagination::has_prev_or_more(conn, "series", sort_name, &SortOrder::DESC)
|
Pagination::has_prev_or_more(conn, "series", sort_name, &SortOrder::DESC)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Check if there are more series after the specified cursor.
|
||||||
pub fn has_more_series(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
pub fn has_more_series(conn: &Connection, sort_name: &str) -> Result<bool, DataStoreError> {
|
||||||
Pagination::has_prev_or_more(conn, "series", sort_name, &SortOrder::ASC)
|
Pagination::has_prev_or_more(conn, "series", sort_name, &SortOrder::ASC)
|
||||||
}
|
}
|
||||||
|
@ -1,4 +1,7 @@
|
|||||||
|
//! Read data from a calibre library, leveraging its SQLite metadata database.
|
||||||
|
|
||||||
pub mod calibre;
|
pub mod calibre;
|
||||||
|
/// Data structs for the calibre database.
|
||||||
pub mod data {
|
pub mod data {
|
||||||
pub mod author;
|
pub mod author;
|
||||||
pub mod book;
|
pub mod book;
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
{
|
{
|
||||||
description = "rusty-library project";
|
description = "little-hesinde project";
|
||||||
inputs = {
|
inputs = {
|
||||||
nixpkgs.url = "github:NixOS/nixpkgs/nixos-unstable";
|
nixpkgs.url = "github:NixOS/nixpkgs/nixos-unstable";
|
||||||
flake-utils.url = "github:numtide/flake-utils";
|
flake-utils.url = "github:numtide/flake-utils";
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "rusty-library"
|
name = "little-hesinde"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
license = { workspace = true }
|
license = { workspace = true }
|
||||||
@ -16,7 +16,8 @@ serde_with = "3.8.1"
|
|||||||
tera = "1.19.1"
|
tera = "1.19.1"
|
||||||
thiserror = { workspace = true }
|
thiserror = { workspace = true }
|
||||||
time = { workspace = true }
|
time = { workspace = true }
|
||||||
tokio = { version = "1.37.0", features = ["rt-multi-thread", "macros"] }
|
tokio = { version = "1.37.0", features = ["signal", "rt-multi-thread", "macros"] }
|
||||||
|
tokio-util = "0.7.11"
|
||||||
tracing = "0.1.40"
|
tracing = "0.1.40"
|
||||||
tracing-subscriber = "0.3.18"
|
tracing-subscriber = "0.3.18"
|
||||||
uuid = { version = "1.8.0", features = ["v4", "fast-rng"] }
|
uuid = { version = "1.8.0", features = ["v4", "fast-rng"] }
|
13
little-hesinde/src/app_state.rs
Normal file
13
little-hesinde/src/app_state.rs
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
//! Data for global app state.
|
||||||
|
|
||||||
|
use calibre_db::calibre::Calibre;
|
||||||
|
|
||||||
|
use crate::config::Config;
|
||||||
|
|
||||||
|
/// Global application state, meant to be used in request handlers.
|
||||||
|
pub struct AppState {
|
||||||
|
/// Access calibre database.
|
||||||
|
pub calibre: Calibre,
|
||||||
|
/// Access application configuration.
|
||||||
|
pub config: Config,
|
||||||
|
}
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Cli interface.
|
||||||
|
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
|
|
||||||
/// Simple opds server for calibre
|
/// Simple opds server for calibre
|
@ -1,23 +1,32 @@
|
|||||||
|
//! Configuration data.
|
||||||
|
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
|
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
|
|
||||||
use crate::cli::Cli;
|
use crate::cli::Cli;
|
||||||
|
|
||||||
|
/// Errors when dealing with application configuration.
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
pub enum ConfigError {
|
pub enum ConfigError {
|
||||||
|
/// Calibre library path does not exist.
|
||||||
#[error("no folder at {0}")]
|
#[error("no folder at {0}")]
|
||||||
LibraryPathNotFound(String),
|
LibraryPathNotFound(String),
|
||||||
|
/// Calibre database does not exist.
|
||||||
#[error("no metadata.db in {0}")]
|
#[error("no metadata.db in {0}")]
|
||||||
MetadataNotFound(String),
|
MetadataNotFound(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Application configuration.
|
||||||
pub struct Config {
|
pub struct Config {
|
||||||
|
/// Calibre library folder.
|
||||||
pub library_path: PathBuf,
|
pub library_path: PathBuf,
|
||||||
|
/// Calibre metadata file path.
|
||||||
pub metadata_path: PathBuf,
|
pub metadata_path: PathBuf,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Config {
|
impl Config {
|
||||||
|
/// Check if the calibre library from `args` exists and if the calibre database can be found.
|
||||||
pub fn load(args: &Cli) -> Result<Self, ConfigError> {
|
pub fn load(args: &Cli) -> Result<Self, ConfigError> {
|
||||||
let library_path = Path::new(&args.library_path).to_path_buf();
|
let library_path = Path::new(&args.library_path).to_path_buf();
|
||||||
|
|
@ -1,17 +1,22 @@
|
|||||||
|
//! Enrich the [`Book`](struct@calibre_db::data::book::Book) type with additional information.
|
||||||
|
|
||||||
use std::{collections::HashMap, fmt::Display, path::Path};
|
use std::{collections::HashMap, fmt::Display, path::Path};
|
||||||
|
|
||||||
use calibre_db::data::{
|
use calibre_db::data::{
|
||||||
author::Author as DbAuthor, book::Book as DbBook, series::Series as DbSeries,
|
author::Author as DbAuthor, book::Book as DbBook, series::Series as DbSeries,
|
||||||
};
|
};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use time::OffsetDateTime;
|
|
||||||
|
|
||||||
use crate::app_state::AppState;
|
use crate::app_state::AppState;
|
||||||
|
|
||||||
|
/// Wrapper type for a file format string (must be a struct in order to implement traits).
|
||||||
#[derive(Debug, Clone, Serialize, Eq, PartialEq, Hash)]
|
#[derive(Debug, Clone, Serialize, Eq, PartialEq, Hash)]
|
||||||
pub struct Format(pub String);
|
pub struct Format(pub String);
|
||||||
|
/// Wrapper type for a collection of formats, [`Format`](struct@Format) on the left and a String
|
||||||
|
/// signifying its file path on the right.
|
||||||
pub type Formats = HashMap<Format, String>;
|
pub type Formats = HashMap<Format, String>;
|
||||||
|
|
||||||
|
/// Recognize `pdf` and `epub` and return their value, everything else transforms to `unknown`.
|
||||||
impl Display for Format {
|
impl Display for Format {
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
match self.0.as_ref() {
|
match self.0.as_ref() {
|
||||||
@ -22,21 +27,30 @@ impl Display for Format {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Wrapper around [`Book`](struct@calibre_db::data::book::Book) from the
|
||||||
|
/// [`calibre-db`](mod@calibre_db) crate. Adding
|
||||||
|
/// [`Author`](struct@calibre_db::data::author::Author),
|
||||||
|
/// [`Series`](struct@calibre_db::data::series::Series) as well as
|
||||||
|
/// [`Format`](struct@Format) information.
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct Book {
|
pub struct Book {
|
||||||
pub id: u64,
|
/// Book data from the database.
|
||||||
pub title: String,
|
pub data: DbBook,
|
||||||
pub sort: String,
|
/// Author information.
|
||||||
pub path: String,
|
|
||||||
pub uuid: String,
|
|
||||||
pub last_modified: OffsetDateTime,
|
|
||||||
pub description: Option<String>,
|
|
||||||
pub author: DbAuthor,
|
pub author: DbAuthor,
|
||||||
|
/// Series information. Series on the left and the index of the book within
|
||||||
|
/// that series on the right.
|
||||||
pub series: Option<(DbSeries, f64)>,
|
pub series: Option<(DbSeries, f64)>,
|
||||||
|
/// Format information.
|
||||||
pub formats: Formats,
|
pub formats: Formats,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Book {
|
impl Book {
|
||||||
|
/// Wrap a [`DbBook`](struct@calibre_db::data::book::Book) in a [`Book`](struct@Book) and
|
||||||
|
/// enrich it with additional information.
|
||||||
|
///
|
||||||
|
/// `db_series` is an Option tuple with the series on the left and the index of the book within
|
||||||
|
/// that series on the right.
|
||||||
pub fn from_db_book(
|
pub fn from_db_book(
|
||||||
db_book: &DbBook,
|
db_book: &DbBook,
|
||||||
db_series: Option<(DbSeries, f64)>,
|
db_series: Option<(DbSeries, f64)>,
|
||||||
@ -44,19 +58,15 @@ impl Book {
|
|||||||
formats: Formats,
|
formats: Formats,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
Self {
|
Self {
|
||||||
id: db_book.id,
|
data: db_book.clone(),
|
||||||
title: db_book.title.clone(),
|
|
||||||
sort: db_book.sort.clone(),
|
|
||||||
path: db_book.path.clone(),
|
|
||||||
uuid: db_book.uuid.clone(),
|
|
||||||
description: db_book.description.clone(),
|
|
||||||
last_modified: db_book.last_modified,
|
|
||||||
author: author.clone(),
|
author: author.clone(),
|
||||||
series: db_series.map(|x| (x.0, x.1)),
|
series: db_series.map(|x| (x.0, x.1)),
|
||||||
formats,
|
formats,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Find all pdf and epub formats of a book on disk. Search their directory in the library for
|
||||||
|
/// it.
|
||||||
fn formats(book: &DbBook, library_path: &Path) -> Formats {
|
fn formats(book: &DbBook, library_path: &Path) -> Formats {
|
||||||
let book_path = library_path.join(&book.path);
|
let book_path = library_path.join(&book.path);
|
||||||
let mut formats = HashMap::new();
|
let mut formats = HashMap::new();
|
||||||
@ -80,6 +90,8 @@ impl Book {
|
|||||||
formats
|
formats
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Wrap a [`DbBook`](struct@calibre_db::data::book::Book) in a [`Book`](struct@Book) by
|
||||||
|
/// fetching additional information about author, formats and series.
|
||||||
pub fn full_book(book: &DbBook, state: &AppState) -> Option<Book> {
|
pub fn full_book(book: &DbBook, state: &AppState) -> Option<Book> {
|
||||||
let formats = Book::formats(book, &state.config.library_path);
|
let formats = Book::formats(book, &state.config.library_path);
|
||||||
let author = state.calibre.book_author(book.id).ok()?;
|
let author = state.calibre.book_author(book.id).ok()?;
|
@ -1,22 +1,23 @@
|
|||||||
|
//! Handle requests for a single author.
|
||||||
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use calibre_db::data::pagination::SortOrder;
|
use calibre_db::data::pagination::SortOrder;
|
||||||
use poem::{
|
use poem::{
|
||||||
error::InternalServerError,
|
|
||||||
handler,
|
handler,
|
||||||
web::{Data, Html, Path},
|
web::{Data, Path},
|
||||||
};
|
Response,
|
||||||
use tera::Context;
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
app_state::AppState, data::book::Book, handlers::error::HandlerError, templates::TEMPLATES,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, data::book::Book, handlers::error::HandlerError, Accept};
|
||||||
|
|
||||||
|
/// Handle a request for an author with `id` and decide whether to render to html or OPDS.
|
||||||
#[handler]
|
#[handler]
|
||||||
pub async fn handler(
|
pub async fn handler(
|
||||||
id: Path<u64>,
|
id: Path<u64>,
|
||||||
|
accept: Data<&Accept>,
|
||||||
state: Data<&Arc<AppState>>,
|
state: Data<&Arc<AppState>>,
|
||||||
) -> Result<Html<String>, poem::Error> {
|
) -> Result<Response, poem::Error> {
|
||||||
let author = state
|
let author = state
|
||||||
.calibre
|
.calibre
|
||||||
.scalar_author(*id)
|
.scalar_author(*id)
|
||||||
@ -30,13 +31,8 @@ pub async fn handler(
|
|||||||
.filter_map(|x| Book::full_book(x, &state))
|
.filter_map(|x| Book::full_book(x, &state))
|
||||||
.collect::<Vec<Book>>();
|
.collect::<Vec<Book>>();
|
||||||
|
|
||||||
let mut context = Context::new();
|
match accept.0 {
|
||||||
context.insert("title", &author.name);
|
Accept::Html => crate::handlers::html::author::handler(author, books).await,
|
||||||
context.insert("nav", "authors");
|
Accept::Opds => crate::handlers::opds::author::handler(author, books).await,
|
||||||
context.insert("books", &books);
|
}
|
||||||
|
|
||||||
TEMPLATES
|
|
||||||
.render("book_list", &context)
|
|
||||||
.map_err(InternalServerError)
|
|
||||||
.map(Html)
|
|
||||||
}
|
}
|
44
little-hesinde/src/handlers/authors.rs
Normal file
44
little-hesinde/src/handlers/authors.rs
Normal file
@ -0,0 +1,44 @@
|
|||||||
|
//! Handle requests for multiple authors.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use calibre_db::{calibre::Calibre, data::pagination::SortOrder};
|
||||||
|
use poem::{
|
||||||
|
handler,
|
||||||
|
web::{Data, Path},
|
||||||
|
Response,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, Accept};
|
||||||
|
|
||||||
|
/// Handle a request for multiple authors, starting at the first.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler_init(
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
authors(&accept, &state.calibre, None, &SortOrder::ASC).await
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Handle a request for multiple authors, starting at the `cursor` and going in the direction of
|
||||||
|
/// `sort_order`.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
authors(&accept, &state.calibre, Some(&cursor), &sort_order).await
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn authors(
|
||||||
|
acccept: &Accept,
|
||||||
|
calibre: &Calibre,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
match acccept {
|
||||||
|
Accept::Html => crate::handlers::html::authors::handler(calibre, cursor, sort_order).await,
|
||||||
|
Accept::Opds => crate::handlers::opds::authors::handler(calibre, cursor, sort_order).await,
|
||||||
|
}
|
||||||
|
}
|
44
little-hesinde/src/handlers/books.rs
Normal file
44
little-hesinde/src/handlers/books.rs
Normal file
@ -0,0 +1,44 @@
|
|||||||
|
//! Handle requests for multiple books.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use calibre_db::data::pagination::SortOrder;
|
||||||
|
use poem::{
|
||||||
|
handler,
|
||||||
|
web::{Data, Path},
|
||||||
|
Response,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, Accept};
|
||||||
|
|
||||||
|
/// Handle a request for multiple books, starting at the first.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler_init(
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
books(&accept, &state, None, &SortOrder::ASC).await
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Handle a request for multiple books, starting at the `cursor` and going in the direction of
|
||||||
|
/// `sort_order`.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
books(&accept, &state, Some(&cursor), &sort_order).await
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn books(
|
||||||
|
accept: &Accept,
|
||||||
|
state: &Arc<AppState>,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
match accept {
|
||||||
|
Accept::Html => crate::handlers::html::books::handler(state, cursor, sort_order).await,
|
||||||
|
Accept::Opds => crate::handlers::opds::books::handler(state, cursor, sort_order).await,
|
||||||
|
}
|
||||||
|
}
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Handle requests for cover images.
|
||||||
|
|
||||||
use std::{fs::File, io::Read, sync::Arc};
|
use std::{fs::File, io::Read, sync::Arc};
|
||||||
|
|
||||||
use poem::{
|
use poem::{
|
||||||
@ -9,6 +11,7 @@ use poem::{
|
|||||||
|
|
||||||
use crate::{app_state::AppState, handlers::error::HandlerError};
|
use crate::{app_state::AppState, handlers::error::HandlerError};
|
||||||
|
|
||||||
|
/// Handle a request for the cover image of book with id `id`.
|
||||||
#[handler]
|
#[handler]
|
||||||
pub async fn handler(
|
pub async fn handler(
|
||||||
id: Path<u64>,
|
id: Path<u64>,
|
50
little-hesinde/src/handlers/download.rs
Normal file
50
little-hesinde/src/handlers/download.rs
Normal file
@ -0,0 +1,50 @@
|
|||||||
|
//! Handle requests for specific formats of a book.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use tokio::fs::File;
|
||||||
|
|
||||||
|
use poem::{
|
||||||
|
error::NotFoundError,
|
||||||
|
handler,
|
||||||
|
web::{Data, Path},
|
||||||
|
Body, IntoResponse, Response,
|
||||||
|
};
|
||||||
|
use tokio_util::io::ReaderStream;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
app_state::AppState,
|
||||||
|
data::book::{Book, Format},
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::media_type::MediaType,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Handle a request for a book with id `id` in format `format`.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
Path((id, format)): Path<(u64, String)>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let book = state
|
||||||
|
.calibre
|
||||||
|
.scalar_book(id)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
let book = Book::full_book(&book, &state).ok_or(NotFoundError)?;
|
||||||
|
let format = Format(format);
|
||||||
|
let file_name = book.formats.get(&format).ok_or(NotFoundError)?;
|
||||||
|
let file_path = state
|
||||||
|
.config
|
||||||
|
.library_path
|
||||||
|
.join(book.data.path)
|
||||||
|
.join(file_name);
|
||||||
|
|
||||||
|
let mut file = File::open(file_path).await.map_err(|_| NotFoundError)?;
|
||||||
|
let stream = ReaderStream::new(file);
|
||||||
|
let body = Body::from_bytes_stream(stream);
|
||||||
|
|
||||||
|
let content_type: MediaType = format.into();
|
||||||
|
Ok(body
|
||||||
|
.with_content_type(format!("{content_type}"))
|
||||||
|
.with_header("Content-Disposition", format!("filename=\"{file_name}\""))
|
||||||
|
.into_response())
|
||||||
|
}
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Error handling for requests handlers.
|
||||||
|
|
||||||
use calibre_db::data::error::DataStoreError;
|
use calibre_db::data::error::DataStoreError;
|
||||||
use poem::{error::ResponseError, http::StatusCode, Body, Response};
|
use poem::{error::ResponseError, http::StatusCode, Body, Response};
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
@ -6,15 +8,22 @@ use uuid::Uuid;
|
|||||||
|
|
||||||
use crate::opds::error::OpdsError;
|
use crate::opds::error::OpdsError;
|
||||||
|
|
||||||
|
/// Errors happening during handling of requests.
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
#[error("opds error")]
|
#[error("opds error")]
|
||||||
pub enum HandlerError {
|
pub enum HandlerError {
|
||||||
|
/// Error rendering OPDS.
|
||||||
#[error("opds error")]
|
#[error("opds error")]
|
||||||
OpdsError(#[from] OpdsError),
|
OpdsError(#[from] OpdsError),
|
||||||
|
/// Error fetching data from calibre.
|
||||||
#[error("data error")]
|
#[error("data error")]
|
||||||
DataError(#[from] DataStoreError),
|
DataError(#[from] DataStoreError),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert a [`HandlerError`](enum@HandlerError) into a suitable response error.
|
||||||
|
///
|
||||||
|
/// Log the real error (internal) with an uuid and send a suitable error message to the user with
|
||||||
|
/// the same uuid (for correlation purposes).
|
||||||
impl ResponseError for HandlerError {
|
impl ResponseError for HandlerError {
|
||||||
fn status(&self) -> StatusCode {
|
fn status(&self) -> StatusCode {
|
||||||
match &self {
|
match &self {
|
21
little-hesinde/src/handlers/html/author.rs
Normal file
21
little-hesinde/src/handlers/html/author.rs
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
//! Handle a single author for html.
|
||||||
|
|
||||||
|
use calibre_db::data::author::Author;
|
||||||
|
use poem::{error::InternalServerError, web::Html, IntoResponse, Response};
|
||||||
|
use tera::Context;
|
||||||
|
|
||||||
|
use crate::{data::book::Book, templates::TEMPLATES};
|
||||||
|
|
||||||
|
/// Render a single author in html.
|
||||||
|
pub async fn handler(author: Author, books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let mut context = Context::new();
|
||||||
|
context.insert("title", &author.name);
|
||||||
|
context.insert("nav", "authors");
|
||||||
|
context.insert("books", &books);
|
||||||
|
|
||||||
|
Ok(TEMPLATES
|
||||||
|
.render("book_list", &context)
|
||||||
|
.map_err(InternalServerError)
|
||||||
|
.map(Html)?
|
||||||
|
.into_response())
|
||||||
|
}
|
21
little-hesinde/src/handlers/html/authors.rs
Normal file
21
little-hesinde/src/handlers/html/authors.rs
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
//! Handle multiple authors in html.
|
||||||
|
|
||||||
|
use calibre_db::{calibre::Calibre, data::pagination::SortOrder};
|
||||||
|
use poem::Response;
|
||||||
|
|
||||||
|
use crate::handlers::paginated;
|
||||||
|
|
||||||
|
/// Render all authors paginated by cursor in html.
|
||||||
|
pub async fn handler(
|
||||||
|
calibre: &Calibre,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
paginated::render(
|
||||||
|
"authors",
|
||||||
|
|| calibre.authors(25, cursor, sort_order),
|
||||||
|
|author| author.sort.clone(),
|
||||||
|
|cursor| calibre.has_previous_authors(cursor),
|
||||||
|
|cursor| calibre.has_more_authors(cursor),
|
||||||
|
)
|
||||||
|
}
|
26
little-hesinde/src/handlers/html/books.rs
Normal file
26
little-hesinde/src/handlers/html/books.rs
Normal file
@ -0,0 +1,26 @@
|
|||||||
|
//! Handle multiple books in html.
|
||||||
|
|
||||||
|
use calibre_db::data::pagination::SortOrder;
|
||||||
|
use poem::Response;
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, data::book::Book, handlers::paginated};
|
||||||
|
|
||||||
|
/// Render all books paginated by cursor in html.
|
||||||
|
pub async fn handler(
|
||||||
|
state: &AppState,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
paginated::render(
|
||||||
|
"books",
|
||||||
|
|| {
|
||||||
|
state
|
||||||
|
.calibre
|
||||||
|
.books(25, cursor, sort_order)
|
||||||
|
.map(|x| x.iter().filter_map(|y| Book::full_book(y, state)).collect())
|
||||||
|
},
|
||||||
|
|book| book.data.sort.clone(),
|
||||||
|
|cursor| state.calibre.has_previous_books(cursor),
|
||||||
|
|cursor| state.calibre.has_more_books(cursor),
|
||||||
|
)
|
||||||
|
}
|
20
little-hesinde/src/handlers/html/recent.rs
Normal file
20
little-hesinde/src/handlers/html/recent.rs
Normal file
@ -0,0 +1,20 @@
|
|||||||
|
//! Handle recent books in html.
|
||||||
|
|
||||||
|
use poem::{error::InternalServerError, web::Html, IntoResponse, Response};
|
||||||
|
use tera::Context;
|
||||||
|
|
||||||
|
use crate::{data::book::Book, templates::TEMPLATES};
|
||||||
|
|
||||||
|
/// Render recent books as html.
|
||||||
|
pub async fn handler(recent_books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let mut context = Context::new();
|
||||||
|
context.insert("title", "Recent Books");
|
||||||
|
context.insert("nav", "recent");
|
||||||
|
context.insert("books", &recent_books);
|
||||||
|
|
||||||
|
Ok(TEMPLATES
|
||||||
|
.render("book_list", &context)
|
||||||
|
.map_err(InternalServerError)
|
||||||
|
.map(Html)?
|
||||||
|
.into_response())
|
||||||
|
}
|
21
little-hesinde/src/handlers/html/series.rs
Normal file
21
little-hesinde/src/handlers/html/series.rs
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
//! Handle multiple series in html.
|
||||||
|
|
||||||
|
use calibre_db::{calibre::Calibre, data::pagination::SortOrder};
|
||||||
|
use poem::Response;
|
||||||
|
|
||||||
|
use crate::handlers::paginated;
|
||||||
|
|
||||||
|
/// Render all series paginated by cursor as html.
|
||||||
|
pub async fn handler(
|
||||||
|
calibre: &Calibre,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
paginated::render(
|
||||||
|
"series",
|
||||||
|
|| calibre.series(25, cursor, sort_order),
|
||||||
|
|series| series.sort.clone(),
|
||||||
|
|cursor| calibre.has_previous_series(cursor),
|
||||||
|
|cursor| calibre.has_more_series(cursor),
|
||||||
|
)
|
||||||
|
}
|
21
little-hesinde/src/handlers/html/series_single.rs
Normal file
21
little-hesinde/src/handlers/html/series_single.rs
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
//! Handle a single series in html.
|
||||||
|
|
||||||
|
use calibre_db::data::series::Series;
|
||||||
|
use poem::{error::InternalServerError, web::Html, IntoResponse, Response};
|
||||||
|
use tera::Context;
|
||||||
|
|
||||||
|
use crate::{data::book::Book, templates::TEMPLATES};
|
||||||
|
|
||||||
|
/// Render a single series as html.
|
||||||
|
pub async fn handler(series: Series, books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let mut context = Context::new();
|
||||||
|
context.insert("title", &series.name);
|
||||||
|
context.insert("nav", "series");
|
||||||
|
context.insert("books", &books);
|
||||||
|
|
||||||
|
Ok(TEMPLATES
|
||||||
|
.render("book_list", &context)
|
||||||
|
.map_err(InternalServerError)
|
||||||
|
.map(Html)?
|
||||||
|
.into_response())
|
||||||
|
}
|
38
little-hesinde/src/handlers/opds/author.rs
Normal file
38
little-hesinde/src/handlers/opds/author.rs
Normal file
@ -0,0 +1,38 @@
|
|||||||
|
//! Handle a single author for opds.
|
||||||
|
|
||||||
|
use calibre_db::data::author::Author;
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::book::Book,
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render a single author as an OPDS entry embedded in a feed.
|
||||||
|
pub async fn handler(author: Author, books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: format!("/opds/authors/{}", author.id),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
&format!("little-hesinde:author:{}", author.id),
|
||||||
|
&author.name,
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
48
little-hesinde/src/handlers/opds/authors.rs
Normal file
48
little-hesinde/src/handlers/opds/authors.rs
Normal file
@ -0,0 +1,48 @@
|
|||||||
|
//! Handle multiple authors for opds.
|
||||||
|
|
||||||
|
use calibre_db::{
|
||||||
|
calibre::Calibre,
|
||||||
|
data::{author::Author as DbAuthor, pagination::SortOrder},
|
||||||
|
};
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render all authors as OPDS entries embedded in a feed.
|
||||||
|
pub async fn handler(
|
||||||
|
calibre: &Calibre,
|
||||||
|
_cursor: Option<&str>,
|
||||||
|
_sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let authors: Vec<DbAuthor> = calibre
|
||||||
|
.authors(u32::MAX.into(), None, &SortOrder::ASC)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
|
||||||
|
let entries: Vec<Entry> = authors.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: "/opds/authors".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
"little-hesinde:authors",
|
||||||
|
"All Authors",
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
49
little-hesinde/src/handlers/opds/books.rs
Normal file
49
little-hesinde/src/handlers/opds/books.rs
Normal file
@ -0,0 +1,49 @@
|
|||||||
|
//! Handle multiple books for opds.
|
||||||
|
|
||||||
|
use calibre_db::data::pagination::SortOrder;
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
app_state::AppState,
|
||||||
|
data::book::Book,
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render all books as OPDS entries embedded in a feed.
|
||||||
|
pub async fn handler(
|
||||||
|
state: &AppState,
|
||||||
|
_cursor: Option<&str>,
|
||||||
|
_sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let books: Vec<Book> = state
|
||||||
|
.calibre
|
||||||
|
.books(u32::MAX.into(), None, &SortOrder::ASC)
|
||||||
|
.map(|x| x.iter().filter_map(|y| Book::full_book(y, state)).collect())
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
|
||||||
|
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: "/opds/books".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
"little-hesinde:books",
|
||||||
|
"All Books",
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
109
little-hesinde/src/handlers/opds/feed.rs
Normal file
109
little-hesinde/src/handlers/opds/feed.rs
Normal file
@ -0,0 +1,109 @@
|
|||||||
|
//! Handle the OPDS root feed.
|
||||||
|
|
||||||
|
use poem::{handler, web::WithContentType, IntoResponse};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{
|
||||||
|
content::Content, entry::Entry, feed::Feed, link::Link, media_type::MediaType,
|
||||||
|
relation::Relation,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render a root OPDS feed with links to the subsections (authors, books, series and recent).
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler() -> Result<WithContentType<String>, poem::Error> {
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: "/opds".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let books_entry = Entry {
|
||||||
|
title: "Books".to_string(),
|
||||||
|
id: "little-hesinde:books".to_string(),
|
||||||
|
updated: now,
|
||||||
|
content: Some(Content {
|
||||||
|
media_type: MediaType::Text,
|
||||||
|
content: "Index of all books".to_string(),
|
||||||
|
}),
|
||||||
|
author: None,
|
||||||
|
links: vec![Link {
|
||||||
|
href: "/opds/books".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Subsection,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
}],
|
||||||
|
};
|
||||||
|
|
||||||
|
let authors_entry = Entry {
|
||||||
|
title: "Authors".to_string(),
|
||||||
|
id: "little-hesinde:authors".to_string(),
|
||||||
|
updated: now,
|
||||||
|
content: Some(Content {
|
||||||
|
media_type: MediaType::Text,
|
||||||
|
content: "Index of all authors".to_string(),
|
||||||
|
}),
|
||||||
|
author: None,
|
||||||
|
links: vec![Link {
|
||||||
|
href: "/opds/authors".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Subsection,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
}],
|
||||||
|
};
|
||||||
|
|
||||||
|
let series_entry = Entry {
|
||||||
|
title: "Series".to_string(),
|
||||||
|
id: "little-hesinde:series".to_string(),
|
||||||
|
updated: now,
|
||||||
|
content: Some(Content {
|
||||||
|
media_type: MediaType::Text,
|
||||||
|
content: "Index of all series".to_string(),
|
||||||
|
}),
|
||||||
|
author: None,
|
||||||
|
links: vec![Link {
|
||||||
|
href: "/opds/series".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Subsection,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
}],
|
||||||
|
};
|
||||||
|
|
||||||
|
let recents_entry = Entry {
|
||||||
|
title: "Recent Additions".to_string(),
|
||||||
|
id: "little-hesinde:recentbooks".to_string(),
|
||||||
|
updated: now,
|
||||||
|
content: Some(Content {
|
||||||
|
media_type: MediaType::Text,
|
||||||
|
content: "Recently added books".to_string(),
|
||||||
|
}),
|
||||||
|
author: None,
|
||||||
|
links: vec![Link {
|
||||||
|
href: "/opds/recent".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Subsection,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
}],
|
||||||
|
};
|
||||||
|
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
"little-hesinde:catalog",
|
||||||
|
"Little Hesinde",
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
vec![authors_entry, series_entry, books_entry, recents_entry],
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml.with_content_type("application/atom+xml"))
|
||||||
|
}
|
37
little-hesinde/src/handlers/opds/recent.rs
Normal file
37
little-hesinde/src/handlers/opds/recent.rs
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
//! Handle recent books for OPDS.
|
||||||
|
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::book::Book,
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render recent books as OPDS entries embedded in a feed.
|
||||||
|
pub async fn handler(recent_books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let entries: Vec<Entry> = recent_books.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: "/opds/recent".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
"little-hesinde:recentbooks",
|
||||||
|
"Recent Books",
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
45
little-hesinde/src/handlers/opds/series.rs
Normal file
45
little-hesinde/src/handlers/opds/series.rs
Normal file
@ -0,0 +1,45 @@
|
|||||||
|
//! Handle multiple series for OPDS.
|
||||||
|
|
||||||
|
use calibre_db::{calibre::Calibre, data::pagination::SortOrder};
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render all series as OPDS entries embedded in a feed.
|
||||||
|
pub async fn handler(
|
||||||
|
calibre: &Calibre,
|
||||||
|
_cursor: Option<&str>,
|
||||||
|
_sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let series = calibre
|
||||||
|
.series(u32::MAX.into(), None, &SortOrder::ASC)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
|
||||||
|
let entries: Vec<Entry> = series.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: "/opds/series".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
"little-hesinde:series",
|
||||||
|
"All Series",
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
38
little-hesinde/src/handlers/opds/series_single.rs
Normal file
38
little-hesinde/src/handlers/opds/series_single.rs
Normal file
@ -0,0 +1,38 @@
|
|||||||
|
//! Handle a single series for opds.
|
||||||
|
|
||||||
|
use calibre_db::data::series::Series;
|
||||||
|
use poem::{IntoResponse, Response};
|
||||||
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::book::Book,
|
||||||
|
handlers::error::HandlerError,
|
||||||
|
opds::{entry::Entry, feed::Feed, link::Link, media_type::MediaType, relation::Relation},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Render a single series as an OPDS entry embedded in a feed.
|
||||||
|
pub async fn handler(series: Series, books: Vec<Book>) -> Result<Response, poem::Error> {
|
||||||
|
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
||||||
|
let now = OffsetDateTime::now_utc();
|
||||||
|
|
||||||
|
let self_link = Link {
|
||||||
|
href: format!("/opds/series/{}", series.id),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Myself,
|
||||||
|
title: None,
|
||||||
|
count: None,
|
||||||
|
};
|
||||||
|
let feed = Feed::create(
|
||||||
|
now,
|
||||||
|
&format!("little-hesinde:series:{}", series.id),
|
||||||
|
&series.name,
|
||||||
|
self_link,
|
||||||
|
vec![],
|
||||||
|
entries,
|
||||||
|
);
|
||||||
|
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
||||||
|
|
||||||
|
Ok(xml
|
||||||
|
.with_content_type("application/atom+xml")
|
||||||
|
.into_response())
|
||||||
|
}
|
@ -1,7 +1,9 @@
|
|||||||
|
//! Deal with cursor pagination.
|
||||||
|
|
||||||
use std::fmt::Debug;
|
use std::fmt::Debug;
|
||||||
|
|
||||||
use calibre_db::data::error::DataStoreError;
|
use calibre_db::data::error::DataStoreError;
|
||||||
use poem::{error::InternalServerError, web::Html};
|
use poem::{error::InternalServerError, web::Html, IntoResponse, Response};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use tera::Context;
|
use tera::Context;
|
||||||
|
|
||||||
@ -9,13 +11,14 @@ use crate::templates::TEMPLATES;
|
|||||||
|
|
||||||
use super::error::HandlerError;
|
use super::error::HandlerError;
|
||||||
|
|
||||||
|
/// Render a tera template with paginated items and generate back and forth links.
|
||||||
pub fn render<T: Serialize + Debug, F, S, P, M>(
|
pub fn render<T: Serialize + Debug, F, S, P, M>(
|
||||||
template: &str,
|
template: &str,
|
||||||
fetcher: F,
|
fetcher: F,
|
||||||
sort_field: S,
|
sort_field: S,
|
||||||
has_previous: P,
|
has_previous: P,
|
||||||
has_more: M,
|
has_more: M,
|
||||||
) -> Result<Html<String>, poem::Error>
|
) -> Result<Response, poem::Error>
|
||||||
where
|
where
|
||||||
F: Fn() -> Result<Vec<T>, DataStoreError>,
|
F: Fn() -> Result<Vec<T>, DataStoreError>,
|
||||||
S: Fn(&T) -> String,
|
S: Fn(&T) -> String,
|
||||||
@ -42,8 +45,9 @@ where
|
|||||||
context.insert("nav", template);
|
context.insert("nav", template);
|
||||||
context.insert(template, &items);
|
context.insert(template, &items);
|
||||||
|
|
||||||
TEMPLATES
|
Ok(TEMPLATES
|
||||||
.render(template, &context)
|
.render(template, &context)
|
||||||
.map_err(InternalServerError)
|
.map_err(InternalServerError)
|
||||||
.map(Html)
|
.map(Html)?
|
||||||
|
.into_response())
|
||||||
}
|
}
|
28
little-hesinde/src/handlers/recent.rs
Normal file
28
little-hesinde/src/handlers/recent.rs
Normal file
@ -0,0 +1,28 @@
|
|||||||
|
//! Handle requests for recent books.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use poem::{handler, web::Data, Response};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, data::book::Book, handlers::error::HandlerError, Accept};
|
||||||
|
|
||||||
|
/// Handle a request recent books and decide whether to render to html or OPDS.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let recent_books = state
|
||||||
|
.calibre
|
||||||
|
.recent_books(25)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
let recent_books = recent_books
|
||||||
|
.iter()
|
||||||
|
.filter_map(|x| Book::full_book(x, &state))
|
||||||
|
.collect::<Vec<Book>>();
|
||||||
|
|
||||||
|
match accept.0 {
|
||||||
|
Accept::Html => crate::handlers::html::recent::handler(recent_books).await,
|
||||||
|
Accept::Opds => crate::handlers::opds::recent::handler(recent_books).await,
|
||||||
|
}
|
||||||
|
}
|
48
little-hesinde/src/handlers/series.rs
Normal file
48
little-hesinde/src/handlers/series.rs
Normal file
@ -0,0 +1,48 @@
|
|||||||
|
//! Handle requests for multiple series.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use calibre_db::data::pagination::SortOrder;
|
||||||
|
use poem::{
|
||||||
|
handler,
|
||||||
|
web::{Data, Path},
|
||||||
|
Response,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, Accept};
|
||||||
|
|
||||||
|
/// Handle a request for multiple series, starting at the first.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler_init(
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
series(&accept, &state, None, &SortOrder::ASC).await
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Handle a request for multiple series, starting at the `cursor` and going in the direction of
|
||||||
|
/// `sort_order`.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
series(&accept, &state, Some(&cursor), &sort_order).await
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn series(
|
||||||
|
accept: &Accept,
|
||||||
|
state: &Arc<AppState>,
|
||||||
|
cursor: Option<&str>,
|
||||||
|
sort_order: &SortOrder,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
match accept {
|
||||||
|
Accept::Html => {
|
||||||
|
crate::handlers::html::series::handler(&state.calibre, cursor, sort_order).await
|
||||||
|
}
|
||||||
|
Accept::Opds => {
|
||||||
|
crate::handlers::opds::series::handler(&state.calibre, cursor, sort_order).await
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
37
little-hesinde/src/handlers/series_single.rs
Normal file
37
little-hesinde/src/handlers/series_single.rs
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
//! Handle requests for a single series.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use poem::{
|
||||||
|
handler,
|
||||||
|
web::{Data, Path},
|
||||||
|
Response,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::{app_state::AppState, data::book::Book, handlers::error::HandlerError, Accept};
|
||||||
|
|
||||||
|
/// Handle a request for a series with `id` and decide whether to render to html or OPDS.
|
||||||
|
#[handler]
|
||||||
|
pub async fn handler(
|
||||||
|
id: Path<u64>,
|
||||||
|
accept: Data<&Accept>,
|
||||||
|
state: Data<&Arc<AppState>>,
|
||||||
|
) -> Result<Response, poem::Error> {
|
||||||
|
let series = state
|
||||||
|
.calibre
|
||||||
|
.scalar_series(*id)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
let books = state
|
||||||
|
.calibre
|
||||||
|
.series_books(*id)
|
||||||
|
.map_err(HandlerError::DataError)?;
|
||||||
|
let books = books
|
||||||
|
.iter()
|
||||||
|
.filter_map(|x| Book::full_book(x, &state))
|
||||||
|
.collect::<Vec<Book>>();
|
||||||
|
|
||||||
|
match accept.0 {
|
||||||
|
Accept::Html => crate::handlers::html::series_single::handler(series, books).await,
|
||||||
|
Accept::Opds => crate::handlers::opds::series_single::handler(series, books).await,
|
||||||
|
}
|
||||||
|
}
|
139
little-hesinde/src/lib.rs
Normal file
139
little-hesinde/src/lib.rs
Normal file
@ -0,0 +1,139 @@
|
|||||||
|
//! A very simple ebook server for a calibre library, providing a html interface as well as an OPDS feed.
|
||||||
|
//!
|
||||||
|
//! Shamelessly written to scratch my own itches.
|
||||||
|
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use app_state::AppState;
|
||||||
|
use calibre_db::calibre::Calibre;
|
||||||
|
use config::Config;
|
||||||
|
use poem::{
|
||||||
|
endpoint::EmbeddedFilesEndpoint, get, listener::TcpListener, middleware::Tracing, EndpointExt,
|
||||||
|
Route, Server,
|
||||||
|
};
|
||||||
|
use rust_embed::RustEmbed;
|
||||||
|
use tokio::signal;
|
||||||
|
use tracing::info;
|
||||||
|
|
||||||
|
pub mod app_state;
|
||||||
|
pub mod cli;
|
||||||
|
pub mod config;
|
||||||
|
/// Data structs and their functions.
|
||||||
|
pub mod data {
|
||||||
|
pub mod book;
|
||||||
|
}
|
||||||
|
/// Request handlers. Because it can not be guaranteed that a proper accept header is sent, the
|
||||||
|
/// routes are doubled and the decision on whether to render html or OPDS is made with internal
|
||||||
|
/// data on the respective routes.
|
||||||
|
pub mod handlers {
|
||||||
|
/// Handle requests for html.
|
||||||
|
pub mod html {
|
||||||
|
pub mod author;
|
||||||
|
pub mod authors;
|
||||||
|
pub mod books;
|
||||||
|
pub mod recent;
|
||||||
|
pub mod series;
|
||||||
|
pub mod series_single;
|
||||||
|
}
|
||||||
|
/// Handle requests for OPDS.
|
||||||
|
pub mod opds {
|
||||||
|
pub mod author;
|
||||||
|
pub mod authors;
|
||||||
|
pub mod books;
|
||||||
|
pub mod feed;
|
||||||
|
pub mod recent;
|
||||||
|
pub mod series;
|
||||||
|
pub mod series_single;
|
||||||
|
}
|
||||||
|
pub mod author;
|
||||||
|
pub mod authors;
|
||||||
|
pub mod books;
|
||||||
|
pub mod cover;
|
||||||
|
pub mod download;
|
||||||
|
pub mod error;
|
||||||
|
pub mod paginated;
|
||||||
|
pub mod recent;
|
||||||
|
pub mod series;
|
||||||
|
pub mod series_single;
|
||||||
|
}
|
||||||
|
/// OPDS data structs.
|
||||||
|
pub mod opds {
|
||||||
|
pub mod author;
|
||||||
|
pub mod content;
|
||||||
|
pub mod entry;
|
||||||
|
pub mod error;
|
||||||
|
pub mod feed;
|
||||||
|
pub mod link;
|
||||||
|
pub mod media_type;
|
||||||
|
pub mod relation;
|
||||||
|
}
|
||||||
|
pub mod templates;
|
||||||
|
|
||||||
|
/// Internal marker data in lieu of a proper `Accept` header.
|
||||||
|
#[derive(Debug, Clone, Copy)]
|
||||||
|
pub enum Accept {
|
||||||
|
/// Render as html.
|
||||||
|
Html,
|
||||||
|
/// Render as OPDS.
|
||||||
|
Opds,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Embedd static files.
|
||||||
|
#[derive(RustEmbed)]
|
||||||
|
#[folder = "static"]
|
||||||
|
pub struct Files;
|
||||||
|
|
||||||
|
/// Main entry point to run the ebook server with a calibre library specified in `config`.
|
||||||
|
pub async fn run(config: Config) -> Result<(), std::io::Error> {
|
||||||
|
let calibre = Calibre::load(&config.metadata_path).expect("failed to load calibre database");
|
||||||
|
let app_state = Arc::new(AppState { calibre, config });
|
||||||
|
|
||||||
|
let html_routes = Route::new()
|
||||||
|
.at("/", get(handlers::recent::handler))
|
||||||
|
.at("/books", get(handlers::books::handler_init))
|
||||||
|
.at("/books/:cursor/:sort_order", get(handlers::books::handler))
|
||||||
|
.at("/series", get(handlers::series::handler_init))
|
||||||
|
.at(
|
||||||
|
"/series/:cursor/:sort_order",
|
||||||
|
get(handlers::series::handler),
|
||||||
|
)
|
||||||
|
.at("/series/:id", get(handlers::series_single::handler))
|
||||||
|
.at("/authors", get(handlers::authors::handler_init))
|
||||||
|
.at("/authors/:id", get(handlers::author::handler))
|
||||||
|
.at(
|
||||||
|
"/authors/:cursor/:sort_order",
|
||||||
|
get(handlers::authors::handler),
|
||||||
|
)
|
||||||
|
.at("/cover/:id", get(handlers::cover::handler))
|
||||||
|
.at("/book/:id/:format", get(handlers::download::handler))
|
||||||
|
.nest("/static", EmbeddedFilesEndpoint::<Files>::new())
|
||||||
|
.data(Accept::Html);
|
||||||
|
|
||||||
|
let opds_routes = Route::new()
|
||||||
|
.at("/", get(handlers::opds::feed::handler))
|
||||||
|
.at("/recent", get(handlers::recent::handler))
|
||||||
|
.at("/books", get(handlers::books::handler_init))
|
||||||
|
.at("/authors", get(handlers::authors::handler_init))
|
||||||
|
.at("/authors/:id", get(handlers::author::handler))
|
||||||
|
.at("/series", get(handlers::series::handler_init))
|
||||||
|
.at("/series/:id", get(handlers::series_single::handler))
|
||||||
|
.data(Accept::Opds);
|
||||||
|
|
||||||
|
let app = Route::new()
|
||||||
|
.nest("/", html_routes)
|
||||||
|
.nest("/opds", opds_routes)
|
||||||
|
.data(app_state)
|
||||||
|
.with(Tracing);
|
||||||
|
|
||||||
|
let server = Server::new(TcpListener::bind("[::]:3000"))
|
||||||
|
.name("cops-web")
|
||||||
|
.run(app);
|
||||||
|
|
||||||
|
tokio::select! {
|
||||||
|
_ = server => {},
|
||||||
|
_ = signal::ctrl_c() => {
|
||||||
|
info!("Received Ctrl+C, shutting down...");
|
||||||
|
},
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
@ -1,5 +1,5 @@
|
|||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use rusty_library::{cli::Cli, config::Config};
|
use little_hesinde::{cli::Cli, config::Config};
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() -> Result<(), std::io::Error> {
|
async fn main() -> Result<(), std::io::Error> {
|
||||||
@ -11,5 +11,5 @@ async fn main() -> Result<(), std::io::Error> {
|
|||||||
let args = Cli::parse();
|
let args = Cli::parse();
|
||||||
let config = Config::load(&args).expect("failed to load configuration");
|
let config = Config::load(&args).expect("failed to load configuration");
|
||||||
|
|
||||||
rusty_library::run(config).await
|
little_hesinde::run(config).await
|
||||||
}
|
}
|
@ -1,10 +1,16 @@
|
|||||||
|
//! Author data.
|
||||||
|
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
|
/// Author information.
|
||||||
#[derive(Debug, Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
#[serde(rename = "author")]
|
#[serde(rename = "author")]
|
||||||
pub struct Author {
|
pub struct Author {
|
||||||
|
/// Full name.
|
||||||
pub name: String,
|
pub name: String,
|
||||||
|
/// Where to find the author.
|
||||||
pub uri: String,
|
pub uri: String,
|
||||||
|
/// Optional email address.
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub email: Option<String>,
|
pub email: Option<String>,
|
||||||
}
|
}
|
@ -1,12 +1,17 @@
|
|||||||
|
//! Content data.
|
||||||
|
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
use super::media_type::MediaType;
|
use super::media_type::MediaType;
|
||||||
|
|
||||||
|
/// Content of different types, used for example for description of an entry.
|
||||||
#[derive(Debug, Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
#[serde(rename = "content")]
|
#[serde(rename = "content")]
|
||||||
pub struct Content {
|
pub struct Content {
|
||||||
|
/// Media type of this content.
|
||||||
#[serde(rename = "@type")]
|
#[serde(rename = "@type")]
|
||||||
pub media_type: MediaType,
|
pub media_type: MediaType,
|
||||||
|
/// Actual content.
|
||||||
#[serde(rename = "$value")]
|
#[serde(rename = "$value")]
|
||||||
pub content: String,
|
pub content: String,
|
||||||
}
|
}
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Entry data.
|
||||||
|
|
||||||
use calibre_db::data::{author::Author as DbAuthor, series::Series};
|
use calibre_db::data::{author::Author as DbAuthor, series::Series};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
@ -8,20 +10,32 @@ use super::{
|
|||||||
author::Author, content::Content, link::Link, media_type::MediaType, relation::Relation,
|
author::Author, content::Content, link::Link, media_type::MediaType, relation::Relation,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/// Fundamental piece of OPDS, holding information about entries (for example a book).
|
||||||
#[derive(Debug, Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
#[serde(rename = "entry")]
|
#[serde(rename = "entry")]
|
||||||
pub struct Entry {
|
pub struct Entry {
|
||||||
|
/// Title of the entry.
|
||||||
pub title: String,
|
pub title: String,
|
||||||
|
/// Id, for example a uuid.
|
||||||
pub id: String,
|
pub id: String,
|
||||||
|
/// When was this entry updated last.
|
||||||
#[serde(with = "time::serde::rfc3339")]
|
#[serde(with = "time::serde::rfc3339")]
|
||||||
pub updated: OffsetDateTime,
|
pub updated: OffsetDateTime,
|
||||||
|
/// Optional content.
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
/// Optional author information.
|
||||||
pub content: Option<Content>,
|
pub content: Option<Content>,
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub author: Option<Author>,
|
pub author: Option<Author>,
|
||||||
|
/// List of links, for example to download an entry.
|
||||||
#[serde(rename = "link")]
|
#[serde(rename = "link")]
|
||||||
pub links: Vec<Link>,
|
pub links: Vec<Link>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert a book into an OPDS entry.
|
||||||
|
///
|
||||||
|
/// Add the cover and formats as link with a proper media type.
|
||||||
|
/// Add author and content information.
|
||||||
impl From<Book> for Entry {
|
impl From<Book> for Entry {
|
||||||
fn from(value: Book) -> Self {
|
fn from(value: Book) -> Self {
|
||||||
let author = Author {
|
let author = Author {
|
||||||
@ -30,7 +44,7 @@ impl From<Book> for Entry {
|
|||||||
email: None,
|
email: None,
|
||||||
};
|
};
|
||||||
let mut links = vec![Link {
|
let mut links = vec![Link {
|
||||||
href: format!("/cover/{}", value.id),
|
href: format!("/cover/{}", value.data.id),
|
||||||
media_type: MediaType::Jpeg,
|
media_type: MediaType::Jpeg,
|
||||||
rel: Relation::Image,
|
rel: Relation::Image,
|
||||||
title: None,
|
title: None,
|
||||||
@ -43,15 +57,15 @@ impl From<Book> for Entry {
|
|||||||
.collect();
|
.collect();
|
||||||
links.append(&mut format_links);
|
links.append(&mut format_links);
|
||||||
|
|
||||||
let content = value.description.map(|desc| Content {
|
let content = value.data.description.map(|desc| Content {
|
||||||
media_type: MediaType::Html,
|
media_type: MediaType::Html,
|
||||||
content: desc,
|
content: desc,
|
||||||
});
|
});
|
||||||
|
|
||||||
Self {
|
Self {
|
||||||
title: value.title.clone(),
|
title: value.data.title.clone(),
|
||||||
id: format!("urn:uuid:{}", value.uuid),
|
id: format!("urn:uuid:{}", value.data.uuid),
|
||||||
updated: value.last_modified,
|
updated: value.data.last_modified,
|
||||||
content,
|
content,
|
||||||
author: Some(author),
|
author: Some(author),
|
||||||
links,
|
links,
|
||||||
@ -59,6 +73,9 @@ impl From<Book> for Entry {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert author information into an OPDS entry.
|
||||||
|
///
|
||||||
|
/// Add the author link.
|
||||||
impl From<DbAuthor> for Entry {
|
impl From<DbAuthor> for Entry {
|
||||||
fn from(value: DbAuthor) -> Self {
|
fn from(value: DbAuthor) -> Self {
|
||||||
let links = vec![Link {
|
let links = vec![Link {
|
||||||
@ -71,7 +88,7 @@ impl From<DbAuthor> for Entry {
|
|||||||
|
|
||||||
Self {
|
Self {
|
||||||
title: value.name.clone(),
|
title: value.name.clone(),
|
||||||
id: format!("rusty:authors:{}", value.id),
|
id: format!("little-hesinde:authors:{}", value.id),
|
||||||
updated: OffsetDateTime::now_utc(),
|
updated: OffsetDateTime::now_utc(),
|
||||||
content: None,
|
content: None,
|
||||||
author: None,
|
author: None,
|
||||||
@ -80,6 +97,9 @@ impl From<DbAuthor> for Entry {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert series information into an OPDS entry.
|
||||||
|
///
|
||||||
|
/// Add the series link.
|
||||||
impl From<Series> for Entry {
|
impl From<Series> for Entry {
|
||||||
fn from(value: Series) -> Self {
|
fn from(value: Series) -> Self {
|
||||||
let links = vec![Link {
|
let links = vec![Link {
|
||||||
@ -92,7 +112,7 @@ impl From<Series> for Entry {
|
|||||||
|
|
||||||
Self {
|
Self {
|
||||||
title: value.name.clone(),
|
title: value.name.clone(),
|
||||||
id: format!("rusty:series:{}", value.id),
|
id: format!("little-hesinde:series:{}", value.id),
|
||||||
updated: OffsetDateTime::now_utc(),
|
updated: OffsetDateTime::now_utc(),
|
||||||
content: None,
|
content: None,
|
||||||
author: None,
|
author: None,
|
@ -1,15 +1,21 @@
|
|||||||
|
//! Error handling for OPDS data.
|
||||||
|
|
||||||
use std::string::FromUtf8Error;
|
use std::string::FromUtf8Error;
|
||||||
|
|
||||||
use quick_xml::DeError;
|
use quick_xml::DeError;
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
|
|
||||||
|
/// Errors happening during handling OPDS data.
|
||||||
#[derive(Error, Debug)]
|
#[derive(Error, Debug)]
|
||||||
#[error("opds error")]
|
#[error("opds error")]
|
||||||
pub enum OpdsError {
|
pub enum OpdsError {
|
||||||
|
/// Error serializing OPDS data.
|
||||||
#[error("failed to serialize struct")]
|
#[error("failed to serialize struct")]
|
||||||
SerializingError(#[from] DeError),
|
SerializingError(#[from] DeError),
|
||||||
|
/// Error parsing OPDS xml structure.
|
||||||
#[error("xml failure")]
|
#[error("xml failure")]
|
||||||
XmlError(#[from] quick_xml::Error),
|
XmlError(#[from] quick_xml::Error),
|
||||||
|
/// Error decoding xml as UTF-8.
|
||||||
#[error("failed to decode as utf-8")]
|
#[error("failed to decode as utf-8")]
|
||||||
Utf8Error(#[from] FromUtf8Error),
|
Utf8Error(#[from] FromUtf8Error),
|
||||||
}
|
}
|
@ -1,3 +1,5 @@
|
|||||||
|
//! Root feed data.
|
||||||
|
|
||||||
use std::io::Cursor;
|
use std::io::Cursor;
|
||||||
|
|
||||||
use quick_xml::{
|
use quick_xml::{
|
||||||
@ -8,24 +10,73 @@ use quick_xml::{
|
|||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
use super::{author::Author, entry::Entry, error::OpdsError, link::Link};
|
use super::{
|
||||||
|
author::Author, entry::Entry, error::OpdsError, link::Link, media_type::MediaType,
|
||||||
|
relation::Relation,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Root feed element of OPDS.
|
||||||
#[derive(Debug, Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
#[serde(rename = "feed")]
|
#[serde(rename = "feed")]
|
||||||
pub struct Feed {
|
pub struct Feed {
|
||||||
|
/// Title, often shown in OPDS clients.
|
||||||
pub title: String,
|
pub title: String,
|
||||||
|
/// Feed id.
|
||||||
pub id: String,
|
pub id: String,
|
||||||
|
/// When was the feed updated last.
|
||||||
#[serde(with = "time::serde::rfc3339")]
|
#[serde(with = "time::serde::rfc3339")]
|
||||||
pub updated: OffsetDateTime,
|
pub updated: OffsetDateTime,
|
||||||
|
/// Icon for the feed.
|
||||||
pub icon: String,
|
pub icon: String,
|
||||||
|
/// Feed author.
|
||||||
pub author: Author,
|
pub author: Author,
|
||||||
|
/// Links, for example home or self.
|
||||||
#[serde(rename = "link")]
|
#[serde(rename = "link")]
|
||||||
pub links: Vec<Link>,
|
pub links: Vec<Link>,
|
||||||
|
/// Entries inside the feed (books, series, subsections, ...)
|
||||||
#[serde(rename = "entry")]
|
#[serde(rename = "entry")]
|
||||||
pub entries: Vec<Entry>,
|
pub entries: Vec<Entry>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Feed {
|
impl Feed {
|
||||||
|
/// Create a feed with the specified data.
|
||||||
|
pub fn create(
|
||||||
|
now: OffsetDateTime,
|
||||||
|
id: &str,
|
||||||
|
title: &str,
|
||||||
|
self_link: Link,
|
||||||
|
mut additional_links: Vec<Link>,
|
||||||
|
entries: Vec<Entry>,
|
||||||
|
) -> Self {
|
||||||
|
let author = Author {
|
||||||
|
name: "Thallian".to_string(),
|
||||||
|
uri: "https://code.vanwa.ch/shu/little-hesinde".to_string(),
|
||||||
|
email: None,
|
||||||
|
};
|
||||||
|
let mut links = vec![
|
||||||
|
Link {
|
||||||
|
href: "/opds".to_string(),
|
||||||
|
media_type: MediaType::Navigation,
|
||||||
|
rel: Relation::Start,
|
||||||
|
title: Some("Home".to_string()),
|
||||||
|
count: None,
|
||||||
|
},
|
||||||
|
self_link,
|
||||||
|
];
|
||||||
|
links.append(&mut additional_links);
|
||||||
|
|
||||||
|
Self {
|
||||||
|
title: title.to_string(),
|
||||||
|
id: id.to_string(),
|
||||||
|
updated: now,
|
||||||
|
icon: "favicon.ico".to_string(),
|
||||||
|
author,
|
||||||
|
links,
|
||||||
|
entries,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Serialize a feed to OPDS xml.
|
||||||
pub fn as_xml(&self) -> Result<String, OpdsError> {
|
pub fn as_xml(&self) -> Result<String, OpdsError> {
|
||||||
let xml = to_string(&self)?;
|
let xml = to_string(&self)?;
|
||||||
let mut reader = Reader::from_str(&xml);
|
let mut reader = Reader::from_str(&xml);
|
@ -1,32 +1,41 @@
|
|||||||
|
//! Link data.
|
||||||
|
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
use crate::data::book::{Book, Format};
|
use crate::data::book::{Book, Format};
|
||||||
|
|
||||||
use super::{media_type::MediaType, relation::Relation};
|
use super::{media_type::MediaType, relation::Relation};
|
||||||
|
|
||||||
|
/// Link element in OPDS.
|
||||||
#[derive(Debug, Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
#[serde(rename = "link")]
|
#[serde(rename = "link")]
|
||||||
pub struct Link {
|
pub struct Link {
|
||||||
|
/// Actual hyperlink.
|
||||||
#[serde(rename = "@href")]
|
#[serde(rename = "@href")]
|
||||||
pub href: String,
|
pub href: String,
|
||||||
|
/// Type of the target.
|
||||||
#[serde(rename = "@type")]
|
#[serde(rename = "@type")]
|
||||||
pub media_type: MediaType,
|
pub media_type: MediaType,
|
||||||
|
/// Relation of the target.
|
||||||
#[serde(rename = "@rel")]
|
#[serde(rename = "@rel")]
|
||||||
pub rel: Relation,
|
pub rel: Relation,
|
||||||
|
/// Optional link title.
|
||||||
#[serde(rename = "@title")]
|
#[serde(rename = "@title")]
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub title: Option<String>,
|
pub title: Option<String>,
|
||||||
|
/// Optional count (how many entries at the target).
|
||||||
#[serde(rename = "@thr:count")]
|
#[serde(rename = "@thr:count")]
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub count: Option<u64>,
|
pub count: Option<u64>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert a format from a book into a link where it is downloadable.
|
||||||
impl From<(&Book, (&Format, &str))> for Link {
|
impl From<(&Book, (&Format, &str))> for Link {
|
||||||
fn from(value: (&Book, (&Format, &str))) -> Self {
|
fn from(value: (&Book, (&Format, &str))) -> Self {
|
||||||
let format = value.1 .0.clone();
|
let format = value.1 .0.clone();
|
||||||
let media_type: MediaType = format.into();
|
let media_type: MediaType = format.into();
|
||||||
Self {
|
Self {
|
||||||
href: format!("/book/{}/{}", value.0.id, value.1 .0),
|
href: format!("/book/{}/{}", value.0.data.id, value.1 .0),
|
||||||
media_type,
|
media_type,
|
||||||
rel: media_type.into(),
|
rel: media_type.into(),
|
||||||
title: Some(value.1 .0 .0.clone()),
|
title: Some(value.1 .0 .0.clone()),
|
@ -1,18 +1,24 @@
|
|||||||
|
//! Media types for OPDS elements.
|
||||||
|
|
||||||
use serde_with::SerializeDisplay;
|
use serde_with::SerializeDisplay;
|
||||||
|
|
||||||
use crate::data::book::Format;
|
use crate::data::book::Format;
|
||||||
|
|
||||||
|
/// Valid media types.
|
||||||
#[derive(Debug, Copy, Clone, SerializeDisplay)]
|
#[derive(Debug, Copy, Clone, SerializeDisplay)]
|
||||||
pub enum MediaType {
|
pub enum MediaType {
|
||||||
|
/// A link with this type is meant to acquire a certain thing, for example an entry.
|
||||||
Acquisition,
|
Acquisition,
|
||||||
Epub,
|
Epub,
|
||||||
Html,
|
Html,
|
||||||
Jpeg,
|
Jpeg,
|
||||||
|
/// A link with this type is meant for navigation around a feed.
|
||||||
Navigation,
|
Navigation,
|
||||||
Pdf,
|
Pdf,
|
||||||
Text,
|
Text,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert `epub` and `pdf` formats to their respective media type. Everything else is `Text`.
|
||||||
impl From<Format> for MediaType {
|
impl From<Format> for MediaType {
|
||||||
fn from(value: Format) -> Self {
|
fn from(value: Format) -> Self {
|
||||||
match value.0.as_ref() {
|
match value.0.as_ref() {
|
||||||
@ -23,6 +29,7 @@ impl From<Format> for MediaType {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Display the respective mime types of the respective media types.
|
||||||
impl std::fmt::Display for MediaType {
|
impl std::fmt::Display for MediaType {
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
match self {
|
match self {
|
@ -1,10 +1,14 @@
|
|||||||
|
//! Relation data.
|
||||||
|
|
||||||
use serde_with::SerializeDisplay;
|
use serde_with::SerializeDisplay;
|
||||||
|
|
||||||
use super::media_type::MediaType;
|
use super::media_type::MediaType;
|
||||||
|
|
||||||
|
/// Types of relations for links.
|
||||||
#[derive(Debug, SerializeDisplay)]
|
#[derive(Debug, SerializeDisplay)]
|
||||||
pub enum Relation {
|
pub enum Relation {
|
||||||
Image,
|
Image,
|
||||||
|
/// Refer to the self feed.
|
||||||
Myself,
|
Myself,
|
||||||
Start,
|
Start,
|
||||||
Subsection,
|
Subsection,
|
||||||
@ -12,6 +16,9 @@ pub enum Relation {
|
|||||||
Acquisition,
|
Acquisition,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Convert a media type int a relation.
|
||||||
|
///
|
||||||
|
/// This is not always deterministic but for the ones I actually use so far it is correct.
|
||||||
impl From<MediaType> for Relation {
|
impl From<MediaType> for Relation {
|
||||||
fn from(value: MediaType) -> Self {
|
fn from(value: MediaType) -> Self {
|
||||||
match value {
|
match value {
|
||||||
@ -26,6 +33,7 @@ impl From<MediaType> for Relation {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Specify how to represent all relations in OPDS.
|
||||||
impl std::fmt::Display for Relation {
|
impl std::fmt::Display for Relation {
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
match self {
|
match self {
|
@ -1,6 +1,9 @@
|
|||||||
|
//! Tera templates.
|
||||||
|
|
||||||
use once_cell::sync::Lazy;
|
use once_cell::sync::Lazy;
|
||||||
use tera::Tera;
|
use tera::Tera;
|
||||||
|
|
||||||
|
/// All tera templates, globally accessible.
|
||||||
pub static TEMPLATES: Lazy<Tera> = Lazy::new(|| {
|
pub static TEMPLATES: Lazy<Tera> = Lazy::new(|| {
|
||||||
let mut tera = Tera::default();
|
let mut tera = Tera::default();
|
||||||
tera.add_raw_templates(vec![
|
tera.add_raw_templates(vec![
|
@ -39,3 +39,8 @@ nav ul li {
|
|||||||
text-align: center;
|
text-align: center;
|
||||||
height: 6rem;
|
height: 6rem;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
footer small {
|
||||||
|
display: flex;
|
||||||
|
justify-content: space-between;
|
||||||
|
}
|
@ -6,7 +6,7 @@
|
|||||||
<meta name="color-scheme" content="light dark" />
|
<meta name="color-scheme" content="light dark" />
|
||||||
<link rel="stylesheet" href="/static/pico.min.css" />
|
<link rel="stylesheet" href="/static/pico.min.css" />
|
||||||
<link rel="stylesheet" href="/static/style.css" />
|
<link rel="stylesheet" href="/static/style.css" />
|
||||||
<title>Rusty Library</title>
|
<title>Little Hesinde</title>
|
||||||
</head>
|
</head>
|
||||||
<body>
|
<body>
|
||||||
<header class="container fixed">
|
<header class="container fixed">
|
||||||
@ -27,7 +27,11 @@
|
|||||||
</main>
|
</main>
|
||||||
<footer class="container">
|
<footer class="container">
|
||||||
<hr />
|
<hr />
|
||||||
<small>From <a href="https://code.vanwa.ch/shu/rusty-library">https://code.vanwa.ch</a> under <a href="https://www.gnu.org/licenses/agpl-3.0.txt">AGPL-3</a></small>
|
<small>
|
||||||
|
<div>From <a href="https://code.vanwa.ch/shu/little-hesinde">https://code.vanwa.ch</a>
|
||||||
|
under <a href="https://www.gnu.org/licenses/agpl-3.0.txt">AGPL-3</a></div>
|
||||||
|
<div><a href="/opds">opds feed</a></div>
|
||||||
|
</small>
|
||||||
</footer>
|
</footer>
|
||||||
</body>
|
</body>
|
||||||
</html>
|
</html>
|
@ -1,7 +1,7 @@
|
|||||||
<article class="book-card">
|
<article class="book-card">
|
||||||
<header class="grid-item">
|
<header class="grid-item">
|
||||||
<hgroup>
|
<hgroup>
|
||||||
<h5>{{ book.title }}</h5>
|
<h5>{{ book.data.title }}</h5>
|
||||||
<p>
|
<p>
|
||||||
<a class="secondary" href="/authors/{{ book.author.id }}">{{ book.author.name }}</a>
|
<a class="secondary" href="/authors/{{ book.author.id }}">{{ book.author.name }}</a>
|
||||||
</p>
|
</p>
|
||||||
@ -14,7 +14,7 @@
|
|||||||
{% endif %}
|
{% endif %}
|
||||||
</hgroup>
|
</hgroup>
|
||||||
</header>
|
</header>
|
||||||
<img class="cover" src="/cover/{{ book.id }}" alt="book cover">
|
<img class="cover" src="/cover/{{ book.data.id }}" alt="book cover">
|
||||||
<footer>
|
<footer>
|
||||||
<form>
|
<form>
|
||||||
<fieldset role="group">
|
<fieldset role="group">
|
||||||
@ -24,7 +24,7 @@
|
|||||||
</summary>
|
</summary>
|
||||||
<ul>
|
<ul>
|
||||||
{% for format, _ in book.formats %}
|
{% for format, _ in book.formats %}
|
||||||
<li><a href="/book/{{ book.id }}/{{ format }}">{{ format }}</a></li>
|
<li><a href="/book/{{ book.data.id }}/{{ format }}">{{ format }}</a></li>
|
||||||
{% endfor %}
|
{% endfor %}
|
||||||
</ul>
|
</ul>
|
||||||
</details>
|
</details>
|
@ -1,8 +0,0 @@
|
|||||||
use calibre_db::calibre::Calibre;
|
|
||||||
|
|
||||||
use crate::config::Config;
|
|
||||||
|
|
||||||
pub struct AppState {
|
|
||||||
pub calibre: Calibre,
|
|
||||||
pub config: Config,
|
|
||||||
}
|
|
@ -1,44 +0,0 @@
|
|||||||
use poem::{
|
|
||||||
http::StatusCode,
|
|
||||||
web::{
|
|
||||||
headers,
|
|
||||||
headers::{authorization::Basic, HeaderMapExt},
|
|
||||||
},
|
|
||||||
Endpoint, Error, Middleware, Request, Result,
|
|
||||||
};
|
|
||||||
|
|
||||||
pub struct BasicAuth {
|
|
||||||
pub username: String,
|
|
||||||
pub password: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<E: Endpoint> Middleware<E> for BasicAuth {
|
|
||||||
type Output = BasicAuthEndpoint<E>;
|
|
||||||
|
|
||||||
fn transform(&self, ep: E) -> Self::Output {
|
|
||||||
BasicAuthEndpoint {
|
|
||||||
ep,
|
|
||||||
username: self.username.clone(),
|
|
||||||
password: self.password.clone(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct BasicAuthEndpoint<E> {
|
|
||||||
ep: E,
|
|
||||||
username: String,
|
|
||||||
password: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<E: Endpoint> Endpoint for BasicAuthEndpoint<E> {
|
|
||||||
type Output = E::Output;
|
|
||||||
|
|
||||||
async fn call(&self, req: Request) -> Result<Self::Output> {
|
|
||||||
if let Some(auth) = req.headers().typed_get::<headers::Authorization<Basic>>() {
|
|
||||||
if auth.0.username() == self.username && auth.0.password() == self.password {
|
|
||||||
return self.ep.call(req).await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(Error::from_status(StatusCode::UNAUTHORIZED))
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,36 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use calibre_db::{calibre::Calibre, data::pagination::SortOrder};
|
|
||||||
use poem::{
|
|
||||||
handler,
|
|
||||||
web::{Data, Html, Path},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::{app_state::AppState, handlers::paginated};
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler_init(state: Data<&Arc<AppState>>) -> Result<Html<String>, poem::Error> {
|
|
||||||
authors(&state.calibre, None, &SortOrder::ASC)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(
|
|
||||||
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
authors(&state.calibre, Some(&cursor), &sort_order)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn authors(
|
|
||||||
calibre: &Calibre,
|
|
||||||
cursor: Option<&str>,
|
|
||||||
sort_order: &SortOrder,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
paginated::render(
|
|
||||||
"authors",
|
|
||||||
|| calibre.authors(25, cursor, sort_order),
|
|
||||||
|author| author.sort.clone(),
|
|
||||||
|cursor| calibre.has_previous_authors(cursor),
|
|
||||||
|cursor| calibre.has_more_authors(cursor),
|
|
||||||
)
|
|
||||||
}
|
|
@ -1,43 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use calibre_db::data::pagination::SortOrder;
|
|
||||||
use poem::{
|
|
||||||
handler,
|
|
||||||
web::{Data, Html, Path},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::{app_state::AppState, data::book::Book};
|
|
||||||
|
|
||||||
use super::paginated;
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler_init(state: Data<&Arc<AppState>>) -> Result<Html<String>, poem::Error> {
|
|
||||||
books(&state, None, &SortOrder::ASC)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(
|
|
||||||
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
books(&state, Some(&cursor), &sort_order)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn books(
|
|
||||||
state: &Arc<AppState>,
|
|
||||||
cursor: Option<&str>,
|
|
||||||
sort_order: &SortOrder,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
paginated::render(
|
|
||||||
"books",
|
|
||||||
|| {
|
|
||||||
state
|
|
||||||
.calibre
|
|
||||||
.books(25, cursor, sort_order)
|
|
||||||
.map(|x| x.iter().filter_map(|y| Book::full_book(y, state)).collect())
|
|
||||||
},
|
|
||||||
|book| book.sort.clone(),
|
|
||||||
|cursor| state.calibre.has_previous_books(cursor),
|
|
||||||
|cursor| state.calibre.has_more_books(cursor),
|
|
||||||
)
|
|
||||||
}
|
|
@ -1,38 +0,0 @@
|
|||||||
use std::{fs::File, io::Read, sync::Arc};
|
|
||||||
|
|
||||||
use poem::{
|
|
||||||
error::NotFoundError,
|
|
||||||
handler,
|
|
||||||
web::{Data, Path, WithContentType, WithHeader},
|
|
||||||
IntoResponse,
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
app_state::AppState,
|
|
||||||
data::book::{Book, Format},
|
|
||||||
handlers::error::HandlerError,
|
|
||||||
};
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(
|
|
||||||
Path((id, format)): Path<(u64, String)>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithHeader<WithContentType<Vec<u8>>>, poem::Error> {
|
|
||||||
let book = state
|
|
||||||
.calibre
|
|
||||||
.scalar_book(id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let book = Book::full_book(&book, &state).ok_or(NotFoundError)?;
|
|
||||||
let format = Format(format);
|
|
||||||
let file_name = book.formats.get(&format).ok_or(NotFoundError)?;
|
|
||||||
let file_path = state.config.library_path.join(book.path).join(file_name);
|
|
||||||
let mut file = File::open(file_path).map_err(|_| NotFoundError)?;
|
|
||||||
|
|
||||||
let mut data = Vec::new();
|
|
||||||
file.read_to_end(&mut data).map_err(|_| NotFoundError)?;
|
|
||||||
let content_type = format.0;
|
|
||||||
|
|
||||||
Ok(data
|
|
||||||
.with_content_type(content_type)
|
|
||||||
.with_header("Content-Disposition", format!("filename={file_name};")))
|
|
||||||
}
|
|
@ -1,363 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use calibre_db::data::{author::Author as DbAuthor, pagination::SortOrder};
|
|
||||||
use poem::{
|
|
||||||
handler,
|
|
||||||
web::{Data, Path, WithContentType},
|
|
||||||
IntoResponse,
|
|
||||||
};
|
|
||||||
use time::OffsetDateTime;
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
app_state::AppState,
|
|
||||||
data::book::Book,
|
|
||||||
handlers::error::HandlerError,
|
|
||||||
opds::{
|
|
||||||
author::Author, content::Content, entry::Entry, feed::Feed, link::Link,
|
|
||||||
media_type::MediaType, relation::Relation,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
fn create_feed(
|
|
||||||
now: OffsetDateTime,
|
|
||||||
id: &str,
|
|
||||||
title: &str,
|
|
||||||
self_link: Link,
|
|
||||||
mut additional_links: Vec<Link>,
|
|
||||||
entries: Vec<Entry>,
|
|
||||||
) -> Feed {
|
|
||||||
let author = Author {
|
|
||||||
name: "Thallian".to_string(),
|
|
||||||
uri: "https://code.vanwa.ch/shu/rusty-library".to_string(),
|
|
||||||
email: None,
|
|
||||||
};
|
|
||||||
let mut links = vec![
|
|
||||||
Link {
|
|
||||||
href: "/opds".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Start,
|
|
||||||
title: Some("Home".to_string()),
|
|
||||||
count: None,
|
|
||||||
},
|
|
||||||
self_link,
|
|
||||||
];
|
|
||||||
links.append(&mut additional_links);
|
|
||||||
|
|
||||||
Feed {
|
|
||||||
title: title.to_string(),
|
|
||||||
id: id.to_string(),
|
|
||||||
updated: now,
|
|
||||||
icon: "favicon.ico".to_string(),
|
|
||||||
author,
|
|
||||||
links,
|
|
||||||
entries,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn recents_handler(
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let books = state
|
|
||||||
.calibre
|
|
||||||
.recent_books(25)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = books
|
|
||||||
.iter()
|
|
||||||
.filter_map(|x| Book::full_book(x, &state))
|
|
||||||
.collect::<Vec<Book>>();
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: "/opds/recent".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
"rusty:recentbooks",
|
|
||||||
"Recent Books",
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
entries,
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn series_single_handler(
|
|
||||||
id: Path<u64>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let series = state
|
|
||||||
.calibre
|
|
||||||
.scalar_series(*id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = state
|
|
||||||
.calibre
|
|
||||||
.series_books(*id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = books
|
|
||||||
.iter()
|
|
||||||
.filter_map(|x| Book::full_book(x, &state))
|
|
||||||
.collect::<Vec<Book>>();
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: format!("/opds/series/{}", *id),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
&format!("rusty:series:{}", *id),
|
|
||||||
&series.name,
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
entries,
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn series_handler(
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let series = state
|
|
||||||
.calibre
|
|
||||||
.series(u32::MAX.into(), None, &SortOrder::ASC)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = series.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: "/opds/series".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
"rusty:series",
|
|
||||||
"All Series",
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
entries,
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn author_handler(
|
|
||||||
id: Path<u64>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let author = state
|
|
||||||
.calibre
|
|
||||||
.scalar_author(*id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = state
|
|
||||||
.calibre
|
|
||||||
.author_books(*id, u32::MAX.into(), None, SortOrder::ASC)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = books
|
|
||||||
.iter()
|
|
||||||
.filter_map(|x| Book::full_book(x, &state))
|
|
||||||
.collect::<Vec<Book>>();
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: format!("/opds/authors/{}", author.id),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
&format!("rusty:author:{}", author.id),
|
|
||||||
&author.name,
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
entries,
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn authors_handler(
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let authors: Vec<DbAuthor> = state
|
|
||||||
.calibre
|
|
||||||
.authors(u32::MAX.into(), None, &SortOrder::ASC)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = authors.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: "/opds/authors".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
"rusty:authors",
|
|
||||||
"All Authors",
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
entries,
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn books_handler(
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let books: Vec<Book> = state
|
|
||||||
.calibre
|
|
||||||
.books(u32::MAX.into(), None, &SortOrder::ASC)
|
|
||||||
.map(|x| {
|
|
||||||
x.iter()
|
|
||||||
.filter_map(|y| Book::full_book(y, &state))
|
|
||||||
.collect()
|
|
||||||
})
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
|
|
||||||
let entries: Vec<Entry> = books.into_iter().map(Entry::from).collect();
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: "/opds/books".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let feed = create_feed(now, "rusty:books", "All Books", self_link, vec![], entries);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler() -> Result<WithContentType<String>, poem::Error> {
|
|
||||||
let now = OffsetDateTime::now_utc();
|
|
||||||
|
|
||||||
let self_link = Link {
|
|
||||||
href: "/opds".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Myself,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
};
|
|
||||||
let books_entry = Entry {
|
|
||||||
title: "Books".to_string(),
|
|
||||||
id: "rusty:books".to_string(),
|
|
||||||
updated: now,
|
|
||||||
content: Some(Content {
|
|
||||||
media_type: MediaType::Text,
|
|
||||||
content: "Index of all books".to_string(),
|
|
||||||
}),
|
|
||||||
author: None,
|
|
||||||
links: vec![Link {
|
|
||||||
href: "/opds/books".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Subsection,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
}],
|
|
||||||
};
|
|
||||||
|
|
||||||
let authors_entry = Entry {
|
|
||||||
title: "Authors".to_string(),
|
|
||||||
id: "rusty:authors".to_string(),
|
|
||||||
updated: now,
|
|
||||||
content: Some(Content {
|
|
||||||
media_type: MediaType::Text,
|
|
||||||
content: "Index of all authors".to_string(),
|
|
||||||
}),
|
|
||||||
author: None,
|
|
||||||
links: vec![Link {
|
|
||||||
href: "/opds/authors".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Subsection,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
}],
|
|
||||||
};
|
|
||||||
|
|
||||||
let series_entry = Entry {
|
|
||||||
title: "Series".to_string(),
|
|
||||||
id: "rusty:series".to_string(),
|
|
||||||
updated: now,
|
|
||||||
content: Some(Content {
|
|
||||||
media_type: MediaType::Text,
|
|
||||||
content: "Index of all series".to_string(),
|
|
||||||
}),
|
|
||||||
author: None,
|
|
||||||
links: vec![Link {
|
|
||||||
href: "/opds/series".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Subsection,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
}],
|
|
||||||
};
|
|
||||||
|
|
||||||
let recents_entry = Entry {
|
|
||||||
title: "Recent Additions".to_string(),
|
|
||||||
id: "rusty:recentbooks".to_string(),
|
|
||||||
updated: now,
|
|
||||||
content: Some(Content {
|
|
||||||
media_type: MediaType::Text,
|
|
||||||
content: "Recently added books".to_string(),
|
|
||||||
}),
|
|
||||||
author: None,
|
|
||||||
links: vec![Link {
|
|
||||||
href: "/opds/recent".to_string(),
|
|
||||||
media_type: MediaType::Navigation,
|
|
||||||
rel: Relation::Subsection,
|
|
||||||
title: None,
|
|
||||||
count: None,
|
|
||||||
}],
|
|
||||||
};
|
|
||||||
|
|
||||||
let feed = create_feed(
|
|
||||||
now,
|
|
||||||
"rusty:catalog",
|
|
||||||
"Rusty-Library",
|
|
||||||
self_link,
|
|
||||||
vec![],
|
|
||||||
vec![authors_entry, series_entry, books_entry, recents_entry],
|
|
||||||
);
|
|
||||||
let xml = feed.as_xml().map_err(HandlerError::OpdsError)?;
|
|
||||||
|
|
||||||
Ok(xml.with_content_type("application/atom+xml"))
|
|
||||||
}
|
|
@ -1,33 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use poem::{
|
|
||||||
error::InternalServerError,
|
|
||||||
handler,
|
|
||||||
web::{Data, Html},
|
|
||||||
};
|
|
||||||
use tera::Context;
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
app_state::AppState, data::book::Book, handlers::error::HandlerError, templates::TEMPLATES,
|
|
||||||
};
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(state: Data<&Arc<AppState>>) -> Result<Html<String>, poem::Error> {
|
|
||||||
let recent_books = state
|
|
||||||
.calibre
|
|
||||||
.recent_books(25)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let recent_books = recent_books
|
|
||||||
.iter()
|
|
||||||
.filter_map(|x| Book::full_book(x, &state))
|
|
||||||
.collect::<Vec<Book>>();
|
|
||||||
|
|
||||||
let mut context = Context::new();
|
|
||||||
context.insert("title", "Recent Books");
|
|
||||||
context.insert("nav", "recent");
|
|
||||||
context.insert("books", &recent_books);
|
|
||||||
TEMPLATES
|
|
||||||
.render("book_list", &context)
|
|
||||||
.map_err(InternalServerError)
|
|
||||||
.map(Html)
|
|
||||||
}
|
|
@ -1,38 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use calibre_db::data::pagination::SortOrder;
|
|
||||||
use poem::{
|
|
||||||
handler,
|
|
||||||
web::{Data, Html, Path},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::app_state::AppState;
|
|
||||||
|
|
||||||
use super::paginated;
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler_init(state: Data<&Arc<AppState>>) -> Result<Html<String>, poem::Error> {
|
|
||||||
series(&state, None, &SortOrder::ASC)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(
|
|
||||||
Path((cursor, sort_order)): Path<(String, SortOrder)>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
series(&state, Some(&cursor), &sort_order)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn series(
|
|
||||||
state: &Arc<AppState>,
|
|
||||||
cursor: Option<&str>,
|
|
||||||
sort_order: &SortOrder,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
paginated::render(
|
|
||||||
"series",
|
|
||||||
|| state.calibre.series(25, cursor, sort_order),
|
|
||||||
|series| series.sort.clone(),
|
|
||||||
|cursor| state.calibre.has_previous_series(cursor),
|
|
||||||
|cursor| state.calibre.has_more_series(cursor),
|
|
||||||
)
|
|
||||||
}
|
|
@ -1,41 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use poem::{
|
|
||||||
error::InternalServerError,
|
|
||||||
handler,
|
|
||||||
web::{Data, Html, Path},
|
|
||||||
};
|
|
||||||
use tera::Context;
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
app_state::AppState, data::book::Book, handlers::error::HandlerError, templates::TEMPLATES,
|
|
||||||
};
|
|
||||||
|
|
||||||
#[handler]
|
|
||||||
pub async fn handler(
|
|
||||||
id: Path<u64>,
|
|
||||||
state: Data<&Arc<AppState>>,
|
|
||||||
) -> Result<Html<String>, poem::Error> {
|
|
||||||
let series = state
|
|
||||||
.calibre
|
|
||||||
.scalar_series(*id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = state
|
|
||||||
.calibre
|
|
||||||
.series_books(*id)
|
|
||||||
.map_err(HandlerError::DataError)?;
|
|
||||||
let books = books
|
|
||||||
.iter()
|
|
||||||
.filter_map(|x| Book::full_book(x, &state))
|
|
||||||
.collect::<Vec<Book>>();
|
|
||||||
|
|
||||||
let mut context = Context::new();
|
|
||||||
context.insert("title", &series.name);
|
|
||||||
context.insert("nav", "series");
|
|
||||||
context.insert("books", &books);
|
|
||||||
|
|
||||||
TEMPLATES
|
|
||||||
.render("book_list", &context)
|
|
||||||
.map_err(InternalServerError)
|
|
||||||
.map(Html)
|
|
||||||
}
|
|
@ -1,79 +0,0 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use app_state::AppState;
|
|
||||||
use calibre_db::calibre::Calibre;
|
|
||||||
use config::Config;
|
|
||||||
use poem::{
|
|
||||||
endpoint::EmbeddedFilesEndpoint, get, listener::TcpListener, middleware::Tracing, EndpointExt,
|
|
||||||
Route, Server,
|
|
||||||
};
|
|
||||||
use rust_embed::RustEmbed;
|
|
||||||
|
|
||||||
pub mod app_state;
|
|
||||||
pub mod basic_auth;
|
|
||||||
pub mod cli;
|
|
||||||
pub mod config;
|
|
||||||
pub mod data {
|
|
||||||
pub mod book;
|
|
||||||
}
|
|
||||||
pub mod handlers {
|
|
||||||
pub mod author;
|
|
||||||
pub mod authors;
|
|
||||||
pub mod books;
|
|
||||||
pub mod cover;
|
|
||||||
pub mod download;
|
|
||||||
pub mod error;
|
|
||||||
pub mod opds;
|
|
||||||
pub mod paginated;
|
|
||||||
pub mod recents;
|
|
||||||
pub mod series;
|
|
||||||
pub mod series_single;
|
|
||||||
}
|
|
||||||
pub mod opds;
|
|
||||||
pub mod templates;
|
|
||||||
|
|
||||||
#[derive(RustEmbed)]
|
|
||||||
#[folder = "static"]
|
|
||||||
pub struct Files;
|
|
||||||
|
|
||||||
pub async fn run(config: Config) -> Result<(), std::io::Error> {
|
|
||||||
let calibre = Calibre::load(&config.metadata_path).expect("failed to load calibre database");
|
|
||||||
let app_state = Arc::new(AppState { calibre, config });
|
|
||||||
|
|
||||||
let app = Route::new()
|
|
||||||
.at("/", get(handlers::recents::handler))
|
|
||||||
.at("/opds", get(handlers::opds::handler))
|
|
||||||
.at("/opds/recent", get(handlers::opds::recents_handler))
|
|
||||||
.at("/opds/books", get(handlers::opds::books_handler))
|
|
||||||
.at("/opds/authors", get(handlers::opds::authors_handler))
|
|
||||||
.at("/opds/authors/:id", get(handlers::opds::author_handler))
|
|
||||||
.at("/opds/series", get(handlers::opds::series_handler))
|
|
||||||
.at(
|
|
||||||
"/opds/series/:id",
|
|
||||||
get(handlers::opds::series_single_handler),
|
|
||||||
)
|
|
||||||
.at("/books", get(handlers::books::handler_init))
|
|
||||||
.at("/books/:cursor/:sort_order", get(handlers::books::handler))
|
|
||||||
.at("/series", get(handlers::series::handler_init))
|
|
||||||
.at(
|
|
||||||
"/series/:cursor/:sort_order",
|
|
||||||
get(handlers::series::handler),
|
|
||||||
)
|
|
||||||
.at("/series/:id", get(handlers::series_single::handler))
|
|
||||||
.at("/authors", get(handlers::authors::handler_init))
|
|
||||||
.at("/authors/:id", get(handlers::author::handler))
|
|
||||||
.at(
|
|
||||||
"/authors/:cursor/:sort_order",
|
|
||||||
get(handlers::authors::handler),
|
|
||||||
)
|
|
||||||
.at("/cover/:id", get(handlers::cover::handler))
|
|
||||||
.at("/book/:id/:format", get(handlers::download::handler))
|
|
||||||
.nest("/static", EmbeddedFilesEndpoint::<Files>::new())
|
|
||||||
.data(app_state)
|
|
||||||
.with(Tracing);
|
|
||||||
|
|
||||||
Server::new(TcpListener::bind("[::]:3000"))
|
|
||||||
.name("cops-web")
|
|
||||||
.run(app)
|
|
||||||
.await
|
|
||||||
}
|
|
@ -1,8 +0,0 @@
|
|||||||
pub mod author;
|
|
||||||
pub mod content;
|
|
||||||
pub mod entry;
|
|
||||||
pub mod error;
|
|
||||||
pub mod feed;
|
|
||||||
pub mod link;
|
|
||||||
pub mod media_type;
|
|
||||||
pub mod relation;
|
|
Loading…
Reference in New Issue
Block a user