Compare commits
3 commits
6793f0fdc9
...
7d14bccdee
Author | SHA1 | Date | |
---|---|---|---|
7d14bccdee | |||
bae7aab3df | |||
12073bfed4 |
9 changed files with 81 additions and 49 deletions
|
@ -311,7 +311,7 @@ pub fn set_maybe_node(obj: &mut serde_json::Value, key: &str, node: crate::Node<
|
|||
if node.is_nothing() {
|
||||
set_maybe_value(obj, key, None)
|
||||
} else {
|
||||
set_maybe_value(obj, key, Some(node.into()))
|
||||
set_maybe_value(obj, key, Some(node.into_inner()))
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -133,6 +133,20 @@ impl<T : super::Base> Node<T> {
|
|||
|
||||
#[cfg(feature = "unstructured")]
|
||||
impl Node<serde_json::Value> {
|
||||
pub fn into_inner(self) -> serde_json::Value {
|
||||
match self {
|
||||
Self::Object(x) => *x,
|
||||
Self::Link(l) => serde_json::Value::String(l.href().unwrap_or_default().to_string()),
|
||||
Self::Empty => serde_json::Value::Null,
|
||||
Self::Array(arr) => serde_json::Value::Array(
|
||||
arr
|
||||
.into_iter()
|
||||
.map(|x| x.into_inner())
|
||||
.collect()
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn link(uri: String) -> Self {
|
||||
Node::Link(Box::new(uri))
|
||||
}
|
||||
|
|
|
@ -13,18 +13,13 @@ pub fn ActorHeader() -> impl IntoView {
|
|||
move |id| {
|
||||
async move {
|
||||
match cache::OBJECTS.get(&Uri::full(U::Actor, &id)) {
|
||||
Some(x) => Ok::<_, String>(x.clone()),
|
||||
Some(x) => Some(x.clone()),
|
||||
None => {
|
||||
let user : serde_json::Value = Http::fetch(&Uri::api(U::Actor, &id, true), auth)
|
||||
.await
|
||||
.map_err(|e| e.to_string())?;
|
||||
let user = std::sync::Arc::new(user);
|
||||
let uid = Uri::full(U::Actor, &id);
|
||||
cache::OBJECTS.store(&uid, user.clone());
|
||||
let user = cache::OBJECTS.resolve(&id, U::Actor, auth).await?;
|
||||
if let Some(url) = user.url().id().str() {
|
||||
cache::WEBFINGER.store(&url, uid);
|
||||
cache::WEBFINGER.store(&url, user.id().unwrap_or_default().to_string());
|
||||
}
|
||||
Ok(user)
|
||||
Some(user)
|
||||
},
|
||||
}
|
||||
}
|
||||
|
@ -32,8 +27,8 @@ pub fn ActorHeader() -> impl IntoView {
|
|||
);
|
||||
move || match actor.get() {
|
||||
None => view! { <Loader /> }.into_view(),
|
||||
Some(Err(e)) => view! { <code class="center cw color">"could not resolve user: "{e}</code> }.into_view(),
|
||||
Some(Ok(actor)) => {
|
||||
Some(None) => view! { <code class="center cw color">"could not resolve user"</code> }.into_view(),
|
||||
Some(Some(actor)) => {
|
||||
let avatar_url = actor.icon().get().map(|x| x.url().id().str().unwrap_or(FALLBACK_IMAGE_URL.into())).unwrap_or(FALLBACK_IMAGE_URL.into());
|
||||
let background_url = actor.image().get().map(|x| x.url().id().str().unwrap_or(FALLBACK_IMAGE_URL.into())).unwrap_or(FALLBACK_IMAGE_URL.into());
|
||||
let username = actor.preferred_username().unwrap_or_default().to_string();
|
||||
|
|
|
@ -10,6 +10,12 @@ use leptos_use::{
|
|||
UseCookieOptions, UseElementSizeReturn
|
||||
};
|
||||
|
||||
// TODO this is getting out of hand
|
||||
// when we will add lists there will have to potentially be multiple timelines (one per list)
|
||||
// per user, which doesn't scale with this model. we should either take the "go back to where
|
||||
// you were" into our own hands (maybe with timeline "segments"? would also solve slow load,
|
||||
// but infinite-scroll upwards too may be hard to do) or understand how it works (with page
|
||||
// stacks?) and keep timelines local to views.
|
||||
#[derive(Clone, Copy)]
|
||||
pub struct Feeds {
|
||||
pub home: Timeline,
|
||||
|
@ -19,6 +25,7 @@ pub struct Feeds {
|
|||
pub user: Timeline,
|
||||
pub server: Timeline,
|
||||
pub context: Timeline,
|
||||
pub replies: Timeline,
|
||||
pub tag: Timeline,
|
||||
}
|
||||
|
||||
|
@ -32,6 +39,7 @@ impl Feeds {
|
|||
server: Timeline::new(format!("{URL_BASE}/outbox/page")),
|
||||
tag: Timeline::new(format!("{URL_BASE}/tags/upub/page")),
|
||||
context: Timeline::new(format!("{URL_BASE}/outbox/page")), // TODO ehhh
|
||||
replies: Timeline::new(format!("{URL_BASE}/outbox/page")), // TODO ehhh
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -42,6 +50,7 @@ impl Feeds {
|
|||
self.user.reset(None);
|
||||
self.server.reset(None);
|
||||
self.context.reset(None);
|
||||
self.replies.reset(None);
|
||||
self.tag.reset(None);
|
||||
}
|
||||
}
|
||||
|
@ -155,6 +164,7 @@ pub fn App() -> impl IntoView {
|
|||
<Route path="global" view=move || view! { <Feed tl=feeds.global /> } />
|
||||
<Route path="local" view=move || view! { <Feed tl=feeds.server /> } />
|
||||
<Route path="notifications" view=move || view! { <Feed tl=feeds.notifications ignore_filters=true /> } />
|
||||
<Route path="tags/:id" view=move || view! { <HashtagFeed tl=feeds.tag /> } />
|
||||
|
||||
<Route path="about" view=AboutPage />
|
||||
<Route path="config" view=move || view! { <ConfigPage setter=set_config /> } />
|
||||
|
@ -166,9 +176,14 @@ pub fn App() -> impl IntoView {
|
|||
<Route path="followers" view=move || view! { <FollowList outgoing=false /> } />
|
||||
</Route>
|
||||
|
||||
<Route path="tags/:id" view=move || view! { <HashtagFeed tl=feeds.tag /> } />
|
||||
|
||||
<Route path="objects/:id" view=ObjectView />
|
||||
<Route path="objects/:id" view=ObjectView >
|
||||
<Route path="" view=ObjectContext />
|
||||
<Route path="replies" view=ObjectReplies />
|
||||
// <Route path="liked" view=ObjectLiked />
|
||||
// <Route path="announced" view=ObjectAnnounced />
|
||||
</Route>
|
||||
|
||||
// <Route path="/web/activities/:id" view=move || view! { <ActivityPage tl=context_tl /> } />
|
||||
|
||||
<Route path="search" view=SearchPage />
|
||||
|
|
|
@ -109,6 +109,8 @@ pub fn PostBox(advanced: WriteSignal<bool>) -> impl IntoView {
|
|||
on:input=move |ev| set_content.set(event_target_value(&ev))
|
||||
></textarea>
|
||||
|
||||
|
||||
|
||||
<table class="align rev w-100">
|
||||
<tr>
|
||||
<td><input id="priv-public" type="radio" name="privacy" value="public" title="public" node_ref=public_ref /></td>
|
||||
|
|
|
@ -41,7 +41,7 @@ pub mod cache {
|
|||
|
||||
#[derive(Debug)]
|
||||
pub enum LookupStatus<T> {
|
||||
Resolving,
|
||||
Resolving, // TODO use this to avoid fetching twice!
|
||||
Found(T),
|
||||
NotFound,
|
||||
}
|
||||
|
@ -90,6 +90,26 @@ impl<T> Cache for DashmapCache<T> {
|
|||
}
|
||||
}
|
||||
|
||||
impl DashmapCache<Object> {
|
||||
pub async fn resolve(&self, key: &str, kind: UriClass, auth: Auth) -> Option<Object> {
|
||||
let full_key = Uri::full(kind, key);
|
||||
match self.get(&full_key) {
|
||||
Some(x) => Some(x),
|
||||
None => {
|
||||
let obj = match Http::fetch::<serde_json::Value>(&Uri::api(kind, key, true), auth).await {
|
||||
Ok(obj) => Arc::new(obj),
|
||||
Err(e) => {
|
||||
tracing::error!("failed loading object from backend: {e}");
|
||||
return None;
|
||||
},
|
||||
};
|
||||
cache::OBJECTS.store(&full_key, obj.clone());
|
||||
Some(obj)
|
||||
},
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// TODO would be cool unifying a bit the fetch code too
|
||||
|
||||
impl DashmapCache<Object> {
|
||||
|
|
|
@ -1,4 +1,6 @@
|
|||
pub mod view;
|
||||
pub mod item;
|
||||
pub mod context;
|
||||
pub mod replies;
|
||||
|
||||
pub mod item;
|
||||
pub mod attachment;
|
||||
|
|
|
@ -1,5 +1,3 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use leptos::*;
|
||||
use leptos_router::*;
|
||||
use crate::prelude::*;
|
||||
|
@ -10,39 +8,21 @@ use apb::{Base, Object};
|
|||
pub fn ObjectView() -> impl IntoView {
|
||||
let params = use_params_map();
|
||||
let auth = use_context::<Auth>().expect("missing auth context");
|
||||
let feeds = use_context::<Feeds>().expect("missing feeds context");
|
||||
let object = create_local_resource(
|
||||
move || params.get().get("id").cloned().unwrap_or_default(),
|
||||
move |oid| async move {
|
||||
let obj = match cache::OBJECTS.get(&Uri::full(U::Object, &oid)) {
|
||||
Some(x) => x.clone(),
|
||||
None => {
|
||||
let obj = match Http::fetch::<serde_json::Value>(&Uri::api(U::Object, &oid, true), auth).await {
|
||||
Ok(obj) => Arc::new(obj),
|
||||
Err(e) => {
|
||||
tracing::error!("failed loading object from backend: {e}");
|
||||
return None;
|
||||
},
|
||||
};
|
||||
if let Ok(author) = obj.attributed_to().id() {
|
||||
if let Ok(user) = Http::fetch::<serde_json::Value>(
|
||||
&Uri::api(U::Actor, author, true), auth
|
||||
).await {
|
||||
cache::OBJECTS.store(&Uri::full(U::Actor, author), Arc::new(user));
|
||||
}
|
||||
}
|
||||
cache::OBJECTS.store(&Uri::full(U::Object, &oid), obj.clone());
|
||||
obj
|
||||
}
|
||||
};
|
||||
if let Ok(ctx) = obj.context().id() {
|
||||
let tl_url = format!("{}/context/page", Uri::api(U::Object, ctx, false));
|
||||
if !feeds.context.next.get_untracked().starts_with(&tl_url) {
|
||||
feeds.context.reset(Some(tl_url));
|
||||
}
|
||||
let obj = cache::OBJECTS.resolve(&oid, U::Object, auth).await?;
|
||||
if let Ok(author) = obj.attributed_to().id() {
|
||||
cache::OBJECTS.resolve(author, U::Actor, auth).await;
|
||||
}
|
||||
|
||||
Some(obj)
|
||||
|
||||
// if let Ok(ctx) = obj.context().id() {
|
||||
// let tl_url = format!("{}/context/page", Uri::api(U::Object, ctx, false));
|
||||
// if !feeds.context.next.get_untracked().starts_with(&tl_url) {
|
||||
// feeds.context.reset(Some(tl_url));
|
||||
// }
|
||||
// }
|
||||
}
|
||||
);
|
||||
|
||||
|
@ -55,12 +35,14 @@ pub fn ObjectView() -> impl IntoView {
|
|||
},
|
||||
Some(Some(o)) => {
|
||||
let object = o.clone();
|
||||
let base = Uri::web(U::Object, o.id().unwrap_or_default());
|
||||
view!{
|
||||
<Object object=object />
|
||||
<hr class="color ma-2" />
|
||||
<div class="mr-1-r ml-1-r">
|
||||
<Thread tl=feeds.context root=o.id().unwrap_or_default().to_string() />
|
||||
</div>
|
||||
<code class="cw color center mt-1 mb-1 ml-3 mr-3">
|
||||
<a href=format!("{base}/context")><b>context</b></a> | <a href=format!("{base}/replies")><b>replies</b></a>
|
||||
</code>
|
||||
<Outlet />
|
||||
}.into_view()
|
||||
},
|
||||
}}
|
||||
|
|
|
@ -19,6 +19,8 @@ pub use crate::{
|
|||
view::ObjectView,
|
||||
attachment::Attachment,
|
||||
item::{Object, Summary, LikeButton, RepostButton, ReplyButton},
|
||||
context::ObjectContext,
|
||||
replies::ObjectReplies,
|
||||
},
|
||||
timeline::{
|
||||
Timeline,
|
||||
|
|
Loading…
Reference in a new issue