Improve query performance, instrumentation.

This commit is contained in:
Syfaro 2021-04-22 21:34:02 -04:00
parent 34d32821cb
commit a1f28707ab
8 changed files with 194 additions and 269 deletions

8
Cargo.lock generated
View File

@ -1263,9 +1263,9 @@ checksum = "05842d0d43232b23ccb7060ecb0f0626922c21f30012e97b767b30afd4a5d4b9"
[[package]] [[package]]
name = "hyper" name = "hyper"
version = "0.14.6" version = "0.14.7"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "5f006b8784cfb01fe7aa9c46f5f5cd4cf5c85a8c612a0653ec97642979062665" checksum = "1e5f105c494081baa3bf9e200b279e27ec1623895cd504c7dbef8d0b080fcf54"
dependencies = [ dependencies = [
"bytes", "bytes",
"futures-channel", "futures-channel",
@ -2394,9 +2394,9 @@ dependencies = [
[[package]] [[package]]
name = "regex" name = "regex"
version = "1.4.5" version = "1.4.6"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "957056ecddbeba1b26965114e191d2e8589ce74db242b6ea25fc4062427a5c19" checksum = "2a26af418b574bd56588335b3a3659a65725d4e636eb1016c2f9e3b38c7cc759"
dependencies = [ dependencies = [
"aho-corasick", "aho-corasick",
"memchr", "memchr",

View File

@ -49,10 +49,9 @@ fn main() {
let data = job let data = job
.args() .args()
.into_iter() .iter()
.next() .next()
.ok_or(WebhookError::MissingData)? .ok_or(WebhookError::MissingData)?
.to_owned()
.to_owned(); .to_owned();
let value: fuzzysearch_common::types::WebHookData = serde_json::value::from_value(data)?; let value: fuzzysearch_common::types::WebHookData = serde_json::value::from_value(data)?;

View File

@ -10,7 +10,6 @@ pub fn search(
) -> impl Filter<Extract = impl Reply, Error = Rejection> + Clone { ) -> impl Filter<Extract = impl Reply, Error = Rejection> + Clone {
search_image(db.clone(), tree.clone()) search_image(db.clone(), tree.clone())
.or(search_hashes(db.clone(), tree.clone())) .or(search_hashes(db.clone(), tree.clone()))
.or(stream_search_image(db.clone(), tree.clone()))
.or(search_file(db.clone())) .or(search_file(db.clone()))
.or(search_video(db.clone())) .or(search_video(db.clone()))
.or(check_handle(db.clone())) .or(check_handle(db.clone()))
@ -89,26 +88,6 @@ pub fn search_hashes(
}) })
} }
pub fn stream_search_image(
db: Pool,
tree: Tree,
) -> impl Filter<Extract = impl Reply, Error = Rejection> + Clone {
warp::path("stream")
.and(warp::header::headers_cloned())
.and(warp::post())
.and(warp::multipart::form().max_length(1024 * 1024 * 10))
.and(with_pool(db))
.and(with_tree(tree))
.and(with_api_key())
.and_then(|headers, form, pool, tree, api_key| {
use tracing_opentelemetry::OpenTelemetrySpanExt;
let span = tracing::info_span!("stream_search_image");
span.set_parent(with_telem(headers));
span.in_scope(|| handlers::stream_image(form, pool, tree, api_key).in_current_span())
})
}
pub fn search_video(db: Pool) -> impl Filter<Extract = impl Reply, Error = Rejection> + Clone { pub fn search_video(db: Pool) -> impl Filter<Extract = impl Reply, Error = Rejection> + Clone {
warp::path("video") warp::path("video")
.and(warp::header::headers_cloned()) .and(warp::header::headers_cloned())

View File

@ -5,7 +5,7 @@ use tracing::{span, warn};
use tracing_futures::Instrument; use tracing_futures::Instrument;
use warp::{Rejection, Reply}; use warp::{Rejection, Reply};
use crate::models::{image_query, image_query_sync}; use crate::models::image_query;
use crate::types::*; use crate::types::*;
use crate::{early_return, rate_limit, Pool, Tree}; use crate::{early_return, rate_limit, Pool, Tree};
use fuzzysearch_common::types::{SearchResult, SiteInfo}; use fuzzysearch_common::types::{SearchResult, SiteInfo};
@ -112,7 +112,7 @@ async fn get_field_bytes(form: warp::multipart::FormData, field: &str) -> bytes:
} }
#[tracing::instrument(skip(form))] #[tracing::instrument(skip(form))]
async fn hash_input(form: warp::multipart::FormData) -> (i64, img_hash::ImageHash<[u8; 8]>) { async fn hash_input(form: warp::multipart::FormData) -> i64 {
let bytes = get_field_bytes(form, "image").await; let bytes = get_field_bytes(form, "image").await;
let len = bytes.len(); let len = bytes.len();
@ -131,7 +131,7 @@ async fn hash_input(form: warp::multipart::FormData) -> (i64, img_hash::ImageHas
let mut buf: [u8; 8] = [0; 8]; let mut buf: [u8; 8] = [0; 8];
buf.copy_from_slice(&hash.as_bytes()); buf.copy_from_slice(&hash.as_bytes());
(i64::from_be_bytes(buf), hash) i64::from_be_bytes(buf)
} }
#[tracing::instrument(skip(form))] #[tracing::instrument(skip(form))]
@ -168,37 +168,19 @@ pub async fn search_image(
let image_remaining = rate_limit!(&api_key, &db, image_limit, "image"); let image_remaining = rate_limit!(&api_key, &db, image_limit, "image");
let hash_remaining = rate_limit!(&api_key, &db, hash_limit, "hash"); let hash_remaining = rate_limit!(&api_key, &db, hash_limit, "hash");
let (num, hash) = hash_input(form).await; let num = hash_input(form).await;
let mut items = { let mut items = {
if opts.search_type == Some(ImageSearchType::Force) { if opts.search_type == Some(ImageSearchType::Force) {
image_query( image_query(db.clone(), tree.clone(), vec![num], 10)
db.clone(),
tree.clone(),
vec![num],
10,
Some(hash.as_bytes().to_vec()),
)
.await .await
.unwrap() .unwrap()
} else { } else {
let results = image_query( let results = image_query(db.clone(), tree.clone(), vec![num], 0)
db.clone(),
tree.clone(),
vec![num],
0,
Some(hash.as_bytes().to_vec()),
)
.await .await
.unwrap(); .unwrap();
if results.is_empty() && opts.search_type != Some(ImageSearchType::Exact) { if results.is_empty() && opts.search_type != Some(ImageSearchType::Exact) {
image_query( image_query(db.clone(), tree.clone(), vec![num], 10)
db.clone(),
tree.clone(),
vec![num],
10,
Some(hash.as_bytes().to_vec()),
)
.await .await
.unwrap() .unwrap()
} else { } else {
@ -235,43 +217,6 @@ pub async fn search_image(
Ok(Box::new(resp)) Ok(Box::new(resp))
} }
pub async fn stream_image(
form: warp::multipart::FormData,
pool: Pool,
tree: Tree,
api_key: String,
) -> Result<Box<dyn Reply>, Rejection> {
rate_limit!(&api_key, &pool, image_limit, "image", 2);
rate_limit!(&api_key, &pool, hash_limit, "hash");
let (num, hash) = hash_input(form).await;
let mut query = image_query_sync(
pool.clone(),
tree,
vec![num],
10,
Some(hash.as_bytes().to_vec()),
);
let event_stream = async_stream::stream! {
while let Some(result) = query.recv().await {
yield sse_matches(result);
}
};
Ok(Box::new(warp::sse::reply(event_stream)))
}
#[allow(clippy::unnecessary_wraps)]
fn sse_matches(
matches: Result<Vec<SearchResult>, sqlx::Error>,
) -> Result<warp::sse::Event, core::convert::Infallible> {
let items = matches.unwrap();
Ok(warp::sse::Event::default().json_data(items).unwrap())
}
pub async fn search_hashes( pub async fn search_hashes(
opts: HashSearchOpts, opts: HashSearchOpts,
db: Pool, db: Pool,
@ -293,18 +238,8 @@ pub async fn search_hashes(
let image_remaining = rate_limit!(&api_key, &db, image_limit, "image", hashes.len() as i16); let image_remaining = rate_limit!(&api_key, &db, image_limit, "image", hashes.len() as i16);
let mut results = image_query_sync( let results =
pool, early_return!(image_query(pool, tree, hashes.clone(), opts.distance.unwrap_or(10),).await);
tree,
hashes.clone(),
opts.distance.unwrap_or(10),
None,
);
let mut matches = Vec::new();
while let Some(r) = results.recv().await {
matches.extend(early_return!(r));
}
let resp = warp::http::Response::builder() let resp = warp::http::Response::builder()
.header("x-rate-limit-total-image", image_remaining.1.to_string()) .header("x-rate-limit-total-image", image_remaining.1.to_string())
@ -313,7 +248,7 @@ pub async fn search_hashes(
image_remaining.0.to_string(), image_remaining.0.to_string(),
) )
.header("content-type", "application/json") .header("content-type", "application/json")
.body(serde_json::to_string(&matches).unwrap()) .body(serde_json::to_string(&results).unwrap())
.unwrap(); .unwrap();
Ok(Box::new(resp)) Ok(Box::new(resp))
@ -427,10 +362,10 @@ pub async fn search_file(
.map(|artist| vec![artist]), .map(|artist| vec![artist]),
distance: None, distance: None,
hash: None, hash: None,
searched_hash: None,
site_info: Some(SiteInfo::FurAffinity { site_info: Some(SiteInfo::FurAffinity {
file_id: row.get("file_id"), file_id: row.get("file_id"),
}), }),
searched_hash: None,
rating: row rating: row
.get::<Option<String>, _>("rating") .get::<Option<String>, _>("rating")
.and_then(|rating| rating.parse().ok()), .and_then(|rating| rating.parse().ok()),
@ -452,8 +387,8 @@ pub async fn search_file(
pub async fn search_video( pub async fn search_video(
form: warp::multipart::FormData, form: warp::multipart::FormData,
db: Pool, _db: Pool,
api_key: String, _api_key: String,
) -> Result<impl Reply, Rejection> { ) -> Result<impl Reply, Rejection> {
let hashes = hash_video(form).await; let hashes = hash_video(form).await;
@ -535,7 +470,7 @@ pub async fn search_image_by_url(
let hash: [u8; 8] = hash.as_bytes().try_into().unwrap(); let hash: [u8; 8] = hash.as_bytes().try_into().unwrap();
let num = i64::from_be_bytes(hash); let num = i64::from_be_bytes(hash);
let results = image_query(db.clone(), tree.clone(), vec![num], 3, Some(hash.to_vec())) let results = image_query(db.clone(), tree.clone(), vec![num], 3)
.await .await
.unwrap(); .unwrap();

View File

@ -171,25 +171,30 @@ async fn create_tree(conn: &Pool) -> bk_tree::BKTree<Node, Hamming> {
let mut tree = bk_tree::BKTree::new(Hamming); let mut tree = bk_tree::BKTree::new(Hamming);
let mut rows = sqlx::query!( let mut rows = sqlx::query!(
"SELECT id, hash_int hash FROM submission WHERE hash_int IS NOT NULL "SELECT hash_int hash FROM submission WHERE hash_int IS NOT NULL
UNION ALL UNION
SELECT id, hash FROM e621 WHERE hash IS NOT NULL SELECT hash FROM e621 WHERE hash IS NOT NULL
UNION ALL UNION
SELECT tweet_id, hash FROM tweet_media WHERE hash IS NOT NULL SELECT hash FROM tweet_media WHERE hash IS NOT NULL
UNION ALL UNION
SELECT id, hash FROM weasyl WHERE hash IS NOT NULL" SELECT hash FROM weasyl WHERE hash IS NOT NULL"
) )
.fetch(conn); .fetch(conn);
let mut count = 0;
while let Some(row) = rows.try_next().await.expect("Unable to get row") { while let Some(row) = rows.try_next().await.expect("Unable to get row") {
if let Some(hash) = row.hash { if let Some(hash) = row.hash {
if tree.find_exact(&Node::new(hash)).is_some() { tree.add(Node::new(hash));
continue; count += 1;
if count % 250_000 == 0 {
tracing::debug!(count, "Made progress in loading tree rows");
}
}
} }
tree.add(Node::new(hash)); tracing::info!(count, "Completed loading rows for tree");
}
}
tree tree
} }

View File

@ -1,18 +1,14 @@
use std::collections::HashSet;
use lazy_static::lazy_static; use lazy_static::lazy_static;
use prometheus::{register_histogram, Histogram}; use prometheus::{register_histogram, Histogram};
use tracing_futures::Instrument;
use crate::types::*; use crate::types::*;
use crate::{Pool, Tree}; use crate::{Pool, Tree};
use futures::TryStreamExt;
use fuzzysearch_common::types::{SearchResult, SiteInfo}; use fuzzysearch_common::types::{SearchResult, SiteInfo};
lazy_static! { lazy_static! {
static ref IMAGE_LOOKUP_DURATION: Histogram = register_histogram!( static ref IMAGE_TREE_DURATION: Histogram = register_histogram!(
"fuzzysearch_api_image_lookup_seconds", "fuzzysearch_api_image_tree_seconds",
"Duration to perform an image lookup" "Duration to search for hashes in tree"
) )
.unwrap(); .unwrap();
static ref IMAGE_QUERY_DURATION: Histogram = register_histogram!( static ref IMAGE_QUERY_DURATION: Histogram = register_histogram!(
@ -48,62 +44,43 @@ pub async fn lookup_api_key(key: &str, db: &sqlx::PgPool) -> Option<ApiKey> {
.flatten() .flatten()
} }
#[derive(serde::Serialize)]
struct HashSearch {
searched_hash: i64,
found_hash: i64,
distance: u64,
}
#[tracing::instrument(skip(pool, tree))] #[tracing::instrument(skip(pool, tree))]
pub async fn image_query( pub async fn image_query(
pool: Pool, pool: Pool,
tree: Tree, tree: Tree,
hashes: Vec<i64>, hashes: Vec<i64>,
distance: i64, distance: i64,
hash: Option<Vec<u8>>,
) -> Result<Vec<SearchResult>, sqlx::Error> { ) -> Result<Vec<SearchResult>, sqlx::Error> {
let mut results = image_query_sync(pool, tree, hashes, distance, hash); let timer = IMAGE_TREE_DURATION.start_timer();
let mut matches = Vec::new();
while let Some(r) = results.recv().await {
matches.extend(r?);
}
Ok(matches)
}
#[tracing::instrument(skip(pool, tree))]
pub fn image_query_sync(
pool: Pool,
tree: Tree,
hashes: Vec<i64>,
distance: i64,
hash: Option<Vec<u8>>,
) -> tokio::sync::mpsc::Receiver<Result<Vec<SearchResult>, sqlx::Error>> {
let (tx, rx) = tokio::sync::mpsc::channel(50);
tokio::spawn(
async move {
let db = pool;
for query_hash in hashes {
tracing::trace!(query_hash, "Evaluating hash");
let mut seen: HashSet<[u8; 8]> = HashSet::new();
let _timer = IMAGE_LOOKUP_DURATION.start_timer();
let node = crate::Node::query(query_hash.to_be_bytes());
let lock = tree.read().await; let lock = tree.read().await;
let items = lock.find(&node, distance as u64); let found_hashes: Vec<HashSearch> = hashes
.iter()
.flat_map(|hash| {
lock.find(&crate::Node::new(*hash), distance as u64)
.map(|(dist, node)| HashSearch {
searched_hash: *hash,
found_hash: node.num(),
distance: dist,
})
.collect::<Vec<_>>()
})
.collect();
timer.stop_and_record();
for (dist, item) in items { let timer = IMAGE_QUERY_DURATION.start_timer();
if seen.contains(&item.0) { let matches = sqlx::query!(
tracing::trace!("Already searched for hash"); "WITH hashes AS (
continue; SELECT * FROM jsonb_to_recordset($1::jsonb)
} AS hashes(searched_hash bigint, found_hash bigint, distance bigint)
seen.insert(item.0); )
SELECT
let _timer = IMAGE_QUERY_DURATION.start_timer();
tracing::debug!(num = item.num(), "Searching database for hash in tree");
let mut row = sqlx::query!(
"SELECT
'FurAffinity' site, 'FurAffinity' site,
submission.id, submission.id,
submission.hash_int hash, submission.hash_int hash,
@ -112,10 +89,13 @@ pub fn image_query_sync(
ARRAY(SELECT artist.name) artists, ARRAY(SELECT artist.name) artists,
submission.file_id, submission.file_id,
null sources, null sources,
submission.rating submission.rating,
FROM submission hashes.searched_hash,
hashes.distance
FROM hashes
JOIN submission ON hashes.found_hash = submission.hash_int
JOIN artist ON submission.artist_id = artist.id JOIN artist ON submission.artist_id = artist.id
WHERE hash_int <@ ($1, 0) WHERE hash_int IN (SELECT hashes.found_hash)
UNION ALL UNION ALL
SELECT SELECT
'e621' site, 'e621' site,
@ -126,9 +106,12 @@ pub fn image_query_sync(
ARRAY(SELECT jsonb_array_elements_text(e621.data->'tags'->'artist')) artists, ARRAY(SELECT jsonb_array_elements_text(e621.data->'tags'->'artist')) artists,
null file_id, null file_id,
ARRAY(SELECT jsonb_array_elements_text(e621.data->'sources')) sources, ARRAY(SELECT jsonb_array_elements_text(e621.data->'sources')) sources,
e621.data->>'rating' rating e621.data->>'rating' rating,
FROM e621 hashes.searched_hash,
WHERE hash <@ ($1, 0) hashes.distance
FROM hashes
JOIN e621 ON hashes.found_hash = e621.hash
WHERE e621.hash IN (SELECT hashes.found_hash)
UNION ALL UNION ALL
SELECT SELECT
'Weasyl' site, 'Weasyl' site,
@ -139,9 +122,12 @@ pub fn image_query_sync(
ARRAY(SELECT weasyl.data->>'owner_login') artists, ARRAY(SELECT weasyl.data->>'owner_login') artists,
null file_id, null file_id,
null sources, null sources,
weasyl.data->>'rating' rating weasyl.data->>'rating' rating,
FROM weasyl hashes.searched_hash,
WHERE hash <@ ($1, 0) hashes.distance
FROM hashes
JOIN weasyl ON hashes.found_hash = weasyl.hash
WHERE weasyl.hash IN (SELECT hashes.found_hash)
UNION ALL UNION ALL
SELECT SELECT
'Twitter' site, 'Twitter' site,
@ -155,46 +141,49 @@ pub fn image_query_sync(
CASE CASE
WHEN (tweet.data->'possibly_sensitive')::boolean IS true THEN 'adult' WHEN (tweet.data->'possibly_sensitive')::boolean IS true THEN 'adult'
WHEN (tweet.data->'possibly_sensitive')::boolean IS false THEN 'general' WHEN (tweet.data->'possibly_sensitive')::boolean IS false THEN 'general'
END rating END rating,
FROM tweet_media hashes.searched_hash,
hashes.distance
FROM hashes
JOIN tweet_media ON hashes.found_hash = tweet_media.hash
JOIN tweet ON tweet_media.tweet_id = tweet.id JOIN tweet ON tweet_media.tweet_id = tweet.id
WHERE hash <@ ($1, 0)", WHERE tweet_media.hash IN (SELECT hashes.found_hash)",
&item.num() serde_json::to_value(&found_hashes).unwrap()
) )
.map(|row| { .map(|row| {
use std::convert::TryFrom;
let site_info = match row.site.as_deref() { let site_info = match row.site.as_deref() {
Some("FurAffinity") => SiteInfo::FurAffinity { file_id: row.file_id.unwrap_or(-1) }, Some("FurAffinity") => SiteInfo::FurAffinity {
Some("e621") => SiteInfo::E621 { sources: row.sources }, file_id: row.file_id.unwrap_or(-1),
},
Some("e621") => SiteInfo::E621 {
sources: row.sources,
},
Some("Twitter") => SiteInfo::Twitter, Some("Twitter") => SiteInfo::Twitter,
Some("Weasyl") => SiteInfo::Weasyl, Some("Weasyl") => SiteInfo::Weasyl,
_ => panic!("Got unknown site"), _ => panic!("Got unknown site"),
}; };
let file = SearchResult { SearchResult {
site_id: row.id.unwrap_or_default(), site_id: row.id.unwrap_or_default(),
site_info: Some(site_info), site_info: Some(site_info),
rating: row.rating.and_then(|rating| rating.parse().ok()), rating: row.rating.and_then(|rating| rating.parse().ok()),
site_id_str: row.id.unwrap_or_default().to_string(), site_id_str: row.id.unwrap_or_default().to_string(),
url: row.url.unwrap_or_default(), url: row.url.unwrap_or_default(),
hash: row.hash, hash: row.hash,
distance: Some(dist), distance: row
.distance
.map(|distance| u64::try_from(distance).ok())
.flatten(),
artists: row.artists, artists: row.artists,
filename: row.filename.unwrap_or_default(), filename: row.filename.unwrap_or_default(),
searched_hash: Some(query_hash), searched_hash: row.searched_hash,
}; }
vec![file]
}) })
.fetch(&db); .fetch_all(&pool)
.await?;
timer.stop_and_record();
while let Some(row) = row.try_next().await.ok().flatten() { Ok(matches)
tx.send(Ok(row)).await.unwrap();
}
}
}
}
.in_current_span(),
);
rx
} }

View File

@ -0,0 +1,9 @@
DROP INDEX submission_hash_int_idx;
DROP INDEX e621_hash_idx;
DROP INDEX tweet_media_hash_idx;
DROP INDEX weasyl_hash_idx;
CREATE INDEX bk_furaffinity_hash ON submission USING spgist (hash_int bktree_ops);
CREATE INDEX bk_e621_hash ON e621 USING spgist (hash bktree_ops);
CREATE INDEX bk_twitter_hash ON tweet_media USING spgist (hash bktree_ops);
CREATE INDEX bk_weasyl_hash ON weasyl USING spgist (hash bktree_ops);

View File

@ -0,0 +1,9 @@
DROP INDEX bk_furaffinity_hash;
DROP INDEX bk_e621_hash;
DROP INDEX bk_twitter_hash;
DROP INDEX bk_weasyl_hash;
CREATE INDEX submission_hash_int_idx ON submission (hash_int);
CREATE INDEX e621_hash_idx ON e621 (hash);
CREATE INDEX tweet_media_hash_idx ON tweet_media (hash);
CREATE INDEX weasyl_hash_idx ON weasyl (hash);