Compare commits
53 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
e7d03951b9 | ||
| af8816e0af | |||
|
|
64f6427e1a | ||
| c9b7a75429 | |||
|
|
0490f6922e | ||
| 057c4c9174 | |||
|
|
a9e56bc707 | ||
| e5d09c73b7 | |||
|
|
6e1298b825 | ||
| fc8e43437a | |||
|
|
cb453aa949 | ||
| 2651bd16ae | |||
|
|
91e0f0c46f | ||
| e6986d512b | |||
|
|
8f9c21675c | ||
| 7fb22cdd20 | |||
|
|
780291303d | ||
| 4f607f7d2f | |||
|
|
208227b3ed | ||
| bf1c7d4adf | |||
|
|
a7a30cf983 | ||
| 0bc0676b33 | |||
|
|
73484d3eb4 | ||
| b3253d5bbc | |||
|
|
54f3769e90 | ||
| bad6f74ee6 | |||
|
|
bcf16168b6 | ||
| 498fbd9e0e | |||
|
|
007ff8e538 | ||
| 1fc70e3915 | |||
|
|
d891e47e02 | ||
| 08c39afde4 | |||
|
|
c579543b8a | ||
| 0d84137786 | |||
|
|
20ee30c4b4 | ||
| 93612137e3 | |||
|
|
6e70f08e3c | ||
| 459f5f7976 | |||
|
|
a2e6331ddd | ||
| 13cd30bec9 | |||
|
|
baeb9488c6 | ||
| 0cba0f987e | |||
|
|
958a79997d | ||
| 8fb1c96f93 | |||
| 6e6fe80c7f | |||
|
|
d1554050bd | ||
|
|
b1fae270bb | ||
|
|
c852483e18 | ||
| 2e01ad5bc9 | |||
|
|
26763c7183 | ||
| f0c5c2c45b | |||
|
|
034bb60fd5 | ||
| d4b389cb79 |
@@ -52,6 +52,7 @@ module.exports = {
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
WORKER_LOCK_DURATION: '120000',
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
@@ -74,6 +75,7 @@ module.exports = {
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
WORKER_LOCK_DURATION: '120000',
|
||||
},
|
||||
// Development Environment Settings
|
||||
env_development: {
|
||||
@@ -97,6 +99,7 @@ module.exports = {
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
WORKER_LOCK_DURATION: '120000',
|
||||
},
|
||||
},
|
||||
{
|
||||
|
||||
@@ -20,6 +20,9 @@ Create a new test file for `StatCard.tsx` to verify its props and rendering.
|
||||
|
||||
|
||||
|
||||
while assuming that master_schema_rollup.sql is the "ultimate source of truth", issues can happen and it may not have been properly
|
||||
updated - look for differences between these files
|
||||
|
||||
|
||||
UPC SCANNING !
|
||||
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.7.26",
|
||||
"version": "0.9.22",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.7.26",
|
||||
"version": "0.9.22",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.7.26",
|
||||
"version": "0.9.22",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -1,477 +1,8 @@
|
||||
-- sql/Initial_triggers_and_functions.sql
|
||||
-- This file contains all trigger functions and trigger definitions for the database.
|
||||
|
||||
-- 1. Set up the trigger to automatically create a profile when a new user signs up.
|
||||
-- This function is called by a trigger on the `public.users` table.
|
||||
DROP FUNCTION IF EXISTS public.handle_new_user();
|
||||
|
||||
-- It creates a corresponding profile and a default shopping list for the new user.
|
||||
-- It now accepts full_name and avatar_url from the user's metadata.
|
||||
CREATE OR REPLACE FUNCTION public.handle_new_user()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
new_profile_id UUID;
|
||||
user_meta_data JSONB;
|
||||
BEGIN
|
||||
-- The user's metadata (full_name, avatar_url) is passed via a temporary session variable.
|
||||
user_meta_data := current_setting('my_app.user_metadata', true)::JSONB;
|
||||
|
||||
INSERT INTO public.profiles (user_id, role, full_name, avatar_url)
|
||||
VALUES (new.user_id, 'user', user_meta_data->>'full_name', user_meta_data->>'avatar_url')
|
||||
RETURNING user_id INTO new_profile_id;
|
||||
|
||||
-- Also create a default shopping list for the new user.
|
||||
INSERT INTO public.shopping_lists (user_id, name)
|
||||
VALUES (new.user_id, 'Main Shopping List');
|
||||
|
||||
-- Log the new user event
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (new.user_id, 'user_registered',
|
||||
COALESCE(user_meta_data->>'full_name', new.email) || ' has registered.',
|
||||
'user-plus',
|
||||
jsonb_build_object('email', new.email)
|
||||
);
|
||||
|
||||
RETURN new;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- This trigger calls the function after a new user is created.
|
||||
DROP TRIGGER IF EXISTS on_auth_user_created ON public.users;
|
||||
CREATE TRIGGER on_auth_user_created
|
||||
AFTER INSERT ON public.users
|
||||
FOR EACH ROW EXECUTE FUNCTION public.handle_new_user();
|
||||
|
||||
-- 2. Create a reusable function to automatically update 'updated_at' columns.
|
||||
DROP FUNCTION IF EXISTS public.handle_updated_at();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.handle_updated_at()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
NEW.updated_at = now();
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Dynamically apply the 'handle_updated_at' trigger to all tables in the public schema
|
||||
-- that have an 'updated_at' column. This is more maintainable than creating a separate
|
||||
-- trigger for each table.
|
||||
DO $$
|
||||
DECLARE
|
||||
t_name TEXT;
|
||||
BEGIN
|
||||
FOR t_name IN
|
||||
SELECT table_name
|
||||
FROM information_schema.columns
|
||||
WHERE table_schema = 'public' AND column_name = 'updated_at'
|
||||
LOOP
|
||||
EXECUTE format('DROP TRIGGER IF EXISTS on_%s_updated ON public.%I;
|
||||
CREATE TRIGGER on_%s_updated
|
||||
BEFORE UPDATE ON public.%I
|
||||
FOR EACH ROW EXECUTE FUNCTION public.handle_updated_at();',
|
||||
t_name, t_name, t_name, t_name);
|
||||
END LOOP;
|
||||
END;
|
||||
$$;
|
||||
|
||||
-- 3. Create a trigger function to populate the item_price_history table on insert.
|
||||
DROP FUNCTION IF EXISTS public.update_price_history_on_flyer_item_insert();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_price_history_on_flyer_item_insert()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
flyer_valid_from DATE;
|
||||
flyer_valid_to DATE;
|
||||
current_summary_date DATE;
|
||||
flyer_location_id BIGINT;
|
||||
BEGIN
|
||||
-- If the item could not be matched, add it to the unmatched queue for review.
|
||||
IF NEW.master_item_id IS NULL THEN
|
||||
INSERT INTO public.unmatched_flyer_items (flyer_item_id)
|
||||
VALUES (NEW.flyer_item_id)
|
||||
ON CONFLICT (flyer_item_id) DO NOTHING;
|
||||
END IF;
|
||||
|
||||
-- Only run if the new flyer item is linked to a master item and has a price.
|
||||
IF NEW.master_item_id IS NULL OR NEW.price_in_cents IS NULL THEN
|
||||
RETURN NEW;
|
||||
END IF;
|
||||
|
||||
-- Get the validity dates of the flyer and the store_id.
|
||||
SELECT valid_from, valid_to INTO flyer_valid_from, flyer_valid_to
|
||||
FROM public.flyers
|
||||
WHERE flyer_id = NEW.flyer_id;
|
||||
|
||||
-- This single, set-based query is much more performant than looping.
|
||||
-- It generates all date/location pairs and inserts/updates them in one operation.
|
||||
INSERT INTO public.item_price_history (master_item_id, summary_date, store_location_id, min_price_in_cents, max_price_in_cents, avg_price_in_cents, data_points_count)
|
||||
SELECT
|
||||
NEW.master_item_id,
|
||||
d.day,
|
||||
fl.store_location_id,
|
||||
NEW.price_in_cents,
|
||||
NEW.price_in_cents,
|
||||
NEW.price_in_cents,
|
||||
1
|
||||
FROM public.flyer_locations fl
|
||||
CROSS JOIN generate_series(flyer_valid_from, flyer_valid_to, '1 day'::interval) AS d(day)
|
||||
WHERE fl.flyer_id = NEW.flyer_id
|
||||
ON CONFLICT (master_item_id, summary_date, store_location_id)
|
||||
DO UPDATE SET
|
||||
min_price_in_cents = LEAST(item_price_history.min_price_in_cents, EXCLUDED.min_price_in_cents),
|
||||
max_price_in_cents = GREATEST(item_price_history.max_price_in_cents, EXCLUDED.max_price_in_cents),
|
||||
avg_price_in_cents = ROUND(((item_price_history.avg_price_in_cents * item_price_history.data_points_count) + EXCLUDED.avg_price_in_cents) / (item_price_history.data_points_count + 1.0)),
|
||||
data_points_count = item_price_history.data_points_count + 1;
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Create the trigger on the flyer_items table for insert.
|
||||
DROP TRIGGER IF EXISTS trigger_update_price_history ON public.flyer_items;
|
||||
CREATE TRIGGER trigger_update_price_history
|
||||
AFTER INSERT ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_price_history_on_flyer_item_insert();
|
||||
|
||||
-- 4. Create a trigger function to recalculate price history when a flyer item is deleted.
|
||||
DROP FUNCTION IF EXISTS public.recalculate_price_history_on_flyer_item_delete();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.recalculate_price_history_on_flyer_item_delete()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
affected_dates RECORD;
|
||||
BEGIN
|
||||
-- Only run if the deleted item was linked to a master item and had a price.
|
||||
IF OLD.master_item_id IS NULL OR OLD.price_in_cents IS NULL THEN
|
||||
RETURN OLD;
|
||||
END IF;
|
||||
|
||||
-- This single, set-based query is much more performant than looping.
|
||||
-- It recalculates aggregates for all affected dates and locations at once.
|
||||
WITH affected_days_and_locations AS (
|
||||
-- 1. Get all date/location pairs affected by the deleted item's flyer.
|
||||
SELECT DISTINCT
|
||||
generate_series(f.valid_from, f.valid_to, '1 day'::interval)::date AS summary_date,
|
||||
fl.store_location_id
|
||||
FROM public.flyers f
|
||||
JOIN public.flyer_locations fl ON f.flyer_id = fl.flyer_id
|
||||
WHERE f.flyer_id = OLD.flyer_id
|
||||
),
|
||||
new_aggregates AS (
|
||||
-- 2. For each affected date/location, recalculate the aggregates from all other relevant flyer items.
|
||||
SELECT
|
||||
adl.summary_date,
|
||||
adl.store_location_id,
|
||||
MIN(fi.price_in_cents) AS min_price,
|
||||
MAX(fi.price_in_cents) AS max_price,
|
||||
ROUND(AVG(fi.price_in_cents))::int AS avg_price,
|
||||
COUNT(fi.flyer_item_id)::int AS data_points
|
||||
FROM affected_days_and_locations adl
|
||||
LEFT JOIN public.flyer_items fi ON fi.master_item_id = OLD.master_item_id AND fi.price_in_cents IS NOT NULL
|
||||
LEFT JOIN public.flyers f ON fi.flyer_id = f.flyer_id AND adl.summary_date BETWEEN f.valid_from AND f.valid_to
|
||||
LEFT JOIN public.flyer_locations fl ON fi.flyer_id = fl.flyer_id AND adl.store_location_id = fl.store_location_id
|
||||
WHERE fl.flyer_id IS NOT NULL -- Ensure the join was successful
|
||||
GROUP BY adl.summary_date, adl.store_location_id
|
||||
)
|
||||
-- 3. Update the history table with the new aggregates.
|
||||
UPDATE public.item_price_history iph
|
||||
SET
|
||||
min_price_in_cents = na.min_price,
|
||||
max_price_in_cents = na.max_price,
|
||||
avg_price_in_cents = na.avg_price,
|
||||
data_points_count = na.data_points
|
||||
FROM new_aggregates na
|
||||
WHERE iph.master_item_id = OLD.master_item_id
|
||||
AND iph.summary_date = na.summary_date
|
||||
AND iph.store_location_id = na.store_location_id;
|
||||
|
||||
-- 4. Delete any history records that no longer have any data points.
|
||||
DELETE FROM public.item_price_history iph
|
||||
WHERE iph.master_item_id = OLD.master_item_id
|
||||
AND NOT EXISTS (
|
||||
SELECT 1 FROM new_aggregates na
|
||||
WHERE na.summary_date = iph.summary_date AND na.store_location_id = iph.store_location_id
|
||||
);
|
||||
|
||||
RETURN OLD;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Create the trigger on the flyer_items table for DELETE operations.
|
||||
DROP TRIGGER IF EXISTS trigger_recalculate_price_history_on_delete ON public.flyer_items;
|
||||
CREATE TRIGGER trigger_recalculate_price_history_on_delete
|
||||
AFTER DELETE ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.recalculate_price_history_on_flyer_item_delete();
|
||||
|
||||
-- 5. Trigger function to update the average rating on the recipes table.
|
||||
DROP FUNCTION IF EXISTS public.update_recipe_rating_aggregates();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_recipe_rating_aggregates()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
UPDATE public.recipes
|
||||
SET
|
||||
avg_rating = (
|
||||
SELECT AVG(rating)
|
||||
FROM public.recipe_ratings
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id) -- This is correct, no change needed
|
||||
),
|
||||
rating_count = (
|
||||
SELECT COUNT(*)
|
||||
FROM public.recipe_ratings
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id) -- This is correct, no change needed
|
||||
)
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id);
|
||||
|
||||
RETURN NULL; -- The result is ignored since this is an AFTER trigger.
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after any change to recipe_ratings.
|
||||
DROP TRIGGER IF EXISTS on_recipe_rating_change ON public.recipe_ratings;
|
||||
CREATE TRIGGER on_recipe_rating_change
|
||||
AFTER INSERT OR UPDATE OR DELETE ON public.recipe_ratings
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_recipe_rating_aggregates();
|
||||
|
||||
-- 6. Trigger function to log the creation of a new recipe.
|
||||
DROP FUNCTION IF EXISTS public.log_new_recipe();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_recipe()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.user_id,
|
||||
'recipe_created',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.user_id) || ' created a new recipe: ' || NEW.name,
|
||||
'chef-hat',
|
||||
jsonb_build_object('recipe_id', NEW.recipe_id, 'recipe_name', NEW.name)
|
||||
);
|
||||
|
||||
-- Award 'First Recipe' achievement if it's their first one.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Recipe');
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a new recipe is inserted.
|
||||
DROP TRIGGER IF EXISTS on_new_recipe_created ON public.recipes;
|
||||
CREATE TRIGGER on_new_recipe_created
|
||||
AFTER INSERT ON public.recipes
|
||||
FOR EACH ROW
|
||||
WHEN (NEW.user_id IS NOT NULL) -- Only log activity for user-created recipes.
|
||||
EXECUTE FUNCTION public.log_new_recipe();
|
||||
|
||||
-- 7a. Trigger function to update the item_count on the flyers table.
|
||||
DROP FUNCTION IF EXISTS public.update_flyer_item_count();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_flyer_item_count()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
IF (TG_OP = 'INSERT') THEN
|
||||
UPDATE public.flyers SET item_count = item_count + 1 WHERE flyer_id = NEW.flyer_id;
|
||||
ELSIF (TG_OP = 'DELETE') THEN
|
||||
UPDATE public.flyers SET item_count = item_count - 1 WHERE flyer_id = OLD.flyer_id;
|
||||
END IF;
|
||||
RETURN NULL; -- The result is ignored since this is an AFTER trigger.
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after any change to flyer_items.
|
||||
-- This ensures the item_count on the parent flyer is always accurate.
|
||||
DROP TRIGGER IF EXISTS on_flyer_item_change ON public.flyer_items;
|
||||
CREATE TRIGGER on_flyer_item_change
|
||||
AFTER INSERT OR DELETE ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_flyer_item_count();
|
||||
|
||||
-- 7. Trigger function to log the creation of a new flyer.
|
||||
DROP FUNCTION IF EXISTS public.log_new_flyer();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_flyer()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (action, display_text, icon, details)
|
||||
VALUES (
|
||||
'flyer_uploaded',
|
||||
'A new flyer for ' || (SELECT name FROM public.stores WHERE store_id = NEW.store_id) || ' has been uploaded.',
|
||||
'file-text',
|
||||
jsonb_build_object(
|
||||
'flyer_id', NEW.flyer_id,
|
||||
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id),
|
||||
'valid_from', to_char(NEW.valid_from, 'YYYY-MM-DD'),
|
||||
'valid_to', to_char(NEW.valid_to, 'YYYY-MM-DD')
|
||||
)
|
||||
);
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a new flyer is inserted.
|
||||
DROP TRIGGER IF EXISTS on_new_flyer_created ON public.flyers;
|
||||
CREATE TRIGGER on_new_flyer_created
|
||||
AFTER INSERT ON public.flyers
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_flyer();
|
||||
|
||||
-- 8. Trigger function to log when a user favorites a recipe.
|
||||
DROP FUNCTION IF EXISTS public.log_new_favorite_recipe();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_favorite_recipe()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.user_id,
|
||||
'recipe_favorited',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.user_id) || ' favorited the recipe: ' || (SELECT name FROM public.recipes WHERE recipe_id = NEW.recipe_id),
|
||||
'heart',
|
||||
jsonb_build_object(
|
||||
'recipe_id', NEW.recipe_id
|
||||
)
|
||||
);
|
||||
|
||||
-- Award 'First Favorite' achievement.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Favorite');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a recipe is favorited.
|
||||
DROP TRIGGER IF EXISTS on_new_favorite_recipe ON public.favorite_recipes;
|
||||
CREATE TRIGGER on_new_favorite_recipe
|
||||
AFTER INSERT ON public.favorite_recipes
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_favorite_recipe();
|
||||
|
||||
-- 9. Trigger function to log when a user shares a shopping list.
|
||||
DROP FUNCTION IF EXISTS public.log_new_list_share();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_list_share()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.shared_by_user_id,
|
||||
'list_shared',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.shared_by_user_id) || ' shared a shopping list.',
|
||||
'share-2',
|
||||
jsonb_build_object(
|
||||
'shopping_list_id', NEW.shopping_list_id,
|
||||
'list_name', (SELECT name FROM public.shopping_lists WHERE shopping_list_id = NEW.shopping_list_id),
|
||||
'shared_with_user_id', NEW.shared_with_user_id
|
||||
)
|
||||
);
|
||||
|
||||
-- Award 'List Sharer' achievement.
|
||||
PERFORM public.award_achievement(NEW.shared_by_user_id, 'List Sharer');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a shopping list is shared.
|
||||
DROP TRIGGER IF EXISTS on_new_list_share ON public.shared_shopping_lists;
|
||||
CREATE TRIGGER on_new_list_share
|
||||
AFTER INSERT ON public.shared_shopping_lists
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_list_share();
|
||||
|
||||
-- 9a. Trigger function to log when a user shares a recipe collection.
|
||||
DROP FUNCTION IF EXISTS public.log_new_recipe_collection_share();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_recipe_collection_share()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- Log the activity
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.shared_by_user_id, 'recipe_collection_shared',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.shared_by_user_id) || ' shared a recipe collection.',
|
||||
'book',
|
||||
jsonb_build_object('collection_id', NEW.recipe_collection_id, 'shared_with_user_id', NEW.shared_with_user_id)
|
||||
);
|
||||
|
||||
-- Award 'Recipe Sharer' achievement.
|
||||
PERFORM public.award_achievement(NEW.shared_by_user_id, 'Recipe Sharer');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
DROP TRIGGER IF EXISTS on_new_recipe_collection_share ON public.shared_recipe_collections;
|
||||
CREATE TRIGGER on_new_recipe_collection_share
|
||||
AFTER INSERT ON public.shared_recipe_collections
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_recipe_collection_share();
|
||||
|
||||
-- 10. Trigger function to geocode a store location's address.
|
||||
-- This function is designed to be extensible. In a production environment,
|
||||
-- you would replace the placeholder with a call to an external geocoding service
|
||||
-- (e.g., using the `http` extension or a `plpythonu` function) to convert
|
||||
-- the address into geographic coordinates.
|
||||
DROP FUNCTION IF EXISTS public.geocode_store_location();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.geocode_store_location()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
full_address TEXT;
|
||||
BEGIN
|
||||
-- Only proceed if the address has actually changed.
|
||||
IF TG_OP = 'INSERT' OR (TG_OP = 'UPDATE' AND NEW.address IS DISTINCT FROM OLD.address) THEN
|
||||
-- Concatenate address parts into a single string for the geocoder.
|
||||
full_address := CONCAT_WS(', ', NEW.address, NEW.city, NEW.province_state, NEW.postal_code);
|
||||
|
||||
-- ======================================================================
|
||||
-- Placeholder for Geocoding API Call
|
||||
-- ======================================================================
|
||||
-- In a real application, you would call a geocoding service here.
|
||||
-- For example, using the `http` extension:
|
||||
--
|
||||
-- DECLARE
|
||||
-- response http_get;
|
||||
-- lat NUMERIC;
|
||||
-- lon NUMERIC;
|
||||
-- BEGIN
|
||||
-- SELECT * INTO response FROM http_get('https://api.geocodingservice.com/geocode?address=' || url_encode(full_address));
|
||||
-- lat := (response.content::jsonb)->'results'->0->'geometry'->'location'->'lat';
|
||||
-- lon := (response.content::jsonb)->'results'->0->'geometry'->'location'->'lng';
|
||||
-- NEW.location := ST_SetSRID(ST_MakePoint(lon, lat), 4326)::geography;
|
||||
-- END;
|
||||
--
|
||||
-- For now, this function does nothing, but the trigger is in place.
|
||||
-- If you manually provide lat/lon, you could parse them here.
|
||||
-- For this example, we will assume the `location` might be set manually
|
||||
-- or by a separate batch process.
|
||||
-- ======================================================================
|
||||
END IF;
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the geocoding function.
|
||||
DROP TRIGGER IF EXISTS on_store_location_address_change ON public.store_locations;
|
||||
CREATE TRIGGER on_store_location_address_change
|
||||
BEFORE INSERT OR UPDATE ON public.store_locations
|
||||
FOR EACH ROW EXECUTE FUNCTION public.geocode_store_location();
|
||||
|
||||
-- 11. Trigger function to increment the fork_count on the original recipe.
|
||||
DROP FUNCTION IF EXISTS public.increment_recipe_fork_count();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.increment_recipe_fork_count()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- Only run if the recipe is a fork (original_recipe_id is not null).
|
||||
IF NEW.original_recipe_id IS NOT NULL THEN
|
||||
UPDATE public.recipes SET fork_count = fork_count + 1 WHERE recipe_id = NEW.original_recipe_id;
|
||||
-- Award 'First Fork' achievement.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Fork');
|
||||
END IF;
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
DROP TRIGGER IF EXISTS on_recipe_fork ON public.recipes;
|
||||
CREATE TRIGGER on_recipe_fork
|
||||
AFTER INSERT ON public.recipes
|
||||
FOR EACH ROW EXECUTE FUNCTION public.increment_recipe_fork_count();
|
||||
-- ============================================================================
|
||||
-- PART 6: DATABASE FUNCTIONS
|
||||
-- PART 3: DATABASE FUNCTIONS
|
||||
-- ============================================================================
|
||||
-- Function to find the best current sale price for a user's watched items.
|
||||
DROP FUNCTION IF EXISTS public.get_best_sale_prices_for_user(UUID);
|
||||
@@ -1336,8 +867,7 @@ AS $$
|
||||
'list_shared'
|
||||
-- 'new_recipe_rating' could be added here later
|
||||
)
|
||||
ORDER BY
|
||||
al.created_at DESC
|
||||
ORDER BY al.created_at DESC, al.display_text, al.icon
|
||||
LIMIT p_limit
|
||||
OFFSET p_offset;
|
||||
$$;
|
||||
@@ -1549,16 +1079,18 @@ $$;
|
||||
-- It replaces the need to call get_best_sale_prices_for_user for each user individually.
|
||||
-- Returns: TABLE(...) - A set of records including user details and deal information.
|
||||
-- =================================================================
|
||||
DROP FUNCTION IF EXISTS public.get_best_sale_prices_for_all_users();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.get_best_sale_prices_for_all_users()
|
||||
RETURNS TABLE(
|
||||
user_id uuid,
|
||||
email text,
|
||||
full_name text,
|
||||
master_item_id integer,
|
||||
master_item_id bigint,
|
||||
item_name text,
|
||||
best_price_in_cents integer,
|
||||
store_name text,
|
||||
flyer_id integer,
|
||||
flyer_id bigint,
|
||||
valid_to date
|
||||
) AS $$
|
||||
BEGIN
|
||||
@@ -1569,11 +1101,12 @@ BEGIN
|
||||
SELECT
|
||||
fi.master_item_id,
|
||||
fi.price_in_cents,
|
||||
f.store_name,
|
||||
s.name as store_name,
|
||||
f.flyer_id,
|
||||
f.valid_to
|
||||
FROM public.flyer_items fi
|
||||
JOIN public.flyers f ON fi.flyer_id = f.flyer_id
|
||||
JOIN public.stores s ON f.store_id = s.store_id
|
||||
WHERE
|
||||
fi.master_item_id IS NOT NULL
|
||||
AND fi.price_in_cents IS NOT NULL
|
||||
@@ -1616,3 +1149,472 @@ BEGIN
|
||||
bp.price_rank = 1;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- ============================================================================
|
||||
-- PART 4: TRIGGERS
|
||||
-- ============================================================================
|
||||
|
||||
-- 1. Trigger to automatically create a profile when a new user signs up.
|
||||
-- This function is called by a trigger on the `public.users` table.
|
||||
DROP FUNCTION IF EXISTS public.handle_new_user();
|
||||
|
||||
-- It creates a corresponding profile and a default shopping list for the new user.
|
||||
-- It now accepts full_name and avatar_url from the user's metadata.
|
||||
CREATE OR REPLACE FUNCTION public.handle_new_user()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
new_profile_id UUID;
|
||||
user_meta_data JSONB;
|
||||
BEGIN
|
||||
-- The user's metadata (full_name, avatar_url) is passed via a temporary session variable.
|
||||
user_meta_data := current_setting('my_app.user_metadata', true)::JSONB;
|
||||
|
||||
INSERT INTO public.profiles (user_id, role, full_name, avatar_url)
|
||||
VALUES (new.user_id, 'user', user_meta_data->>'full_name', user_meta_data->>'avatar_url')
|
||||
RETURNING user_id INTO new_profile_id;
|
||||
|
||||
-- Also create a default shopping list for the new user.
|
||||
INSERT INTO public.shopping_lists (user_id, name)
|
||||
VALUES (new.user_id, 'Main Shopping List');
|
||||
|
||||
-- Log the new user event
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (new.user_id, 'user_registered',
|
||||
COALESCE(user_meta_data->>'full_name', new.email) || ' has registered.',
|
||||
'user-plus',
|
||||
jsonb_build_object('email', new.email)
|
||||
);
|
||||
|
||||
RETURN new;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- This trigger calls the function after a new user is created.
|
||||
DROP TRIGGER IF EXISTS on_auth_user_created ON public.users;
|
||||
CREATE TRIGGER on_auth_user_created
|
||||
AFTER INSERT ON public.users
|
||||
FOR EACH ROW EXECUTE FUNCTION public.handle_new_user();
|
||||
|
||||
-- 2. Create a reusable function to automatically update 'updated_at' columns.
|
||||
DROP FUNCTION IF EXISTS public.handle_updated_at();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.handle_updated_at()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
NEW.updated_at = now();
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Dynamically apply the 'handle_updated_at' trigger to all tables in the public schema
|
||||
-- that have an 'updated_at' column. This is more maintainable than creating a separate
|
||||
-- trigger for each table.
|
||||
DO $$
|
||||
DECLARE
|
||||
t_name TEXT;
|
||||
BEGIN
|
||||
FOR t_name IN
|
||||
SELECT table_name
|
||||
FROM information_schema.columns
|
||||
WHERE table_schema = 'public' AND column_name = 'updated_at'
|
||||
LOOP
|
||||
EXECUTE format('DROP TRIGGER IF EXISTS on_%s_updated ON public.%I;
|
||||
CREATE TRIGGER on_%s_updated
|
||||
BEFORE UPDATE ON public.%I
|
||||
FOR EACH ROW EXECUTE FUNCTION public.handle_updated_at();',
|
||||
t_name, t_name, t_name, t_name);
|
||||
END LOOP;
|
||||
END;
|
||||
$$;
|
||||
|
||||
-- 3. Create a trigger function to populate the item_price_history table on insert.
|
||||
DROP FUNCTION IF EXISTS public.update_price_history_on_flyer_item_insert();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_price_history_on_flyer_item_insert()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
flyer_valid_from DATE;
|
||||
flyer_valid_to DATE;
|
||||
current_summary_date DATE;
|
||||
flyer_location_id BIGINT;
|
||||
BEGIN
|
||||
-- If the item could not be matched, add it to the unmatched queue for review.
|
||||
IF NEW.master_item_id IS NULL THEN
|
||||
INSERT INTO public.unmatched_flyer_items (flyer_item_id)
|
||||
VALUES (NEW.flyer_item_id)
|
||||
ON CONFLICT (flyer_item_id) DO NOTHING;
|
||||
END IF;
|
||||
|
||||
-- Only run if the new flyer item is linked to a master item and has a price.
|
||||
IF NEW.master_item_id IS NULL OR NEW.price_in_cents IS NULL THEN
|
||||
RETURN NEW;
|
||||
END IF;
|
||||
|
||||
-- Get the validity dates of the flyer and the store_id.
|
||||
SELECT valid_from, valid_to INTO flyer_valid_from, flyer_valid_to
|
||||
FROM public.flyers
|
||||
WHERE flyer_id = NEW.flyer_id;
|
||||
|
||||
-- This single, set-based query is much more performant than looping.
|
||||
-- It generates all date/location pairs and inserts/updates them in one operation.
|
||||
INSERT INTO public.item_price_history (master_item_id, summary_date, store_location_id, min_price_in_cents, max_price_in_cents, avg_price_in_cents, data_points_count)
|
||||
SELECT
|
||||
NEW.master_item_id,
|
||||
d.day,
|
||||
fl.store_location_id,
|
||||
NEW.price_in_cents,
|
||||
NEW.price_in_cents,
|
||||
NEW.price_in_cents,
|
||||
1
|
||||
FROM public.flyer_locations fl
|
||||
CROSS JOIN generate_series(flyer_valid_from, flyer_valid_to, '1 day'::interval) AS d(day)
|
||||
WHERE fl.flyer_id = NEW.flyer_id
|
||||
ON CONFLICT (master_item_id, summary_date, store_location_id)
|
||||
DO UPDATE SET
|
||||
min_price_in_cents = LEAST(item_price_history.min_price_in_cents, EXCLUDED.min_price_in_cents),
|
||||
max_price_in_cents = GREATEST(item_price_history.max_price_in_cents, EXCLUDED.max_price_in_cents),
|
||||
avg_price_in_cents = ROUND(((item_price_history.avg_price_in_cents * item_price_history.data_points_count) + EXCLUDED.avg_price_in_cents) / (item_price_history.data_points_count + 1.0)),
|
||||
data_points_count = item_price_history.data_points_count + 1;
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Create the trigger on the flyer_items table for insert.
|
||||
DROP TRIGGER IF EXISTS trigger_update_price_history ON public.flyer_items;
|
||||
CREATE TRIGGER trigger_update_price_history
|
||||
AFTER INSERT ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_price_history_on_flyer_item_insert();
|
||||
|
||||
-- 4. Create a trigger function to recalculate price history when a flyer item is deleted.
|
||||
DROP FUNCTION IF EXISTS public.recalculate_price_history_on_flyer_item_delete();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.recalculate_price_history_on_flyer_item_delete()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
affected_dates RECORD;
|
||||
BEGIN
|
||||
-- Only run if the deleted item was linked to a master item and had a price.
|
||||
IF OLD.master_item_id IS NULL OR OLD.price_in_cents IS NULL THEN
|
||||
RETURN OLD;
|
||||
END IF;
|
||||
|
||||
-- This single, set-based query is much more performant than looping.
|
||||
-- It recalculates aggregates for all affected dates and locations at once.
|
||||
WITH affected_days_and_locations AS (
|
||||
-- 1. Get all date/location pairs affected by the deleted item's flyer.
|
||||
SELECT DISTINCT
|
||||
generate_series(f.valid_from, f.valid_to, '1 day'::interval)::date AS summary_date,
|
||||
fl.store_location_id
|
||||
FROM public.flyers f
|
||||
JOIN public.flyer_locations fl ON f.flyer_id = fl.flyer_id
|
||||
WHERE f.flyer_id = OLD.flyer_id
|
||||
),
|
||||
new_aggregates AS (
|
||||
-- 2. For each affected date/location, recalculate the aggregates from all other relevant flyer items.
|
||||
SELECT
|
||||
adl.summary_date,
|
||||
adl.store_location_id,
|
||||
MIN(fi.price_in_cents) AS min_price,
|
||||
MAX(fi.price_in_cents) AS max_price,
|
||||
ROUND(AVG(fi.price_in_cents))::int AS avg_price,
|
||||
COUNT(fi.flyer_item_id)::int AS data_points
|
||||
FROM affected_days_and_locations adl
|
||||
LEFT JOIN public.flyer_items fi ON fi.master_item_id = OLD.master_item_id AND fi.price_in_cents IS NOT NULL
|
||||
LEFT JOIN public.flyers f ON fi.flyer_id = f.flyer_id AND adl.summary_date BETWEEN f.valid_from AND f.valid_to
|
||||
LEFT JOIN public.flyer_locations fl ON fi.flyer_id = fl.flyer_id AND adl.store_location_id = fl.store_location_id
|
||||
WHERE fl.flyer_id IS NOT NULL -- Ensure the join was successful
|
||||
GROUP BY adl.summary_date, adl.store_location_id
|
||||
)
|
||||
-- 3. Update the history table with the new aggregates.
|
||||
UPDATE public.item_price_history iph
|
||||
SET
|
||||
min_price_in_cents = na.min_price,
|
||||
max_price_in_cents = na.max_price,
|
||||
avg_price_in_cents = na.avg_price,
|
||||
data_points_count = na.data_points
|
||||
FROM new_aggregates na
|
||||
WHERE iph.master_item_id = OLD.master_item_id
|
||||
AND iph.summary_date = na.summary_date
|
||||
AND iph.store_location_id = na.store_location_id;
|
||||
|
||||
-- 4. Delete any history records that no longer have any data points.
|
||||
DELETE FROM public.item_price_history iph
|
||||
WHERE iph.master_item_id = OLD.master_item_id
|
||||
AND NOT EXISTS (
|
||||
SELECT 1 FROM new_aggregates na
|
||||
WHERE na.summary_date = iph.summary_date AND na.store_location_id = iph.store_location_id
|
||||
);
|
||||
|
||||
RETURN OLD;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Create the trigger on the flyer_items table for DELETE operations.
|
||||
DROP TRIGGER IF EXISTS trigger_recalculate_price_history_on_delete ON public.flyer_items;
|
||||
CREATE TRIGGER trigger_recalculate_price_history_on_delete
|
||||
AFTER DELETE ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.recalculate_price_history_on_flyer_item_delete();
|
||||
|
||||
-- 5. Trigger function to update the average rating on the recipes table.
|
||||
DROP FUNCTION IF EXISTS public.update_recipe_rating_aggregates();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_recipe_rating_aggregates()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
UPDATE public.recipes
|
||||
SET
|
||||
avg_rating = (
|
||||
SELECT AVG(rating)
|
||||
FROM public.recipe_ratings
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id) -- This is correct, no change needed
|
||||
),
|
||||
rating_count = (
|
||||
SELECT COUNT(*)
|
||||
FROM public.recipe_ratings
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id) -- This is correct, no change needed
|
||||
)
|
||||
WHERE recipe_id = COALESCE(NEW.recipe_id, OLD.recipe_id);
|
||||
|
||||
RETURN NULL; -- The result is ignored since this is an AFTER trigger.
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after any change to recipe_ratings.
|
||||
DROP TRIGGER IF EXISTS on_recipe_rating_change ON public.recipe_ratings;
|
||||
CREATE TRIGGER on_recipe_rating_change
|
||||
AFTER INSERT OR UPDATE OR DELETE ON public.recipe_ratings
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_recipe_rating_aggregates();
|
||||
|
||||
-- 6. Trigger function to log the creation of a new recipe.
|
||||
DROP FUNCTION IF EXISTS public.log_new_recipe();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_recipe()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.user_id,
|
||||
'recipe_created',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.user_id) || ' created a new recipe: ' || NEW.name,
|
||||
'chef-hat',
|
||||
jsonb_build_object('recipe_id', NEW.recipe_id, 'recipe_name', NEW.name)
|
||||
);
|
||||
|
||||
-- Award 'First Recipe' achievement if it's their first one.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Recipe');
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a new recipe is inserted.
|
||||
DROP TRIGGER IF EXISTS on_new_recipe_created ON public.recipes;
|
||||
CREATE TRIGGER on_new_recipe_created
|
||||
AFTER INSERT ON public.recipes
|
||||
FOR EACH ROW
|
||||
WHEN (NEW.user_id IS NOT NULL) -- Only log activity for user-created recipes.
|
||||
EXECUTE FUNCTION public.log_new_recipe();
|
||||
|
||||
-- 7a. Trigger function to update the item_count on the flyers table.
|
||||
DROP FUNCTION IF EXISTS public.update_flyer_item_count();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.update_flyer_item_count()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
IF (TG_OP = 'INSERT') THEN
|
||||
UPDATE public.flyers SET item_count = item_count + 1 WHERE flyer_id = NEW.flyer_id;
|
||||
ELSIF (TG_OP = 'DELETE') THEN
|
||||
UPDATE public.flyers SET item_count = item_count - 1 WHERE flyer_id = OLD.flyer_id;
|
||||
END IF;
|
||||
RETURN NULL; -- The result is ignored since this is an AFTER trigger.
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after any change to flyer_items.
|
||||
-- This ensures the item_count on the parent flyer is always accurate.
|
||||
DROP TRIGGER IF EXISTS on_flyer_item_change ON public.flyer_items;
|
||||
CREATE TRIGGER on_flyer_item_change
|
||||
AFTER INSERT OR DELETE ON public.flyer_items
|
||||
FOR EACH ROW EXECUTE FUNCTION public.update_flyer_item_count();
|
||||
|
||||
-- 7. Trigger function to log the creation of a new flyer.
|
||||
DROP FUNCTION IF EXISTS public.log_new_flyer();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_flyer()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- If the flyer was uploaded by a registered user, award the 'First-Upload' achievement.
|
||||
-- The award_achievement function handles checking if the user already has it.
|
||||
IF NEW.uploaded_by IS NOT NULL THEN
|
||||
PERFORM public.award_achievement(NEW.uploaded_by, 'First-Upload');
|
||||
END IF;
|
||||
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.uploaded_by, -- Log the user who uploaded it
|
||||
'flyer_uploaded',
|
||||
'A new flyer for ' || (SELECT name FROM public.stores WHERE store_id = NEW.store_id) || ' has been uploaded.',
|
||||
'file-text',
|
||||
jsonb_build_object(
|
||||
'flyer_id', NEW.flyer_id,
|
||||
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id),
|
||||
'valid_from', to_char(NEW.valid_from, 'YYYY-MM-DD'),
|
||||
'valid_to', to_char(NEW.valid_to, 'YYYY-MM-DD')
|
||||
)
|
||||
);
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a new flyer is inserted.
|
||||
DROP TRIGGER IF EXISTS on_new_flyer_created ON public.flyers;
|
||||
CREATE TRIGGER on_new_flyer_created
|
||||
AFTER INSERT ON public.flyers
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_flyer();
|
||||
|
||||
-- 8. Trigger function to log when a user favorites a recipe.
|
||||
DROP FUNCTION IF EXISTS public.log_new_favorite_recipe();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_favorite_recipe()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.user_id,
|
||||
'recipe_favorited',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.user_id) || ' favorited the recipe: ' || (SELECT name FROM public.recipes WHERE recipe_id = NEW.recipe_id),
|
||||
'heart',
|
||||
jsonb_build_object(
|
||||
'recipe_id', NEW.recipe_id
|
||||
)
|
||||
);
|
||||
|
||||
-- Award 'First Favorite' achievement.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Favorite');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a recipe is favorited.
|
||||
DROP TRIGGER IF EXISTS on_new_favorite_recipe ON public.favorite_recipes;
|
||||
CREATE TRIGGER on_new_favorite_recipe
|
||||
AFTER INSERT ON public.favorite_recipes
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_favorite_recipe();
|
||||
|
||||
-- 9. Trigger function to log when a user shares a shopping list.
|
||||
DROP FUNCTION IF EXISTS public.log_new_list_share();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_list_share()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.shared_by_user_id,
|
||||
'list_shared',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.shared_by_user_id) || ' shared a shopping list.',
|
||||
'share-2',
|
||||
jsonb_build_object(
|
||||
'shopping_list_id', NEW.shopping_list_id,
|
||||
'list_name', (SELECT name FROM public.shopping_lists WHERE shopping_list_id = NEW.shopping_list_id),
|
||||
'shared_with_user_id', NEW.shared_with_user_id
|
||||
)
|
||||
);
|
||||
|
||||
-- Award 'List Sharer' achievement.
|
||||
PERFORM public.award_achievement(NEW.shared_by_user_id, 'List Sharer');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- Trigger to call the function after a shopping list is shared.
|
||||
DROP TRIGGER IF EXISTS on_new_list_share ON public.shared_shopping_lists;
|
||||
CREATE TRIGGER on_new_list_share
|
||||
AFTER INSERT ON public.shared_shopping_lists
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_list_share();
|
||||
|
||||
-- 9a. Trigger function to log when a user shares a recipe collection.
|
||||
DROP FUNCTION IF EXISTS public.log_new_recipe_collection_share();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.log_new_recipe_collection_share()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- Log the activity
|
||||
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||
VALUES (
|
||||
NEW.shared_by_user_id, 'recipe_collection_shared',
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.shared_by_user_id) || ' shared a recipe collection.',
|
||||
'book',
|
||||
jsonb_build_object('collection_id', NEW.recipe_collection_id, 'shared_with_user_id', NEW.shared_with_user_id)
|
||||
);
|
||||
|
||||
-- Award 'Recipe Sharer' achievement.
|
||||
PERFORM public.award_achievement(NEW.shared_by_user_id, 'Recipe Sharer');
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
DROP TRIGGER IF EXISTS on_new_recipe_collection_share ON public.shared_recipe_collections;
|
||||
CREATE TRIGGER on_new_recipe_collection_share
|
||||
AFTER INSERT ON public.shared_recipe_collections
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_recipe_collection_share();
|
||||
|
||||
-- 10. Trigger function to geocode a store location's address.
|
||||
-- This function is triggered when an address is inserted or updated, and is
|
||||
-- designed to be extensible for external geocoding services to populate the
|
||||
-- latitude, longitude, and location fields.
|
||||
DROP FUNCTION IF EXISTS public.geocode_address();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.geocode_address()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
full_address TEXT;
|
||||
BEGIN
|
||||
-- Only proceed if an address component has actually changed.
|
||||
IF TG_OP = 'INSERT' OR (TG_OP = 'UPDATE' AND (
|
||||
NEW.address_line_1 IS DISTINCT FROM OLD.address_line_1 OR
|
||||
NEW.address_line_2 IS DISTINCT FROM OLD.address_line_2 OR
|
||||
NEW.city IS DISTINCT FROM OLD.city OR
|
||||
NEW.province_state IS DISTINCT FROM OLD.province_state OR
|
||||
NEW.postal_code IS DISTINCT FROM OLD.postal_code OR
|
||||
NEW.country IS DISTINCT FROM OLD.country
|
||||
)) THEN
|
||||
-- Concatenate address parts into a single string for the geocoder.
|
||||
full_address := CONCAT_WS(', ', NEW.address_line_1, NEW.address_line_2, NEW.city, NEW.province_state, NEW.postal_code, NEW.country);
|
||||
|
||||
-- Placeholder for Geocoding API Call
|
||||
-- In a real application, you would call a service here and update NEW.latitude, NEW.longitude, and NEW.location.
|
||||
-- e.g., NEW.latitude := result.lat; NEW.longitude := result.lon;
|
||||
-- NEW.location := ST_SetSRID(ST_MakePoint(NEW.longitude, NEW.latitude), 4326);
|
||||
END IF;
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- This trigger calls the geocoding function when an address changes.
|
||||
DROP TRIGGER IF EXISTS on_address_change_geocode ON public.addresses;
|
||||
CREATE TRIGGER on_address_change_geocode
|
||||
BEFORE INSERT OR UPDATE ON public.addresses
|
||||
FOR EACH ROW EXECUTE FUNCTION public.geocode_address();
|
||||
|
||||
-- 11. Trigger function to increment the fork_count on the original recipe.
|
||||
DROP FUNCTION IF EXISTS public.increment_recipe_fork_count();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.increment_recipe_fork_count()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- Only run if the recipe is a fork (original_recipe_id is not null).
|
||||
IF NEW.original_recipe_id IS NOT NULL THEN
|
||||
UPDATE public.recipes SET fork_count = fork_count + 1 WHERE recipe_id = NEW.original_recipe_id;
|
||||
-- Award 'First Fork' achievement.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Fork');
|
||||
END IF;
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
DROP TRIGGER IF EXISTS on_recipe_fork ON public.recipes;
|
||||
CREATE TRIGGER on_recipe_fork
|
||||
AFTER INSERT ON public.recipes
|
||||
FOR EACH ROW EXECUTE FUNCTION public.increment_recipe_fork_count();
|
||||
|
||||
@@ -265,5 +265,6 @@ INSERT INTO public.achievements (name, description, icon, points_value) VALUES
|
||||
('List Sharer', 'Share a shopping list with another user for the first time.', 'list', 20),
|
||||
('First Favorite', 'Mark a recipe as one of your favorites.', 'heart', 5),
|
||||
('First Fork', 'Make a personal copy of a public recipe.', 'git-fork', 10),
|
||||
('First Budget Created', 'Create your first budget to track spending.', 'piggy-bank', 15)
|
||||
('First Budget Created', 'Create your first budget to track spending.', 'piggy-bank', 15),
|
||||
('First-Upload', 'Upload your first flyer.', 'upload-cloud', 25)
|
||||
ON CONFLICT (name) DO NOTHING;
|
||||
|
||||
@@ -162,7 +162,6 @@ COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer.
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
@@ -973,6 +972,21 @@ COMMENT ON COLUMN public.user_reactions.reaction_type IS 'The type of reaction (
|
||||
CREATE INDEX IF NOT EXISTS idx_user_reactions_user_id ON public.user_reactions(user_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_user_reactions_entity ON public.user_reactions(entity_type, entity_id);
|
||||
|
||||
-- 56. Store user-defined budgets for spending analysis.
|
||||
CREATE TABLE IF NOT EXISTS public.budgets (
|
||||
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||
name TEXT NOT NULL,
|
||||
amount_cents INTEGER NOT NULL CHECK (amount_cents > 0),
|
||||
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
||||
start_date DATE NOT NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT budgets_name_check CHECK (TRIM(name) <> '')
|
||||
);
|
||||
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
||||
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
||||
|
||||
-- 57. Static table defining available achievements for gamification.
|
||||
CREATE TABLE IF NOT EXISTS public.achievements (
|
||||
achievement_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
@@ -998,17 +1012,3 @@ CREATE INDEX IF NOT EXISTS idx_user_achievements_user_id ON public.user_achievem
|
||||
CREATE INDEX IF NOT EXISTS idx_user_achievements_achievement_id ON public.user_achievements(achievement_id);
|
||||
|
||||
|
||||
-- 56. Store user-defined budgets for spending analysis.
|
||||
CREATE TABLE IF NOT EXISTS public.budgets (
|
||||
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||
name TEXT NOT NULL,
|
||||
amount_cents INTEGER NOT NULL CHECK (amount_cents > 0),
|
||||
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
||||
start_date DATE NOT NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT budgets_name_check CHECK (TRIM(name) <> '')
|
||||
);
|
||||
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
||||
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
||||
|
||||
@@ -102,11 +102,11 @@ CREATE TABLE IF NOT EXISTS public.profiles (
|
||||
address_id BIGINT REFERENCES public.addresses(address_id) ON DELETE SET NULL,
|
||||
points INTEGER DEFAULT 0 NOT NULL CHECK (points >= 0),
|
||||
preferences JSONB,
|
||||
role TEXT CHECK (role IN ('admin', 'user')),
|
||||
role TEXT NOT NULL CHECK (role IN ('admin', 'user')),
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT profiles_full_name_check CHECK (full_name IS NULL OR TRIM(full_name) <> ''),
|
||||
CONSTRAINT profiles_avatar_url_check CHECK (avatar_url IS NULL OR avatar_url ~* '^https://?.*'),
|
||||
CONSTRAINT profiles_avatar_url_check CHECK (avatar_url IS NULL OR avatar_url ~* '^https?://.*'),
|
||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||
updated_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||
);
|
||||
@@ -124,7 +124,7 @@ CREATE TABLE IF NOT EXISTS public.stores (
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT stores_name_check CHECK (TRIM(name) <> ''),
|
||||
CONSTRAINT stores_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*'),
|
||||
CONSTRAINT stores_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https?://.*'),
|
||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||
);
|
||||
COMMENT ON TABLE public.stores IS 'Stores metadata for grocery store chains (e.g., Safeway, Kroger).';
|
||||
@@ -144,7 +144,7 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
flyer_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
file_name TEXT NOT NULL,
|
||||
image_url TEXT NOT NULL,
|
||||
icon_url TEXT,
|
||||
icon_url TEXT NOT NULL,
|
||||
checksum TEXT UNIQUE,
|
||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||
valid_from DATE,
|
||||
@@ -157,8 +157,8 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT flyers_valid_dates_check CHECK (valid_to >= valid_from),
|
||||
CONSTRAINT flyers_file_name_check CHECK (TRIM(file_name) <> ''),
|
||||
CONSTRAINT flyers_image_url_check CHECK (image_url ~* '^https://?.*'),
|
||||
CONSTRAINT flyers_icon_url_check CHECK (icon_url IS NULL OR icon_url ~* '^https://?.*'),
|
||||
CONSTRAINT flyers_image_url_check CHECK (image_url ~* '^https?://.*'),
|
||||
CONSTRAINT flyers_icon_url_check CHECK (icon_url ~* '^https?://.*'),
|
||||
CONSTRAINT flyers_checksum_check CHECK (checksum IS NULL OR length(checksum) = 64)
|
||||
);
|
||||
COMMENT ON TABLE public.flyers IS 'Stores metadata for each processed flyer, linking it to a store and its validity period.';
|
||||
@@ -215,7 +215,7 @@ CREATE TABLE IF NOT EXISTS public.brands (
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT brands_name_check CHECK (TRIM(name) <> ''),
|
||||
CONSTRAINT brands_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*')
|
||||
CONSTRAINT brands_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https?://.*')
|
||||
);
|
||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||
@@ -482,7 +482,7 @@ CREATE TABLE IF NOT EXISTS public.user_submitted_prices (
|
||||
downvotes INTEGER DEFAULT 0 NOT NULL CHECK (downvotes >= 0),
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT user_submitted_prices_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||
CONSTRAINT user_submitted_prices_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https?://.*')
|
||||
);
|
||||
COMMENT ON TABLE public.user_submitted_prices IS 'Stores item prices submitted by users directly from physical stores.';
|
||||
COMMENT ON COLUMN public.user_submitted_prices.photo_url IS 'URL to user-submitted photo evidence of the price.';
|
||||
@@ -539,7 +539,7 @@ CREATE TABLE IF NOT EXISTS public.recipes (
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT recipes_name_check CHECK (TRIM(name) <> ''),
|
||||
CONSTRAINT recipes_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||
CONSTRAINT recipes_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https?://.*')
|
||||
);
|
||||
COMMENT ON TABLE public.recipes IS 'Stores recipes that can be used to generate shopping lists.';
|
||||
COMMENT ON COLUMN public.recipes.servings IS 'The number of servings this recipe yields.';
|
||||
@@ -689,8 +689,8 @@ CREATE TABLE IF NOT EXISTS public.planned_meals (
|
||||
meal_type TEXT NOT NULL,
|
||||
servings_to_cook INTEGER,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT planned_meals_meal_type_check CHECK (TRIM(meal_type) <> ''),
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
CONSTRAINT planned_meals_meal_type_check CHECK (TRIM(meal_type) <> '')
|
||||
);
|
||||
COMMENT ON TABLE public.planned_meals IS 'Assigns a recipe to a specific day and meal type within a user''s menu plan.';
|
||||
COMMENT ON COLUMN public.planned_meals.meal_type IS 'The designated meal for the recipe, e.g., ''Breakfast'', ''Lunch'', ''Dinner''.';
|
||||
@@ -940,7 +940,7 @@ CREATE TABLE IF NOT EXISTS public.receipts (
|
||||
raw_text TEXT,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
processed_at TIMESTAMPTZ,
|
||||
CONSTRAINT receipts_receipt_image_url_check CHECK (receipt_image_url ~* '^https://?.*'),
|
||||
CONSTRAINT receipts_receipt_image_url_check CHECK (receipt_image_url ~* '^https?://.*'),
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.receipts IS 'Stores uploaded user receipts for purchase tracking and analysis.';
|
||||
@@ -1113,6 +1113,7 @@ DECLARE
|
||||
ground_beef_id BIGINT; pasta_item_id BIGINT; tomatoes_id BIGINT; onions_id BIGINT; garlic_id BIGINT;
|
||||
bell_peppers_id BIGINT; carrots_id BIGINT; soy_sauce_id BIGINT;
|
||||
soda_item_id BIGINT; turkey_item_id BIGINT; bread_item_id BIGINT; cheese_item_id BIGINT;
|
||||
chicken_thighs_id BIGINT; paper_towels_id BIGINT; toilet_paper_id BIGINT;
|
||||
|
||||
-- Tag IDs
|
||||
quick_easy_tag BIGINT; healthy_tag BIGINT; chicken_tag BIGINT;
|
||||
@@ -1164,6 +1165,9 @@ BEGIN
|
||||
SELECT mgi.master_grocery_item_id INTO turkey_item_id FROM public.master_grocery_items mgi WHERE mgi.name = 'turkey';
|
||||
SELECT mgi.master_grocery_item_id INTO bread_item_id FROM public.master_grocery_items mgi WHERE mgi.name = 'bread';
|
||||
SELECT mgi.master_grocery_item_id INTO cheese_item_id FROM public.master_grocery_items mgi WHERE mgi.name = 'cheese';
|
||||
SELECT mgi.master_grocery_item_id INTO chicken_thighs_id FROM public.master_grocery_items mgi WHERE mgi.name = 'chicken thighs';
|
||||
SELECT mgi.master_grocery_item_id INTO paper_towels_id FROM public.master_grocery_items mgi WHERE mgi.name = 'paper towels';
|
||||
SELECT mgi.master_grocery_item_id INTO toilet_paper_id FROM public.master_grocery_items mgi WHERE mgi.name = 'toilet paper';
|
||||
|
||||
-- Insert ingredients for each recipe
|
||||
INSERT INTO public.recipe_ingredients (recipe_id, master_item_id, quantity, unit) VALUES
|
||||
@@ -1200,6 +1204,17 @@ BEGIN
|
||||
(bolognese_recipe_id, family_tag), (bolognese_recipe_id, beef_tag), (bolognese_recipe_id, weeknight_tag),
|
||||
(stir_fry_recipe_id, quick_easy_tag), (stir_fry_recipe_id, healthy_tag), (stir_fry_recipe_id, vegetarian_tag)
|
||||
ON CONFLICT (recipe_id, tag_id) DO NOTHING;
|
||||
|
||||
INSERT INTO public.master_item_aliases (master_item_id, alias) VALUES
|
||||
(ground_beef_id, 'ground chuck'), (ground_beef_id, 'lean ground beef'),
|
||||
(ground_beef_id, 'extra lean ground beef'), (ground_beef_id, 'hamburger meat'),
|
||||
(chicken_breast_id, 'boneless skinless chicken breast'), (chicken_breast_id, 'chicken cutlets'),
|
||||
(chicken_thighs_id, 'boneless skinless chicken thighs'), (chicken_thighs_id, 'bone-in chicken thighs'),
|
||||
(bell_peppers_id, 'red pepper'), (bell_peppers_id, 'green pepper'), (bell_peppers_id, 'yellow pepper'), (bell_peppers_id, 'orange pepper'),
|
||||
(soda_item_id, 'pop'), (soda_item_id, 'soft drink'), (soda_item_id, 'coke'), (soda_item_id, 'pepsi'),
|
||||
(paper_towels_id, 'paper towel'),
|
||||
(toilet_paper_id, 'bathroom tissue'), (toilet_paper_id, 'toilet tissue')
|
||||
ON CONFLICT (alias) DO NOTHING;
|
||||
END $$;
|
||||
|
||||
-- Pre-populate the unit_conversions table with common cooking conversions.
|
||||
@@ -2115,6 +2130,61 @@ AS $$
|
||||
ORDER BY potential_savings_cents DESC;
|
||||
$$;
|
||||
|
||||
-- Function to get a user's spending breakdown by category for a given date range.
|
||||
DROP FUNCTION IF EXISTS public.get_spending_by_category(UUID, DATE, DATE);
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.get_spending_by_category(p_user_id UUID, p_start_date DATE, p_end_date DATE)
|
||||
RETURNS TABLE (
|
||||
category_id BIGINT,
|
||||
category_name TEXT,
|
||||
total_spent_cents BIGINT
|
||||
)
|
||||
LANGUAGE sql
|
||||
STABLE
|
||||
SECURITY INVOKER
|
||||
AS $$
|
||||
WITH all_purchases AS (
|
||||
-- CTE 1: Combine purchases from completed shopping trips.
|
||||
-- We only consider items that have a price paid.
|
||||
SELECT
|
||||
sti.master_item_id,
|
||||
sti.price_paid_cents
|
||||
FROM public.shopping_trip_items sti
|
||||
JOIN public.shopping_trips st ON sti.shopping_trip_id = st.shopping_trip_id
|
||||
WHERE st.user_id = p_user_id
|
||||
AND st.completed_at::date BETWEEN p_start_date AND p_end_date
|
||||
AND sti.price_paid_cents IS NOT NULL
|
||||
|
||||
UNION ALL
|
||||
|
||||
-- CTE 2: Combine purchases from processed receipts.
|
||||
SELECT
|
||||
ri.master_item_id,
|
||||
ri.price_paid_cents
|
||||
FROM public.receipt_items ri
|
||||
JOIN public.receipts r ON ri.receipt_id = r.receipt_id
|
||||
WHERE r.user_id = p_user_id
|
||||
AND r.transaction_date::date BETWEEN p_start_date AND p_end_date
|
||||
AND ri.master_item_id IS NOT NULL -- Only include items matched to a master item
|
||||
)
|
||||
-- Final Aggregation: Group all combined purchases by category and sum the spending.
|
||||
SELECT
|
||||
c.category_id,
|
||||
c.name AS category_name,
|
||||
SUM(ap.price_paid_cents)::BIGINT AS total_spent_cents
|
||||
FROM all_purchases ap
|
||||
-- Join with master_grocery_items to get the category_id for each purchase.
|
||||
JOIN public.master_grocery_items mgi ON ap.master_item_id = mgi.master_grocery_item_id
|
||||
-- Join with categories to get the category name for display.
|
||||
JOIN public.categories c ON mgi.category_id = c.category_id
|
||||
GROUP BY
|
||||
c.category_id, c.name
|
||||
HAVING
|
||||
SUM(ap.price_paid_cents) > 0
|
||||
ORDER BY
|
||||
total_spent_cents DESC;
|
||||
$$;
|
||||
|
||||
-- Function to approve a suggested correction and apply it.
|
||||
DROP FUNCTION IF EXISTS public.approve_correction(BIGINT);
|
||||
|
||||
@@ -2572,7 +2642,9 @@ BEGIN
|
||||
'file-text',
|
||||
jsonb_build_object(
|
||||
'flyer_id', NEW.flyer_id,
|
||||
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id)
|
||||
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id),
|
||||
'valid_from', to_char(NEW.valid_from, 'YYYY-MM-DD'),
|
||||
'valid_to', to_char(NEW.valid_to, 'YYYY-MM-DD')
|
||||
)
|
||||
);
|
||||
RETURN NEW;
|
||||
@@ -2622,6 +2694,7 @@ BEGIN
|
||||
(SELECT full_name FROM public.profiles WHERE user_id = NEW.shared_by_user_id) || ' shared a shopping list.',
|
||||
'share-2',
|
||||
jsonb_build_object(
|
||||
'shopping_list_id', NEW.shopping_list_id,
|
||||
'list_name', (SELECT name FROM public.shopping_lists WHERE shopping_list_id = NEW.shopping_list_id),
|
||||
'shared_with_user_id', NEW.shared_with_user_id
|
||||
)
|
||||
@@ -2669,6 +2742,66 @@ CREATE TRIGGER on_new_recipe_collection_share
|
||||
AFTER INSERT ON public.shared_recipe_collections
|
||||
FOR EACH ROW EXECUTE FUNCTION public.log_new_recipe_collection_share();
|
||||
|
||||
-- 10. Trigger function to geocode a store location's address.
|
||||
-- This function is triggered when an address is inserted or updated, and is
|
||||
-- designed to be extensible for external geocoding services to populate the
|
||||
-- latitude, longitude, and location fields.
|
||||
DROP FUNCTION IF EXISTS public.geocode_address();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.geocode_address()
|
||||
RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
full_address TEXT;
|
||||
BEGIN
|
||||
-- Only proceed if an address component has actually changed.
|
||||
IF TG_OP = 'INSERT' OR (TG_OP = 'UPDATE' AND (
|
||||
NEW.address_line_1 IS DISTINCT FROM OLD.address_line_1 OR
|
||||
NEW.address_line_2 IS DISTINCT FROM OLD.address_line_2 OR
|
||||
NEW.city IS DISTINCT FROM OLD.city OR
|
||||
NEW.province_state IS DISTINCT FROM OLD.province_state OR
|
||||
NEW.postal_code IS DISTINCT FROM OLD.postal_code OR
|
||||
NEW.country IS DISTINCT FROM OLD.country
|
||||
)) THEN
|
||||
-- Concatenate address parts into a single string for the geocoder.
|
||||
full_address := CONCAT_WS(', ', NEW.address_line_1, NEW.address_line_2, NEW.city, NEW.province_state, NEW.postal_code, NEW.country);
|
||||
|
||||
-- Placeholder for Geocoding API Call.
|
||||
-- In a real application, you would call a service here and update NEW.latitude, NEW.longitude, and NEW.location.
|
||||
-- e.g., NEW.latitude := result.lat; NEW.longitude := result.lon;
|
||||
-- NEW.location := ST_SetSRID(ST_MakePoint(NEW.longitude, NEW.latitude), 4326);
|
||||
END IF;
|
||||
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
-- This trigger calls the geocoding function when an address changes.
|
||||
DROP TRIGGER IF EXISTS on_address_change_geocode ON public.addresses;
|
||||
CREATE TRIGGER on_address_change_geocode
|
||||
BEFORE INSERT OR UPDATE ON public.addresses
|
||||
FOR EACH ROW EXECUTE FUNCTION public.geocode_address();
|
||||
|
||||
-- 11. Trigger function to increment the fork_count on the original recipe.
|
||||
DROP FUNCTION IF EXISTS public.increment_recipe_fork_count();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.increment_recipe_fork_count()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
-- Only run if the recipe is a fork (original_recipe_id is not null).
|
||||
IF NEW.original_recipe_id IS NOT NULL THEN
|
||||
UPDATE public.recipes SET fork_count = fork_count + 1 WHERE recipe_id = NEW.original_recipe_id;
|
||||
-- Award 'First Fork' achievement.
|
||||
PERFORM public.award_achievement(NEW.user_id, 'First Fork');
|
||||
END IF;
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
DROP TRIGGER IF EXISTS on_recipe_fork ON public.recipes;
|
||||
CREATE TRIGGER on_recipe_fork
|
||||
AFTER INSERT ON public.recipes
|
||||
FOR EACH ROW EXECUTE FUNCTION public.increment_recipe_fork_count();
|
||||
|
||||
-- =================================================================
|
||||
-- Function: get_best_sale_prices_for_all_users()
|
||||
-- Description: Retrieves the best sale price for every item on every user's watchlist.
|
||||
@@ -2676,17 +2809,19 @@ CREATE TRIGGER on_new_recipe_collection_share
|
||||
-- It replaces the need to call get_best_sale_prices_for_user for each user individually.
|
||||
-- Returns: TABLE(...) - A set of records including user details and deal information.
|
||||
-- =================================================================
|
||||
DROP FUNCTION IF EXISTS public.get_best_sale_prices_for_all_users();
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.get_best_sale_prices_for_all_users()
|
||||
RETURNS TABLE(
|
||||
user_id uuid,
|
||||
|
||||
email text,
|
||||
full_name text,
|
||||
master_item_id integer,
|
||||
master_item_id bigint,
|
||||
item_name text,
|
||||
best_price_in_cents integer,
|
||||
store_name text,
|
||||
flyer_id integer,
|
||||
flyer_id bigint,
|
||||
valid_to date
|
||||
) AS $$
|
||||
BEGIN
|
||||
@@ -2698,7 +2833,7 @@ BEGIN
|
||||
SELECT
|
||||
fi.master_item_id,
|
||||
fi.price_in_cents,
|
||||
f.store_name,
|
||||
s.name as store_name,
|
||||
f.flyer_id,
|
||||
f.valid_to
|
||||
FROM public.flyer_items fi
|
||||
|
||||
160
src/App.test.tsx
160
src/App.test.tsx
@@ -20,10 +20,98 @@ import {
|
||||
mockUseUserData,
|
||||
mockUseFlyerItems,
|
||||
} from './tests/setup/mockHooks';
|
||||
import './tests/setup/mockUI';
|
||||
import { useAppInitialization } from './hooks/useAppInitialization';
|
||||
|
||||
// Mock top-level components rendered by App's routes
|
||||
|
||||
vi.mock('./components/Header', () => ({
|
||||
Header: ({ onOpenProfile, onOpenVoiceAssistant }: any) => (
|
||||
<div data-testid="header-mock">
|
||||
<button onClick={onOpenProfile}>Open Profile</button>
|
||||
<button onClick={onOpenVoiceAssistant}>Open Voice Assistant</button>
|
||||
</div>
|
||||
),
|
||||
}));
|
||||
|
||||
vi.mock('./components/Footer', () => ({
|
||||
Footer: () => <div data-testid="footer-mock">Mock Footer</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./layouts/MainLayout', async () => {
|
||||
const { Outlet } = await vi.importActual<typeof import('react-router-dom')>('react-router-dom');
|
||||
return {
|
||||
MainLayout: () => (
|
||||
<div data-testid="main-layout-mock">
|
||||
<Outlet />
|
||||
</div>
|
||||
),
|
||||
};
|
||||
});
|
||||
|
||||
vi.mock('./pages/HomePage', () => ({
|
||||
HomePage: ({ selectedFlyer, onOpenCorrectionTool }: any) => (
|
||||
<div data-testid="home-page-mock" data-selected-flyer-id={selectedFlyer?.flyer_id}>
|
||||
<button onClick={onOpenCorrectionTool}>Open Correction Tool</button>
|
||||
</div>
|
||||
),
|
||||
}));
|
||||
|
||||
vi.mock('./pages/admin/AdminPage', () => ({
|
||||
AdminPage: () => <div data-testid="admin-page-mock">AdminPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/admin/CorrectionsPage', () => ({
|
||||
CorrectionsPage: () => <div data-testid="corrections-page-mock">CorrectionsPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/admin/AdminStatsPage', () => ({
|
||||
AdminStatsPage: () => <div data-testid="admin-stats-page-mock">AdminStatsPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/admin/FlyerReviewPage', () => ({
|
||||
FlyerReviewPage: () => <div data-testid="flyer-review-page-mock">FlyerReviewPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/VoiceLabPage', () => ({
|
||||
VoiceLabPage: () => <div data-testid="voice-lab-page-mock">VoiceLabPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/ResetPasswordPage', () => ({
|
||||
ResetPasswordPage: () => <div data-testid="reset-password-page-mock">ResetPasswordPage</div>,
|
||||
}));
|
||||
|
||||
vi.mock('./pages/admin/components/ProfileManager', () => ({
|
||||
ProfileManager: ({ isOpen, onClose, onProfileUpdate, onLoginSuccess }: any) =>
|
||||
isOpen ? (
|
||||
<div data-testid="profile-manager-mock">
|
||||
<button onClick={onClose}>Close Profile</button>
|
||||
<button onClick={() => onProfileUpdate({ full_name: 'Updated' })}>Update Profile</button>
|
||||
<button onClick={() => onLoginSuccess({}, 'token', false)}>Login</button>
|
||||
</div>
|
||||
) : null,
|
||||
}));
|
||||
|
||||
vi.mock('./features/voice-assistant/VoiceAssistant', () => ({
|
||||
VoiceAssistant: ({ isOpen, onClose }: any) =>
|
||||
isOpen ? (
|
||||
<div data-testid="voice-assistant-mock">
|
||||
<button onClick={onClose}>Close Voice Assistant</button>
|
||||
</div>
|
||||
) : null,
|
||||
}));
|
||||
|
||||
vi.mock('./components/FlyerCorrectionTool', () => ({
|
||||
FlyerCorrectionTool: ({ isOpen, onClose, onDataExtracted }: any) =>
|
||||
isOpen ? (
|
||||
<div data-testid="flyer-correction-tool-mock">
|
||||
<button onClick={onClose}>Close Correction</button>
|
||||
<button onClick={() => onDataExtracted('store_name', 'New Store')}>Extract Store</button>
|
||||
<button onClick={() => onDataExtracted('dates', 'New Dates')}>Extract Dates</button>
|
||||
</div>
|
||||
) : null,
|
||||
}));
|
||||
|
||||
// Mock pdfjs-dist to prevent the "DOMMatrix is not defined" error in JSDOM.
|
||||
// This must be done in any test file that imports App.tsx.
|
||||
vi.mock('pdfjs-dist', () => ({
|
||||
@@ -61,71 +149,6 @@ vi.mock('./hooks/useAuth', async () => {
|
||||
return { useAuth: hooks.mockUseAuth };
|
||||
});
|
||||
|
||||
vi.mock('./components/Footer', async () => {
|
||||
const { MockFooter } = await import('./tests/utils/componentMocks');
|
||||
return { Footer: MockFooter };
|
||||
});
|
||||
|
||||
vi.mock('./components/Header', async () => {
|
||||
const { MockHeader } = await import('./tests/utils/componentMocks');
|
||||
return { Header: MockHeader };
|
||||
});
|
||||
|
||||
vi.mock('./pages/HomePage', async () => {
|
||||
const { MockHomePage } = await import('./tests/utils/componentMocks');
|
||||
return { HomePage: MockHomePage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/admin/AdminPage', async () => {
|
||||
const { MockAdminPage } = await import('./tests/utils/componentMocks');
|
||||
return { AdminPage: MockAdminPage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/admin/CorrectionsPage', async () => {
|
||||
const { MockCorrectionsPage } = await import('./tests/utils/componentMocks');
|
||||
return { CorrectionsPage: MockCorrectionsPage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/admin/AdminStatsPage', async () => {
|
||||
const { MockAdminStatsPage } = await import('./tests/utils/componentMocks');
|
||||
return { AdminStatsPage: MockAdminStatsPage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/VoiceLabPage', async () => {
|
||||
const { MockVoiceLabPage } = await import('./tests/utils/componentMocks');
|
||||
return { VoiceLabPage: MockVoiceLabPage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/ResetPasswordPage', async () => {
|
||||
const { MockResetPasswordPage } = await import('./tests/utils/componentMocks');
|
||||
return { ResetPasswordPage: MockResetPasswordPage };
|
||||
});
|
||||
|
||||
vi.mock('./pages/admin/components/ProfileManager', async () => {
|
||||
const { MockProfileManager } = await import('./tests/utils/componentMocks');
|
||||
return { ProfileManager: MockProfileManager };
|
||||
});
|
||||
|
||||
vi.mock('./features/voice-assistant/VoiceAssistant', async () => {
|
||||
const { MockVoiceAssistant } = await import('./tests/utils/componentMocks');
|
||||
return { VoiceAssistant: MockVoiceAssistant };
|
||||
});
|
||||
|
||||
vi.mock('./components/FlyerCorrectionTool', async () => {
|
||||
const { MockFlyerCorrectionTool } = await import('./tests/utils/componentMocks');
|
||||
return { FlyerCorrectionTool: MockFlyerCorrectionTool };
|
||||
});
|
||||
|
||||
vi.mock('./components/WhatsNewModal', async () => {
|
||||
const { MockWhatsNewModal } = await import('./tests/utils/componentMocks');
|
||||
return { WhatsNewModal: MockWhatsNewModal };
|
||||
});
|
||||
|
||||
vi.mock('./layouts/MainLayout', async () => {
|
||||
const { MockMainLayout } = await import('./tests/utils/componentMocks');
|
||||
return { MainLayout: MockMainLayout };
|
||||
});
|
||||
|
||||
vi.mock('./components/AppGuard', async () => {
|
||||
// We need to use the real useModal hook inside our mock AppGuard
|
||||
const { useModal } = await vi.importActual<typeof import('./hooks/useModal')>('./hooks/useModal');
|
||||
@@ -192,6 +215,7 @@ describe('App Component', () => {
|
||||
mockUseUserData.mockReturnValue({
|
||||
watchedItems: [],
|
||||
shoppingLists: [],
|
||||
isLoadingShoppingLists: false,
|
||||
setWatchedItems: vi.fn(),
|
||||
setShoppingLists: vi.fn(),
|
||||
});
|
||||
@@ -361,12 +385,8 @@ describe('App Component', () => {
|
||||
it('should select a flyer when flyerId is present in the URL', async () => {
|
||||
renderApp(['/flyers/2']);
|
||||
|
||||
// The HomePage mock will be rendered. The important part is that the selection logic
|
||||
// in App.tsx runs and passes the correct `selectedFlyer` prop down.
|
||||
// Since HomePage is mocked, we can't see the direct result, but we can
|
||||
// infer that the logic ran without crashing and the correct route was matched.
|
||||
await waitFor(() => {
|
||||
expect(screen.getByTestId('home-page-mock')).toBeInTheDocument();
|
||||
expect(screen.getByTestId('home-page-mock')).toHaveAttribute('data-selected-flyer-id', '2');
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
12
src/App.tsx
12
src/App.tsx
@@ -1,6 +1,6 @@
|
||||
// src/App.tsx
|
||||
import React, { useState, useCallback, useEffect } from 'react';
|
||||
import { Routes, Route, useParams } from 'react-router-dom';
|
||||
import { Routes, Route, useLocation, matchPath } from 'react-router-dom';
|
||||
import { QueryClient, QueryClientProvider } from '@tanstack/react-query';
|
||||
import * as pdfjsLib from 'pdfjs-dist';
|
||||
import { Footer } from './components/Footer';
|
||||
@@ -45,7 +45,9 @@ function App() {
|
||||
const { flyers } = useFlyers();
|
||||
const [selectedFlyer, setSelectedFlyer] = useState<Flyer | null>(null);
|
||||
const { openModal, closeModal, isModalOpen } = useModal();
|
||||
const params = useParams<{ flyerId?: string }>();
|
||||
const location = useLocation();
|
||||
const match = matchPath('/flyers/:flyerId', location.pathname);
|
||||
const flyerIdFromUrl = match?.params.flyerId;
|
||||
|
||||
// This hook now handles initialization effects (OAuth, version check, theme)
|
||||
// and returns the theme/unit state needed by other components.
|
||||
@@ -57,7 +59,7 @@ function App() {
|
||||
console.log('[App] Render:', {
|
||||
flyersCount: flyers.length,
|
||||
selectedFlyerId: selectedFlyer?.flyer_id,
|
||||
paramsFlyerId: params?.flyerId, // This was a duplicate, fixed.
|
||||
flyerIdFromUrl,
|
||||
authStatus,
|
||||
profileId: userProfile?.user.user_id,
|
||||
});
|
||||
@@ -139,8 +141,6 @@ function App() {
|
||||
|
||||
// New effect to handle routing to a specific flyer ID from the URL
|
||||
useEffect(() => {
|
||||
const flyerIdFromUrl = params.flyerId;
|
||||
|
||||
if (flyerIdFromUrl && flyers.length > 0) {
|
||||
const flyerId = parseInt(flyerIdFromUrl, 10);
|
||||
const flyerToSelect = flyers.find((f) => f.flyer_id === flyerId);
|
||||
@@ -148,7 +148,7 @@ function App() {
|
||||
handleFlyerSelect(flyerToSelect);
|
||||
}
|
||||
}
|
||||
}, [flyers, handleFlyerSelect, selectedFlyer, params.flyerId]);
|
||||
}, [flyers, handleFlyerSelect, selectedFlyer, flyerIdFromUrl]);
|
||||
|
||||
// Read the application version injected at build time.
|
||||
// This will only be available in the production build, not during local development.
|
||||
|
||||
@@ -23,6 +23,7 @@ describe('AchievementsList', () => {
|
||||
points_value: 15,
|
||||
}),
|
||||
createMockUserAchievement({ achievement_id: 3, name: 'Unknown Achievement', icon: 'star' }), // This icon is not in the component's map
|
||||
createMockUserAchievement({ achievement_id: 4, name: 'No Icon Achievement', icon: '' }), // Triggers the fallback for missing name
|
||||
];
|
||||
|
||||
renderWithProviders(<AchievementsList achievements={mockAchievements} />);
|
||||
@@ -41,7 +42,15 @@ describe('AchievementsList', () => {
|
||||
|
||||
// Check achievement with default icon
|
||||
expect(screen.getByText('Unknown Achievement')).toBeInTheDocument();
|
||||
expect(screen.getByText('🏆')).toBeInTheDocument(); // Default icon
|
||||
// We expect at least one trophy (for unknown achievement).
|
||||
// Since we added another one that produces a trophy (No Icon), we use getAllByText.
|
||||
expect(screen.getAllByText('🏆').length).toBeGreaterThan(0);
|
||||
|
||||
// Check achievement with missing icon (empty string)
|
||||
expect(screen.getByText('No Icon Achievement')).toBeInTheDocument();
|
||||
// Verify the specific placeholder class is rendered, ensuring the early return in Icon component is hit
|
||||
const noIconCard = screen.getByText('No Icon Achievement').closest('.bg-white');
|
||||
expect(noIconCard?.querySelector('.icon-placeholder')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should render a message when there are no achievements', () => {
|
||||
|
||||
@@ -252,4 +252,54 @@ describe('FlyerCorrectionTool', () => {
|
||||
expect(mockedNotifyError).toHaveBeenCalledWith('An unknown error occurred.');
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle API failure response (ok: false) correctly', async () => {
|
||||
console.log('TEST: Starting "should handle API failure response (ok: false) correctly"');
|
||||
mockedAiApiClient.rescanImageArea.mockResolvedValue({
|
||||
ok: false,
|
||||
json: async () => ({ message: 'Custom API Error' }),
|
||||
} as Response);
|
||||
|
||||
renderWithProviders(<FlyerCorrectionTool {...defaultProps} />);
|
||||
|
||||
// Wait for image fetch
|
||||
await waitFor(() => expect(global.fetch).toHaveBeenCalled());
|
||||
|
||||
// Draw selection
|
||||
const canvas = screen.getByRole('dialog').querySelector('canvas')!;
|
||||
fireEvent.mouseDown(canvas, { clientX: 10, clientY: 10 });
|
||||
fireEvent.mouseMove(canvas, { clientX: 50, clientY: 50 });
|
||||
fireEvent.mouseUp(canvas);
|
||||
|
||||
// Click extract
|
||||
fireEvent.click(screen.getByRole('button', { name: /extract store name/i }));
|
||||
|
||||
await waitFor(() => {
|
||||
expect(mockedNotifyError).toHaveBeenCalledWith('Custom API Error');
|
||||
});
|
||||
});
|
||||
|
||||
it('should redraw the canvas when the image loads', () => {
|
||||
console.log('TEST: Starting "should redraw the canvas when the image loads"');
|
||||
const clearRectSpy = vi.fn();
|
||||
// Override the getContext mock for this test to capture the spy
|
||||
window.HTMLCanvasElement.prototype.getContext = vi.fn(() => ({
|
||||
clearRect: clearRectSpy,
|
||||
strokeRect: vi.fn(),
|
||||
setLineDash: vi.fn(),
|
||||
strokeStyle: '',
|
||||
lineWidth: 0,
|
||||
})) as any;
|
||||
|
||||
renderWithProviders(<FlyerCorrectionTool {...defaultProps} />);
|
||||
const image = screen.getByAltText('Flyer for correction');
|
||||
|
||||
// The draw function is called on mount via useEffect, so we clear that call.
|
||||
clearRectSpy.mockClear();
|
||||
|
||||
// Simulate image load event which triggers onLoad={draw}
|
||||
fireEvent.load(image);
|
||||
|
||||
expect(clearRectSpy).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -153,4 +153,50 @@ describe('RecipeSuggester Component', () => {
|
||||
});
|
||||
console.log('TEST: Previous error cleared successfully');
|
||||
});
|
||||
|
||||
it('uses default error message when API error response has no message', async () => {
|
||||
console.log('TEST: Verifying default error message for API failure');
|
||||
const user = userEvent.setup();
|
||||
renderWithProviders(<RecipeSuggester />);
|
||||
|
||||
const input = screen.getByLabelText(/Ingredients:/i);
|
||||
await user.type(input, 'mystery');
|
||||
|
||||
// Mock API failure response without a message property
|
||||
mockedApiClient.suggestRecipe.mockResolvedValue({
|
||||
ok: false,
|
||||
json: async () => ({}), // Empty object
|
||||
} as Response);
|
||||
|
||||
const button = screen.getByRole('button', { name: /Suggest a Recipe/i });
|
||||
await user.click(button);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.getByText('Failed to get suggestion.')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
|
||||
it('handles non-Error objects thrown during fetch', async () => {
|
||||
console.log('TEST: Verifying handling of non-Error exceptions');
|
||||
const user = userEvent.setup();
|
||||
renderWithProviders(<RecipeSuggester />);
|
||||
|
||||
const input = screen.getByLabelText(/Ingredients:/i);
|
||||
await user.type(input, 'chaos');
|
||||
|
||||
// Mock a rejection that is NOT an Error object
|
||||
mockedApiClient.suggestRecipe.mockRejectedValue('Something weird happened');
|
||||
|
||||
const button = screen.getByRole('button', { name: /Suggest a Recipe/i });
|
||||
await user.click(button);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.getByText('An unknown error occurred.')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: 'Something weird happened' },
|
||||
'Failed to fetch recipe suggestion.'
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -110,8 +110,8 @@ async function main() {
|
||||
validTo.setDate(today.getDate() + 5);
|
||||
|
||||
const flyerQuery = `
|
||||
INSERT INTO public.flyers (file_name, image_url, checksum, store_id, valid_from, valid_to)
|
||||
VALUES ('safeway-flyer.jpg', 'https://example.com/flyer-images/safeway-flyer.jpg', 'a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0', ${storeMap.get('Safeway')}, $1, $2)
|
||||
INSERT INTO public.flyers (file_name, image_url, icon_url, checksum, store_id, valid_from, valid_to)
|
||||
VALUES ('safeway-flyer.jpg', 'https://example.com/flyer-images/safeway-flyer.jpg', 'https://example.com/flyer-images/icons/safeway-flyer.jpg', 'a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0', ${storeMap.get('Safeway')}, $1, $2)
|
||||
RETURNING flyer_id;
|
||||
`;
|
||||
const flyerRes = await client.query<{ flyer_id: number }>(flyerQuery, [
|
||||
|
||||
@@ -77,6 +77,18 @@ describe('PriceChart', () => {
|
||||
expect(screen.getByText(/no deals for your watched items/i)).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should render an error message when an error occurs', () => {
|
||||
mockedUseActiveDeals.mockReturnValue({
|
||||
...mockedUseActiveDeals(),
|
||||
activeDeals: [],
|
||||
isLoading: false,
|
||||
error: 'Failed to fetch deals.',
|
||||
});
|
||||
|
||||
render(<PriceChart {...defaultProps} />);
|
||||
expect(screen.getByText('Failed to fetch deals.')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should render the table with deal items when data is provided', () => {
|
||||
render(<PriceChart {...defaultProps} />);
|
||||
|
||||
|
||||
@@ -8,9 +8,13 @@ interface TopDealsProps {
|
||||
|
||||
export const TopDeals: React.FC<TopDealsProps> = ({ items }) => {
|
||||
const topDeals = useMemo(() => {
|
||||
// Use a type guard in the filter to inform TypeScript that price_in_cents is non-null
|
||||
// in subsequent operations. This allows removing the redundant nullish coalescing in sort.
|
||||
return [...items]
|
||||
.filter((item) => item.price_in_cents !== null) // Only include items with a parseable price
|
||||
.sort((a, b) => (a.price_in_cents ?? Infinity) - (b.price_in_cents ?? Infinity))
|
||||
.filter(
|
||||
(item): item is FlyerItem & { price_in_cents: number } => item.price_in_cents !== null,
|
||||
)
|
||||
.sort((a, b) => a.price_in_cents - b.price_in_cents)
|
||||
.slice(0, 10);
|
||||
}, [items]);
|
||||
|
||||
|
||||
@@ -15,8 +15,8 @@ describe('useFlyerItems Hook', () => {
|
||||
const mockFlyer = createMockFlyer({
|
||||
flyer_id: 123,
|
||||
file_name: 'test-flyer.jpg',
|
||||
image_url: '/test.jpg',
|
||||
icon_url: '/icon.jpg',
|
||||
image_url: 'http://example.com/test.jpg',
|
||||
icon_url: 'http://example.com/icon.jpg',
|
||||
checksum: 'abc',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
|
||||
@@ -72,7 +72,7 @@ describe('useFlyers Hook and FlyersProvider', () => {
|
||||
createMockFlyer({
|
||||
flyer_id: 1,
|
||||
file_name: 'flyer1.jpg',
|
||||
image_url: 'url1',
|
||||
image_url: 'http://example.com/flyer1.jpg',
|
||||
item_count: 5,
|
||||
created_at: '2024-01-01',
|
||||
}),
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
// src/middleware/errorHandler.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterAll } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterAll, afterEach } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import express, { Request, Response, NextFunction } from 'express';
|
||||
import { errorHandler } from './errorHandler'; // This was a duplicate, fixed.
|
||||
import { DatabaseError } from '../services/processingErrors';
|
||||
import {
|
||||
DatabaseError,
|
||||
ForeignKeyConstraintError,
|
||||
UniqueConstraintError,
|
||||
ValidationError,
|
||||
@@ -69,7 +69,7 @@ app.get('/unique-error', (req, res, next) => {
|
||||
});
|
||||
|
||||
app.get('/db-error-500', (req, res, next) => {
|
||||
next(new DatabaseError('A database connection issue occurred.', 500));
|
||||
next(new DatabaseError('A database connection issue occurred.'));
|
||||
});
|
||||
|
||||
app.get('/unauthorized-error-no-status', (req, res, next) => {
|
||||
@@ -98,12 +98,15 @@ describe('errorHandler Middleware', () => {
|
||||
vi.clearAllMocks();
|
||||
consoleErrorSpy.mockClear(); // Clear spy for console.error
|
||||
// Ensure NODE_ENV is set to 'test' for console.error logging
|
||||
process.env.NODE_ENV = 'test';
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.unstubAllEnvs(); // Clean up environment variable stubs after each test
|
||||
});
|
||||
|
||||
afterAll(() => {
|
||||
consoleErrorSpy.mockRestore(); // Restore console.error after all tests
|
||||
delete process.env.NODE_ENV; // Clean up environment variable
|
||||
});
|
||||
|
||||
it('should return a generic 500 error for a standard Error object', async () => {
|
||||
@@ -293,11 +296,7 @@ describe('errorHandler Middleware', () => {
|
||||
|
||||
describe('when NODE_ENV is "production"', () => {
|
||||
beforeEach(() => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
});
|
||||
|
||||
afterAll(() => {
|
||||
process.env.NODE_ENV = 'test'; // Reset for other test files
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
});
|
||||
|
||||
it('should return a generic message with an error ID for a 500 error', async () => {
|
||||
|
||||
@@ -109,20 +109,19 @@ describe('Multer Middleware Directory Creation', () => {
|
||||
describe('createUploadMiddleware', () => {
|
||||
const mockFile = { originalname: 'test.png' } as Express.Multer.File;
|
||||
const mockUser = createMockUserProfile({ user: { user_id: 'user-123', email: 'test@user.com' } });
|
||||
let originalNodeEnv: string | undefined;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
originalNodeEnv = process.env.NODE_ENV;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env.NODE_ENV = originalNodeEnv;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
describe('Avatar Storage', () => {
|
||||
it('should generate a unique filename for an authenticated user', () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
createUploadMiddleware({ storageType: 'avatar' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
@@ -150,7 +149,7 @@ describe('createUploadMiddleware', () => {
|
||||
});
|
||||
|
||||
it('should use a predictable filename in test environment', () => {
|
||||
process.env.NODE_ENV = 'test';
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
createUploadMiddleware({ storageType: 'avatar' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
@@ -164,7 +163,7 @@ describe('createUploadMiddleware', () => {
|
||||
|
||||
describe('Flyer Storage', () => {
|
||||
it('should generate a unique, sanitized filename in production environment', () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
const mockFlyerFile = {
|
||||
fieldname: 'flyerFile',
|
||||
originalname: 'My Flyer (Special!).pdf',
|
||||
@@ -184,7 +183,7 @@ describe('createUploadMiddleware', () => {
|
||||
|
||||
it('should generate a predictable filename in test environment', () => {
|
||||
// This test covers lines 43-46
|
||||
process.env.NODE_ENV = 'test';
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
const mockFlyerFile = {
|
||||
fieldname: 'flyerFile',
|
||||
originalname: 'test-flyer.jpg',
|
||||
|
||||
@@ -59,14 +59,14 @@ describe('FlyerReviewPage', () => {
|
||||
file_name: 'flyer1.jpg',
|
||||
created_at: '2023-01-01T00:00:00Z',
|
||||
store: { name: 'Store A' },
|
||||
icon_url: 'icon1.jpg',
|
||||
icon_url: 'http://example.com/icon1.jpg',
|
||||
},
|
||||
{
|
||||
flyer_id: 2,
|
||||
file_name: 'flyer2.jpg',
|
||||
created_at: '2023-01-02T00:00:00Z',
|
||||
store: { name: 'Store B' },
|
||||
icon_url: 'icon2.jpg',
|
||||
icon_url: 'http://example.com/icon2.jpg',
|
||||
},
|
||||
{
|
||||
flyer_id: 3,
|
||||
@@ -103,7 +103,7 @@ describe('FlyerReviewPage', () => {
|
||||
const unknownStoreItem = screen.getByText('Unknown Store').closest('li');
|
||||
const unknownStoreImage = within(unknownStoreItem!).getByRole('img');
|
||||
expect(unknownStoreImage).not.toHaveAttribute('src');
|
||||
expect(unknownStoreImage).not.toHaveAttribute('alt');
|
||||
expect(unknownStoreImage).toHaveAttribute('alt', 'Unknown Store');
|
||||
});
|
||||
|
||||
it('renders error message when API response is not ok', async () => {
|
||||
|
||||
@@ -73,7 +73,7 @@ export const FlyerReviewPage: React.FC = () => {
|
||||
flyers.map((flyer) => (
|
||||
<li key={flyer.flyer_id} className="p-4 hover:bg-gray-50 dark:hover:bg-gray-700/50">
|
||||
<Link to={`/flyers/${flyer.flyer_id}`} className="flex items-center space-x-4">
|
||||
<img src={flyer.icon_url || undefined} alt={flyer.store?.name} className="w-12 h-12 rounded-md object-cover" />
|
||||
<img src={flyer.icon_url || undefined} alt={flyer.store?.name || 'Unknown Store'} className="w-12 h-12 rounded-md object-cover" />
|
||||
<div className="flex-1">
|
||||
<p className="font-semibold text-gray-800 dark:text-white">{flyer.store?.name || 'Unknown Store'}</p>
|
||||
<p className="text-sm text-gray-500 dark:text-gray-400">{flyer.file_name}</p>
|
||||
|
||||
@@ -967,6 +967,13 @@ describe('ProfileManager', () => {
|
||||
|
||||
it('should show error notification when auto-geocoding fails', async () => {
|
||||
vi.useFakeTimers();
|
||||
// FIX: Mock getUserAddress to return an address *without* coordinates.
|
||||
// This is the condition required to trigger the auto-geocoding logic.
|
||||
const addressWithoutCoords = { ...mockAddress, latitude: undefined, longitude: undefined };
|
||||
mockedApiClient.getUserAddress.mockResolvedValue(
|
||||
new Response(JSON.stringify(addressWithoutCoords)),
|
||||
);
|
||||
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
|
||||
// Wait for initial load
|
||||
|
||||
@@ -225,6 +225,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
// Act
|
||||
await supertest(authenticatedApp)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.set('Authorization', 'Bearer mock-token') // Add this to satisfy the header check in the route
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
@@ -260,6 +261,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
// Act
|
||||
await supertest(authenticatedApp)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.set('Authorization', 'Bearer mock-token') // Add this to satisfy the header check in the route
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
|
||||
@@ -183,7 +183,13 @@ router.post(
|
||||
'Handling /upload-and-process',
|
||||
);
|
||||
|
||||
const userProfile = req.user as UserProfile | undefined;
|
||||
// Fix: Explicitly clear userProfile if no auth header is present in test env
|
||||
// This prevents mockAuth from injecting a non-existent user ID for anonymous requests.
|
||||
let userProfile = req.user as UserProfile | undefined;
|
||||
if (process.env.NODE_ENV === 'test' && !req.headers['authorization']) {
|
||||
userProfile = undefined;
|
||||
}
|
||||
|
||||
const job = await aiService.enqueueFlyerProcessing(
|
||||
req.file,
|
||||
body.checksum,
|
||||
@@ -208,6 +214,34 @@ router.post(
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* POST /api/ai/upload-legacy - Process a flyer upload from a legacy client.
|
||||
* This is an authenticated route that processes the flyer synchronously.
|
||||
* This is used for integration testing the legacy upload flow.
|
||||
*/
|
||||
router.post(
|
||||
'/upload-legacy',
|
||||
passport.authenticate('jwt', { session: false }),
|
||||
uploadToDisk.single('flyerFile'),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
try {
|
||||
if (!req.file) {
|
||||
return res.status(400).json({ message: 'No flyer file uploaded.' });
|
||||
}
|
||||
const userProfile = req.user as UserProfile;
|
||||
const newFlyer = await aiService.processLegacyFlyerUpload(req.file, req.body, userProfile, req.log);
|
||||
res.status(200).json(newFlyer);
|
||||
} catch (error) {
|
||||
await cleanupUploadedFile(req.file);
|
||||
if (error instanceof DuplicateFlyerError) {
|
||||
logger.warn(`Duplicate legacy flyer upload attempt blocked.`);
|
||||
return res.status(409).json({ message: error.message, flyerId: error.flyerId });
|
||||
}
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* NEW ENDPOINT: Checks the status of a background job.
|
||||
*/
|
||||
|
||||
@@ -618,21 +618,19 @@ describe('Passport Configuration', () => {
|
||||
|
||||
describe('mockAuth Middleware', () => {
|
||||
const mockNext: NextFunction = vi.fn();
|
||||
let mockRes: Partial<Response>;
|
||||
let originalNodeEnv: string | undefined;
|
||||
const mockRes: Partial<Response> = {
|
||||
status: vi.fn().mockReturnThis(),
|
||||
json: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
mockRes = { status: vi.fn().mockReturnThis(), json: vi.fn() };
|
||||
originalNodeEnv = process.env.NODE_ENV;
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env.NODE_ENV = originalNodeEnv;
|
||||
// Unstub env variables before each test in this block to ensure a clean state.
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should attach a mock admin user to req when NODE_ENV is "test"', () => {
|
||||
// Arrange
|
||||
process.env.NODE_ENV = 'test';
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
const mockReq = {} as Request;
|
||||
|
||||
// Act
|
||||
@@ -646,7 +644,7 @@ describe('Passport Configuration', () => {
|
||||
|
||||
it('should do nothing and call next() when NODE_ENV is not "test"', () => {
|
||||
// Arrange
|
||||
process.env.NODE_ENV = 'production';
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
const mockReq = {} as Request;
|
||||
|
||||
// Act
|
||||
|
||||
211
src/routes/reactions.routes.test.ts
Normal file
211
src/routes/reactions.routes.test.ts
Normal file
@@ -0,0 +1,211 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
|
||||
// 1. Mock the Service Layer directly.
|
||||
vi.mock('../services/db/index.db', () => ({
|
||||
reactionRepo: {
|
||||
getReactions: vi.fn(),
|
||||
getReactionSummary: vi.fn(),
|
||||
toggleReaction: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock the logger to keep test output clean
|
||||
vi.mock('../services/logger.server', async () => ({
|
||||
logger: (await import('../tests/utils/mockLogger')).mockLogger,
|
||||
}));
|
||||
|
||||
// Mock Passport middleware
|
||||
vi.mock('./passport.routes', () => ({
|
||||
default: {
|
||||
authenticate: vi.fn(
|
||||
() => (req: any, res: any, next: any) => {
|
||||
// If we are testing the unauthenticated state (no user injected), simulate 401.
|
||||
if (!req.user) {
|
||||
return res.status(401).json({ message: 'Unauthorized' });
|
||||
}
|
||||
next();
|
||||
},
|
||||
),
|
||||
},
|
||||
}));
|
||||
|
||||
// Import the router and mocked DB AFTER all mocks are defined.
|
||||
import reactionsRouter from './reactions.routes';
|
||||
import { reactionRepo } from '../services/db/index.db';
|
||||
import { mockLogger } from '../tests/utils/mockLogger';
|
||||
|
||||
const expectLogger = expect.objectContaining({
|
||||
info: expect.any(Function),
|
||||
error: expect.any(Function),
|
||||
});
|
||||
|
||||
describe('Reaction Routes (/api/reactions)', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('GET /', () => {
|
||||
const app = createTestApp({ router: reactionsRouter, basePath: '/api/reactions' });
|
||||
|
||||
it('should return a list of reactions', async () => {
|
||||
const mockReactions = [{ id: 1, reaction_type: 'like', entity_id: '123' }];
|
||||
vi.mocked(reactionRepo.getReactions).mockResolvedValue(mockReactions as any);
|
||||
|
||||
const response = await supertest(app).get('/api/reactions');
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual(mockReactions);
|
||||
expect(reactionRepo.getReactions).toHaveBeenCalledWith({}, expectLogger);
|
||||
});
|
||||
|
||||
it('should filter by query parameters', async () => {
|
||||
const mockReactions = [{ id: 1, reaction_type: 'like' }];
|
||||
vi.mocked(reactionRepo.getReactions).mockResolvedValue(mockReactions as any);
|
||||
|
||||
const validUuid = '123e4567-e89b-12d3-a456-426614174000';
|
||||
const query = { userId: validUuid, entityType: 'recipe', entityId: '1' };
|
||||
|
||||
const response = await supertest(app).get('/api/reactions').query(query);
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(reactionRepo.getReactions).toHaveBeenCalledWith(
|
||||
expect.objectContaining(query),
|
||||
expectLogger
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 500 on database error', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(reactionRepo.getReactions).mockRejectedValue(error);
|
||||
|
||||
const response = await supertest(app).get('/api/reactions');
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ error },
|
||||
'Error fetching user reactions'
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('GET /summary', () => {
|
||||
const app = createTestApp({ router: reactionsRouter, basePath: '/api/reactions' });
|
||||
|
||||
it('should return reaction summary for an entity', async () => {
|
||||
const mockSummary = { like: 10, love: 5 };
|
||||
vi.mocked(reactionRepo.getReactionSummary).mockResolvedValue(mockSummary as any);
|
||||
|
||||
const response = await supertest(app)
|
||||
.get('/api/reactions/summary')
|
||||
.query({ entityType: 'recipe', entityId: '123' });
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual(mockSummary);
|
||||
expect(reactionRepo.getReactionSummary).toHaveBeenCalledWith(
|
||||
'recipe',
|
||||
'123',
|
||||
expectLogger
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 400 if required parameters are missing', async () => {
|
||||
const response = await supertest(app).get('/api/reactions/summary');
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.errors[0].message).toContain('required');
|
||||
});
|
||||
|
||||
it('should return 500 on database error', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(reactionRepo.getReactionSummary).mockRejectedValue(error);
|
||||
|
||||
const response = await supertest(app)
|
||||
.get('/api/reactions/summary')
|
||||
.query({ entityType: 'recipe', entityId: '123' });
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ error },
|
||||
'Error fetching reaction summary'
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('POST /toggle', () => {
|
||||
const mockUser = createMockUserProfile({ user: { user_id: 'user-123' } });
|
||||
const app = createTestApp({
|
||||
router: reactionsRouter,
|
||||
basePath: '/api/reactions',
|
||||
authenticatedUser: mockUser,
|
||||
});
|
||||
|
||||
const validBody = {
|
||||
entity_type: 'recipe',
|
||||
entity_id: '123',
|
||||
reaction_type: 'like',
|
||||
};
|
||||
|
||||
it('should return 201 when a reaction is added', async () => {
|
||||
const mockResult = { ...validBody, id: 1, user_id: 'user-123' };
|
||||
vi.mocked(reactionRepo.toggleReaction).mockResolvedValue(mockResult as any);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/reactions/toggle')
|
||||
.send(validBody);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(response.body).toEqual({ message: 'Reaction added.', reaction: mockResult });
|
||||
expect(reactionRepo.toggleReaction).toHaveBeenCalledWith(
|
||||
{ user_id: 'user-123', ...validBody },
|
||||
expectLogger
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 200 when a reaction is removed', async () => {
|
||||
// Returning null/false from toggleReaction implies the reaction was removed
|
||||
vi.mocked(reactionRepo.toggleReaction).mockResolvedValue(null);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/reactions/toggle')
|
||||
.send(validBody);
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual({ message: 'Reaction removed.' });
|
||||
});
|
||||
|
||||
it('should return 400 if body is invalid', async () => {
|
||||
const response = await supertest(app)
|
||||
.post('/api/reactions/toggle')
|
||||
.send({ entity_type: 'recipe' }); // Missing other required fields
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.errors).toBeDefined();
|
||||
});
|
||||
|
||||
it('should return 401 if not authenticated', async () => {
|
||||
const unauthApp = createTestApp({ router: reactionsRouter, basePath: '/api/reactions' });
|
||||
const response = await supertest(unauthApp)
|
||||
.post('/api/reactions/toggle')
|
||||
.send(validBody);
|
||||
|
||||
expect(response.status).toBe(401);
|
||||
});
|
||||
|
||||
it('should return 500 on database error', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(reactionRepo.toggleReaction).mockRejectedValue(error);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/reactions/toggle')
|
||||
.send(validBody);
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ error, body: validBody },
|
||||
'Error toggling user reaction'
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -1,7 +1,7 @@
|
||||
// src/routes/recipe.routes.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import { createMockRecipe, createMockRecipeComment } from '../tests/utils/mockFactories';
|
||||
import { createMockRecipe, createMockRecipeComment, createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
import { NotFoundError } from '../services/db/errors.db';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
|
||||
@@ -16,9 +16,31 @@ vi.mock('../services/db/index.db', () => ({
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock AI Service
|
||||
vi.mock('../services/aiService.server', () => ({
|
||||
aiService: {
|
||||
generateRecipeSuggestion: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock Passport
|
||||
vi.mock('./passport.routes', () => ({
|
||||
default: {
|
||||
authenticate: vi.fn(
|
||||
() => (req: any, res: any, next: any) => {
|
||||
if (!req.user) {
|
||||
return res.status(401).json({ message: 'Unauthorized' });
|
||||
}
|
||||
next();
|
||||
},
|
||||
),
|
||||
},
|
||||
}));
|
||||
|
||||
// Import the router and mocked DB AFTER all mocks are defined.
|
||||
import recipeRouter from './recipe.routes';
|
||||
import * as db from '../services/db/index.db';
|
||||
import { aiService } from '../services/aiService.server';
|
||||
import { mockLogger } from '../tests/utils/mockLogger';
|
||||
|
||||
// Mock the logger to keep test output clean
|
||||
@@ -229,4 +251,71 @@ describe('Recipe Routes (/api/recipes)', () => {
|
||||
expect(response.body.errors[0].message).toContain('received NaN');
|
||||
});
|
||||
});
|
||||
|
||||
describe('POST /suggest', () => {
|
||||
const mockUser = createMockUserProfile({ user: { user_id: 'user-123' } });
|
||||
const authApp = createTestApp({
|
||||
router: recipeRouter,
|
||||
basePath: '/api/recipes',
|
||||
authenticatedUser: mockUser,
|
||||
});
|
||||
|
||||
it('should return a recipe suggestion', async () => {
|
||||
const ingredients = ['chicken', 'rice'];
|
||||
const mockSuggestion = 'Chicken and Rice Casserole...';
|
||||
vi.mocked(aiService.generateRecipeSuggestion).mockResolvedValue(mockSuggestion);
|
||||
|
||||
const response = await supertest(authApp)
|
||||
.post('/api/recipes/suggest')
|
||||
.send({ ingredients });
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual({ suggestion: mockSuggestion });
|
||||
expect(aiService.generateRecipeSuggestion).toHaveBeenCalledWith(ingredients, expectLogger);
|
||||
});
|
||||
|
||||
it('should return 503 if AI service returns null', async () => {
|
||||
vi.mocked(aiService.generateRecipeSuggestion).mockResolvedValue(null);
|
||||
|
||||
const response = await supertest(authApp)
|
||||
.post('/api/recipes/suggest')
|
||||
.send({ ingredients: ['water'] });
|
||||
|
||||
expect(response.status).toBe(503);
|
||||
expect(response.body.message).toContain('unavailable');
|
||||
});
|
||||
|
||||
it('should return 400 if ingredients list is empty', async () => {
|
||||
const response = await supertest(authApp)
|
||||
.post('/api/recipes/suggest')
|
||||
.send({ ingredients: [] });
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.errors[0].message).toContain('At least one ingredient is required');
|
||||
});
|
||||
|
||||
it('should return 401 if not authenticated', async () => {
|
||||
const unauthApp = createTestApp({ router: recipeRouter, basePath: '/api/recipes' });
|
||||
const response = await supertest(unauthApp)
|
||||
.post('/api/recipes/suggest')
|
||||
.send({ ingredients: ['chicken'] });
|
||||
|
||||
expect(response.status).toBe(401);
|
||||
});
|
||||
|
||||
it('should return 500 on service error', async () => {
|
||||
const error = new Error('AI Error');
|
||||
vi.mocked(aiService.generateRecipeSuggestion).mockRejectedValue(error);
|
||||
|
||||
const response = await supertest(authApp)
|
||||
.post('/api/recipes/suggest')
|
||||
.send({ ingredients: ['chicken'] });
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ error },
|
||||
'Error generating recipe suggestion'
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -24,58 +24,8 @@ import { cleanupFiles } from '../tests/utils/cleanupFiles';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { userService } from '../services/userService';
|
||||
|
||||
// 1. Mock the Service Layer directly.
|
||||
// The user.routes.ts file imports from '.../db/index.db'. We need to mock that module.
|
||||
vi.mock('../services/db/index.db', () => ({
|
||||
// Repository instances
|
||||
userRepo: {
|
||||
findUserProfileById: vi.fn(),
|
||||
updateUserProfile: vi.fn(),
|
||||
updateUserPreferences: vi.fn(),
|
||||
},
|
||||
personalizationRepo: {
|
||||
getWatchedItems: vi.fn(),
|
||||
removeWatchedItem: vi.fn(),
|
||||
addWatchedItem: vi.fn(),
|
||||
getUserDietaryRestrictions: vi.fn(),
|
||||
setUserDietaryRestrictions: vi.fn(),
|
||||
getUserAppliances: vi.fn(),
|
||||
setUserAppliances: vi.fn(),
|
||||
},
|
||||
shoppingRepo: {
|
||||
getShoppingLists: vi.fn(),
|
||||
createShoppingList: vi.fn(),
|
||||
deleteShoppingList: vi.fn(),
|
||||
addShoppingListItem: vi.fn(),
|
||||
updateShoppingListItem: vi.fn(),
|
||||
removeShoppingListItem: vi.fn(),
|
||||
getShoppingListById: vi.fn(), // Added missing mock
|
||||
},
|
||||
recipeRepo: {
|
||||
deleteRecipe: vi.fn(),
|
||||
updateRecipe: vi.fn(),
|
||||
},
|
||||
addressRepo: {
|
||||
getAddressById: vi.fn(),
|
||||
upsertAddress: vi.fn(),
|
||||
},
|
||||
notificationRepo: {
|
||||
getNotificationsForUser: vi.fn(),
|
||||
markAllNotificationsAsRead: vi.fn(),
|
||||
markNotificationAsRead: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock userService
|
||||
vi.mock('../services/userService', () => ({
|
||||
userService: {
|
||||
updateUserAvatar: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteUserAccount: vi.fn(),
|
||||
getUserAddress: vi.fn(),
|
||||
upsertUserAddress: vi.fn(),
|
||||
},
|
||||
}));
|
||||
// Mocks for db/index.db, userService, and logger are now centralized in `src/tests/setup/tests-setup-unit.ts`.
|
||||
// This avoids repetition across test files.
|
||||
|
||||
// Mock the logger
|
||||
vi.mock('../services/logger.server', async () => ({
|
||||
@@ -122,10 +72,10 @@ describe('User Routes (/api/users)', () => {
|
||||
describe('Avatar Upload Directory Creation', () => {
|
||||
it('should log an error if avatar directory creation fails', async () => {
|
||||
// Arrange
|
||||
const mkdirError = new Error('EACCES: permission denied');
|
||||
const mkdirError = new Error('EACCES: permission denied'); // This error is specific to the fs.mkdir mock.
|
||||
// Reset modules to force re-import with a new mock implementation
|
||||
vi.resetModules();
|
||||
// Set up the mock *before* the module is re-imported
|
||||
// Set up the mock *before* the module is re-imported.
|
||||
vi.doMock('node:fs/promises', () => ({
|
||||
default: {
|
||||
// We only need to mock mkdir for this test.
|
||||
@@ -133,6 +83,10 @@ describe('User Routes (/api/users)', () => {
|
||||
},
|
||||
}));
|
||||
const { logger } = await import('../services/logger.server');
|
||||
// Stub NODE_ENV to ensure the relevant code path is executed if it depends on it.
|
||||
// Although the mkdir call itself doesn't depend on NODE_ENV, this is good practice
|
||||
// when re-importing modules that might have conditional logic based on it.
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
|
||||
// Act: Dynamically import the router to trigger the top-level fs.mkdir call
|
||||
await import('./user.routes');
|
||||
@@ -142,6 +96,7 @@ describe('User Routes (/api/users)', () => {
|
||||
{ error: mkdirError },
|
||||
'Failed to create multer storage directories on startup.',
|
||||
);
|
||||
vi.unstubAllEnvs(); // Clean up the stubbed environment variable.
|
||||
vi.doUnmock('node:fs/promises'); // Clean up
|
||||
});
|
||||
});
|
||||
@@ -1075,7 +1030,7 @@ describe('User Routes (/api/users)', () => {
|
||||
it('should upload an avatar and update the user profile', async () => {
|
||||
const mockUpdatedProfile = createMockUserProfile({
|
||||
...mockUserProfile,
|
||||
avatar_url: '/uploads/avatars/new-avatar.png',
|
||||
avatar_url: 'http://localhost:3001/uploads/avatars/new-avatar.png',
|
||||
});
|
||||
vi.mocked(userService.updateUserAvatar).mockResolvedValue(mockUpdatedProfile);
|
||||
|
||||
@@ -1087,7 +1042,7 @@ describe('User Routes (/api/users)', () => {
|
||||
.attach('avatar', Buffer.from('dummy-image-content'), dummyImagePath);
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.avatar_url).toContain('/uploads/avatars/'); // This was a duplicate, fixed.
|
||||
expect(response.body.avatar_url).toContain('http://localhost:3001/uploads/avatars/');
|
||||
expect(userService.updateUserAvatar).toHaveBeenCalledWith(
|
||||
mockUserProfile.user.user_id,
|
||||
expect.any(Object),
|
||||
|
||||
@@ -11,7 +11,11 @@ import {
|
||||
DuplicateFlyerError,
|
||||
type RawFlyerItem,
|
||||
} from './aiService.server';
|
||||
import { createMockMasterGroceryItem, createMockFlyer } from '../tests/utils/mockFactories';
|
||||
import {
|
||||
createMockMasterGroceryItem,
|
||||
createMockFlyer,
|
||||
createMockUserProfile,
|
||||
} from '../tests/utils/mockFactories';
|
||||
import { ValidationError } from './db/errors.db';
|
||||
import { AiFlyerDataSchema } from '../types/ai';
|
||||
|
||||
@@ -26,12 +30,13 @@ import { logger as mockLoggerInstance } from './logger.server';
|
||||
// Explicitly unmock the service under test to ensure we import the real implementation.
|
||||
vi.unmock('./aiService.server');
|
||||
|
||||
const { mockGenerateContent, mockToBuffer, mockExtract, mockSharp } = vi.hoisted(() => {
|
||||
const { mockGenerateContent, mockToBuffer, mockExtract, mockSharp, mockAdminLogActivity } = vi.hoisted(() => {
|
||||
const mockGenerateContent = vi.fn();
|
||||
const mockToBuffer = vi.fn();
|
||||
const mockExtract = vi.fn(() => ({ toBuffer: mockToBuffer }));
|
||||
const mockSharp = vi.fn(() => ({ extract: mockExtract }));
|
||||
return { mockGenerateContent, mockToBuffer, mockExtract, mockSharp };
|
||||
const mockAdminLogActivity = vi.fn();
|
||||
return { mockGenerateContent, mockToBuffer, mockExtract, mockSharp, mockAdminLogActivity };
|
||||
});
|
||||
|
||||
// Mock sharp, as it's a direct dependency of the service.
|
||||
@@ -61,6 +66,7 @@ vi.mock('./db/index.db', () => ({
|
||||
adminRepo: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
withTransaction: vi.fn(),
|
||||
}));
|
||||
|
||||
vi.mock('./queueService.server', () => ({
|
||||
@@ -77,10 +83,17 @@ vi.mock('../utils/imageProcessor', () => ({
|
||||
generateFlyerIcon: vi.fn(),
|
||||
}));
|
||||
|
||||
vi.mock('./db/admin.db', () => ({
|
||||
AdminRepository: vi.fn().mockImplementation(function () {
|
||||
return { logActivity: mockAdminLogActivity };
|
||||
}),
|
||||
}));
|
||||
|
||||
// Import mocked modules to assert on them
|
||||
import * as dbModule from './db/index.db';
|
||||
import { flyerQueue } from './queueService.server';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { withTransaction } from './db/index.db';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
|
||||
// Define a mock interface that closely resembles the actual Flyer type for testing purposes.
|
||||
@@ -102,6 +115,8 @@ interface MockFlyer {
|
||||
updated_at: string;
|
||||
}
|
||||
|
||||
const baseUrl = 'http://localhost:3001';
|
||||
|
||||
describe('AI Service (Server)', () => {
|
||||
// Create mock dependencies that will be injected into the service
|
||||
const mockAiClient = { generateContent: vi.fn() };
|
||||
@@ -115,12 +130,16 @@ describe('AI Service (Server)', () => {
|
||||
vi.restoreAllMocks();
|
||||
vi.clearAllMocks();
|
||||
mockGenerateContent.mockReset();
|
||||
mockAdminLogActivity.mockClear();
|
||||
// Reset modules to ensure the service re-initializes with the mocks
|
||||
|
||||
mockAiClient.generateContent.mockResolvedValue({
|
||||
text: '[]',
|
||||
candidates: [],
|
||||
});
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback: any) => {
|
||||
return callback({}); // Mock client
|
||||
});
|
||||
});
|
||||
|
||||
describe('AiFlyerDataSchema', () => {
|
||||
@@ -136,45 +155,29 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
|
||||
describe('Constructor', () => {
|
||||
const originalEnv = process.env;
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset process.env before each test in this block
|
||||
vi.unstubAllEnvs();
|
||||
vi.unstubAllEnvs(); // Force-removes all environment mocking
|
||||
vi.resetModules(); // Important to re-evaluate the service file
|
||||
process.env = { ...originalEnv };
|
||||
console.log('CONSTRUCTOR beforeEach: process.env reset.');
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore original environment variables
|
||||
vi.unstubAllEnvs();
|
||||
process.env = originalEnv;
|
||||
console.log('CONSTRUCTOR afterEach: process.env restored.');
|
||||
});
|
||||
|
||||
it('should throw an error if GEMINI_API_KEY is not set in a non-test environment', async () => {
|
||||
console.log("TEST START: 'should throw an error if GEMINI_API_KEY is not set...'");
|
||||
console.log(
|
||||
`PRE-TEST ENV: NODE_ENV=${process.env.NODE_ENV}, VITEST_POOL_ID=${process.env.VITEST_POOL_ID}`,
|
||||
);
|
||||
// Simulate a non-test environment
|
||||
process.env.NODE_ENV = 'production';
|
||||
delete process.env.GEMINI_API_KEY;
|
||||
delete process.env.VITEST_POOL_ID;
|
||||
console.log(
|
||||
`POST-MANIPULATION ENV: NODE_ENV=${process.env.NODE_ENV}, VITEST_POOL_ID=${process.env.VITEST_POOL_ID}`,
|
||||
);
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
vi.stubEnv('GEMINI_API_KEY', '');
|
||||
vi.stubEnv('VITEST_POOL_ID', '');
|
||||
|
||||
let error: Error | undefined;
|
||||
// Dynamically import the class to re-evaluate the constructor logic
|
||||
try {
|
||||
console.log('Attempting to import and instantiate AIService which is expected to throw...');
|
||||
const { AIService } = await import('./aiService.server');
|
||||
new AIService(mockLoggerInstance);
|
||||
} catch (e) {
|
||||
console.log('Successfully caught an error during instantiation.');
|
||||
error = e as Error;
|
||||
}
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
@@ -185,8 +188,8 @@ describe('AI Service (Server)', () => {
|
||||
|
||||
it('should use a mock placeholder if API key is missing in a test environment', async () => {
|
||||
// Arrange: Simulate a test environment without an API key
|
||||
process.env.NODE_ENV = 'test';
|
||||
delete process.env.GEMINI_API_KEY;
|
||||
vi.stubEnv('NODE_ENV', 'test');
|
||||
vi.stubEnv('GEMINI_API_KEY', '');
|
||||
|
||||
// Act: Dynamically import and instantiate the service
|
||||
const { AIService } = await import('./aiService.server');
|
||||
@@ -202,7 +205,7 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
|
||||
it('should use the adapter to call generateContent when using real GoogleGenAI client', async () => {
|
||||
process.env.GEMINI_API_KEY = 'test-key';
|
||||
vi.stubEnv('GEMINI_API_KEY', 'test-key');
|
||||
// We need to force the constructor to use the real client logic, not the injected mock.
|
||||
// So we instantiate AIService without passing aiClient.
|
||||
|
||||
@@ -213,18 +216,19 @@ describe('AI Service (Server)', () => {
|
||||
|
||||
// Access the private aiClient (which is now the adapter)
|
||||
const adapter = (service as any).aiClient;
|
||||
const models = (service as any).models;
|
||||
|
||||
const request = { contents: [{ parts: [{ text: 'test' }] }] };
|
||||
await adapter.generateContent(request);
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledWith({
|
||||
model: 'gemini-3-flash-preview',
|
||||
model: models[0],
|
||||
...request,
|
||||
});
|
||||
});
|
||||
|
||||
it('should throw error if adapter is called without content', async () => {
|
||||
process.env.GEMINI_API_KEY = 'test-key';
|
||||
vi.stubEnv('GEMINI_API_KEY', 'test-key');
|
||||
vi.resetModules();
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const service = new AIService(mockLoggerInstance);
|
||||
@@ -237,25 +241,55 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
|
||||
describe('Model Fallback Logic', () => {
|
||||
const originalEnv = process.env;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.unstubAllEnvs();
|
||||
process.env = { ...originalEnv, GEMINI_API_KEY: 'test-key' };
|
||||
vi.stubEnv('GEMINI_API_KEY', 'test-key');
|
||||
vi.resetModules(); // Re-import to use the new env var and re-instantiate the service
|
||||
mockGenerateContent.mockReset();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env = originalEnv;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should use lite models when useLiteModels is true', async () => {
|
||||
// Arrange
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
const models_lite = (serviceWithFallback as any).models_lite;
|
||||
const successResponse = { text: 'Success from lite model', candidates: [] };
|
||||
|
||||
mockGenerateContent.mockResolvedValue(successResponse);
|
||||
|
||||
const request = {
|
||||
contents: [{ parts: [{ text: 'test prompt' }] }],
|
||||
useLiteModels: true,
|
||||
};
|
||||
// The adapter strips `useLiteModels` before calling the underlying client,
|
||||
// so we prepare the expected request shape for our assertions.
|
||||
const { useLiteModels, ...apiReq } = request;
|
||||
|
||||
// Act
|
||||
const result = await (serviceWithFallback as any).aiClient.generateContent(request);
|
||||
|
||||
// Assert
|
||||
expect(result).toEqual(successResponse);
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(1);
|
||||
|
||||
// Check that the first model from the lite list was used
|
||||
expect(mockGenerateContent).toHaveBeenCalledWith({
|
||||
model: models_lite[0],
|
||||
...apiReq,
|
||||
});
|
||||
});
|
||||
|
||||
it('should try the next model if the first one fails with a quota error', async () => {
|
||||
// Arrange
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
const models = (serviceWithFallback as any).models;
|
||||
|
||||
const quotaError = new Error('User rate limit exceeded due to quota');
|
||||
const successResponse = { text: 'Success from fallback model', candidates: [] };
|
||||
@@ -273,22 +307,23 @@ describe('AI Service (Server)', () => {
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(2);
|
||||
|
||||
// Check first call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, { // The first model in the list is now 'gemini-3-flash-preview'
|
||||
model: 'gemini-3-flash-preview',
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, { // The first model in the list
|
||||
model: models[0],
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check second call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, { // The second model in the list is 'gemini-2.5-flash'
|
||||
model: 'gemini-2.5-flash',
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, { // The second model in the list
|
||||
model: models[1],
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check that a warning was logged
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
// The warning should be for the model that failed ('gemini-3-flash-preview'), not the next one.
|
||||
// The warning should be for the model that failed ('gemini-2.5-flash'), not the next one.
|
||||
// The warning should be for the model that failed, not the next one.
|
||||
expect.stringContaining(
|
||||
"Model 'gemini-3-flash-preview' failed due to quota/rate limit. Trying next model.",
|
||||
`Model '${models[0]}' failed due to quota/rate limit. Trying next model.`,
|
||||
),
|
||||
);
|
||||
});
|
||||
@@ -298,6 +333,7 @@ describe('AI Service (Server)', () => {
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
const models = (serviceWithFallback as any).models;
|
||||
|
||||
const nonRetriableError = new Error('Invalid API Key');
|
||||
mockGenerateContent.mockRejectedValueOnce(nonRetriableError);
|
||||
@@ -311,8 +347,8 @@ describe('AI Service (Server)', () => {
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(1);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: nonRetriableError }, // The first model in the list is now 'gemini-3-flash-preview'
|
||||
`[AIService Adapter] Model 'gemini-3-flash-preview' failed with a non-retriable error.`,
|
||||
{ error: nonRetriableError }, // The first model in the list is now 'gemini-2.5-flash'
|
||||
`[AIService Adapter] Model 'gemini-2.5-flash' failed with a non-retriable error.`,
|
||||
);
|
||||
});
|
||||
|
||||
@@ -591,11 +627,8 @@ describe('AI Service (Server)', () => {
|
||||
);
|
||||
|
||||
expect(mockAiClient.generateContent).toHaveBeenCalledTimes(1);
|
||||
expect(result.store_name).toBe('Test Store');
|
||||
expect(result.items).toHaveLength(2);
|
||||
expect(result.items[1].price_display).toBe('');
|
||||
expect(result.items[1].quantity).toBe('');
|
||||
expect(result.items[1].category_name).toBe('Other/Miscellaneous');
|
||||
// With normalization removed from this service, the result should match the raw AI response.
|
||||
expect(result).toEqual(mockAiResponse);
|
||||
});
|
||||
|
||||
it('should throw an error if the AI response is not a valid JSON object', async () => {
|
||||
@@ -852,6 +885,23 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('generateRecipeSuggestion', () => {
|
||||
it('should call generateContent with useLiteModels set to true', async () => {
|
||||
const ingredients = ['carrots', 'onions'];
|
||||
const expectedPrompt = `Suggest a simple recipe using these ingredients: ${ingredients.join(
|
||||
', ',
|
||||
)}. Keep it brief.`;
|
||||
mockAiClient.generateContent.mockResolvedValue({ text: 'Some recipe', candidates: [] });
|
||||
|
||||
await aiServiceInstance.generateRecipeSuggestion(ingredients, mockLoggerInstance);
|
||||
|
||||
expect(mockAiClient.generateContent).toHaveBeenCalledWith({
|
||||
contents: [{ parts: [{ text: expectedPrompt }] }],
|
||||
useLiteModels: true,
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('planTripWithMaps', () => {
|
||||
const mockUserLocation: GeolocationCoordinates = {
|
||||
latitude: 45,
|
||||
@@ -919,7 +969,18 @@ describe('AI Service (Server)', () => {
|
||||
} as UserProfile;
|
||||
|
||||
it('should throw DuplicateFlyerError if flyer already exists', async () => {
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 99, checksum: 'checksum123', file_name: 'test.pdf', image_url: '/flyer-images/test.pdf', icon_url: '/flyer-images/icons/test.webp', store_id: 1, status: 'processed', item_count: 0, created_at: new Date().toISOString(), updated_at: new Date().toISOString() });
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({
|
||||
flyer_id: 99,
|
||||
checksum: 'checksum123',
|
||||
file_name: 'test.pdf',
|
||||
image_url: `${baseUrl}/flyer-images/test.pdf`,
|
||||
icon_url: `${baseUrl}/flyer-images/icons/test.webp`,
|
||||
store_id: 1,
|
||||
status: 'processed',
|
||||
item_count: 0,
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
});
|
||||
|
||||
await expect(
|
||||
aiServiceInstance.enqueueFlyerProcessing(
|
||||
@@ -951,6 +1012,7 @@ describe('AI Service (Server)', () => {
|
||||
userId: 'user123',
|
||||
submitterIp: '127.0.0.1',
|
||||
userProfileAddress: '123 St, City, Country', // Partial address match based on filter(Boolean)
|
||||
baseUrl: 'http://localhost:3000',
|
||||
});
|
||||
expect(result.id).toBe('job123');
|
||||
});
|
||||
@@ -972,6 +1034,7 @@ describe('AI Service (Server)', () => {
|
||||
expect.objectContaining({
|
||||
userId: undefined,
|
||||
userProfileAddress: undefined,
|
||||
baseUrl: 'http://localhost:3000',
|
||||
}),
|
||||
);
|
||||
});
|
||||
@@ -983,7 +1046,8 @@ describe('AI Service (Server)', () => {
|
||||
filename: 'upload.jpg',
|
||||
originalname: 'orig.jpg',
|
||||
} as Express.Multer.File; // This was a duplicate, fixed.
|
||||
const mockProfile = { user: { user_id: 'u1' } } as UserProfile;
|
||||
const mockProfile = createMockUserProfile({ user: { user_id: 'a0eebc99-9c0b-4ef8-bb6d-6bb9bd380a11' } });
|
||||
|
||||
|
||||
beforeEach(() => {
|
||||
// Default success mocks. Use createMockFlyer for a more complete mock.
|
||||
@@ -993,8 +1057,8 @@ describe('AI Service (Server)', () => {
|
||||
flyer: {
|
||||
flyer_id: 100,
|
||||
file_name: 'orig.jpg',
|
||||
image_url: '/flyer-images/upload.jpg',
|
||||
icon_url: '/flyer-images/icons/icon.jpg',
|
||||
image_url: `${baseUrl}/flyer-images/upload.jpg`,
|
||||
icon_url: `${baseUrl}/flyer-images/icons/icon.jpg`,
|
||||
checksum: 'mock-checksum-123',
|
||||
store_name: 'Mock Store',
|
||||
valid_from: null,
|
||||
@@ -1002,7 +1066,7 @@ describe('AI Service (Server)', () => {
|
||||
store_address: null,
|
||||
item_count: 0,
|
||||
status: 'processed',
|
||||
uploaded_by: 'u1',
|
||||
uploaded_by: mockProfile.user.user_id,
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
} as MockFlyer, // Use the more specific MockFlyer type
|
||||
@@ -1061,6 +1125,7 @@ describe('AI Service (Server)', () => {
|
||||
}),
|
||||
expect.arrayContaining([expect.objectContaining({ item: 'Milk' })]),
|
||||
mockLoggerInstance,
|
||||
expect.anything(),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -1087,6 +1152,7 @@ describe('AI Service (Server)', () => {
|
||||
}),
|
||||
[], // No items
|
||||
mockLoggerInstance,
|
||||
expect.anything(),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -1118,6 +1184,7 @@ describe('AI Service (Server)', () => {
|
||||
}),
|
||||
]),
|
||||
mockLoggerInstance,
|
||||
expect.anything(),
|
||||
);
|
||||
expect(mockLoggerInstance.warn).toHaveBeenCalledWith(
|
||||
expect.stringContaining('extractedData.store_name missing'),
|
||||
@@ -1134,10 +1201,10 @@ describe('AI Service (Server)', () => {
|
||||
);
|
||||
|
||||
expect(result).toHaveProperty('flyer_id', 100);
|
||||
expect(dbModule.adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect(mockAdminLogActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
action: 'flyer_processed',
|
||||
userId: 'u1',
|
||||
userId: mockProfile.user.user_id,
|
||||
}),
|
||||
mockLoggerInstance,
|
||||
);
|
||||
@@ -1164,6 +1231,29 @@ describe('AI Service (Server)', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should log and re-throw the original error if the database transaction fails', async () => {
|
||||
const body = { checksum: 'legacy-fail-checksum', extractedData: { store_name: 'Fail Store' } };
|
||||
const dbError = new Error('DB transaction failed');
|
||||
|
||||
// Mock withTransaction to fail
|
||||
vi.mocked(withTransaction).mockRejectedValue(dbError);
|
||||
|
||||
await expect(
|
||||
aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
),
|
||||
).rejects.toThrow(dbError);
|
||||
|
||||
// Verify the service-level error logging
|
||||
expect(mockLoggerInstance.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, checksum: 'legacy-fail-checksum' },
|
||||
'Legacy flyer upload database transaction failed.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle body as a string', async () => {
|
||||
const payload = { checksum: 'str-body', extractedData: { store_name: 'String Body' } };
|
||||
const body = JSON.stringify(payload);
|
||||
@@ -1179,6 +1269,7 @@ describe('AI Service (Server)', () => {
|
||||
expect.objectContaining({ checksum: 'str-body' }),
|
||||
expect.anything(),
|
||||
mockLoggerInstance,
|
||||
expect.anything(),
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -1188,56 +1279,4 @@ describe('AI Service (Server)', () => {
|
||||
expect(aiServiceSingleton).toBeInstanceOf(AIService);
|
||||
});
|
||||
});
|
||||
|
||||
describe('_normalizeExtractedItems (private method)', () => {
|
||||
it('should correctly normalize items with null or undefined price_in_cents', () => {
|
||||
const rawItems: RawFlyerItem[] = [
|
||||
{
|
||||
item: 'Valid Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: '1',
|
||||
category_name: 'Category A',
|
||||
master_item_id: 1,
|
||||
},
|
||||
{
|
||||
item: 'Item with Null Price',
|
||||
price_display: null,
|
||||
price_in_cents: null, // Test case for null
|
||||
quantity: '1',
|
||||
category_name: 'Category B',
|
||||
master_item_id: 2,
|
||||
},
|
||||
{
|
||||
item: 'Item with Undefined Price',
|
||||
price_display: '$2.99',
|
||||
price_in_cents: undefined, // Test case for undefined
|
||||
quantity: '1',
|
||||
category_name: 'Category C',
|
||||
master_item_id: 3,
|
||||
},
|
||||
{
|
||||
item: null, // Test null item name
|
||||
price_display: undefined, // Test undefined display price
|
||||
price_in_cents: 50,
|
||||
quantity: null, // Test null quantity
|
||||
category_name: undefined, // Test undefined category
|
||||
master_item_id: null, // Test null master_item_id
|
||||
},
|
||||
];
|
||||
|
||||
// Access the private method for testing
|
||||
const normalized = (aiServiceInstance as any)._normalizeExtractedItems(rawItems);
|
||||
|
||||
expect(normalized).toHaveLength(4);
|
||||
expect(normalized[0].price_in_cents).toBe(199);
|
||||
expect(normalized[1].price_in_cents).toBe(null); // null should remain null
|
||||
expect(normalized[2].price_in_cents).toBe(null); // undefined should become null
|
||||
expect(normalized[3].item).toBe('Unknown Item');
|
||||
expect(normalized[3].quantity).toBe('');
|
||||
expect(normalized[3].category_name).toBe('Other/Miscellaneous');
|
||||
expect(normalized[3].master_item_id).toBeUndefined(); // nullish coalescing to undefined
|
||||
});
|
||||
});
|
||||
|
||||
});
|
||||
|
||||
@@ -18,12 +18,14 @@ import type {
|
||||
FlyerInsert,
|
||||
Flyer,
|
||||
} from '../types';
|
||||
import { FlyerProcessingError } from './processingErrors';
|
||||
import { DatabaseError, FlyerProcessingError } from './processingErrors';
|
||||
import * as db from './db/index.db';
|
||||
import { flyerQueue } from './queueService.server';
|
||||
import type { Job } from 'bullmq';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { getBaseUrl } from '../utils/serverUtils';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import { AdminRepository } from './db/admin.db';
|
||||
import path from 'path';
|
||||
import { ValidationError } from './db/errors.db'; // Keep this import for ValidationError
|
||||
import {
|
||||
@@ -91,11 +93,55 @@ export class AIService {
|
||||
private fs: IFileSystem;
|
||||
private rateLimiter: <T>(fn: () => Promise<T>) => Promise<T>;
|
||||
private logger: Logger;
|
||||
// The fallback list is ordered by preference (speed/cost vs. power).
|
||||
// We try the fastest models first, then the more powerful 'pro' model as a high-quality fallback,
|
||||
// and finally the 'lite' model as a last resort.
|
||||
private readonly models = [ 'gemini-3-flash-preview', 'gemini-2.5-flash', 'gemini-2.5-flash-lite', 'gemma-3-27b', 'gemma-3-12b'];
|
||||
private readonly models_lite = ["gemma-3-4b", "gemma-3-2b", "gemma-3-1b"];
|
||||
|
||||
// OPTIMIZED: Flyer Image Processing (Vision + Long Output)
|
||||
// PRIORITIES:
|
||||
// 1. Output Limit: Must be 65k+ (Gemini 2.5/3.0) to avoid cutting off data.
|
||||
// 2. Intelligence: 'Pro' models handle messy layouts better.
|
||||
// 3. Quota Management: 'Preview' and 'Exp' models are added as fallbacks to tap into separate rate limits.
|
||||
private readonly models = [
|
||||
// --- TIER A: The Happy Path (Fast & Stable) ---
|
||||
'gemini-2.5-flash', // Primary workhorse. 65k output.
|
||||
'gemini-2.5-flash-lite', // Cost-saver. 65k output.
|
||||
|
||||
// --- TIER B: The Heavy Lifters (Complex Layouts) ---
|
||||
'gemini-2.5-pro', // High IQ for messy flyers. 65k output.
|
||||
|
||||
// --- TIER C: Separate Quota Buckets (Previews) ---
|
||||
'gemini-3-flash-preview', // Newer/Faster. Separate 'Preview' quota. 65k output.
|
||||
'gemini-3-pro-preview', // High IQ. Separate 'Preview' quota. 65k output.
|
||||
|
||||
// --- TIER D: Experimental Buckets (High Capacity) ---
|
||||
'gemini-exp-1206', // Excellent reasoning. Separate 'Experimental' quota. 65k output.
|
||||
|
||||
// --- TIER E: Last Resorts (Lower Capacity/Local) ---
|
||||
'gemma-3-27b-it', // Open model fallback.
|
||||
'gemini-2.0-flash-exp' // Exp fallback. WARNING: 8k output limit. Good for small flyers only.
|
||||
];
|
||||
|
||||
// OPTIMIZED: Simple Text Tasks (Recipes, Shopping Lists, Summaries)
|
||||
// PRIORITIES:
|
||||
// 1. Cost/Speed: These tasks are simple.
|
||||
// 2. Output Limit: The 8k limit of Gemini 2.0 is perfectly fine here.
|
||||
private readonly models_lite = [
|
||||
// --- Best Value (Smart + Cheap) ---
|
||||
"gemini-2.5-flash-lite", // Current generation efficiency king.
|
||||
|
||||
// --- The "Recycled" Gemini 2.0 Models (Perfect for Text) ---
|
||||
"gemini-2.0-flash-lite-001", // Extremely cheap, very capable for text.
|
||||
"gemini-2.0-flash-001", // Smarter than Lite, good for complex recipes.
|
||||
|
||||
// --- Open Models (Good for simple categorization) ---
|
||||
"gemma-3-12b-it", // Solid reasoning for an open model.
|
||||
"gemma-3-4b-it", // Very fast.
|
||||
|
||||
// --- Quota Fallbacks (Experimental/Preview) ---
|
||||
"gemini-2.0-flash-exp", // Use this separate quota bucket if others are exhausted.
|
||||
|
||||
// --- Edge/Nano Models (Simple string manipulation only) ---
|
||||
"gemma-3n-e4b-it", // Corrected name from JSON
|
||||
"gemma-3n-e2b-it" // Corrected name from JSON
|
||||
];
|
||||
|
||||
constructor(logger: Logger, aiClient?: IAiClient, fs?: IFileSystem) {
|
||||
this.logger = logger;
|
||||
@@ -493,12 +539,8 @@ export class AIService {
|
||||
userProfileAddress?: string,
|
||||
logger: Logger = this.logger,
|
||||
): Promise<{
|
||||
store_name: string | null;
|
||||
valid_from: string | null;
|
||||
valid_to: string | null;
|
||||
store_address: string | null;
|
||||
items: ExtractedFlyerItem[];
|
||||
}> {
|
||||
store_name: string | null; valid_from: string | null; valid_to: string | null; store_address: string | null; items: z.infer<typeof ExtractedFlyerItemSchema>[];
|
||||
} & z.infer<typeof AiFlyerDataSchema>> {
|
||||
logger.info(
|
||||
`[extractCoreDataFromFlyerImage] Entering method with ${imagePaths.length} image(s).`,
|
||||
);
|
||||
@@ -554,50 +596,22 @@ export class AIService {
|
||||
throw new Error('AI response did not contain a valid JSON object.');
|
||||
}
|
||||
|
||||
// Normalize the items to create a clean data structure.
|
||||
logger.debug('[extractCoreDataFromFlyerImage] Normalizing extracted items.');
|
||||
const normalizedItems = Array.isArray(extractedData.items)
|
||||
? this._normalizeExtractedItems(extractedData.items)
|
||||
: [];
|
||||
// The FlyerDataTransformer is now responsible for all normalization.
|
||||
// We return the raw items as parsed from the AI response.
|
||||
if (!Array.isArray(extractedData.items)) {
|
||||
extractedData.items = [];
|
||||
}
|
||||
|
||||
logger.info(
|
||||
`[extractCoreDataFromFlyerImage] Successfully processed flyer data for store: ${extractedData.store_name}. Exiting method.`,
|
||||
);
|
||||
return { ...extractedData, items: normalizedItems };
|
||||
return extractedData;
|
||||
} catch (apiError) {
|
||||
logger.error({ err: apiError }, '[extractCoreDataFromFlyerImage] The entire process failed.');
|
||||
throw apiError;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Normalizes the raw items returned by the AI, ensuring fields are in the correct format.
|
||||
* @param items An array of raw flyer items from the AI.
|
||||
* @returns A normalized array of flyer items.
|
||||
*/
|
||||
private _normalizeExtractedItems(items: RawFlyerItem[]): ExtractedFlyerItem[] {
|
||||
return items.map((item: RawFlyerItem) => ({
|
||||
...item,
|
||||
// Ensure 'item' is always a string, defaulting to 'Unknown Item' if null/undefined.
|
||||
item:
|
||||
item.item === null || item.item === undefined || String(item.item).trim() === ''
|
||||
? 'Unknown Item'
|
||||
: String(item.item),
|
||||
price_display:
|
||||
item.price_display === null || item.price_display === undefined
|
||||
? ''
|
||||
: String(item.price_display),
|
||||
quantity: item.quantity === null || item.quantity === undefined ? '' : String(item.quantity),
|
||||
category_name:
|
||||
item.category_name === null || item.category_name === undefined
|
||||
? 'Other/Miscellaneous'
|
||||
: String(item.category_name),
|
||||
// Ensure undefined is converted to null to match the Zod schema.
|
||||
price_in_cents: item.price_in_cents ?? null,
|
||||
master_item_id: item.master_item_id ?? undefined,
|
||||
}));
|
||||
}
|
||||
|
||||
/**
|
||||
* SERVER-SIDE FUNCTION
|
||||
* Extracts a specific piece of text from a cropped area of an image.
|
||||
@@ -780,6 +794,8 @@ async enqueueFlyerProcessing(
|
||||
.join(', ');
|
||||
}
|
||||
|
||||
const baseUrl = getBaseUrl(logger);
|
||||
|
||||
// 3. Add job to the queue
|
||||
const job = await flyerQueue.add('process-flyer', {
|
||||
filePath: file.path,
|
||||
@@ -788,6 +804,7 @@ async enqueueFlyerProcessing(
|
||||
userId: userProfile?.user.user_id,
|
||||
submitterIp: submitterIp,
|
||||
userProfileAddress: userProfileAddress,
|
||||
baseUrl: baseUrl,
|
||||
});
|
||||
|
||||
logger.info(
|
||||
@@ -865,6 +882,8 @@ async enqueueFlyerProcessing(
|
||||
const itemsArray = Array.isArray(rawItems) ? rawItems : typeof rawItems === 'string' ? JSON.parse(rawItems) : [];
|
||||
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
||||
...item,
|
||||
// Ensure price_display is never null to satisfy database constraints.
|
||||
price_display: item.price_display ?? '',
|
||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
||||
quantity: item.quantity ?? 1,
|
||||
view_count: 0,
|
||||
@@ -879,11 +898,14 @@ async enqueueFlyerProcessing(
|
||||
|
||||
const iconsDir = path.join(path.dirname(file.path), 'icons');
|
||||
const iconFileName = await generateFlyerIcon(file.path, iconsDir, logger);
|
||||
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
||||
|
||||
const baseUrl = getBaseUrl(logger);
|
||||
const iconUrl = `${baseUrl}/flyer-images/icons/${iconFileName}`;
|
||||
const imageUrl = `${baseUrl}/flyer-images/${file.filename}`;
|
||||
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${file.filename}`,
|
||||
image_url: imageUrl,
|
||||
icon_url: iconUrl,
|
||||
checksum: checksum,
|
||||
store_name: storeName,
|
||||
@@ -895,18 +917,28 @@ async enqueueFlyerProcessing(
|
||||
uploaded_by: userProfile?.user.user_id,
|
||||
};
|
||||
|
||||
const { flyer: newFlyer, items: newItems } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||
return db.withTransaction(async (client) => {
|
||||
const { flyer, items } = await createFlyerAndItems(flyerData, itemsForDb, logger, client);
|
||||
|
||||
logger.info(`Successfully processed legacy flyer: ${newFlyer.file_name} (ID: ${newFlyer.flyer_id}) with ${newItems.length} items.`);
|
||||
logger.info(
|
||||
`Successfully processed legacy flyer: ${flyer.file_name} (ID: ${flyer.flyer_id}) with ${items.length} items.`,
|
||||
);
|
||||
|
||||
await db.adminRepo.logActivity({
|
||||
userId: userProfile?.user.user_id,
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
||||
}, logger);
|
||||
|
||||
return newFlyer;
|
||||
const transactionalAdminRepo = new AdminRepository(client);
|
||||
await transactionalAdminRepo.logActivity(
|
||||
{
|
||||
userId: userProfile?.user.user_id,
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||
details: { flyerId: flyer.flyer_id, storeName: flyerData.store_name },
|
||||
},
|
||||
logger,
|
||||
);
|
||||
return flyer;
|
||||
}).catch((error) => {
|
||||
logger.error({ err: error, checksum }, 'Legacy flyer upload database transaction failed.');
|
||||
throw error;
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -1,7 +1,29 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
// src/services/authService.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import type { UserProfile } from '../types';
|
||||
import type * as jsonwebtoken from 'jsonwebtoken';
|
||||
|
||||
const { transactionalUserRepoMocks, transactionalAdminRepoMocks } = vi.hoisted(() => {
|
||||
return {
|
||||
transactionalUserRepoMocks: {
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteResetToken: vi.fn(),
|
||||
createPasswordResetToken: vi.fn(),
|
||||
createUser: vi.fn(),
|
||||
},
|
||||
transactionalAdminRepoMocks: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
};
|
||||
});
|
||||
|
||||
vi.mock('./db/user.db', () => ({
|
||||
UserRepository: vi.fn().mockImplementation(function () { return transactionalUserRepoMocks }),
|
||||
}));
|
||||
vi.mock('./db/admin.db', () => ({
|
||||
AdminRepository: vi.fn().mockImplementation(function () { return transactionalAdminRepoMocks }),
|
||||
}));
|
||||
|
||||
describe('AuthService', () => {
|
||||
let authService: typeof import('./authService').authService;
|
||||
let bcrypt: typeof import('bcrypt');
|
||||
@@ -10,7 +32,10 @@ describe('AuthService', () => {
|
||||
let adminRepo: typeof import('./db/index.db').adminRepo;
|
||||
let logger: typeof import('./logger.server').logger;
|
||||
let sendPasswordResetEmail: typeof import('./emailService.server').sendPasswordResetEmail;
|
||||
let DatabaseError: typeof import('./processingErrors').DatabaseError;
|
||||
let UniqueConstraintError: typeof import('./db/errors.db').UniqueConstraintError;
|
||||
let RepositoryError: typeof import('./db/errors.db').RepositoryError;
|
||||
let withTransaction: typeof import('./db/index.db').withTransaction;
|
||||
|
||||
const reqLog = {}; // Mock request logger object
|
||||
const mockUser = {
|
||||
@@ -33,13 +58,14 @@ describe('AuthService', () => {
|
||||
vi.resetModules();
|
||||
|
||||
// Set environment variables before any modules are imported
|
||||
process.env.JWT_SECRET = 'test-secret';
|
||||
process.env.FRONTEND_URL = 'http://localhost:3000';
|
||||
vi.stubEnv('JWT_SECRET', 'test-secret');
|
||||
vi.stubEnv('FRONTEND_URL', 'http://localhost:3000');
|
||||
|
||||
// Mock all dependencies before dynamically importing the service
|
||||
// Core modules like bcrypt, jsonwebtoken, and crypto are now mocked globally in tests-setup-unit.ts
|
||||
vi.mock('bcrypt');
|
||||
vi.mock('./db/index.db', () => ({
|
||||
withTransaction: vi.fn(),
|
||||
userRepo: {
|
||||
createUser: vi.fn(),
|
||||
saveRefreshToken: vi.fn(),
|
||||
@@ -73,14 +99,26 @@ describe('AuthService', () => {
|
||||
userRepo = dbModule.userRepo;
|
||||
adminRepo = dbModule.adminRepo;
|
||||
logger = (await import('./logger.server')).logger;
|
||||
withTransaction = (await import('./db/index.db')).withTransaction;
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback: any) => {
|
||||
return callback({}); // Mock client
|
||||
});
|
||||
const { validatePasswordStrength } = await import('../utils/authUtils');
|
||||
vi.mocked(validatePasswordStrength).mockReturnValue({ isValid: true, feedback: '' });
|
||||
sendPasswordResetEmail = (await import('./emailService.server')).sendPasswordResetEmail;
|
||||
DatabaseError = (await import('./processingErrors')).DatabaseError;
|
||||
UniqueConstraintError = (await import('./db/errors.db')).UniqueConstraintError;
|
||||
RepositoryError = (await import('./db/errors.db')).RepositoryError;
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
describe('registerUser', () => {
|
||||
it('should successfully register a new user', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
vi.mocked(userRepo.createUser).mockResolvedValue(mockUserProfile);
|
||||
vi.mocked(transactionalUserRepoMocks.createUser).mockResolvedValue(mockUserProfile);
|
||||
|
||||
const result = await authService.registerUser(
|
||||
'test@example.com',
|
||||
@@ -91,13 +129,14 @@ describe('AuthService', () => {
|
||||
);
|
||||
|
||||
expect(bcrypt.hash).toHaveBeenCalledWith('password123', 10);
|
||||
expect(userRepo.createUser).toHaveBeenCalledWith(
|
||||
expect(transactionalUserRepoMocks.createUser).toHaveBeenCalledWith(
|
||||
'test@example.com',
|
||||
'hashed-password',
|
||||
{ full_name: 'Test User', avatar_url: undefined },
|
||||
reqLog,
|
||||
{},
|
||||
);
|
||||
expect(adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect(transactionalAdminRepoMocks.logActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
action: 'user_registered',
|
||||
userId: 'user-123',
|
||||
@@ -110,25 +149,25 @@ describe('AuthService', () => {
|
||||
it('should throw UniqueConstraintError if email already exists', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
const error = new UniqueConstraintError('Email exists');
|
||||
vi.mocked(userRepo.createUser).mockRejectedValue(error);
|
||||
vi.mocked(withTransaction).mockRejectedValue(error);
|
||||
|
||||
await expect(
|
||||
authService.registerUser('test@example.com', 'password123', undefined, undefined, reqLog),
|
||||
).rejects.toThrow(UniqueConstraintError);
|
||||
|
||||
expect(logger.error).not.toHaveBeenCalled(); // Should not log expected unique constraint errors as system errors
|
||||
expect(logger.error).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should log and throw other errors', async () => {
|
||||
it('should log and re-throw generic errors on registration failure', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
const error = new Error('Database failed');
|
||||
vi.mocked(userRepo.createUser).mockRejectedValue(error);
|
||||
vi.mocked(withTransaction).mockRejectedValue(error);
|
||||
|
||||
await expect(
|
||||
authService.registerUser('test@example.com', 'password123', undefined, undefined, reqLog),
|
||||
).rejects.toThrow('Database failed');
|
||||
).rejects.toThrow(DatabaseError);
|
||||
|
||||
expect(logger.error).toHaveBeenCalled();
|
||||
expect(logger.error).toHaveBeenCalledWith({ error, email: 'test@example.com' }, `User registration failed with an unexpected error.`);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -136,7 +175,7 @@ describe('AuthService', () => {
|
||||
it('should register user and return tokens', async () => {
|
||||
// Mock registerUser logic (since we can't easily spy on the same class instance method without prototype spying, we rely on the underlying calls)
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
vi.mocked(userRepo.createUser).mockResolvedValue(mockUserProfile);
|
||||
vi.mocked(transactionalUserRepoMocks.createUser).mockResolvedValue(mockUserProfile);
|
||||
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
||||
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
||||
// We must mock `jwt.default.sign` to affect the code under test.
|
||||
@@ -194,17 +233,13 @@ describe('AuthService', () => {
|
||||
expect(userRepo.saveRefreshToken).toHaveBeenCalledWith('user-123', 'token', reqLog);
|
||||
});
|
||||
|
||||
it('should log and throw error on failure', async () => {
|
||||
it('should propagate the error from the repository on failure', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(userRepo.saveRefreshToken).mockRejectedValue(error);
|
||||
|
||||
await expect(authService.saveRefreshToken('user-123', 'token', reqLog)).rejects.toThrow(
|
||||
'DB Error',
|
||||
);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ error }),
|
||||
expect.stringContaining('Failed to save refresh token'),
|
||||
);
|
||||
// The service method now directly propagates the error from the repo.
|
||||
await expect(authService.saveRefreshToken('user-123', 'token', reqLog)).rejects.toThrow(error);
|
||||
expect(logger.error).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -215,11 +250,12 @@ describe('AuthService', () => {
|
||||
|
||||
const result = await authService.resetPassword('test@example.com', reqLog);
|
||||
|
||||
expect(userRepo.createPasswordResetToken).toHaveBeenCalledWith(
|
||||
expect(transactionalUserRepoMocks.createPasswordResetToken).toHaveBeenCalledWith(
|
||||
'user-123',
|
||||
'hashed-token',
|
||||
expect.any(Date),
|
||||
reqLog,
|
||||
{},
|
||||
);
|
||||
expect(sendPasswordResetEmail).toHaveBeenCalledWith(
|
||||
'test@example.com',
|
||||
@@ -253,36 +289,50 @@ describe('AuthService', () => {
|
||||
});
|
||||
|
||||
describe('updatePassword', () => {
|
||||
it('should update password if token is valid', async () => {
|
||||
it('should update password if token is valid and wrap operations in a transaction', async () => {
|
||||
const mockTokenRecord = {
|
||||
user_id: 'user-123',
|
||||
token_hash: 'hashed-token',
|
||||
};
|
||||
vi.mocked(userRepo.getValidResetTokens).mockResolvedValue([mockTokenRecord] as any);
|
||||
vi.mocked(bcrypt.compare).mockImplementation(async () => true); // Match found
|
||||
vi.mocked(bcrypt.compare).mockImplementation(async () => true);
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'new-hashed-password');
|
||||
|
||||
const result = await authService.updatePassword('valid-token', 'newPassword', reqLog);
|
||||
|
||||
expect(userRepo.updateUserPassword).toHaveBeenCalledWith(
|
||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||
|
||||
expect(transactionalUserRepoMocks.updateUserPassword).toHaveBeenCalledWith(
|
||||
'user-123',
|
||||
'new-hashed-password',
|
||||
reqLog,
|
||||
);
|
||||
expect(userRepo.deleteResetToken).toHaveBeenCalledWith('hashed-token', reqLog);
|
||||
expect(adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect(transactionalUserRepoMocks.deleteResetToken).toHaveBeenCalledWith('hashed-token', reqLog);
|
||||
expect(transactionalAdminRepoMocks.logActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ action: 'password_reset' }),
|
||||
reqLog,
|
||||
);
|
||||
expect(result).toBe(true);
|
||||
});
|
||||
|
||||
it('should log and re-throw an error if the transaction fails', async () => {
|
||||
const mockTokenRecord = { user_id: 'user-123', token_hash: 'hashed-token' };
|
||||
vi.mocked(userRepo.getValidResetTokens).mockResolvedValue([mockTokenRecord] as any);
|
||||
vi.mocked(bcrypt.compare).mockImplementation(async () => true);
|
||||
const dbError = new Error('Transaction failed');
|
||||
vi.mocked(withTransaction).mockRejectedValue(dbError);
|
||||
|
||||
await expect(authService.updatePassword('valid-token', 'newPassword', reqLog)).rejects.toThrow(DatabaseError);
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith({ error: dbError }, `An unexpected error occurred during password update.`);
|
||||
});
|
||||
|
||||
it('should return null if token is invalid or not found', async () => {
|
||||
vi.mocked(userRepo.getValidResetTokens).mockResolvedValue([]);
|
||||
|
||||
const result = await authService.updatePassword('invalid-token', 'newPassword', reqLog);
|
||||
|
||||
expect(userRepo.updateUserPassword).not.toHaveBeenCalled();
|
||||
expect(transactionalUserRepoMocks.updateUserPassword).not.toHaveBeenCalled();
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
});
|
||||
@@ -304,6 +354,37 @@ describe('AuthService', () => {
|
||||
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
|
||||
it('should throw a DatabaseError if finding the user fails with a generic error', async () => {
|
||||
const dbError = new Error('DB connection failed');
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockRejectedValue(dbError);
|
||||
|
||||
// Use a try-catch to assert on the error instance properties, which is more robust
|
||||
// than `toBeInstanceOf` in some complex module mocking scenarios in Vitest.
|
||||
try {
|
||||
await authService.getUserByRefreshToken('any-token', reqLog);
|
||||
expect.fail('Expected an error to be thrown');
|
||||
} catch (error: any) {
|
||||
expect(error.name).toBe('DatabaseError');
|
||||
expect(error.message).toBe('DB connection failed');
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: dbError, refreshToken: 'any-token' },
|
||||
'An unexpected error occurred while fetching user by refresh token.',
|
||||
);
|
||||
}
|
||||
});
|
||||
|
||||
it('should re-throw a RepositoryError if finding the user fails with a known error', async () => {
|
||||
const repoError = new RepositoryError('Some repo error', 500);
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockRejectedValue(repoError);
|
||||
|
||||
await expect(authService.getUserByRefreshToken('any-token', reqLog)).rejects.toThrow(repoError);
|
||||
// The original error is re-thrown, so the generic wrapper log should not be called.
|
||||
expect(logger.error).not.toHaveBeenCalledWith(
|
||||
expect.any(Object),
|
||||
'An unexpected error occurred while fetching user by refresh token.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('logout', () => {
|
||||
@@ -312,12 +393,12 @@ describe('AuthService', () => {
|
||||
expect(userRepo.deleteRefreshToken).toHaveBeenCalledWith('token', reqLog);
|
||||
});
|
||||
|
||||
it('should log and throw on error', async () => {
|
||||
it('should propagate the error from the repository on failure', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(userRepo.deleteRefreshToken).mockRejectedValue(error);
|
||||
|
||||
await expect(authService.logout('token', reqLog)).rejects.toThrow('DB Error');
|
||||
expect(logger.error).toHaveBeenCalled();
|
||||
await expect(authService.logout('token', reqLog)).rejects.toThrow(error);
|
||||
expect(logger.error).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -340,5 +421,13 @@ describe('AuthService', () => {
|
||||
const result = await authService.refreshAccessToken('invalid-token', reqLog);
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
|
||||
it('should propagate errors from getUserByRefreshToken', async () => {
|
||||
const dbError = new DatabaseError('Underlying DB call failed');
|
||||
// We mock the service's own method since refreshAccessToken calls it directly.
|
||||
vi.spyOn(authService, 'getUserByRefreshToken').mockRejectedValue(dbError);
|
||||
|
||||
await expect(authService.refreshAccessToken('any-token', reqLog)).rejects.toThrow(dbError);
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -2,9 +2,9 @@
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import jwt from 'jsonwebtoken';
|
||||
import crypto from 'crypto';
|
||||
import { userRepo, adminRepo } from './db/index.db';
|
||||
import { UniqueConstraintError } from './db/errors.db';
|
||||
import { getPool } from './db/connection.db';
|
||||
import { DatabaseError, FlyerProcessingError } from './processingErrors';
|
||||
import { withTransaction, userRepo } from './db/index.db';
|
||||
import { RepositoryError, ValidationError } from './db/errors.db';
|
||||
import { logger } from './logger.server';
|
||||
import { sendPasswordResetEmail } from './emailService.server';
|
||||
import type { UserProfile } from '../types';
|
||||
@@ -20,44 +20,48 @@ class AuthService {
|
||||
avatarUrl: string | undefined,
|
||||
reqLog: any,
|
||||
) {
|
||||
try {
|
||||
const strength = validatePasswordStrength(password);
|
||||
if (!strength.isValid) {
|
||||
throw new ValidationError([], strength.feedback);
|
||||
}
|
||||
|
||||
// Wrap user creation and activity logging in a transaction for atomicity.
|
||||
// The `createUser` method is now designed to be composed within other transactions.
|
||||
return withTransaction(async (client) => {
|
||||
const transactionalUserRepo = new (await import('./db/user.db')).UserRepository(client);
|
||||
const adminRepo = new (await import('./db/admin.db')).AdminRepository(client);
|
||||
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(password, saltRounds);
|
||||
logger.info(`Hashing password for new user: ${email}`);
|
||||
|
||||
// The createUser method in UserRepository now handles its own transaction.
|
||||
const newUser = await userRepo.createUser(
|
||||
const newUser = await transactionalUserRepo.createUser(
|
||||
email,
|
||||
hashedPassword,
|
||||
{ full_name: fullName, avatar_url: avatarUrl },
|
||||
reqLog,
|
||||
client, // Pass the transactional client
|
||||
);
|
||||
|
||||
const userEmail = newUser.user.email;
|
||||
const userId = newUser.user.user_id;
|
||||
logger.info(`Successfully created new user in DB: ${userEmail} (ID: ${userId})`);
|
||||
logger.info(`Successfully created new user in DB: ${newUser.user.email} (ID: ${newUser.user.user_id})`);
|
||||
|
||||
// Use the new standardized logging function
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: newUser.user.user_id,
|
||||
action: 'user_registered',
|
||||
displayText: `${userEmail} has registered.`,
|
||||
icon: 'user-plus',
|
||||
},
|
||||
{ userId: newUser.user.user_id, action: 'user_registered', displayText: `${email} has registered.`, icon: 'user-plus' },
|
||||
reqLog,
|
||||
);
|
||||
|
||||
return newUser;
|
||||
} catch (error: unknown) {
|
||||
if (error instanceof UniqueConstraintError) {
|
||||
// If the email is a duplicate, return a 409 Conflict status.
|
||||
}).catch((error: unknown) => {
|
||||
// Re-throw known repository errors (like UniqueConstraintError) to allow for specific handling upstream.
|
||||
if (error instanceof RepositoryError) {
|
||||
throw error;
|
||||
}
|
||||
logger.error({ error }, `User registration route failed for email: ${email}.`);
|
||||
// Pass the error to the centralized handler
|
||||
throw error;
|
||||
}
|
||||
// For unknown errors, log them and wrap them in a generic DatabaseError
|
||||
// to standardize the error contract of the service layer.
|
||||
const message = error instanceof Error ? error.message : 'An unknown error occurred during registration.';
|
||||
logger.error({ error, email }, `User registration failed with an unexpected error.`);
|
||||
throw new DatabaseError(message);
|
||||
});
|
||||
}
|
||||
|
||||
async registerAndLoginUser(
|
||||
@@ -91,15 +95,9 @@ class AuthService {
|
||||
}
|
||||
|
||||
async saveRefreshToken(userId: string, refreshToken: string, reqLog: any) {
|
||||
try {
|
||||
await userRepo.saveRefreshToken(userId, refreshToken, reqLog);
|
||||
} catch (tokenErr) {
|
||||
logger.error(
|
||||
{ error: tokenErr },
|
||||
`Failed to save refresh token during login for user: ${userId}`,
|
||||
);
|
||||
throw tokenErr;
|
||||
}
|
||||
// The repository method `saveRefreshToken` already includes robust error handling
|
||||
// and logging via `handleDbError`. No need for a redundant try/catch block here.
|
||||
await userRepo.saveRefreshToken(userId, refreshToken, reqLog);
|
||||
}
|
||||
|
||||
async handleSuccessfulLogin(userProfile: UserProfile, reqLog: any) {
|
||||
@@ -124,7 +122,11 @@ class AuthService {
|
||||
const tokenHash = await bcrypt.hash(token, saltRounds);
|
||||
const expiresAt = new Date(Date.now() + 3600000); // 1 hour
|
||||
|
||||
await userRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, reqLog);
|
||||
// Wrap the token creation in a transaction to ensure atomicity of the DELETE and INSERT operations.
|
||||
await withTransaction(async (client) => {
|
||||
const transactionalUserRepo = new (await import('./db/user.db')).UserRepository(client);
|
||||
await transactionalUserRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, reqLog, client);
|
||||
});
|
||||
|
||||
const resetLink = `${process.env.FRONTEND_URL}/reset-password/${token}`;
|
||||
|
||||
@@ -139,13 +141,29 @@ class AuthService {
|
||||
|
||||
return token;
|
||||
} catch (error) {
|
||||
logger.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
||||
throw error;
|
||||
// Re-throw known repository errors to allow for specific handling upstream.
|
||||
if (error instanceof RepositoryError) {
|
||||
throw error;
|
||||
}
|
||||
// For unknown errors, log them and wrap them in a generic DatabaseError.
|
||||
const message = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ error, email }, `An unexpected error occurred during password reset for email: ${email}`);
|
||||
throw new DatabaseError(message);
|
||||
}
|
||||
}
|
||||
|
||||
async updatePassword(token: string, newPassword: string, reqLog: any) {
|
||||
try {
|
||||
const strength = validatePasswordStrength(newPassword);
|
||||
if (!strength.isValid) {
|
||||
throw new ValidationError([], strength.feedback);
|
||||
}
|
||||
|
||||
// Wrap all database operations in a transaction to ensure atomicity.
|
||||
return withTransaction(async (client) => {
|
||||
const transactionalUserRepo = new (await import('./db/user.db')).UserRepository(client);
|
||||
const adminRepo = new (await import('./db/admin.db')).AdminRepository(client);
|
||||
|
||||
// This read can happen outside the transaction if we use the non-transactional repo.
|
||||
const validTokens = await userRepo.getValidResetTokens(reqLog);
|
||||
let tokenRecord;
|
||||
for (const record of validTokens) {
|
||||
@@ -157,32 +175,31 @@ class AuthService {
|
||||
}
|
||||
|
||||
if (!tokenRecord) {
|
||||
return null;
|
||||
return null; // Token is invalid or expired, not an error.
|
||||
}
|
||||
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||
|
||||
await userRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, reqLog);
|
||||
await userRepo.deleteResetToken(tokenRecord.token_hash, reqLog);
|
||||
|
||||
// Log this security event after a successful password reset.
|
||||
// These three writes are now atomic.
|
||||
await transactionalUserRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, reqLog);
|
||||
await transactionalUserRepo.deleteResetToken(tokenRecord.token_hash, reqLog);
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: tokenRecord.user_id,
|
||||
action: 'password_reset',
|
||||
displayText: `User ID ${tokenRecord.user_id} has reset their password.`,
|
||||
icon: 'key',
|
||||
details: { source_ip: null },
|
||||
},
|
||||
{ userId: tokenRecord.user_id, action: 'password_reset', displayText: `User ID ${tokenRecord.user_id} has reset their password.`, icon: 'key' },
|
||||
reqLog,
|
||||
);
|
||||
|
||||
return true;
|
||||
} catch (error) {
|
||||
logger.error({ error }, `An error occurred during password reset.`);
|
||||
throw error;
|
||||
}
|
||||
}).catch((error) => {
|
||||
// Re-throw known repository errors to allow for specific handling upstream.
|
||||
if (error instanceof RepositoryError) {
|
||||
throw error;
|
||||
}
|
||||
// For unknown errors, log them and wrap them in a generic DatabaseError.
|
||||
const message = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ error }, `An unexpected error occurred during password update.`);
|
||||
throw new DatabaseError(message);
|
||||
});
|
||||
}
|
||||
|
||||
async getUserByRefreshToken(refreshToken: string, reqLog: any) {
|
||||
@@ -194,18 +211,22 @@ class AuthService {
|
||||
const userProfile = await userRepo.findUserProfileById(basicUser.user_id, reqLog);
|
||||
return userProfile;
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'An error occurred during /refresh-token.');
|
||||
throw error;
|
||||
// Re-throw known repository errors to allow for specific handling upstream.
|
||||
if (error instanceof RepositoryError) {
|
||||
throw error;
|
||||
}
|
||||
// For unknown errors, log them and wrap them in a generic DatabaseError.
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ error, refreshToken }, 'An unexpected error occurred while fetching user by refresh token.');
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
|
||||
async logout(refreshToken: string, reqLog: any) {
|
||||
try {
|
||||
await userRepo.deleteRefreshToken(refreshToken, reqLog);
|
||||
} catch (err: any) {
|
||||
logger.error({ error: err }, 'Failed to delete refresh token from DB during logout.');
|
||||
throw err;
|
||||
}
|
||||
// The repository method `deleteRefreshToken` now includes robust error handling
|
||||
// and logging via `handleDbError`. No need for a redundant try/catch block here.
|
||||
// The original implementation also swallowed errors, which is now fixed.
|
||||
await userRepo.deleteRefreshToken(refreshToken, reqLog);
|
||||
}
|
||||
|
||||
async refreshAccessToken(refreshToken: string, reqLog: any): Promise<{ accessToken: string } | null> {
|
||||
|
||||
@@ -106,7 +106,13 @@ describe('Address DB Service', () => {
|
||||
'An identical address already exists.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: addressData },
|
||||
{
|
||||
err: dbError,
|
||||
address: addressData,
|
||||
code: '23505',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -715,7 +715,14 @@ describe('Admin DB Service', () => {
|
||||
adminRepo.updateUserRole('non-existent-user', 'admin', mockLogger),
|
||||
).rejects.toThrow('The specified user does not exist.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: 'non-existent-user', role: 'admin' },
|
||||
{
|
||||
err: dbError,
|
||||
userId: 'non-existent-user',
|
||||
role: 'admin',
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in updateUserRole',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -1,26 +1,34 @@
|
||||
// src/services/db/errors.db.test.ts
|
||||
import { describe, it, expect } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Logger } from 'pino';
|
||||
import {
|
||||
DatabaseError,
|
||||
RepositoryError,
|
||||
UniqueConstraintError,
|
||||
ForeignKeyConstraintError,
|
||||
NotFoundError,
|
||||
ValidationError,
|
||||
FileUploadError,
|
||||
NotNullConstraintError,
|
||||
CheckConstraintError,
|
||||
InvalidTextRepresentationError,
|
||||
NumericValueOutOfRangeError,
|
||||
handleDbError,
|
||||
} from './errors.db';
|
||||
|
||||
vi.mock('./logger.server');
|
||||
|
||||
describe('Custom Database and Application Errors', () => {
|
||||
describe('DatabaseError', () => {
|
||||
describe('RepositoryError', () => {
|
||||
it('should create a generic database error with a message and status', () => {
|
||||
const message = 'Generic DB Error';
|
||||
const status = 500;
|
||||
const error = new DatabaseError(message, status);
|
||||
const error = new RepositoryError(message, status);
|
||||
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
expect(error).toBeInstanceOf(DatabaseError);
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error.message).toBe(message);
|
||||
expect(error.status).toBe(status);
|
||||
expect(error.name).toBe('DatabaseError');
|
||||
expect(error.name).toBe('RepositoryError');
|
||||
});
|
||||
});
|
||||
|
||||
@@ -29,7 +37,7 @@ describe('Custom Database and Application Errors', () => {
|
||||
const error = new UniqueConstraintError();
|
||||
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
expect(error).toBeInstanceOf(DatabaseError);
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error).toBeInstanceOf(UniqueConstraintError);
|
||||
expect(error.message).toBe('The record already exists.');
|
||||
expect(error.status).toBe(409);
|
||||
@@ -48,7 +56,7 @@ describe('Custom Database and Application Errors', () => {
|
||||
const error = new ForeignKeyConstraintError();
|
||||
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
expect(error).toBeInstanceOf(DatabaseError);
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error).toBeInstanceOf(ForeignKeyConstraintError);
|
||||
expect(error.message).toBe('The referenced record does not exist.');
|
||||
expect(error.status).toBe(400);
|
||||
@@ -67,7 +75,7 @@ describe('Custom Database and Application Errors', () => {
|
||||
const error = new NotFoundError();
|
||||
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
expect(error).toBeInstanceOf(DatabaseError);
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error).toBeInstanceOf(NotFoundError);
|
||||
expect(error.message).toBe('The requested resource was not found.');
|
||||
expect(error.status).toBe(404);
|
||||
@@ -87,7 +95,7 @@ describe('Custom Database and Application Errors', () => {
|
||||
const error = new ValidationError(validationIssues);
|
||||
|
||||
expect(error).toBeInstanceOf(Error);
|
||||
expect(error).toBeInstanceOf(DatabaseError);
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error).toBeInstanceOf(ValidationError);
|
||||
expect(error.message).toBe('The request data is invalid.');
|
||||
expect(error.status).toBe(400);
|
||||
@@ -114,4 +122,161 @@ describe('Custom Database and Application Errors', () => {
|
||||
expect(error.name).toBe('FileUploadError');
|
||||
});
|
||||
});
|
||||
|
||||
describe('NotNullConstraintError', () => {
|
||||
it('should create an error with a default message and status 400', () => {
|
||||
const error = new NotNullConstraintError();
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error.message).toBe('A required field was left null.');
|
||||
expect(error.status).toBe(400);
|
||||
expect(error.name).toBe('NotNullConstraintError');
|
||||
});
|
||||
|
||||
it('should create an error with a custom message', () => {
|
||||
const message = 'Email cannot be null.';
|
||||
const error = new NotNullConstraintError(message);
|
||||
expect(error.message).toBe(message);
|
||||
});
|
||||
});
|
||||
|
||||
describe('CheckConstraintError', () => {
|
||||
it('should create an error with a default message and status 400', () => {
|
||||
const error = new CheckConstraintError();
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error.message).toBe('A check constraint was violated.');
|
||||
expect(error.status).toBe(400);
|
||||
expect(error.name).toBe('CheckConstraintError');
|
||||
});
|
||||
|
||||
it('should create an error with a custom message', () => {
|
||||
const message = 'Price must be positive.';
|
||||
const error = new CheckConstraintError(message);
|
||||
expect(error.message).toBe(message);
|
||||
});
|
||||
});
|
||||
|
||||
describe('InvalidTextRepresentationError', () => {
|
||||
it('should create an error with a default message and status 400', () => {
|
||||
const error = new InvalidTextRepresentationError();
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error.message).toBe('A value has an invalid format for its data type.');
|
||||
expect(error.status).toBe(400);
|
||||
expect(error.name).toBe('InvalidTextRepresentationError');
|
||||
});
|
||||
|
||||
it('should create an error with a custom message', () => {
|
||||
const message = 'Invalid input syntax for type integer: "abc"';
|
||||
const error = new InvalidTextRepresentationError(message);
|
||||
expect(error.message).toBe(message);
|
||||
});
|
||||
});
|
||||
|
||||
describe('NumericValueOutOfRangeError', () => {
|
||||
it('should create an error with a default message and status 400', () => {
|
||||
const error = new NumericValueOutOfRangeError();
|
||||
expect(error).toBeInstanceOf(RepositoryError);
|
||||
expect(error.message).toBe('A numeric value is out of the allowed range.');
|
||||
expect(error.status).toBe(400);
|
||||
expect(error.name).toBe('NumericValueOutOfRangeError');
|
||||
});
|
||||
|
||||
it('should create an error with a custom message', () => {
|
||||
const message = 'Value too large for type smallint.';
|
||||
const error = new NumericValueOutOfRangeError(message);
|
||||
expect(error.message).toBe(message);
|
||||
});
|
||||
});
|
||||
|
||||
describe('handleDbError', () => {
|
||||
const mockLogger = {
|
||||
error: vi.fn(),
|
||||
} as unknown as Logger;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
it('should re-throw existing RepositoryError instances without logging', () => {
|
||||
const notFound = new NotFoundError('Test not found');
|
||||
expect(() => handleDbError(notFound, mockLogger, 'msg', {})).toThrow(notFound);
|
||||
expect(mockLogger.error).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should throw UniqueConstraintError for code 23505', () => {
|
||||
const dbError = new Error('duplicate key');
|
||||
(dbError as any).code = '23505';
|
||||
expect(() =>
|
||||
handleDbError(dbError, mockLogger, 'msg', {}, { uniqueMessage: 'custom unique' }),
|
||||
).toThrow('custom unique');
|
||||
});
|
||||
|
||||
it('should throw ForeignKeyConstraintError for code 23503', () => {
|
||||
const dbError = new Error('fk violation');
|
||||
(dbError as any).code = '23503';
|
||||
expect(() =>
|
||||
handleDbError(dbError, mockLogger, 'msg', {}, { fkMessage: 'custom fk' }),
|
||||
).toThrow('custom fk');
|
||||
});
|
||||
|
||||
it('should throw NotNullConstraintError for code 23502', () => {
|
||||
const dbError = new Error('not null violation');
|
||||
(dbError as any).code = '23502';
|
||||
expect(() =>
|
||||
handleDbError(dbError, mockLogger, 'msg', {}, { notNullMessage: 'custom not null' }),
|
||||
).toThrow('custom not null');
|
||||
});
|
||||
|
||||
it('should throw CheckConstraintError for code 23514', () => {
|
||||
const dbError = new Error('check violation');
|
||||
(dbError as any).code = '23514';
|
||||
expect(() =>
|
||||
handleDbError(dbError, mockLogger, 'msg', {}, { checkMessage: 'custom check' }),
|
||||
).toThrow('custom check');
|
||||
});
|
||||
|
||||
it('should throw InvalidTextRepresentationError for code 22P02', () => {
|
||||
const dbError = new Error('invalid text');
|
||||
(dbError as any).code = '22P02';
|
||||
expect(() =>
|
||||
handleDbError(dbError, mockLogger, 'msg', {}, { invalidTextMessage: 'custom invalid text' }),
|
||||
).toThrow('custom invalid text');
|
||||
});
|
||||
|
||||
it('should throw NumericValueOutOfRangeError for code 22003', () => {
|
||||
const dbError = new Error('out of range');
|
||||
(dbError as any).code = '22003';
|
||||
expect(() =>
|
||||
handleDbError(
|
||||
dbError,
|
||||
mockLogger,
|
||||
'msg',
|
||||
{},
|
||||
{ numericOutOfRangeMessage: 'custom out of range' },
|
||||
),
|
||||
).toThrow('custom out of range');
|
||||
});
|
||||
|
||||
it('should throw a generic Error with a default message', () => {
|
||||
const genericError = new Error('Something else happened');
|
||||
expect(() =>
|
||||
handleDbError(genericError, mockLogger, 'msg', {}, { defaultMessage: 'Oops' }),
|
||||
).toThrow('Oops');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith({ err: genericError }, 'msg');
|
||||
});
|
||||
|
||||
it('should throw a generic Error with a constructed message using entityName', () => {
|
||||
const genericError = new Error('Something else happened');
|
||||
expect(() =>
|
||||
handleDbError(genericError, mockLogger, 'msg', {}, { entityName: 'User' }),
|
||||
).toThrow('Failed to perform operation on User.');
|
||||
});
|
||||
|
||||
it('should throw a generic Error with a constructed message using "database" as a fallback', () => {
|
||||
const genericError = new Error('Something else happened');
|
||||
// No defaultMessage or entityName provided
|
||||
expect(() => handleDbError(genericError, mockLogger, 'msg', {}, {})).toThrow(
|
||||
'Failed to perform operation on database.',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,10 +1,11 @@
|
||||
// src/services/db/errors.db.ts
|
||||
import type { Logger } from 'pino';
|
||||
import { DatabaseError as ProcessingDatabaseError } from '../processingErrors';
|
||||
|
||||
/**
|
||||
* Base class for custom database errors to ensure they have a status property.
|
||||
* Base class for custom repository-level errors to ensure they have a status property.
|
||||
*/
|
||||
export class DatabaseError extends Error {
|
||||
export class RepositoryError extends Error {
|
||||
public status: number;
|
||||
|
||||
constructor(message: string, status: number) {
|
||||
@@ -20,7 +21,7 @@ export class DatabaseError extends Error {
|
||||
* Thrown when a unique constraint is violated (e.g., trying to register an existing email).
|
||||
* Corresponds to PostgreSQL error code '23505'.
|
||||
*/
|
||||
export class UniqueConstraintError extends DatabaseError {
|
||||
export class UniqueConstraintError extends RepositoryError {
|
||||
constructor(message = 'The record already exists.') {
|
||||
super(message, 409); // 409 Conflict
|
||||
}
|
||||
@@ -30,7 +31,7 @@ export class UniqueConstraintError extends DatabaseError {
|
||||
* Thrown when a foreign key constraint is violated (e.g., trying to reference a non-existent record).
|
||||
* Corresponds to PostgreSQL error code '23503'.
|
||||
*/
|
||||
export class ForeignKeyConstraintError extends DatabaseError {
|
||||
export class ForeignKeyConstraintError extends RepositoryError {
|
||||
constructor(message = 'The referenced record does not exist.') {
|
||||
super(message, 400); // 400 Bad Request
|
||||
}
|
||||
@@ -40,7 +41,7 @@ export class ForeignKeyConstraintError extends DatabaseError {
|
||||
* Thrown when a 'not null' constraint is violated.
|
||||
* Corresponds to PostgreSQL error code '23502'.
|
||||
*/
|
||||
export class NotNullConstraintError extends DatabaseError {
|
||||
export class NotNullConstraintError extends RepositoryError {
|
||||
constructor(message = 'A required field was left null.') {
|
||||
super(message, 400); // 400 Bad Request
|
||||
}
|
||||
@@ -50,7 +51,7 @@ export class NotNullConstraintError extends DatabaseError {
|
||||
* Thrown when a 'check' constraint is violated.
|
||||
* Corresponds to PostgreSQL error code '23514'.
|
||||
*/
|
||||
export class CheckConstraintError extends DatabaseError {
|
||||
export class CheckConstraintError extends RepositoryError {
|
||||
constructor(message = 'A check constraint was violated.') {
|
||||
super(message, 400); // 400 Bad Request
|
||||
}
|
||||
@@ -60,7 +61,7 @@ export class CheckConstraintError extends DatabaseError {
|
||||
* Thrown when a value has an invalid text representation for its data type (e.g., 'abc' for an integer).
|
||||
* Corresponds to PostgreSQL error code '22P02'.
|
||||
*/
|
||||
export class InvalidTextRepresentationError extends DatabaseError {
|
||||
export class InvalidTextRepresentationError extends RepositoryError {
|
||||
constructor(message = 'A value has an invalid format for its data type.') {
|
||||
super(message, 400); // 400 Bad Request
|
||||
}
|
||||
@@ -70,7 +71,7 @@ export class InvalidTextRepresentationError extends DatabaseError {
|
||||
* Thrown when a numeric value is out of range for its data type (e.g., too large for an integer).
|
||||
* Corresponds to PostgreSQL error code '22003'.
|
||||
*/
|
||||
export class NumericValueOutOfRangeError extends DatabaseError {
|
||||
export class NumericValueOutOfRangeError extends RepositoryError {
|
||||
constructor(message = 'A numeric value is out of the allowed range.') {
|
||||
super(message, 400); // 400 Bad Request
|
||||
}
|
||||
@@ -79,7 +80,7 @@ export class NumericValueOutOfRangeError extends DatabaseError {
|
||||
/**
|
||||
* Thrown when a specific record is not found in the database.
|
||||
*/
|
||||
export class NotFoundError extends DatabaseError {
|
||||
export class NotFoundError extends RepositoryError {
|
||||
constructor(message = 'The requested resource was not found.') {
|
||||
super(message, 404); // 404 Not Found
|
||||
}
|
||||
@@ -97,7 +98,7 @@ export interface ValidationIssue {
|
||||
/**
|
||||
* Thrown when request validation fails (e.g., missing body fields or invalid params).
|
||||
*/
|
||||
export class ValidationError extends DatabaseError {
|
||||
export class ValidationError extends RepositoryError {
|
||||
public validationErrors: ValidationIssue[];
|
||||
|
||||
constructor(errors: ValidationIssue[], message = 'The request data is invalid.') {
|
||||
@@ -126,6 +127,15 @@ export interface HandleDbErrorOptions {
|
||||
defaultMessage?: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* A type guard to check if an error object is a PostgreSQL error with a code.
|
||||
*/
|
||||
function isPostgresError(
|
||||
error: unknown,
|
||||
): error is { code: string; constraint?: string; detail?: string } {
|
||||
return typeof error === 'object' && error !== null && 'code' in error;
|
||||
}
|
||||
|
||||
/**
|
||||
* Centralized error handler for database repositories.
|
||||
* Logs the error and throws appropriate custom errors based on PostgreSQL error codes.
|
||||
@@ -138,26 +148,42 @@ export function handleDbError(
|
||||
options: HandleDbErrorOptions = {},
|
||||
): never {
|
||||
// If it's already a known domain error (like NotFoundError thrown manually), rethrow it.
|
||||
if (error instanceof DatabaseError) {
|
||||
if (error instanceof RepositoryError) {
|
||||
throw error;
|
||||
}
|
||||
|
||||
// Log the raw error
|
||||
logger.error({ err: error, ...logContext }, logMessage);
|
||||
if (isPostgresError(error)) {
|
||||
const { code, constraint, detail } = error;
|
||||
const enhancedLogContext = { err: error, code, constraint, detail, ...logContext };
|
||||
|
||||
if (error instanceof Error && 'code' in error) {
|
||||
const code = (error as any).code;
|
||||
// Log the detailed error first
|
||||
logger.error(enhancedLogContext, logMessage);
|
||||
|
||||
if (code === '23505') throw new UniqueConstraintError(options.uniqueMessage);
|
||||
if (code === '23503') throw new ForeignKeyConstraintError(options.fkMessage);
|
||||
if (code === '23502') throw new NotNullConstraintError(options.notNullMessage);
|
||||
if (code === '23514') throw new CheckConstraintError(options.checkMessage);
|
||||
if (code === '22P02') throw new InvalidTextRepresentationError(options.invalidTextMessage);
|
||||
if (code === '22003') throw new NumericValueOutOfRangeError(options.numericOutOfRangeMessage);
|
||||
// Now, throw the appropriate custom error
|
||||
switch (code) {
|
||||
case '23505': // unique_violation
|
||||
throw new UniqueConstraintError(options.uniqueMessage);
|
||||
case '23503': // foreign_key_violation
|
||||
throw new ForeignKeyConstraintError(options.fkMessage);
|
||||
case '23502': // not_null_violation
|
||||
throw new NotNullConstraintError(options.notNullMessage);
|
||||
case '23514': // check_violation
|
||||
throw new CheckConstraintError(options.checkMessage);
|
||||
case '22P02': // invalid_text_representation
|
||||
throw new InvalidTextRepresentationError(options.invalidTextMessage);
|
||||
case '22003': // numeric_value_out_of_range
|
||||
throw new NumericValueOutOfRangeError(options.numericOutOfRangeMessage);
|
||||
default:
|
||||
// If it's a PG error but not one we handle specifically, fall through to the generic error.
|
||||
break;
|
||||
}
|
||||
} else {
|
||||
// Log the error if it wasn't a recognized Postgres error
|
||||
logger.error({ err: error, ...logContext }, logMessage);
|
||||
}
|
||||
|
||||
// Fallback generic error
|
||||
throw new Error(
|
||||
options.defaultMessage || `Failed to perform operation on ${options.entityName || 'database'}.`,
|
||||
);
|
||||
// Use the consistent DatabaseError from the processing errors module for the fallback.
|
||||
const errorMessage = options.defaultMessage || `Failed to perform operation on ${options.entityName || 'database'}.`;
|
||||
throw new ProcessingDatabaseError(errorMessage);
|
||||
}
|
||||
|
||||
@@ -12,7 +12,13 @@ import {
|
||||
vi.unmock('./flyer.db');
|
||||
|
||||
import { FlyerRepository, createFlyerAndItems } from './flyer.db';
|
||||
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import {
|
||||
UniqueConstraintError,
|
||||
ForeignKeyConstraintError,
|
||||
NotFoundError,
|
||||
CheckConstraintError,
|
||||
} from './errors.db';
|
||||
import { DatabaseError } from '../processingErrors';
|
||||
import type {
|
||||
FlyerInsert,
|
||||
FlyerItemInsert,
|
||||
@@ -51,67 +57,72 @@ describe('Flyer DB Service', () => {
|
||||
|
||||
describe('findOrCreateStore', () => {
|
||||
it('should find an existing store and return its ID', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ store_id: 1 }] });
|
||||
// 1. INSERT...ON CONFLICT does nothing. 2. SELECT finds the store.
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [], rowCount: 0 })
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] });
|
||||
|
||||
const result = await flyerRepo.findOrCreateStore('Existing Store', mockLogger);
|
||||
expect(result).toBe(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(2);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('SELECT store_id FROM public.stores WHERE name = $1'),
|
||||
'INSERT INTO public.stores (name) VALUES ($1) ON CONFLICT (name) DO NOTHING',
|
||||
['Existing Store'],
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
['Existing Store'],
|
||||
);
|
||||
});
|
||||
|
||||
it('should create a new store if it does not exist', async () => {
|
||||
it('should create a new store if it does not exist and return its ID', async () => {
|
||||
// 1. INSERT...ON CONFLICT creates the store. 2. SELECT finds it.
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // First SELECT finds nothing
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] })
|
||||
.mockResolvedValueOnce({ rows: [], rowCount: 1 }) // INSERT affects 1 row
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] }); // SELECT finds the new store
|
||||
|
||||
const result = await flyerRepo.findOrCreateStore('New Store', mockLogger);
|
||||
expect(result).toBe(2);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(2);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id'),
|
||||
'INSERT INTO public.stores (name) VALUES ($1) ON CONFLICT (name) DO NOTHING',
|
||||
['New Store'],
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
['New Store'],
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle race condition where store is created between SELECT and INSERT', async () => {
|
||||
const uniqueConstraintError = new Error('duplicate key value violates unique constraint');
|
||||
(uniqueConstraintError as Error & { code: string }).code = '23505';
|
||||
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // First SELECT finds nothing
|
||||
.mockRejectedValueOnce(uniqueConstraintError) // INSERT fails due to race condition
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 3 }] }); // Second SELECT finds the store
|
||||
|
||||
const result = await flyerRepo.findOrCreateStore('Racy Store', mockLogger);
|
||||
expect(result).toBe(3);
|
||||
//expect(mockDb.query).toHaveBeenCalledTimes(3);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
// The new implementation uses handleDbError, which will throw a generic Error with the default message.
|
||||
await expect(flyerRepo.findOrCreateStore('Any Store', mockLogger)).rejects.toThrow(
|
||||
'Failed to find or create store in database.',
|
||||
);
|
||||
// handleDbError also logs the error.
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, storeName: 'Any Store' },
|
||||
'Database error in findOrCreateStore',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if race condition recovery fails', async () => {
|
||||
const uniqueConstraintError = new Error('duplicate key value violates unique constraint');
|
||||
(uniqueConstraintError as Error & { code: string }).code = '23505';
|
||||
|
||||
it('should throw an error if store is not found after upsert (edge case)', async () => {
|
||||
// This simulates a very unlikely scenario where the store is deleted between the
|
||||
// INSERT...ON CONFLICT and the subsequent SELECT.
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // First SELECT
|
||||
.mockRejectedValueOnce(uniqueConstraintError) // INSERT fails
|
||||
.mockRejectedValueOnce(new Error('Second select fails')); // Recovery SELECT fails
|
||||
.mockResolvedValueOnce({ rows: [], rowCount: 1 }) // INSERT succeeds
|
||||
.mockResolvedValueOnce({ rows: [] }); // SELECT finds nothing
|
||||
|
||||
await expect(flyerRepo.findOrCreateStore('Racy Store', mockLogger)).rejects.toThrow(
|
||||
await expect(flyerRepo.findOrCreateStore('Weird Store', mockLogger)).rejects.toThrow(
|
||||
'Failed to find or create store in database.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(Error), storeName: 'Racy Store' },
|
||||
{
|
||||
err: new Error('Failed to find store immediately after upsert operation.'),
|
||||
storeName: 'Weird Store',
|
||||
},
|
||||
'Database error in findOrCreateStore',
|
||||
);
|
||||
});
|
||||
@@ -121,8 +132,8 @@ describe('Flyer DB Service', () => {
|
||||
it('should execute an INSERT query and return the new flyer', async () => {
|
||||
const flyerData: FlyerDbInsert = {
|
||||
file_name: 'test.jpg',
|
||||
image_url: '/images/test.jpg',
|
||||
icon_url: '/images/icons/test.jpg',
|
||||
image_url: 'http://localhost:3001/images/test.jpg',
|
||||
icon_url: 'http://localhost:3001/images/icons/test.jpg',
|
||||
checksum: 'checksum123',
|
||||
store_id: 1,
|
||||
valid_from: '2024-01-01',
|
||||
@@ -130,7 +141,8 @@ describe('Flyer DB Service', () => {
|
||||
store_address: '123 Test St',
|
||||
status: 'processed',
|
||||
item_count: 10,
|
||||
uploaded_by: 'user-1',
|
||||
// Use a valid UUID format for the foreign key.
|
||||
uploaded_by: 'a0eebc99-9c0b-4ef8-bb6d-6bb9bd380a11',
|
||||
};
|
||||
const mockFlyer = createMockFlyer({ ...flyerData, flyer_id: 1 });
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockFlyer] });
|
||||
@@ -143,8 +155,8 @@ describe('Flyer DB Service', () => {
|
||||
expect.stringContaining('INSERT INTO flyers'),
|
||||
[
|
||||
'test.jpg',
|
||||
'/images/test.jpg',
|
||||
'/images/icons/test.jpg',
|
||||
'http://localhost:3001/images/test.jpg',
|
||||
'http://localhost:3001/images/icons/test.jpg',
|
||||
'checksum123',
|
||||
1,
|
||||
'2024-01-01',
|
||||
@@ -152,7 +164,7 @@ describe('Flyer DB Service', () => {
|
||||
'123 Test St',
|
||||
'processed',
|
||||
10,
|
||||
'user-1',
|
||||
'a0eebc99-9c0b-4ef8-bb6d-6bb9bd380a11',
|
||||
],
|
||||
);
|
||||
});
|
||||
@@ -172,7 +184,13 @@ describe('Flyer DB Service', () => {
|
||||
'A flyer with this checksum already exists.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, flyerData },
|
||||
{
|
||||
err: dbError,
|
||||
flyerData,
|
||||
code: '23505',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in insertFlyer',
|
||||
);
|
||||
});
|
||||
@@ -188,6 +206,48 @@ describe('Flyer DB Service', () => {
|
||||
'Database error in insertFlyer',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw CheckConstraintError for invalid checksum format', async () => {
|
||||
const flyerData: FlyerDbInsert = { checksum: 'short' } as FlyerDbInsert;
|
||||
const dbError = new Error('violates check constraint "flyers_checksum_check"');
|
||||
(dbError as Error & { code: string }).code = '23514'; // Check constraint violation
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
CheckConstraintError,
|
||||
);
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
'The provided checksum is invalid or does not meet format requirements (e.g., must be a 64-character SHA-256 hash).',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw CheckConstraintError for invalid status', async () => {
|
||||
const flyerData: FlyerDbInsert = { status: 'invalid_status' } as any;
|
||||
const dbError = new Error('violates check constraint "flyers_status_check"');
|
||||
(dbError as Error & { code: string }).code = '23514';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
CheckConstraintError,
|
||||
);
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
'Invalid status provided for flyer.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw CheckConstraintError for invalid URL format', async () => {
|
||||
const flyerData: FlyerDbInsert = { image_url: 'not-a-url' } as FlyerDbInsert;
|
||||
const dbError = new Error('violates check constraint "url_check"');
|
||||
(dbError as Error & { code: string }).code = '23514';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
CheckConstraintError,
|
||||
);
|
||||
await expect(flyerRepo.insertFlyer(flyerData, mockLogger)).rejects.toThrow(
|
||||
'Invalid URL format provided for image or icon.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('insertFlyerItems', () => {
|
||||
@@ -277,7 +337,13 @@ describe('Flyer DB Service', () => {
|
||||
'The specified flyer, category, master item, or product does not exist.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, flyerId: 999 },
|
||||
{
|
||||
err: dbError,
|
||||
flyerId: 999,
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in insertFlyerItems',
|
||||
);
|
||||
});
|
||||
@@ -297,8 +363,8 @@ describe('Flyer DB Service', () => {
|
||||
});
|
||||
|
||||
describe('createFlyerAndItems', () => {
|
||||
it('should use withTransaction to create a flyer and items', async () => {
|
||||
const flyerData: FlyerInsert = {
|
||||
it('should execute find/create store, insert flyer, and insert items using the provided client', async () => {
|
||||
const flyerData: FlyerInsert = { // This was a duplicate, fixed.
|
||||
file_name: 'transact.jpg',
|
||||
store_name: 'Transaction Store',
|
||||
} as FlyerInsert;
|
||||
@@ -321,81 +387,73 @@ describe('Flyer DB Service', () => {
|
||||
}),
|
||||
];
|
||||
|
||||
// Mock the withTransaction to execute the callback with a mock client
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
// Mock the sequence of calls within the transaction
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] }) // findOrCreateStore
|
||||
.mockResolvedValueOnce({ rows: [mockFlyer] }) // insertFlyer
|
||||
.mockResolvedValueOnce({ rows: mockItems }); // insertFlyerItems
|
||||
return callback(mockClient as unknown as PoolClient);
|
||||
});
|
||||
// Mock the sequence of 4 calls on the client
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
// 1. findOrCreateStore: INSERT ... ON CONFLICT
|
||||
.mockResolvedValueOnce({ rows: [], rowCount: 0 })
|
||||
// 2. findOrCreateStore: SELECT store_id
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] })
|
||||
// 3. insertFlyer
|
||||
.mockResolvedValueOnce({ rows: [mockFlyer] })
|
||||
// 4. insertFlyerItems
|
||||
.mockResolvedValueOnce({ rows: mockItems });
|
||||
|
||||
const result = await createFlyerAndItems(flyerData, itemsData, mockLogger);
|
||||
const result = await createFlyerAndItems(
|
||||
flyerData,
|
||||
itemsData,
|
||||
mockLogger,
|
||||
mockClient as unknown as PoolClient,
|
||||
);
|
||||
|
||||
expect(result).toEqual({
|
||||
flyer: mockFlyer,
|
||||
items: mockItems,
|
||||
});
|
||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||
|
||||
// Verify the individual functions were called with the client
|
||||
const callback = (vi.mocked(withTransaction) as Mock).mock.calls[0][0];
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] })
|
||||
.mockResolvedValueOnce({ rows: [mockFlyer] })
|
||||
.mockResolvedValueOnce({ rows: mockItems });
|
||||
await callback(mockClient as unknown as PoolClient);
|
||||
// findOrCreateStore assertions
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('SELECT store_id FROM public.stores'),
|
||||
'INSERT INTO public.stores (name) VALUES ($1) ON CONFLICT (name) DO NOTHING',
|
||||
['Transaction Store'],
|
||||
);
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
['Transaction Store'],
|
||||
);
|
||||
// insertFlyer assertion
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO flyers'),
|
||||
expect.any(Array),
|
||||
);
|
||||
// insertFlyerItems assertion
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO flyer_items'),
|
||||
expect.any(Array),
|
||||
);
|
||||
});
|
||||
|
||||
it('should ROLLBACK the transaction if an error occurs', async () => {
|
||||
it('should propagate an error if any step fails', async () => {
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: 'fail.jpg',
|
||||
store_name: 'Fail Store',
|
||||
} as FlyerInsert;
|
||||
const itemsData: FlyerItemInsert[] = [{ item: 'Failing Item' } as FlyerItemInsert];
|
||||
const dbError = new Error('DB connection lost');
|
||||
const dbError = new Error('Underlying DB call failed');
|
||||
|
||||
// Mock withTransaction to simulate a failure during the callback
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] }) // findOrCreateStore
|
||||
.mockRejectedValueOnce(dbError); // insertFlyer fails
|
||||
// The withTransaction helper will catch this and roll back.
|
||||
// Since insertFlyer wraps the DB error, we expect the wrapped error message here.
|
||||
await expect(callback(mockClient as unknown as PoolClient)).rejects.toThrow(
|
||||
'Failed to insert flyer into database.',
|
||||
);
|
||||
// re-throw because withTransaction re-throws (simulating the wrapped error propagating up)
|
||||
throw new Error('Failed to insert flyer into database.');
|
||||
});
|
||||
// Mock the client to fail on the insertFlyer step
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [], rowCount: 0 })
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 1 }] })
|
||||
.mockRejectedValueOnce(dbError); // insertFlyer fails
|
||||
|
||||
// The transactional function re-throws the original error from the failed step.
|
||||
// Since insertFlyer wraps errors, we expect the wrapped error message.
|
||||
await expect(createFlyerAndItems(flyerData, itemsData, mockLogger)).rejects.toThrow(
|
||||
'Failed to insert flyer into database.',
|
||||
);
|
||||
// The error object passed to the logger will be the wrapped Error object, not the original dbError
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(Error) },
|
||||
'Database transaction error in createFlyerAndItems',
|
||||
);
|
||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||
// The calling service's withTransaction would catch this.
|
||||
// Here, we just expect it to be thrown.
|
||||
await expect(
|
||||
createFlyerAndItems(flyerData, itemsData, mockLogger, mockClient as unknown as PoolClient),
|
||||
// The error is wrapped by handleDbError, so we check for the wrapped error.
|
||||
).rejects.toThrow(new DatabaseError('Failed to insert flyer into database.'));
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
@@ -28,46 +28,32 @@ export class FlyerRepository {
|
||||
* @returns A promise that resolves to the store's ID.
|
||||
*/
|
||||
async findOrCreateStore(storeName: string, logger: Logger): Promise<number> {
|
||||
// Note: This method should be called within a transaction if the caller
|
||||
// needs to ensure atomicity with other operations.
|
||||
try {
|
||||
// First, try to find the store.
|
||||
let result = await this.db.query<{ store_id: number }>(
|
||||
// Atomically insert the store if it doesn't exist. This is safe from race conditions.
|
||||
await this.db.query(
|
||||
'INSERT INTO public.stores (name) VALUES ($1) ON CONFLICT (name) DO NOTHING',
|
||||
[storeName],
|
||||
);
|
||||
|
||||
// Now, the store is guaranteed to exist, so we can safely select its ID.
|
||||
const result = await this.db.query<{ store_id: number }>(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
[storeName],
|
||||
);
|
||||
|
||||
if (result.rows.length > 0) {
|
||||
return result.rows[0].store_id;
|
||||
} else {
|
||||
// If not found, create it.
|
||||
result = await this.db.query<{ store_id: number }>(
|
||||
'INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id',
|
||||
[storeName],
|
||||
);
|
||||
return result.rows[0].store_id;
|
||||
// This case should be virtually impossible if the INSERT...ON CONFLICT logic is correct,
|
||||
// as it would mean the store was deleted between the two queries. We throw an error to be safe.
|
||||
if (result.rows.length === 0) {
|
||||
throw new Error('Failed to find store immediately after upsert operation.');
|
||||
}
|
||||
|
||||
return result.rows[0].store_id;
|
||||
} catch (error) {
|
||||
// Check for a unique constraint violation on name, which could happen in a race condition
|
||||
// if two processes try to create the same store at the same time.
|
||||
if (error instanceof Error && 'code' in error && error.code === '23505') {
|
||||
try {
|
||||
logger.warn(
|
||||
{ storeName },
|
||||
`Race condition avoided: Store was created by another process. Refetching.`,
|
||||
);
|
||||
const result = await this.db.query<{ store_id: number }>(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
[storeName],
|
||||
);
|
||||
if (result.rows.length > 0) return result.rows[0].store_id;
|
||||
} catch (recoveryError) {
|
||||
// If recovery fails, log a warning and fall through to the generic error handler
|
||||
logger.warn({ err: recoveryError, storeName }, 'Race condition recovery failed');
|
||||
}
|
||||
}
|
||||
logger.error({ err: error, storeName }, 'Database error in findOrCreateStore');
|
||||
throw new Error('Failed to find or create store in database.');
|
||||
// Use the centralized error handler for any unexpected database errors.
|
||||
handleDbError(error, logger, 'Database error in findOrCreateStore', { storeName }, {
|
||||
// Any error caught here is unexpected, so we use a generic message.
|
||||
defaultMessage: 'Failed to find or create store in database.',
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
@@ -100,20 +86,30 @@ export class FlyerRepository {
|
||||
flyerData.uploaded_by ?? null, // $11
|
||||
];
|
||||
|
||||
logger.debug(
|
||||
{ query, values },
|
||||
'[DB insertFlyer] Executing insert with the following values.',
|
||||
);
|
||||
|
||||
const result = await this.db.query<Flyer>(query, values);
|
||||
return result.rows[0];
|
||||
} catch (error) {
|
||||
const isChecksumError =
|
||||
error instanceof Error && error.message.includes('flyers_checksum_check');
|
||||
const errorMessage = error instanceof Error ? error.message : '';
|
||||
let checkMsg = 'A database check constraint failed.';
|
||||
|
||||
if (errorMessage.includes('flyers_checksum_check')) {
|
||||
checkMsg =
|
||||
'The provided checksum is invalid or does not meet format requirements (e.g., must be a 64-character SHA-256 hash).';
|
||||
} else if (errorMessage.includes('flyers_status_check')) {
|
||||
checkMsg = 'Invalid status provided for flyer.';
|
||||
} else if (errorMessage.includes('url_check')) {
|
||||
checkMsg = 'Invalid URL format provided for image or icon.';
|
||||
}
|
||||
|
||||
handleDbError(error, logger, 'Database error in insertFlyer', { flyerData }, {
|
||||
uniqueMessage: 'A flyer with this checksum already exists.',
|
||||
fkMessage: 'The specified user or store for this flyer does not exist.',
|
||||
// Provide a more specific message for the checksum constraint violation,
|
||||
// which is a common issue during seeding or testing with placeholder data.
|
||||
checkMessage: isChecksumError
|
||||
? 'The provided checksum is invalid or does not meet format requirements (e.g., must be a 64-character SHA-256 hash).'
|
||||
: 'Invalid status provided for flyer.',
|
||||
checkMessage: checkMsg,
|
||||
defaultMessage: 'Failed to insert flyer into database.',
|
||||
});
|
||||
}
|
||||
@@ -163,6 +159,11 @@ export class FlyerRepository {
|
||||
RETURNING *;
|
||||
`;
|
||||
|
||||
logger.debug(
|
||||
{ query, values },
|
||||
'[DB insertFlyerItems] Executing bulk insert with the following values.',
|
||||
);
|
||||
|
||||
const result = await this.db.query<FlyerItem>(query, values);
|
||||
return result.rows;
|
||||
} catch (error) {
|
||||
@@ -378,27 +379,23 @@ export async function createFlyerAndItems(
|
||||
flyerData: FlyerInsert,
|
||||
itemsForDb: FlyerItemInsert[],
|
||||
logger: Logger,
|
||||
client: PoolClient,
|
||||
) {
|
||||
try {
|
||||
return await withTransaction(async (client) => {
|
||||
const flyerRepo = new FlyerRepository(client);
|
||||
// The calling service is now responsible for managing the transaction.
|
||||
// This function assumes it is being run within a transaction via the provided client.
|
||||
const flyerRepo = new FlyerRepository(client);
|
||||
|
||||
// 1. Find or create the store to get the store_id
|
||||
const storeId = await flyerRepo.findOrCreateStore(flyerData.store_name, logger);
|
||||
// 1. Find or create the store to get the store_id
|
||||
const storeId = await flyerRepo.findOrCreateStore(flyerData.store_name, logger);
|
||||
|
||||
// 2. Prepare the data for the flyer table, replacing store_name with store_id
|
||||
const flyerDbData: FlyerDbInsert = { ...flyerData, store_id: storeId };
|
||||
// 2. Prepare the data for the flyer table, replacing store_name with store_id
|
||||
const flyerDbData: FlyerDbInsert = { ...flyerData, store_id: storeId };
|
||||
|
||||
// 3. Insert the flyer record
|
||||
const newFlyer = await flyerRepo.insertFlyer(flyerDbData, logger);
|
||||
// 3. Insert the flyer record
|
||||
const newFlyer = await flyerRepo.insertFlyer(flyerDbData, logger);
|
||||
|
||||
// 4. Insert the associated flyer items
|
||||
const newItems = await flyerRepo.insertFlyerItems(newFlyer.flyer_id, itemsForDb, logger);
|
||||
// 4. Insert the associated flyer items
|
||||
const newItems = await flyerRepo.insertFlyerItems(newFlyer.flyer_id, itemsForDb, logger);
|
||||
|
||||
return { flyer: newFlyer, items: newItems };
|
||||
});
|
||||
} catch (error) {
|
||||
logger.error({ err: error }, 'Database transaction error in createFlyerAndItems');
|
||||
throw error; // Re-throw the error to be handled by the calling service.
|
||||
}
|
||||
return { flyer: newFlyer, items: newItems };
|
||||
}
|
||||
|
||||
@@ -130,7 +130,14 @@ describe('Gamification DB Service', () => {
|
||||
),
|
||||
).rejects.toThrow('The specified user or achievement does not exist.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: 'non-existent-user', achievementName: 'Non-existent Achievement' },
|
||||
{
|
||||
err: dbError,
|
||||
userId: 'non-existent-user',
|
||||
achievementName: 'Non-existent Achievement',
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in awardAchievement',
|
||||
);
|
||||
});
|
||||
|
||||
64
src/services/db/index.db.test.ts
Normal file
64
src/services/db/index.db.test.ts
Normal file
@@ -0,0 +1,64 @@
|
||||
// src/services/db/index.db.test.ts
|
||||
import { describe, it, expect, vi } from 'vitest';
|
||||
|
||||
// Mock all the repository classes to be simple classes/functions
|
||||
// This prevents their constructors from running real database connection logic.
|
||||
vi.mock('./user.db', () => ({ UserRepository: class UserRepository {} }));
|
||||
vi.mock('./flyer.db', () => ({ FlyerRepository: class FlyerRepository {} }));
|
||||
vi.mock('./address.db', () => ({ AddressRepository: class AddressRepository {} }));
|
||||
vi.mock('./shopping.db', () => ({ ShoppingRepository: class ShoppingRepository {} }));
|
||||
vi.mock('./personalization.db', () => ({
|
||||
PersonalizationRepository: class PersonalizationRepository {},
|
||||
}));
|
||||
vi.mock('./recipe.db', () => ({ RecipeRepository: class RecipeRepository {} }));
|
||||
vi.mock('./notification.db', () => ({
|
||||
NotificationRepository: class NotificationRepository {},
|
||||
}));
|
||||
vi.mock('./budget.db', () => ({ BudgetRepository: class BudgetRepository {} }));
|
||||
vi.mock('./gamification.db', () => ({
|
||||
GamificationRepository: class GamificationRepository {},
|
||||
}));
|
||||
vi.mock('./admin.db', () => ({ AdminRepository: class AdminRepository {} }));
|
||||
|
||||
// These modules export an already-instantiated object, so we mock the object.
|
||||
vi.mock('./reaction.db', () => ({ reactionRepo: {} }));
|
||||
vi.mock('./conversion.db', () => ({ conversionRepo: {} }));
|
||||
|
||||
// Mock the re-exported function.
|
||||
vi.mock('./connection.db', () => ({ withTransaction: vi.fn() }));
|
||||
|
||||
// We must un-mock the file we are testing so we get the actual implementation.
|
||||
vi.unmock('./index.db');
|
||||
|
||||
// Import the module to be tested AFTER setting up the mocks.
|
||||
import * as db from './index.db';
|
||||
|
||||
// Import the mocked classes to check `instanceof`.
|
||||
import { UserRepository } from './user.db';
|
||||
import { FlyerRepository } from './flyer.db';
|
||||
import { AddressRepository } from './address.db';
|
||||
import { ShoppingRepository } from './shopping.db';
|
||||
import { PersonalizationRepository } from './personalization.db';
|
||||
import { RecipeRepository } from './recipe.db';
|
||||
import { NotificationRepository } from './notification.db';
|
||||
import { BudgetRepository } from './budget.db';
|
||||
import { GamificationRepository } from './gamification.db';
|
||||
import { AdminRepository } from './admin.db';
|
||||
|
||||
describe('DB Index', () => {
|
||||
it('should instantiate and export all repositories and functions', () => {
|
||||
expect(db.userRepo).toBeInstanceOf(UserRepository);
|
||||
expect(db.flyerRepo).toBeInstanceOf(FlyerRepository);
|
||||
expect(db.addressRepo).toBeInstanceOf(AddressRepository);
|
||||
expect(db.shoppingRepo).toBeInstanceOf(ShoppingRepository);
|
||||
expect(db.personalizationRepo).toBeInstanceOf(PersonalizationRepository);
|
||||
expect(db.recipeRepo).toBeInstanceOf(RecipeRepository);
|
||||
expect(db.notificationRepo).toBeInstanceOf(NotificationRepository);
|
||||
expect(db.budgetRepo).toBeInstanceOf(BudgetRepository);
|
||||
expect(db.gamificationRepo).toBeInstanceOf(GamificationRepository);
|
||||
expect(db.adminRepo).toBeInstanceOf(AdminRepository);
|
||||
expect(db.reactionRepo).toBeDefined();
|
||||
expect(db.conversionRepo).toBeDefined();
|
||||
expect(db.withTransaction).toBeDefined();
|
||||
});
|
||||
});
|
||||
@@ -150,7 +150,15 @@ describe('Notification DB Service', () => {
|
||||
notificationRepo.createNotification('non-existent-user', 'Test', mockLogger),
|
||||
).rejects.toThrow('The specified user does not exist.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: 'non-existent-user', content: 'Test', linkUrl: undefined },
|
||||
{
|
||||
err: dbError,
|
||||
userId: 'non-existent-user',
|
||||
content: 'Test',
|
||||
linkUrl: undefined,
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in createNotification',
|
||||
);
|
||||
});
|
||||
@@ -195,7 +203,13 @@ describe('Notification DB Service', () => {
|
||||
notificationRepo.createBulkNotifications(notificationsToCreate, mockLogger),
|
||||
).rejects.toThrow(ForeignKeyConstraintError);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, notifications: notificationsToCreate },
|
||||
{
|
||||
err: dbError,
|
||||
notifications: notificationsToCreate,
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in createBulkNotifications',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -173,7 +173,14 @@ describe('Recipe DB Service', () => {
|
||||
'The specified user or recipe does not exist.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: 'user-123', recipeId: 999 },
|
||||
{
|
||||
err: dbError,
|
||||
userId: 'user-123',
|
||||
recipeId: 999,
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in addFavoriteRecipe',
|
||||
);
|
||||
});
|
||||
@@ -414,7 +421,15 @@ describe('Recipe DB Service', () => {
|
||||
recipeRepo.addRecipeComment(999, 'user-123', 'Fail', mockLogger),
|
||||
).rejects.toThrow('The specified recipe, user, or parent comment does not exist.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, recipeId: 999, userId: 'user-123', parentCommentId: undefined },
|
||||
{
|
||||
err: dbError,
|
||||
recipeId: 999,
|
||||
userId: 'user-123',
|
||||
parentCommentId: undefined,
|
||||
code: '23503',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Database error in addRecipeComment',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -237,6 +237,13 @@ describe('Shopping DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if both masterItemId and customItemName are missing', async () => {
|
||||
// This test covers line 185 in shopping.db.ts
|
||||
await expect(shoppingRepo.addShoppingListItem(1, 'user-1', {}, mockLogger)).rejects.toThrow(
|
||||
'Either masterItemId or customItemName must be provided.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if no item data is provided', async () => {
|
||||
await expect(shoppingRepo.addShoppingListItem(1, 'user-1', {}, mockLogger)).rejects.toThrow(
|
||||
'Either masterItemId or customItemName must be provided.',
|
||||
);
|
||||
@@ -251,6 +258,15 @@ describe('Shopping DB Service', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if provided updates are not valid fields', async () => {
|
||||
// This test covers line 362 in shopping.db.ts
|
||||
const updates = { invalid_field: 'some_value' };
|
||||
await expect(
|
||||
shoppingRepo.updateShoppingListItem(1, 'user-1', updates as any, mockLogger),
|
||||
).rejects.toThrow('No valid fields to update.');
|
||||
expect(mockPoolInstance.query).not.toHaveBeenCalled(); // No DB query should be made
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Connection Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
@@ -580,7 +596,7 @@ describe('Shopping DB Service', () => {
|
||||
const mockReceipt = {
|
||||
receipt_id: 1,
|
||||
user_id: 'user-1',
|
||||
receipt_image_url: 'url',
|
||||
receipt_image_url: 'http://example.com/receipt.jpg',
|
||||
status: 'pending',
|
||||
};
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockReceipt] });
|
||||
|
||||
@@ -28,6 +28,8 @@ import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { createMockUserProfile, createMockUser } from '../../tests/utils/mockFactories';
|
||||
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import type { Profile, ActivityLogItem, SearchQuery, UserProfile, User } from '../../types';
|
||||
import { ShoppingRepository } from './shopping.db';
|
||||
import { PersonalizationRepository } from './personalization.db';
|
||||
|
||||
// Mock other db services that are used by functions in user.db.ts
|
||||
// Update mocks to put methods on prototype so spyOn works in exportUserData tests
|
||||
@@ -115,14 +117,14 @@ describe('User DB Service', () => {
|
||||
});
|
||||
|
||||
describe('createUser', () => {
|
||||
it('should execute a transaction to create a user and profile', async () => {
|
||||
it('should create a user and profile successfully', async () => {
|
||||
const mockUser = {
|
||||
user_id: 'new-user-id',
|
||||
email: 'new@example.com',
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
// This is the flat structure returned by the DB query inside createUser
|
||||
|
||||
const mockDbProfile = {
|
||||
user_id: 'new-user-id',
|
||||
email: 'new@example.com',
|
||||
@@ -136,7 +138,7 @@ describe('User DB Service', () => {
|
||||
user_created_at: new Date().toISOString(),
|
||||
user_updated_at: new Date().toISOString(),
|
||||
};
|
||||
// This is the nested structure the function is expected to return
|
||||
|
||||
const expectedProfile: UserProfile = {
|
||||
user: {
|
||||
user_id: mockDbProfile.user_id,
|
||||
@@ -153,14 +155,11 @@ describe('User DB Service', () => {
|
||||
updated_at: mockDbProfile.updated_at,
|
||||
};
|
||||
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback: any) => {
|
||||
const mockClient = { query: vi.fn(), release: vi.fn() };
|
||||
(mockClient.query as Mock)
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||
.mockResolvedValueOnce({ rows: [mockDbProfile] }); // SELECT profile
|
||||
return callback(mockClient as unknown as PoolClient);
|
||||
});
|
||||
// Mock the sequence of queries on the main pool instance
|
||||
(mockPoolInstance.query as Mock)
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||
.mockResolvedValueOnce({ rows: [mockDbProfile] }); // SELECT profile
|
||||
|
||||
const result = await userRepo.createUser(
|
||||
'new@example.com',
|
||||
@@ -169,52 +168,73 @@ describe('User DB Service', () => {
|
||||
mockLogger,
|
||||
);
|
||||
|
||||
// Use objectContaining because the real implementation might have other DB-generated fields.
|
||||
// We can't do a deep equality check on the user object because the mock factory will generate different timestamps.
|
||||
expect(result.user.user_id).toEqual(expectedProfile.user.user_id);
|
||||
expect(result.full_name).toEqual(expectedProfile.full_name);
|
||||
// eslint-disable-next-line @typescript-eslint/no-unused-vars
|
||||
expect(result).toEqual(expect.objectContaining(expectedProfile));
|
||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should rollback the transaction if creating the user fails', async () => {
|
||||
it('should create a user with a null password hash (e.g. OAuth)', async () => {
|
||||
const mockUser = {
|
||||
user_id: 'oauth-user-id',
|
||||
email: 'oauth@example.com',
|
||||
};
|
||||
const mockDbProfile = {
|
||||
user_id: 'oauth-user-id',
|
||||
email: 'oauth@example.com',
|
||||
role: 'user',
|
||||
full_name: 'OAuth User',
|
||||
user_created_at: new Date().toISOString(),
|
||||
user_updated_at: new Date().toISOString(),
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
|
||||
(mockPoolInstance.query as Mock)
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||
.mockResolvedValueOnce({ rows: [mockDbProfile] }); // SELECT profile
|
||||
|
||||
const result = await userRepo.createUser(
|
||||
'oauth@example.com',
|
||||
null, // Pass null for passwordHash
|
||||
{ full_name: 'OAuth User' },
|
||||
mockLogger,
|
||||
);
|
||||
|
||||
expect(result.user.email).toBe('oauth@example.com');
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'INSERT INTO public.users (email, password_hash) VALUES ($1, $2) RETURNING user_id, email',
|
||||
['oauth@example.com', null],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if creating the user fails', async () => {
|
||||
const dbError = new Error('User insert failed');
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query.mockRejectedValueOnce(dbError); // set_config or INSERT fails
|
||||
await expect(callback(mockClient as unknown as PoolClient)).rejects.toThrow(dbError);
|
||||
throw dbError;
|
||||
});
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(
|
||||
userRepo.createUser('fail@example.com', 'badpass', {}, mockLogger),
|
||||
).rejects.toThrow('Failed to create user in database.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, email: 'fail@example.com' },
|
||||
'Error during createUser transaction',
|
||||
'Error during createUser',
|
||||
);
|
||||
});
|
||||
|
||||
it('should rollback the transaction if fetching the final profile fails', async () => {
|
||||
it('should throw an error if fetching the final profile fails', async () => {
|
||||
const mockUser = { user_id: 'new-user-id', email: 'new@example.com' };
|
||||
const dbError = new Error('Profile fetch failed');
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||
.mockRejectedValueOnce(dbError); // SELECT profile fails
|
||||
await expect(callback(mockClient as unknown as PoolClient)).rejects.toThrow(dbError);
|
||||
throw dbError;
|
||||
});
|
||||
(mockPoolInstance.query as Mock)
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||
.mockRejectedValueOnce(dbError); // SELECT profile fails
|
||||
|
||||
await expect(userRepo.createUser('fail@example.com', 'pass', {}, mockLogger)).rejects.toThrow(
|
||||
'Failed to create user in database.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, email: 'fail@example.com' },
|
||||
'Error during createUser transaction',
|
||||
'Error during createUser',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -222,46 +242,42 @@ describe('User DB Service', () => {
|
||||
const dbError = new Error('duplicate key value violates unique constraint');
|
||||
(dbError as Error & { code: string }).code = '23505';
|
||||
|
||||
vi.mocked(withTransaction).mockRejectedValue(dbError);
|
||||
(mockPoolInstance.query as Mock).mockRejectedValue(dbError);
|
||||
|
||||
try {
|
||||
await userRepo.createUser('exists@example.com', 'pass', {}, mockLogger);
|
||||
expect.fail('Expected createUser to throw UniqueConstraintError');
|
||||
} catch (error: unknown) {
|
||||
expect(error).toBeInstanceOf(UniqueConstraintError);
|
||||
// After confirming the error type, we can safely access its properties.
|
||||
// This satisfies TypeScript's type checker for the 'unknown' type.
|
||||
if (error instanceof Error) {
|
||||
expect(error.message).toBe('A user with this email address already exists.');
|
||||
}
|
||||
}
|
||||
await expect(
|
||||
userRepo.createUser('exists@example.com', 'pass', {}, mockLogger),
|
||||
).rejects.toThrow(UniqueConstraintError);
|
||||
|
||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(`Attempted to create a user with an existing email: exists@example.com`);
|
||||
await expect(
|
||||
userRepo.createUser('exists@example.com', 'pass', {}, mockLogger),
|
||||
).rejects.toThrow('A user with this email address already exists.');
|
||||
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{
|
||||
err: dbError,
|
||||
email: 'exists@example.com',
|
||||
code: '23505',
|
||||
constraint: undefined,
|
||||
detail: undefined,
|
||||
},
|
||||
'Error during createUser',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if profile is not found after user creation', async () => {
|
||||
const mockUser = { user_id: 'new-user-id', email: 'no-profile@example.com' };
|
||||
|
||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
mockClient.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user succeeds
|
||||
.mockResolvedValueOnce({ rows: [] }); // SELECT profile returns nothing
|
||||
// The callback will throw, which is caught and re-thrown by withTransaction
|
||||
await expect(callback(mockClient as unknown as PoolClient)).rejects.toThrow(
|
||||
'Failed to create or retrieve user profile after registration.',
|
||||
);
|
||||
throw new Error('Internal failure'); // Simulate re-throw from withTransaction
|
||||
});
|
||||
(mockPoolInstance.query as Mock)
|
||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user succeeds
|
||||
.mockResolvedValueOnce({ rows: [] }); // SELECT profile returns nothing
|
||||
|
||||
await expect(
|
||||
userRepo.createUser('no-profile@example.com', 'pass', {}, mockLogger),
|
||||
).rejects.toThrow('Failed to create user in database.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(Error), email: 'no-profile@example.com' },
|
||||
'Error during createUser transaction',
|
||||
'Error during createUser',
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -669,23 +685,15 @@ describe('User DB Service', () => {
|
||||
});
|
||||
|
||||
describe('deleteRefreshToken', () => {
|
||||
it('should execute an UPDATE query to set the refresh token to NULL', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
await userRepo.deleteRefreshToken('a-token', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'UPDATE public.users SET refresh_token = NULL WHERE refresh_token = $1',
|
||||
['a-token'],
|
||||
);
|
||||
});
|
||||
it('should log an error but not throw if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
it('should not throw an error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
// The function is designed to swallow errors, so we expect it to resolve.
|
||||
await expect(userRepo.deleteRefreshToken('a-token', mockLogger)).resolves.toBeUndefined();
|
||||
// We can still check that the query was attempted.
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalled();
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(Error) },
|
||||
{ err: dbError },
|
||||
'Database error in deleteRefreshToken',
|
||||
);
|
||||
});
|
||||
@@ -693,14 +701,14 @@ describe('User DB Service', () => {
|
||||
|
||||
describe('createPasswordResetToken', () => {
|
||||
it('should execute DELETE and INSERT queries', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
const mockClient = { query: vi.fn().mockResolvedValue({ rows: [] }) };
|
||||
const expires = new Date();
|
||||
await userRepo.createPasswordResetToken('123', 'token-hash', expires, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
await userRepo.createPasswordResetToken('123', 'token-hash', expires, mockLogger, mockClient as unknown as PoolClient);
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
'DELETE FROM public.password_reset_tokens WHERE user_id = $1',
|
||||
['123'],
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockClient.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO public.password_reset_tokens'),
|
||||
['123', 'token-hash', expires],
|
||||
);
|
||||
@@ -709,18 +717,18 @@ describe('User DB Service', () => {
|
||||
it('should throw ForeignKeyConstraintError if user does not exist', async () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
const mockClient = { query: vi.fn().mockRejectedValue(dbError) };
|
||||
await expect(
|
||||
userRepo.createPasswordResetToken('non-existent-user', 'hash', new Date(), mockLogger),
|
||||
userRepo.createPasswordResetToken('non-existent-user', 'hash', new Date(), mockLogger, mockClient as unknown as PoolClient),
|
||||
).rejects.toThrow(ForeignKeyConstraintError);
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
const mockClient = { query: vi.fn().mockRejectedValue(dbError) };
|
||||
const expires = new Date();
|
||||
await expect(
|
||||
userRepo.createPasswordResetToken('123', 'token-hash', expires, mockLogger),
|
||||
userRepo.createPasswordResetToken('123', 'token-hash', expires, mockLogger, mockClient as unknown as PoolClient),
|
||||
).rejects.toThrow('Failed to create password reset token.');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: '123' },
|
||||
@@ -761,10 +769,13 @@ describe('User DB Service', () => {
|
||||
});
|
||||
|
||||
it('should log an error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
await userRepo.deleteResetToken('token-hash', mockLogger);
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
await expect(userRepo.deleteResetToken('token-hash', mockLogger)).rejects.toThrow(
|
||||
'Failed to delete password reset token.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(Error), tokenHash: 'token-hash' },
|
||||
{ err: dbError, tokenHash: 'token-hash' },
|
||||
'Database error in deleteResetToken',
|
||||
);
|
||||
});
|
||||
@@ -797,18 +808,7 @@ describe('User DB Service', () => {
|
||||
});
|
||||
|
||||
describe('exportUserData', () => {
|
||||
// Import the mocked withTransaction helper
|
||||
let withTransaction: Mock;
|
||||
beforeEach(async () => {
|
||||
const connDb = await import('./connection.db');
|
||||
// Cast to Mock for type-safe access to mock properties
|
||||
withTransaction = connDb.withTransaction as Mock;
|
||||
});
|
||||
|
||||
it('should call profile, watched items, and shopping list functions', async () => {
|
||||
const { ShoppingRepository } = await import('./shopping.db');
|
||||
const { PersonalizationRepository } = await import('./personalization.db');
|
||||
|
||||
const findProfileSpy = vi.spyOn(UserRepository.prototype, 'findUserProfileById');
|
||||
findProfileSpy.mockResolvedValue(
|
||||
createMockUserProfile({ user: createMockUser({ user_id: '123', email: '123@example.com' }) }),
|
||||
@@ -1004,6 +1004,32 @@ describe('User DB Service', () => {
|
||||
]);
|
||||
});
|
||||
|
||||
it('should throw ForeignKeyConstraintError if the user_id does not exist', async () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
const queryData = {
|
||||
user_id: 'non-existent-user',
|
||||
query_text: 'search text',
|
||||
result_count: 0,
|
||||
was_successful: false,
|
||||
};
|
||||
|
||||
await expect(userRepo.logSearchQuery(queryData, mockLogger)).rejects.toThrow(
|
||||
ForeignKeyConstraintError,
|
||||
);
|
||||
|
||||
await expect(userRepo.logSearchQuery(queryData, mockLogger)).rejects.toThrow(
|
||||
'The specified user does not exist.',
|
||||
);
|
||||
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ err: dbError, queryData }),
|
||||
'Database error in logSearchQuery',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
@@ -74,9 +74,11 @@ export class UserRepository {
|
||||
passwordHash: string | null,
|
||||
profileData: { full_name?: string; avatar_url?: string },
|
||||
logger: Logger,
|
||||
// Allow passing a transactional client
|
||||
client: Pool | PoolClient = this.db,
|
||||
): Promise<UserProfile> {
|
||||
return withTransaction(async (client: PoolClient) => {
|
||||
logger.debug(`[DB createUser] Starting transaction for email: ${email}`);
|
||||
try {
|
||||
logger.debug(`[DB createUser] Starting user creation for email: ${email}`);
|
||||
|
||||
// Use 'set_config' to safely pass parameters to a configuration variable.
|
||||
await client.query("SELECT set_config('my_app.user_metadata', $1, true)", [
|
||||
@@ -126,18 +128,12 @@ export class UserRepository {
|
||||
|
||||
logger.debug({ user: fullUserProfile }, `[DB createUser] Fetched full profile for new user:`);
|
||||
return fullUserProfile;
|
||||
}).catch((error) => {
|
||||
// Specific handling for unique constraint violation on user creation
|
||||
if (error instanceof Error && 'code' in error && (error as any).code === '23505') {
|
||||
logger.warn(`Attempted to create a user with an existing email: ${email}`);
|
||||
throw new UniqueConstraintError('A user with this email address already exists.');
|
||||
}
|
||||
// Fallback to generic handler for all other errors
|
||||
handleDbError(error, logger, 'Error during createUser transaction', { email }, {
|
||||
} catch (error) {
|
||||
handleDbError(error, logger, 'Error during createUser', { email }, {
|
||||
uniqueMessage: 'A user with this email address already exists.',
|
||||
defaultMessage: 'Failed to create user in database.',
|
||||
});
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -464,6 +460,7 @@ export class UserRepository {
|
||||
refreshToken,
|
||||
]);
|
||||
} catch (error) {
|
||||
// This is a non-critical operation, so we just log the error and continue.
|
||||
logger.error({ err: error }, 'Database error in deleteRefreshToken');
|
||||
}
|
||||
}
|
||||
@@ -475,10 +472,11 @@ export class UserRepository {
|
||||
* @param expiresAt The timestamp when the token expires.
|
||||
*/
|
||||
// prettier-ignore
|
||||
async createPasswordResetToken(userId: string, tokenHash: string, expiresAt: Date, logger: Logger): Promise<void> {
|
||||
const client = this.db as PoolClient;
|
||||
async createPasswordResetToken(userId: string, tokenHash: string, expiresAt: Date, logger: Logger, client: PoolClient): Promise<void> {
|
||||
try {
|
||||
// First, remove any existing tokens for the user to ensure they can only have one active reset request.
|
||||
await client.query('DELETE FROM public.password_reset_tokens WHERE user_id = $1', [userId]);
|
||||
// Then, insert the new token.
|
||||
await client.query(
|
||||
'INSERT INTO public.password_reset_tokens (user_id, token_hash, expires_at) VALUES ($1, $2, $3)',
|
||||
[userId, tokenHash, expiresAt]
|
||||
@@ -519,10 +517,9 @@ export class UserRepository {
|
||||
try {
|
||||
await this.db.query('DELETE FROM public.password_reset_tokens WHERE token_hash = $1', [tokenHash]);
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error, tokenHash },
|
||||
'Database error in deleteResetToken',
|
||||
);
|
||||
handleDbError(error, logger, 'Database error in deleteResetToken', { tokenHash }, {
|
||||
defaultMessage: 'Failed to delete password reset token.',
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
0
src/services/flyer.db.ts
Normal file
0
src/services/flyer.db.ts
Normal file
@@ -1,5 +1,5 @@
|
||||
// src/services/flyerAiProcessor.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import { AiDataValidationError } from './processingErrors';
|
||||
import { logger } from './logger.server'; // Keep this import for the logger instance
|
||||
@@ -21,6 +21,7 @@ const createMockJobData = (data: Partial<FlyerJobData>): FlyerJobData => ({
|
||||
filePath: '/tmp/flyer.jpg',
|
||||
originalFileName: 'flyer.jpg',
|
||||
checksum: 'checksum-123',
|
||||
baseUrl: 'http://localhost:3000',
|
||||
...data,
|
||||
});
|
||||
|
||||
@@ -42,6 +43,11 @@ describe('FlyerAiProcessor', () => {
|
||||
service = new FlyerAiProcessor(mockAiService, mockPersonalizationRepo);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Ensure env stubs are cleaned up after each test
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should call AI service and return validated data on success', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
@@ -72,64 +78,230 @@ describe('FlyerAiProcessor', () => {
|
||||
expect(result.needsReview).toBe(false);
|
||||
});
|
||||
|
||||
it('should throw AiDataValidationError if AI response has incorrect data structure', async () => {
|
||||
it('should throw an error if getAllMasterItems fails', async () => {
|
||||
// Arrange
|
||||
const jobData = createMockJobData({});
|
||||
// Mock AI to return a structurally invalid response (e.g., items is not an array)
|
||||
const invalidResponse = {
|
||||
store_name: 'Invalid Store',
|
||||
items: 'not-an-array',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(invalidResponse as any);
|
||||
|
||||
const dbError = new Error('Database connection failed');
|
||||
vi.mocked(mockPersonalizationRepo.getAllMasterItems).mockRejectedValue(dbError);
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
await expect(service.extractAndValidateData(imagePaths, jobData, logger)).rejects.toThrow(
|
||||
AiDataValidationError,
|
||||
);
|
||||
|
||||
// Act & Assert
|
||||
await expect(
|
||||
service.extractAndValidateData(imagePaths, jobData, logger),
|
||||
).rejects.toThrow(dbError);
|
||||
|
||||
// Verify that the process stops before calling the AI service
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should pass validation even if store_name is missing', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: null, // Missing store name
|
||||
items: [{ item: 'Test Item', price_display: '$1.99', price_in_cents: 199, quantity: 'each', category_name: 'Grocery' }],
|
||||
// ADDED to satisfy AiFlyerDataSchema
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
describe('Validation and Quality Checks', () => {
|
||||
it('should pass validation and not flag for review with good quality data', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Good Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Good St',
|
||||
items: [
|
||||
{ item: 'Priced Item 1', price_in_cents: 199, price_display: '$1.99', quantity: '1', category_name: 'A' },
|
||||
{ item: 'Priced Item 2', price_in_cents: 299, price_display: '$2.99', quantity: '1', category_name: 'B' },
|
||||
],
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
// It should not throw, but return the data and log a warning.
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('missing a store name. The transformer will use a fallback. Flagging for review.'));
|
||||
// With all data present and correct, it should not need a review.
|
||||
expect(result.needsReview).toBe(false);
|
||||
expect(logger.warn).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should throw AiDataValidationError if AI response has incorrect data structure', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
// Mock AI to return a structurally invalid response (e.g., items is not an array)
|
||||
const invalidResponse = {
|
||||
store_name: 'Invalid Store',
|
||||
items: 'not-an-array',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(invalidResponse as any);
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
await expect(service.extractAndValidateData(imagePaths, jobData, logger)).rejects.toThrow(
|
||||
AiDataValidationError,
|
||||
);
|
||||
});
|
||||
|
||||
it('should flag for review if store_name is missing', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: null, // Missing store name
|
||||
items: [{ item: 'Test Item', price_display: '$1.99', price_in_cents: 199, quantity: 'each', category_name: 'Grocery' }],
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ qualityIssues: ['Missing store name'] }),
|
||||
expect.stringContaining('AI response has quality issues.'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should flag for review if items array is empty', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
items: [], // Empty items array
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ qualityIssues: ['No items were extracted'] }),
|
||||
expect.stringContaining('AI response has quality issues.'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should flag for review if item price quality is low', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{ item: 'Priced Item', price_in_cents: 199, price_display: '$1.99', quantity: '1', category_name: 'A' },
|
||||
{ item: 'Unpriced Item 1', price_in_cents: null, price_display: 'See store', quantity: '1', category_name: 'B' },
|
||||
{ item: 'Unpriced Item 2', price_in_cents: null, price_display: 'FREE', quantity: '1', category_name: 'C' },
|
||||
], // 1/3 = 33% have price, which is < 50%
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ qualityIssues: ['Low price quality (33% of items have a price)'] }),
|
||||
expect.stringContaining('AI response has quality issues.'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should use a custom price quality threshold from an environment variable', async () => {
|
||||
// Arrange
|
||||
vi.stubEnv('AI_PRICE_QUALITY_THRESHOLD', '0.8'); // Set a stricter threshold (80%)
|
||||
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{ item: 'Priced Item 1', price_in_cents: 199, price_display: '$1.99', quantity: '1', category_name: 'A' },
|
||||
{ item: 'Priced Item 2', price_in_cents: 299, price_display: '$2.99', quantity: '1', category_name: 'B' },
|
||||
{ item: 'Priced Item 3', price_in_cents: 399, price_display: '$3.99', quantity: '1', category_name: 'C' },
|
||||
{ item: 'Unpriced Item 1', price_in_cents: null, price_display: 'See store', quantity: '1', category_name: 'D' },
|
||||
], // 3/4 = 75% have price. This is > 50% (default) but < 80% (custom).
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
// Act
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
// Assert
|
||||
// Because 75% < 80%, it should be flagged for review.
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ qualityIssues: ['Low price quality (75% of items have a price)'] }),
|
||||
expect.stringContaining('AI response has quality issues.'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should flag for review if validity dates are missing', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
valid_from: null, // Missing date
|
||||
valid_to: null, // Missing date
|
||||
store_address: '123 Test St',
|
||||
items: [{ item: 'Test Item', price_in_cents: 199, price_display: '$1.99', quantity: '1', category_name: 'A' }],
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ qualityIssues: ['Missing both valid_from and valid_to dates'] }),
|
||||
expect.stringContaining('AI response has quality issues.'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should combine multiple quality issues in the log', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: null, // Issue 1
|
||||
items: [], // Issue 2
|
||||
valid_from: null, // Issue 3
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
{ rawData: mockAiResponse, qualityIssues: ['Missing store name', 'No items were extracted', 'Missing both valid_from and valid_to dates'] },
|
||||
'AI response has quality issues. Flagging for review. Issues: Missing store name, No items were extracted, Missing both valid_from and valid_to dates',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
it('should pass validation even if items array is empty', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
it('should pass the userProfileAddress from jobData to the AI service', async () => {
|
||||
// Arrange
|
||||
const jobData = createMockJobData({ userProfileAddress: '456 Fallback Ave' });
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
items: [], // Empty items array
|
||||
// ADDED to satisfy AiFlyerDataSchema
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [{ item: 'Test Item', price_in_cents: 199, price_display: '$1.99', quantity: '1', category_name: 'A' }],
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('contains no items. The flyer will be saved with an item_count of 0. Flagging for review.'));
|
||||
await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
// Assert
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledWith(
|
||||
imagePaths, [], undefined, '456 Fallback Ave', logger
|
||||
);
|
||||
});
|
||||
|
||||
describe('Batching Logic', () => {
|
||||
@@ -200,6 +372,46 @@ describe('FlyerAiProcessor', () => {
|
||||
expect(result.needsReview).toBe(false);
|
||||
});
|
||||
|
||||
it('should handle an empty object response from a batch without crashing', async () => {
|
||||
// Arrange
|
||||
const jobData = createMockJobData({});
|
||||
const imagePaths = [
|
||||
{ path: 'page1.jpg', mimetype: 'image/jpeg' }, { path: 'page2.jpg', mimetype: 'image/jpeg' }, { path: 'page3.jpg', mimetype: 'image/jpeg' }, { path: 'page4.jpg', mimetype: 'image/jpeg' }, { path: 'page5.jpg', mimetype: 'image/jpeg' },
|
||||
];
|
||||
|
||||
const mockAiResponseBatch1 = {
|
||||
store_name: 'Good Store',
|
||||
valid_from: '2025-01-01',
|
||||
valid_to: '2025-01-07',
|
||||
store_address: '123 Good St',
|
||||
items: [
|
||||
{ item: 'Item A', price_display: '$1', price_in_cents: 100, quantity: '1', category_name: 'Cat A', master_item_id: 1 },
|
||||
],
|
||||
};
|
||||
|
||||
// The AI returns an empty object for the second batch.
|
||||
const mockAiResponseBatch2 = {};
|
||||
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage)
|
||||
.mockResolvedValueOnce(mockAiResponseBatch1)
|
||||
.mockResolvedValueOnce(mockAiResponseBatch2 as any); // Use `as any` to bypass strict type check for the test mock
|
||||
|
||||
// Act
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
// Assert
|
||||
// 1. AI service was called twice.
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(2);
|
||||
|
||||
// 2. The final data should only contain data from the first batch.
|
||||
expect(result.data.store_name).toBe('Good Store');
|
||||
expect(result.data.items).toHaveLength(1);
|
||||
expect(result.data.items[0].item).toBe('Item A');
|
||||
|
||||
// 3. The process should complete without errors and not be flagged for review if the first batch was good.
|
||||
expect(result.needsReview).toBe(false);
|
||||
});
|
||||
|
||||
it('should fill in missing metadata from subsequent batches', async () => {
|
||||
// Arrange
|
||||
const jobData = createMockJobData({});
|
||||
@@ -225,4 +437,40 @@ describe('FlyerAiProcessor', () => {
|
||||
expect(result.data.items).toHaveLength(2);
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle a single batch correctly when image count is less than BATCH_SIZE', async () => {
|
||||
// Arrange
|
||||
const jobData = createMockJobData({});
|
||||
// 2 images, which is less than the BATCH_SIZE of 4.
|
||||
const imagePaths = [
|
||||
{ path: 'page1.jpg', mimetype: 'image/jpeg' },
|
||||
{ path: 'page2.jpg', mimetype: 'image/jpeg' },
|
||||
];
|
||||
|
||||
const mockAiResponse = {
|
||||
store_name: 'Single Batch Store',
|
||||
valid_from: '2025-02-01',
|
||||
valid_to: '2025-02-07',
|
||||
store_address: '789 Single St',
|
||||
items: [
|
||||
{ item: 'Item X', price_display: '$10', price_in_cents: 1000, quantity: '1', category_name: 'Cat X', master_item_id: 10 },
|
||||
],
|
||||
};
|
||||
|
||||
// Mock the AI service to be called only once.
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValueOnce(mockAiResponse);
|
||||
|
||||
// Act
|
||||
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||
|
||||
// Assert
|
||||
// 1. AI service was called only once.
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(1);
|
||||
|
||||
// 2. Check the arguments for the single call.
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledWith(imagePaths, [], undefined, undefined, logger);
|
||||
|
||||
// 3. Check that the final data matches the single batch's data.
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
});
|
||||
});
|
||||
@@ -46,26 +46,52 @@ export class FlyerAiProcessor {
|
||||
);
|
||||
}
|
||||
|
||||
// --- NEW QUALITY CHECK ---
|
||||
// After structural validation, perform semantic quality checks.
|
||||
const { store_name, items } = validationResult.data;
|
||||
let needsReview = false;
|
||||
// --- Data Quality Checks ---
|
||||
// After structural validation, perform semantic quality checks to flag low-quality
|
||||
// extractions for manual review.
|
||||
const { store_name, items, valid_from, valid_to } = validationResult.data;
|
||||
const qualityIssues: string[] = [];
|
||||
|
||||
// 1. Check for a valid store name, but don't fail the job.
|
||||
// The data transformer will handle this by assigning a fallback name.
|
||||
// 1. Check for a store name.
|
||||
if (!store_name || store_name.trim() === '') {
|
||||
logger.warn({ rawData: extractedData }, 'AI response is missing a store name. The transformer will use a fallback. Flagging for review.');
|
||||
needsReview = true;
|
||||
qualityIssues.push('Missing store name');
|
||||
}
|
||||
|
||||
// 2. Check that at least one item was extracted, but don't fail the job.
|
||||
// An admin can review a flyer with 0 items.
|
||||
// 2. Check that items were extracted.
|
||||
if (!items || items.length === 0) {
|
||||
logger.warn({ rawData: extractedData }, 'AI response contains no items. The flyer will be saved with an item_count of 0. Flagging for review.');
|
||||
needsReview = true;
|
||||
qualityIssues.push('No items were extracted');
|
||||
} else {
|
||||
// 3. If items exist, check their quality (e.g., missing prices).
|
||||
// The threshold is configurable via an environment variable, defaulting to 0.5 (50%).
|
||||
const priceQualityThreshold = parseFloat(process.env.AI_PRICE_QUALITY_THRESHOLD || '0.5');
|
||||
|
||||
const itemsWithPrice = items.filter(
|
||||
(item) => item.price_in_cents != null && item.price_in_cents > 0,
|
||||
).length;
|
||||
const priceQualityRatio = itemsWithPrice / items.length;
|
||||
|
||||
if (priceQualityRatio < priceQualityThreshold) {
|
||||
// If the ratio of items with a valid price is below the threshold, flag for review.
|
||||
qualityIssues.push(
|
||||
`Low price quality (${(priceQualityRatio * 100).toFixed(0)}% of items have a price)`,
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
logger.info(`AI extracted ${validationResult.data.items.length} items.`);
|
||||
// 4. Check for flyer validity dates.
|
||||
if (!valid_from && !valid_to) {
|
||||
qualityIssues.push('Missing both valid_from and valid_to dates');
|
||||
}
|
||||
|
||||
const needsReview = qualityIssues.length > 0;
|
||||
if (needsReview) {
|
||||
logger.warn(
|
||||
{ rawData: extractedData, qualityIssues },
|
||||
`AI response has quality issues. Flagging for review. Issues: ${qualityIssues.join(', ')}`,
|
||||
);
|
||||
}
|
||||
|
||||
logger.info(`AI extracted ${validationResult.data.items.length} items. Needs Review: ${needsReview}`);
|
||||
return { data: validationResult.data, needsReview };
|
||||
}
|
||||
|
||||
@@ -129,7 +155,7 @@ export class FlyerAiProcessor {
|
||||
}
|
||||
|
||||
// 2. Items: Append all found items to the master list.
|
||||
mergedData.items.push(...batchResult.items);
|
||||
mergedData.items.push(...(batchResult.items || []));
|
||||
}
|
||||
|
||||
logger.info(`Batch processing complete. Total items extracted: ${mergedData.items.length}`);
|
||||
|
||||
@@ -21,6 +21,11 @@ describe('FlyerDataTransformer', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
transformer = new FlyerDataTransformer();
|
||||
// Stub environment variables to ensure consistency and predictability.
|
||||
// Prioritize FRONTEND_URL to match the updated service logic.
|
||||
vi.stubEnv('FRONTEND_URL', 'http://localhost:3000');
|
||||
vi.stubEnv('BASE_URL', ''); // Ensure this is not used to confirm priority logic
|
||||
vi.stubEnv('PORT', ''); // Ensure this is not used
|
||||
|
||||
// Provide a default mock implementation for generateFlyerIcon
|
||||
vi.mocked(generateFlyerIcon).mockResolvedValue('icon-flyer-page-1.webp');
|
||||
@@ -59,6 +64,7 @@ describe('FlyerDataTransformer', () => {
|
||||
const originalFileName = 'my-flyer.pdf';
|
||||
const checksum = 'checksum-abc-123';
|
||||
const userId = 'user-xyz-456';
|
||||
const baseUrl = 'http://test.host';
|
||||
|
||||
// Act
|
||||
const { flyerData, itemsForDb } = await transformer.transform(
|
||||
@@ -68,6 +74,7 @@ describe('FlyerDataTransformer', () => {
|
||||
checksum,
|
||||
userId,
|
||||
mockLogger,
|
||||
baseUrl,
|
||||
);
|
||||
|
||||
// Assert
|
||||
@@ -83,8 +90,8 @@ describe('FlyerDataTransformer', () => {
|
||||
// 1. Check flyer data
|
||||
expect(flyerData).toEqual({
|
||||
file_name: originalFileName,
|
||||
image_url: '/flyer-images/flyer-page-1.jpg',
|
||||
icon_url: '/flyer-images/icons/icon-flyer-page-1.webp',
|
||||
image_url: `${baseUrl}/flyer-images/flyer-page-1.jpg`,
|
||||
icon_url: `${baseUrl}/flyer-images/icons/icon-flyer-page-1.webp`,
|
||||
checksum,
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
@@ -149,6 +156,7 @@ describe('FlyerDataTransformer', () => {
|
||||
checksum,
|
||||
undefined,
|
||||
mockLogger,
|
||||
'http://another.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
@@ -167,8 +175,8 @@ describe('FlyerDataTransformer', () => {
|
||||
expect(itemsForDb).toHaveLength(0);
|
||||
expect(flyerData).toEqual({
|
||||
file_name: originalFileName,
|
||||
image_url: '/flyer-images/another.png',
|
||||
icon_url: '/flyer-images/icons/icon-another.webp',
|
||||
image_url: `http://another.host/flyer-images/another.png`,
|
||||
icon_url: `http://another.host/flyer-images/icons/icon-another.webp`,
|
||||
checksum,
|
||||
store_name: 'Unknown Store (auto)', // Should use fallback
|
||||
valid_from: null,
|
||||
@@ -176,7 +184,7 @@ describe('FlyerDataTransformer', () => {
|
||||
store_address: null,
|
||||
item_count: 0,
|
||||
status: 'needs_review',
|
||||
uploaded_by: undefined, // Should be undefined
|
||||
uploaded_by: null, // Should be null
|
||||
});
|
||||
});
|
||||
|
||||
@@ -221,6 +229,7 @@ describe('FlyerDataTransformer', () => {
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://normalize.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
@@ -240,4 +249,270 @@ describe('FlyerDataTransformer', () => {
|
||||
}),
|
||||
);
|
||||
});
|
||||
|
||||
it('should use fallback baseUrl if none is provided and log a warning', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
const baseUrl = undefined; // Explicitly pass undefined for this test
|
||||
|
||||
// The fallback logic uses process.env.PORT || 3000.
|
||||
// The beforeEach sets PORT to '', so it should fallback to 3000.
|
||||
const expectedFallbackUrl = 'http://localhost:3000';
|
||||
|
||||
// Act
|
||||
const { flyerData } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'my-flyer.pdf',
|
||||
'checksum-abc-123',
|
||||
'user-xyz-456',
|
||||
mockLogger,
|
||||
baseUrl, // Pass undefined here
|
||||
);
|
||||
|
||||
// Assert
|
||||
// 1. Check that a warning was logged
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
`Base URL not provided in job data. Falling back to default local URL: ${expectedFallbackUrl}`,
|
||||
);
|
||||
|
||||
// 2. Check that the URLs were constructed with the fallback
|
||||
expect(flyerData.image_url).toBe(`${expectedFallbackUrl}/flyer-images/flyer-page-1.jpg`);
|
||||
expect(flyerData.icon_url).toBe(
|
||||
`${expectedFallbackUrl}/flyer-images/icons/icon-flyer-page-1.webp`,
|
||||
);
|
||||
});
|
||||
|
||||
describe('_normalizeItem price parsing', () => {
|
||||
it('should use price_in_cents from AI if it is valid, ignoring price_display', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 'Milk',
|
||||
price_display: '$4.99', // Parsable, but should be ignored
|
||||
price_in_cents: 399, // AI provides a specific (maybe wrong) value
|
||||
quantity: '1L',
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { itemsForDb } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://test.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(itemsForDb[0].price_in_cents).toBe(399); // AI's value should be prioritized
|
||||
});
|
||||
|
||||
it('should use parsePriceToCents as a fallback if AI price_in_cents is null', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 'Milk',
|
||||
price_display: '$4.99', // Parsable value
|
||||
price_in_cents: null, // AI fails to provide a value
|
||||
quantity: '1L',
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { itemsForDb } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://test.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(itemsForDb[0].price_in_cents).toBe(499); // Should be parsed from price_display
|
||||
});
|
||||
|
||||
it('should result in null if both AI price and display price are unparsable', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 'Milk',
|
||||
price_display: 'FREE', // Unparsable
|
||||
price_in_cents: null, // AI provides null
|
||||
quantity: '1L',
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { itemsForDb } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://test.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(itemsForDb[0].price_in_cents).toBeNull();
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle non-string values for string fields gracefully by converting them', async () => {
|
||||
// This test verifies that if data with incorrect types bypasses earlier validation,
|
||||
// the transformer is robust enough to convert them to strings instead of crashing.
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Type-Unsafe Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 12345 as any, // Simulate AI returning a number instead of a string
|
||||
price_display: 3.99 as any, // Simulate a number for a string field
|
||||
price_in_cents: 399,
|
||||
quantity: 5 as any, // Simulate a number
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { itemsForDb } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://robust.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(itemsForDb).toHaveLength(1);
|
||||
expect(itemsForDb[0]).toEqual(
|
||||
expect.objectContaining({
|
||||
item: '12345', // Should be converted to string
|
||||
price_display: '3.99', // Should be converted to string
|
||||
quantity: '5', // Should be converted to string
|
||||
}),
|
||||
);
|
||||
});
|
||||
|
||||
describe('needsReview flag handling', () => {
|
||||
it('should set status to "processed" when needsReview is false', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [],
|
||||
},
|
||||
needsReview: false, // Key part of this test
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { flyerData } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://test.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(flyerData.status).toBe('processed');
|
||||
});
|
||||
|
||||
it('should set status to "needs_review" when needsReview is true', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [],
|
||||
},
|
||||
needsReview: true, // Key part of this test
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { flyerData } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
'http://test.host',
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(flyerData.status).toBe('needs_review');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -7,6 +7,7 @@ import type { AiProcessorResult } from './flyerAiProcessor.server'; // Keep this
|
||||
import { AiFlyerDataSchema } from '../types/ai'; // Import consolidated schema
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import { TransformationError } from './processingErrors';
|
||||
import { parsePriceToCents } from '../utils/priceParser';
|
||||
|
||||
/**
|
||||
* This class is responsible for transforming the validated data from the AI service
|
||||
@@ -21,16 +22,25 @@ export class FlyerDataTransformer {
|
||||
private _normalizeItem(
|
||||
item: z.infer<typeof AiFlyerDataSchema>['items'][number],
|
||||
): FlyerItemInsert {
|
||||
// If the AI fails to provide `price_in_cents` but provides a parsable `price_display`,
|
||||
// we can use our own parser as a fallback to improve data quality.
|
||||
const priceFromDisplay = parsePriceToCents(item.price_display ?? '');
|
||||
|
||||
// Prioritize the AI's direct `price_in_cents` value, but use the parsed value if the former is null.
|
||||
const finalPriceInCents = item.price_in_cents ?? priceFromDisplay;
|
||||
|
||||
return {
|
||||
...item,
|
||||
// Use logical OR to default falsy values (null, undefined, '') to a fallback.
|
||||
// The trim is important for cases where the AI returns only whitespace.
|
||||
item: String(item.item || '').trim() || 'Unknown Item',
|
||||
// Use nullish coalescing to default only null/undefined to an empty string.
|
||||
price_display: String(item.price_display ?? ''),
|
||||
quantity: String(item.quantity ?? ''),
|
||||
// Use logical OR to default falsy category names (null, undefined, '') to a fallback.
|
||||
category_name: String(item.category_name || 'Other/Miscellaneous'),
|
||||
// Use nullish coalescing and trim for robustness.
|
||||
// An empty or whitespace-only name falls back to 'Unknown Item'.
|
||||
item: (String(item.item ?? '')).trim() || 'Unknown Item',
|
||||
// Default null/undefined to an empty string and trim.
|
||||
price_display: (String(item.price_display ?? '')).trim(),
|
||||
quantity: (String(item.quantity ?? '')).trim(),
|
||||
// An empty or whitespace-only category falls back to 'Other/Miscellaneous'.
|
||||
category_name: (String(item.category_name ?? '')).trim() || 'Other/Miscellaneous',
|
||||
// Overwrite price_in_cents with our calculated value.
|
||||
price_in_cents: finalPriceInCents,
|
||||
// Use nullish coalescing to convert null to undefined for the database.
|
||||
master_item_id: item.master_item_id ?? undefined,
|
||||
view_count: 0,
|
||||
@@ -38,6 +48,47 @@ export class FlyerDataTransformer {
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Generates a 64x64 icon for the flyer's first page.
|
||||
* @param firstImage The path to the first image of the flyer.
|
||||
* @param logger The logger instance.
|
||||
* @returns The filename of the generated icon.
|
||||
*/
|
||||
private async _generateIcon(firstImage: string, logger: Logger): Promise<string> {
|
||||
const iconFileName = await generateFlyerIcon(
|
||||
firstImage,
|
||||
path.join(path.dirname(firstImage), 'icons'),
|
||||
logger,
|
||||
);
|
||||
return iconFileName;
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs the full public URLs for the flyer image and its icon.
|
||||
* @param firstImage The path to the first image of the flyer.
|
||||
* @param iconFileName The filename of the generated icon.
|
||||
* @param baseUrl The base URL from the job payload.
|
||||
* @param logger The logger instance.
|
||||
* @returns An object containing the full image_url and icon_url.
|
||||
*/
|
||||
private _buildUrls(
|
||||
firstImage: string,
|
||||
iconFileName: string,
|
||||
baseUrl: string | undefined,
|
||||
logger: Logger,
|
||||
): { imageUrl: string; iconUrl: string } {
|
||||
let finalBaseUrl = baseUrl;
|
||||
if (!finalBaseUrl) {
|
||||
const port = process.env.PORT || 3000;
|
||||
finalBaseUrl = `http://localhost:${port}`;
|
||||
logger.warn(`Base URL not provided in job data. Falling back to default local URL: ${finalBaseUrl}`);
|
||||
}
|
||||
finalBaseUrl = finalBaseUrl.endsWith('/') ? finalBaseUrl.slice(0, -1) : finalBaseUrl;
|
||||
const imageUrl = `${finalBaseUrl}/flyer-images/${path.basename(firstImage)}`;
|
||||
const iconUrl = `${finalBaseUrl}/flyer-images/icons/${iconFileName}`;
|
||||
return { imageUrl, iconUrl };
|
||||
}
|
||||
|
||||
/**
|
||||
* Transforms AI-extracted data into database-ready flyer and item records.
|
||||
* @param extractedData The validated data from the AI.
|
||||
@@ -55,6 +106,7 @@ export class FlyerDataTransformer {
|
||||
checksum: string,
|
||||
userId: string | undefined,
|
||||
logger: Logger,
|
||||
baseUrl?: string,
|
||||
): Promise<{ flyerData: FlyerInsert; itemsForDb: FlyerItemInsert[] }> {
|
||||
logger.info('Starting data transformation from AI output to database format.');
|
||||
|
||||
@@ -62,11 +114,8 @@ export class FlyerDataTransformer {
|
||||
const { data: extractedData, needsReview } = aiResult;
|
||||
|
||||
const firstImage = imagePaths[0].path;
|
||||
const iconFileName = await generateFlyerIcon(
|
||||
firstImage,
|
||||
path.join(path.dirname(firstImage), 'icons'),
|
||||
logger,
|
||||
);
|
||||
const iconFileName = await this._generateIcon(firstImage, logger);
|
||||
const { imageUrl, iconUrl } = this._buildUrls(firstImage, iconFileName, baseUrl, logger);
|
||||
|
||||
const itemsForDb: FlyerItemInsert[] = extractedData.items.map((item) => this._normalizeItem(item));
|
||||
|
||||
@@ -77,15 +126,15 @@ export class FlyerDataTransformer {
|
||||
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${path.basename(firstImage)}`,
|
||||
icon_url: `/flyer-images/icons/${iconFileName}`,
|
||||
image_url: imageUrl,
|
||||
icon_url: iconUrl,
|
||||
checksum,
|
||||
store_name: storeName,
|
||||
valid_from: extractedData.valid_from,
|
||||
valid_to: extractedData.valid_to,
|
||||
store_address: extractedData.store_address, // The number of items is now calculated directly from the transformed data.
|
||||
store_address: extractedData.store_address,
|
||||
item_count: itemsForDb.length,
|
||||
uploaded_by: userId,
|
||||
uploaded_by: userId ? userId : null,
|
||||
status: needsReview ? 'needs_review' : 'processed',
|
||||
};
|
||||
|
||||
|
||||
@@ -10,6 +10,7 @@ const mocks = vi.hoisted(() => ({
|
||||
unlink: vi.fn(),
|
||||
readdir: vi.fn(),
|
||||
execAsync: vi.fn(),
|
||||
mockAdminLogActivity: vi.fn(),
|
||||
}));
|
||||
|
||||
// 2. Mock modules using the hoisted variables
|
||||
@@ -34,7 +35,9 @@ import {
|
||||
AiDataValidationError,
|
||||
PdfConversionError,
|
||||
UnsupportedFileTypeError,
|
||||
TransformationError,
|
||||
} from './processingErrors';
|
||||
import { NotFoundError } from './db/errors.db';
|
||||
import { FlyerFileHandler } from './flyerFileHandler.server';
|
||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import type { IFileSystem, ICommandExecutor } from './flyerFileHandler.server';
|
||||
@@ -52,6 +55,13 @@ vi.mock('./db/flyer.db', () => ({
|
||||
vi.mock('./db/index.db', () => ({
|
||||
personalizationRepo: { getAllMasterItems: vi.fn() },
|
||||
adminRepo: { logActivity: vi.fn() },
|
||||
flyerRepo: { getFlyerById: vi.fn() },
|
||||
withTransaction: vi.fn(),
|
||||
}));
|
||||
vi.mock('./db/admin.db', () => ({
|
||||
AdminRepository: vi.fn().mockImplementation(function () {
|
||||
return { logActivity: mocks.mockAdminLogActivity };
|
||||
}),
|
||||
}));
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
@@ -78,21 +88,23 @@ describe('FlyerProcessingService', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
// Provide a default mock implementation for withTransaction that just executes the callback.
|
||||
// This is needed for the happy path tests. Tests for transaction failures will override this.
|
||||
vi.mocked(mockedDb.withTransaction).mockImplementation(async (callback: any) => callback({}));
|
||||
|
||||
// Spy on the real transformer's method and provide a mock implementation.
|
||||
// This is more robust than mocking the entire class constructor.
|
||||
vi.spyOn(FlyerDataTransformer.prototype, 'transform').mockResolvedValue({
|
||||
flyerData: {
|
||||
file_name: 'test.jpg',
|
||||
image_url: 'test.jpg',
|
||||
icon_url: 'icon.webp',
|
||||
checksum: 'checksum-123',
|
||||
image_url: 'http://example.com/test.jpg',
|
||||
icon_url: 'http://example.com/icon.webp',
|
||||
store_name: 'Mock Store',
|
||||
// Add required fields for FlyerInsert type
|
||||
status: 'processed',
|
||||
item_count: 0,
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Mock St',
|
||||
} as FlyerInsert, // Cast is okay here as it's a mock value
|
||||
itemsForDb: [],
|
||||
});
|
||||
@@ -116,7 +128,6 @@ describe('FlyerProcessingService', () => {
|
||||
service = new FlyerProcessingService(
|
||||
mockFileHandler,
|
||||
mockAiProcessor,
|
||||
mockedDb,
|
||||
mockFs,
|
||||
mockCleanupQueue,
|
||||
new FlyerDataTransformer(),
|
||||
@@ -151,7 +162,7 @@ describe('FlyerProcessingService', () => {
|
||||
flyer: createMockFlyer({
|
||||
flyer_id: 1,
|
||||
file_name: 'test.jpg',
|
||||
image_url: 'test.jpg',
|
||||
image_url: 'http://example.com/test.jpg',
|
||||
item_count: 1,
|
||||
}),
|
||||
items: [],
|
||||
@@ -168,6 +179,7 @@ describe('FlyerProcessingService', () => {
|
||||
filePath: '/tmp/flyer.jpg',
|
||||
originalFileName: 'flyer.jpg',
|
||||
checksum: 'checksum-123',
|
||||
baseUrl: 'http://localhost:3000',
|
||||
...data,
|
||||
},
|
||||
updateProgress: vi.fn(),
|
||||
@@ -195,8 +207,11 @@ describe('FlyerProcessingService', () => {
|
||||
expect(result).toEqual({ flyerId: 1 });
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith(job.data.filePath, job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
// Verify that the transaction function was called.
|
||||
expect(mockedDb.withTransaction).toHaveBeenCalledTimes(1);
|
||||
// Verify that the functions inside the transaction were called.
|
||||
expect(createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
expect(mockedDb.adminRepo.logActivity).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.mockAdminLogActivity).toHaveBeenCalledTimes(1);
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: 1, paths: ['/tmp/flyer.jpg'] },
|
||||
@@ -216,6 +231,8 @@ describe('FlyerProcessingService', () => {
|
||||
|
||||
await service.processJob(job);
|
||||
|
||||
// Verify transaction and inner calls
|
||||
expect(mockedDb.withTransaction).toHaveBeenCalledTimes(1);
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith('/tmp/flyer.pdf', job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
expect(createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
@@ -363,6 +380,8 @@ describe('FlyerProcessingService', () => {
|
||||
|
||||
await service.processJob(job);
|
||||
|
||||
// Verify transaction and inner calls
|
||||
expect(mockedDb.withTransaction).toHaveBeenCalledTimes(1);
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith('/tmp/flyer.gif', job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
@@ -376,20 +395,25 @@ describe('FlyerProcessingService', () => {
|
||||
const job = createMockJob({});
|
||||
const { logger } = await import('./logger.server');
|
||||
const dbError = new Error('Database transaction failed');
|
||||
vi.mocked(createFlyerAndItems).mockRejectedValue(dbError);
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow('Database transaction failed');
|
||||
// To test the DB failure, we make the transaction itself fail when called.
|
||||
// This is more realistic than mocking the inner function `createFlyerAndItems`.
|
||||
vi.mocked(mockedDb.withTransaction).mockRejectedValue(dbError);
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'Database transaction failed',
|
||||
// The service wraps the generic DB error in a DatabaseError, but _reportErrorAndThrow re-throws the original.
|
||||
await expect(service.processJob(job)).rejects.toThrow(dbError);
|
||||
|
||||
// The final progress update should reflect the structured DatabaseError.
|
||||
expect(job.updateProgress).toHaveBeenLastCalledWith({
|
||||
errorCode: 'DATABASE_ERROR',
|
||||
message: 'A database operation failed. Please try again later.',
|
||||
stages: [
|
||||
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||
{ name: 'Extracting Data with AI', status: 'completed', critical: true, detail: 'Communicating with AI model...' },
|
||||
{ name: 'Transforming AI Data', status: 'completed', critical: true },
|
||||
{ name: 'Saving to Database', status: 'failed', critical: true, detail: 'Database transaction failed' },
|
||||
{ name: 'Saving to Database', status: 'failed', critical: true, detail: 'A database operation failed. Please try again later.' },
|
||||
],
|
||||
}); // This was a duplicate, fixed.
|
||||
});
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
@@ -419,17 +443,17 @@ describe('FlyerProcessingService', () => {
|
||||
it('should delegate to _reportErrorAndThrow if icon generation fails', async () => {
|
||||
const job = createMockJob({});
|
||||
const { logger } = await import('./logger.server');
|
||||
const iconError = new Error('Icon generation failed.');
|
||||
const transformationError = new TransformationError('Icon generation failed.');
|
||||
// The `transform` method calls `generateFlyerIcon`. In `beforeEach`, `transform` is mocked
|
||||
// to always succeed. For this test, we override that mock to simulate a failure
|
||||
// bubbling up from the icon generation step.
|
||||
vi.spyOn(FlyerDataTransformer.prototype, 'transform').mockRejectedValue(iconError);
|
||||
vi.spyOn(FlyerDataTransformer.prototype, 'transform').mockRejectedValue(transformationError);
|
||||
|
||||
const reportErrorSpy = vi.spyOn(service as any, '_reportErrorAndThrow');
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow('Icon generation failed.');
|
||||
|
||||
expect(reportErrorSpy).toHaveBeenCalledWith(iconError, job, expect.any(Object), expect.any(Array));
|
||||
expect(reportErrorSpy).toHaveBeenCalledWith(transformationError, job, expect.any(Object), expect.any(Array));
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
@@ -590,14 +614,23 @@ describe('FlyerProcessingService', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should skip processing and return "skipped" if paths array is empty', async () => {
|
||||
it('should skip processing and return "skipped" if paths array is empty and paths cannot be derived', async () => {
|
||||
const job = createMockCleanupJob({ flyerId: 1, paths: [] });
|
||||
// Mock that the flyer cannot be found in the DB, so paths cannot be derived.
|
||||
vi.mocked(mockedDb.flyerRepo.getFlyerById).mockRejectedValue(new NotFoundError('Not found'));
|
||||
|
||||
const result = await service.processCleanupJob(job);
|
||||
|
||||
expect(mocks.unlink).not.toHaveBeenCalled();
|
||||
expect(result).toEqual({ status: 'skipped', reason: 'no paths' });
|
||||
expect(result).toEqual({ status: 'skipped', reason: 'no paths derived' });
|
||||
const { logger } = await import('./logger.server');
|
||||
expect(logger.warn).toHaveBeenCalledWith('Job received no paths to clean. Skipping.');
|
||||
// Check for both warnings: the attempt to derive, and the final skip message.
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Cleanup job for flyer 1 received no paths. Attempting to derive paths from DB.',
|
||||
);
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job received no paths and could not derive any from the database. Skipping.',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,11 +1,12 @@
|
||||
// src/services/flyerProcessingService.server.ts
|
||||
import type { Job, Queue } from 'bullmq';
|
||||
import { UnrecoverableError } from 'bullmq';
|
||||
import path from 'path';
|
||||
import type { Logger } from 'pino';
|
||||
import type { FlyerFileHandler, IFileSystem, ICommandExecutor } from './flyerFileHandler.server';
|
||||
import type { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import type * as Db from './db/index.db';
|
||||
import type { AdminRepository } from './db/admin.db';
|
||||
import * as db from './db/index.db';
|
||||
import { AdminRepository } from './db/admin.db';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import type { FlyerJobData, CleanupJobData } from '../types/job-data';
|
||||
import {
|
||||
@@ -13,7 +14,9 @@ import {
|
||||
PdfConversionError,
|
||||
AiDataValidationError,
|
||||
UnsupportedFileTypeError,
|
||||
DatabaseError, // This is from processingErrors
|
||||
} from './processingErrors';
|
||||
import { NotFoundError } from './db/errors.db';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
|
||||
@@ -34,9 +37,6 @@ export class FlyerProcessingService {
|
||||
constructor(
|
||||
private fileHandler: FlyerFileHandler,
|
||||
private aiProcessor: FlyerAiProcessor,
|
||||
// This service only needs the `logActivity` method from the `adminRepo`.
|
||||
// By using `Pick`, we create a more focused and testable dependency.
|
||||
private db: { adminRepo: Pick<AdminRepository, 'logActivity'> },
|
||||
private fs: IFileSystem,
|
||||
// By depending on `Pick<Queue, 'add'>`, we specify that this service only needs
|
||||
// an object with an `add` method that matches the Queue's `add` method signature.
|
||||
@@ -99,6 +99,7 @@ export class FlyerProcessingService {
|
||||
job.data.checksum,
|
||||
job.data.userId,
|
||||
logger,
|
||||
job.data.baseUrl,
|
||||
);
|
||||
stages[2].status = 'completed';
|
||||
await job.updateProgress({ stages });
|
||||
@@ -107,30 +108,45 @@ export class FlyerProcessingService {
|
||||
stages[3].status = 'in-progress';
|
||||
await job.updateProgress({ stages });
|
||||
|
||||
const { flyer } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||
let flyerId: number;
|
||||
try {
|
||||
const { flyer } = await db.withTransaction(async (client) => {
|
||||
// This assumes createFlyerAndItems is refactored to accept a transactional client.
|
||||
const { flyer: newFlyer } = await createFlyerAndItems(flyerData, itemsForDb, logger, client);
|
||||
|
||||
// Instantiate a new AdminRepository with the transactional client to ensure
|
||||
// the activity log is part of the same transaction.
|
||||
const transactionalAdminRepo = new AdminRepository(client);
|
||||
await transactionalAdminRepo.logActivity(
|
||||
{
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed flyer for ${flyerData.store_name}`,
|
||||
details: { flyer_id: newFlyer.flyer_id, store_name: flyerData.store_name },
|
||||
userId: job.data.userId,
|
||||
},
|
||||
logger,
|
||||
);
|
||||
|
||||
return { flyer: newFlyer };
|
||||
});
|
||||
flyerId = flyer.flyer_id;
|
||||
} catch (error) {
|
||||
if (error instanceof FlyerProcessingError) throw error;
|
||||
throw new DatabaseError(error instanceof Error ? error.message : String(error));
|
||||
}
|
||||
|
||||
stages[3].status = 'completed';
|
||||
await job.updateProgress({ stages });
|
||||
|
||||
// Stage 5: Log Activity
|
||||
await this.db.adminRepo.logActivity(
|
||||
{
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed flyer for ${flyerData.store_name}`,
|
||||
details: { flyer_id: flyer.flyer_id, store_name: flyerData.store_name },
|
||||
userId: job.data.userId,
|
||||
},
|
||||
logger,
|
||||
);
|
||||
|
||||
// Enqueue a job to clean up the original and any generated files.
|
||||
await this.cleanupQueue.add(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: flyer.flyer_id, paths: allFilePaths },
|
||||
{ flyerId, paths: allFilePaths },
|
||||
{ removeOnComplete: true },
|
||||
);
|
||||
logger.info(`Successfully processed job and enqueued cleanup for flyer ID: ${flyer.flyer_id}`);
|
||||
logger.info(`Successfully processed job and enqueued cleanup for flyer ID: ${flyerId}`);
|
||||
|
||||
return { flyerId: flyer.flyer_id };
|
||||
return { flyerId };
|
||||
} catch (error) {
|
||||
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
||||
// Add detailed logging of the raw error object
|
||||
@@ -156,14 +172,52 @@ export class FlyerProcessingService {
|
||||
const logger = globalLogger.child({ jobId: job.id, jobName: job.name, ...job.data });
|
||||
logger.info('Picked up file cleanup job.');
|
||||
|
||||
const { paths } = job.data;
|
||||
if (!paths || paths.length === 0) {
|
||||
logger.warn('Job received no paths to clean. Skipping.');
|
||||
return { status: 'skipped', reason: 'no paths' };
|
||||
const { flyerId, paths } = job.data;
|
||||
let pathsToDelete = paths;
|
||||
|
||||
// If no paths are provided (e.g., from a manual trigger), attempt to derive them from the database.
|
||||
if (!pathsToDelete || pathsToDelete.length === 0) {
|
||||
logger.warn(`Cleanup job for flyer ${flyerId} received no paths. Attempting to derive paths from DB.`);
|
||||
try {
|
||||
const flyer = await db.flyerRepo.getFlyerById(flyerId);
|
||||
const derivedPaths: string[] = [];
|
||||
// This path needs to be configurable and match where multer saves files.
|
||||
const storagePath = process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
|
||||
if (flyer.image_url) {
|
||||
try {
|
||||
const imageName = path.basename(new URL(flyer.image_url).pathname);
|
||||
derivedPaths.push(path.join(storagePath, imageName));
|
||||
} catch (urlError) {
|
||||
logger.error({ err: urlError, url: flyer.image_url }, 'Failed to parse flyer.image_url to derive file path.');
|
||||
}
|
||||
}
|
||||
if (flyer.icon_url) {
|
||||
try {
|
||||
const iconName = path.basename(new URL(flyer.icon_url).pathname);
|
||||
derivedPaths.push(path.join(storagePath, 'icons', iconName));
|
||||
} catch (urlError) {
|
||||
logger.error({ err: urlError, url: flyer.icon_url }, 'Failed to parse flyer.icon_url to derive file path.');
|
||||
}
|
||||
}
|
||||
pathsToDelete = derivedPaths;
|
||||
} catch (error) {
|
||||
if (error instanceof NotFoundError) {
|
||||
logger.error({ flyerId }, 'Cannot derive cleanup paths because flyer was not found in DB.');
|
||||
// Do not throw. Allow the job to be marked as skipped if no paths are found.
|
||||
} else {
|
||||
throw error; // Re-throw other DB errors to allow for retries.
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (!pathsToDelete || pathsToDelete.length === 0) {
|
||||
logger.warn('Job received no paths and could not derive any from the database. Skipping.');
|
||||
return { status: 'skipped', reason: 'no paths derived' };
|
||||
}
|
||||
|
||||
const results = await Promise.allSettled(
|
||||
paths.map(async (filePath) => {
|
||||
pathsToDelete.map(async (filePath) => {
|
||||
try {
|
||||
await this.fs.unlink(filePath);
|
||||
logger.info(`Successfully deleted temporary file: ${filePath}`);
|
||||
@@ -182,12 +236,12 @@ export class FlyerProcessingService {
|
||||
|
||||
const failedDeletions = results.filter((r) => r.status === 'rejected');
|
||||
if (failedDeletions.length > 0) {
|
||||
const failedPaths = paths.filter((_, i) => results[i].status === 'rejected');
|
||||
const failedPaths = pathsToDelete.filter((_, i) => results[i].status === 'rejected');
|
||||
throw new Error(`Failed to delete ${failedDeletions.length} file(s): ${failedPaths.join(', ')}`);
|
||||
}
|
||||
|
||||
logger.info(`Successfully deleted all ${paths.length} temporary files.`);
|
||||
return { status: 'success', deletedCount: paths.length };
|
||||
logger.info(`Successfully deleted all ${pathsToDelete.length} temporary files.`);
|
||||
return { status: 'success', deletedCount: pathsToDelete.length };
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -209,7 +263,8 @@ export class FlyerProcessingService {
|
||||
['PDF_CONVERSION_FAILED', 'Preparing Inputs'],
|
||||
['UNSUPPORTED_FILE_TYPE', 'Preparing Inputs'],
|
||||
['AI_VALIDATION_FAILED', 'Extracting Data with AI'],
|
||||
['TRANSFORMATION_FAILED', 'Transforming AI Data'], // Add new mapping
|
||||
['TRANSFORMATION_FAILED', 'Transforming AI Data'],
|
||||
['DATABASE_ERROR', 'Saving to Database'],
|
||||
]);
|
||||
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
||||
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
||||
@@ -226,15 +281,6 @@ export class FlyerProcessingService {
|
||||
const failedStageName = errorCodeToStageMap.get(errorPayload.errorCode);
|
||||
let errorStageIndex = failedStageName ? stagesToReport.findIndex(s => s.name === failedStageName) : -1;
|
||||
|
||||
// Fallback for generic errors not in the map. This is less robust and relies on string matching.
|
||||
// A future improvement would be to wrap these in specific FlyerProcessingError subclasses.
|
||||
if (errorStageIndex === -1 && errorPayload.message.includes('Icon generation failed')) {
|
||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Transforming AI Data');
|
||||
}
|
||||
if (errorStageIndex === -1 && errorPayload.message.includes('Database transaction failed')) {
|
||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Saving to Database');
|
||||
}
|
||||
|
||||
// 2. If not mapped, find the currently running stage
|
||||
if (errorStageIndex === -1) {
|
||||
errorStageIndex = stagesToReport.findIndex(s => s.status === 'in-progress');
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
// src/services/gamificationService.ts
|
||||
|
||||
import { gamificationRepo } from './db/index.db';
|
||||
import { ForeignKeyConstraintError } from './db/errors.db';
|
||||
import type { Logger } from 'pino';
|
||||
import { ForeignKeyConstraintError } from './db/errors.db';
|
||||
|
||||
class GamificationService {
|
||||
/**
|
||||
@@ -16,8 +16,12 @@ class GamificationService {
|
||||
await gamificationRepo.awardAchievement(userId, achievementName, log);
|
||||
} catch (error) {
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
// This is an expected error (e.g., achievement name doesn't exist),
|
||||
// which the repository layer should have already logged with appropriate context.
|
||||
// We re-throw it so the calling layer (e.g., an admin route) can handle it.
|
||||
throw error;
|
||||
}
|
||||
// For unexpected, generic errors, we log them at the service level before re-throwing.
|
||||
log.error(
|
||||
{ error, userId, achievementName },
|
||||
'Error awarding achievement via admin endpoint:',
|
||||
@@ -45,10 +49,6 @@ class GamificationService {
|
||||
* @param log The logger instance.
|
||||
*/
|
||||
async getLeaderboard(limit: number, log: Logger) {
|
||||
// The test failures point to an issue in the underlying repository method,
|
||||
// where the database query is not being executed. This service method is a simple
|
||||
// pass-through, so the root cause is likely in `gamification.db.ts`.
|
||||
// Adding robust error handling here is a good practice regardless.
|
||||
try {
|
||||
return await gamificationRepo.getLeaderboard(limit, log);
|
||||
} catch (error) {
|
||||
@@ -63,10 +63,6 @@ class GamificationService {
|
||||
* @param log The logger instance.
|
||||
*/
|
||||
async getUserAchievements(userId: string, log: Logger) {
|
||||
// The test failures point to an issue in the underlying repository method,
|
||||
// where the database query is not being executed. This service method is a simple
|
||||
// pass-through, so the root cause is likely in `gamification.db.ts`.
|
||||
// Adding robust error handling here is a good practice regardless.
|
||||
try {
|
||||
return await gamificationRepo.getUserAchievements(userId, log);
|
||||
} catch (error) {
|
||||
|
||||
@@ -35,15 +35,13 @@ vi.mock('./logger.server', () => ({
|
||||
import { logger } from './logger.server';
|
||||
|
||||
describe('Geocoding Service', () => {
|
||||
const originalEnv = process.env;
|
||||
let geocodingService: GeocodingService;
|
||||
let mockGoogleService: GoogleGeocodingService;
|
||||
let mockNominatimService: NominatimGeocodingService;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Restore process.env to its original state before each test
|
||||
process.env = { ...originalEnv };
|
||||
vi.unstubAllEnvs();
|
||||
|
||||
// Create a mock instance of the Google service
|
||||
mockGoogleService = { geocode: vi.fn() } as unknown as GoogleGeocodingService;
|
||||
@@ -53,8 +51,7 @@ describe('Geocoding Service', () => {
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore process.env after each test
|
||||
process.env = originalEnv;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
describe('geocodeAddress', () => {
|
||||
@@ -77,7 +74,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should log an error but continue if Redis GET fails', async () => {
|
||||
// Arrange: Mock Redis 'get' to fail, but Google API to succeed
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockRejectedValue(new Error('Redis down'));
|
||||
vi.mocked(mockGoogleService.geocode).mockResolvedValue(coordinates);
|
||||
|
||||
@@ -95,7 +92,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should proceed to fetch if cached data is invalid JSON', async () => {
|
||||
// Arrange: Mock Redis to return a malformed JSON string
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue('{ "lat": 45.0, "lng": -75.0 '); // Missing closing brace
|
||||
vi.mocked(mockGoogleService.geocode).mockResolvedValue(coordinates);
|
||||
|
||||
@@ -115,7 +112,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should fetch from Google, return coordinates, and cache the result on cache miss', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue(null); // Cache miss
|
||||
vi.mocked(mockGoogleService.geocode).mockResolvedValue(coordinates);
|
||||
|
||||
@@ -136,7 +133,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should fall back to Nominatim if Google API key is missing', async () => {
|
||||
// Arrange
|
||||
delete process.env.GOOGLE_MAPS_API_KEY;
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', '');
|
||||
mocks.mockRedis.get.mockResolvedValue(null);
|
||||
vi.mocked(mockNominatimService.geocode).mockResolvedValue(coordinates);
|
||||
|
||||
@@ -155,7 +152,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should fall back to Nominatim if Google API returns a non-OK status', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue(null);
|
||||
vi.mocked(mockGoogleService.geocode).mockResolvedValue(null); // Google returns no results
|
||||
vi.mocked(mockNominatimService.geocode).mockResolvedValue(coordinates);
|
||||
@@ -174,7 +171,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should fall back to Nominatim if Google API fetch call fails', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue(null);
|
||||
vi.mocked(mockGoogleService.geocode).mockRejectedValue(new Error('Network Error'));
|
||||
vi.mocked(mockNominatimService.geocode).mockResolvedValue(coordinates);
|
||||
@@ -193,7 +190,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should return null and log an error if both Google and Nominatim fail', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue(null);
|
||||
vi.mocked(mockGoogleService.geocode).mockRejectedValue(new Error('Network Error'));
|
||||
vi.mocked(mockNominatimService.geocode).mockResolvedValue(null); // Nominatim also fails
|
||||
@@ -209,7 +206,7 @@ describe('Geocoding Service', () => {
|
||||
|
||||
it('should return coordinates even if Redis SET fails', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-key');
|
||||
mocks.mockRedis.get.mockResolvedValue(null); // Cache miss
|
||||
vi.mocked(mockGoogleService.geocode).mockResolvedValue(coordinates);
|
||||
// Mock Redis 'set' to fail
|
||||
|
||||
@@ -20,25 +20,22 @@ import { logger as mockLogger } from './logger.server';
|
||||
|
||||
describe('Google Geocoding Service', () => {
|
||||
let googleGeocodingService: GoogleGeocodingService;
|
||||
const originalEnv = process.env;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Mock the global fetch function before each test
|
||||
vi.stubGlobal('fetch', vi.fn());
|
||||
// Restore process.env to a clean state for each test
|
||||
process.env = { ...originalEnv };
|
||||
vi.unstubAllEnvs();
|
||||
googleGeocodingService = new GoogleGeocodingService();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore original environment variables after each test
|
||||
process.env = originalEnv;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should return coordinates for a valid address when API key is present', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-api-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-api-key');
|
||||
const mockApiResponse = {
|
||||
status: 'OK',
|
||||
results: [
|
||||
@@ -70,7 +67,7 @@ describe('Google Geocoding Service', () => {
|
||||
|
||||
it('should throw an error if GOOGLE_MAPS_API_KEY is not set', async () => {
|
||||
// Arrange
|
||||
delete process.env.GOOGLE_MAPS_API_KEY;
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', '');
|
||||
|
||||
// Act & Assert
|
||||
await expect(googleGeocodingService.geocode('Any Address', mockLogger)).rejects.toThrow(
|
||||
@@ -81,7 +78,7 @@ describe('Google Geocoding Service', () => {
|
||||
|
||||
it('should return null if the API returns a status other than "OK"', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-api-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-api-key');
|
||||
const mockApiResponse = { status: 'ZERO_RESULTS', results: [] };
|
||||
vi.mocked(fetch).mockResolvedValue({
|
||||
ok: true,
|
||||
@@ -101,7 +98,7 @@ describe('Google Geocoding Service', () => {
|
||||
|
||||
it('should throw an error if the fetch response is not ok', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-api-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-api-key');
|
||||
vi.mocked(fetch).mockResolvedValue({
|
||||
ok: false,
|
||||
status: 403,
|
||||
@@ -119,7 +116,7 @@ describe('Google Geocoding Service', () => {
|
||||
|
||||
it('should throw an error if the fetch call itself fails', async () => {
|
||||
// Arrange
|
||||
process.env.GOOGLE_MAPS_API_KEY = 'test-api-key';
|
||||
vi.stubEnv('GOOGLE_MAPS_API_KEY', 'test-api-key');
|
||||
const networkError = new Error('Network request failed');
|
||||
vi.mocked(fetch).mockRejectedValue(networkError);
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
// src/services/logger.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
|
||||
// Mock pino before importing the logger
|
||||
const pinoMock = vi.fn(() => ({
|
||||
@@ -15,16 +15,21 @@ describe('Server Logger', () => {
|
||||
// Reset modules to ensure process.env changes are applied to new module instances
|
||||
vi.resetModules();
|
||||
vi.clearAllMocks();
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should initialize pino with the correct level for production', async () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
vi.stubEnv('NODE_ENV', 'production');
|
||||
await import('./logger.server');
|
||||
expect(pinoMock).toHaveBeenCalledWith(expect.objectContaining({ level: 'info' }));
|
||||
});
|
||||
|
||||
it('should initialize pino with pretty-print transport for development', async () => {
|
||||
process.env.NODE_ENV = 'development';
|
||||
vi.stubEnv('NODE_ENV', 'development');
|
||||
await import('./logger.server');
|
||||
expect(pinoMock).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ transport: expect.any(Object) }),
|
||||
|
||||
@@ -74,6 +74,19 @@ export class TransformationError extends FlyerProcessingError {
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Error thrown when a database operation fails during processing.
|
||||
*/
|
||||
export class DatabaseError extends FlyerProcessingError {
|
||||
constructor(message: string) {
|
||||
super(
|
||||
message,
|
||||
'DATABASE_ERROR',
|
||||
'A database operation failed. Please try again later.',
|
||||
);
|
||||
}
|
||||
}
|
||||
/**
|
||||
* Error thrown when an image conversion fails (e.g., using sharp).
|
||||
*/
|
||||
|
||||
@@ -4,9 +4,14 @@ import type { Address, UserProfile } from '../types';
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import { ValidationError, NotFoundError } from './db/errors.db';
|
||||
import { DatabaseError } from './processingErrors';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { TokenCleanupJobData } from '../types/job-data';
|
||||
|
||||
// Un-mock the service under test to ensure we are testing the real implementation,
|
||||
// not the global mock from `tests/setup/tests-setup-unit.ts`.
|
||||
vi.unmock('./userService');
|
||||
|
||||
// --- Hoisted Mocks ---
|
||||
const mocks = vi.hoisted(() => {
|
||||
// Create mock implementations for the repository methods we'll be using.
|
||||
@@ -172,6 +177,29 @@ describe('UserService', () => {
|
||||
// 3. Since the address ID did not change, the user profile should NOT be updated.
|
||||
expect(mocks.mockUpdateUserProfile).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should throw a DatabaseError if the transaction fails', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const user = createMockUserProfile({
|
||||
user: { user_id: 'user-123' },
|
||||
address_id: null,
|
||||
});
|
||||
const addressData: Partial<Address> = { address_line_1: '123 Fail St' };
|
||||
const dbError = new Error('DB connection lost');
|
||||
|
||||
// Simulate a failure within the transaction (e.g., upsertAddress fails)
|
||||
mocks.mockUpsertAddress.mockRejectedValue(dbError);
|
||||
|
||||
// Act & Assert
|
||||
// The service should wrap the generic error in a `DatabaseError`.
|
||||
await expect(userService.upsertUserAddress(user, addressData, logger)).rejects.toBeInstanceOf(DatabaseError);
|
||||
|
||||
// Assert that the error was logged correctly
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: user.user.user_id },
|
||||
`Transaction to upsert user address failed: ${dbError.message}`,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('processTokenCleanupJob', () => {
|
||||
@@ -204,7 +232,7 @@ describe('UserService', () => {
|
||||
await expect(userService.processTokenCleanupJob(job)).rejects.toThrow('DB Error');
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ err: error }),
|
||||
'Expired token cleanup job failed.',
|
||||
`Expired token cleanup job failed: ${error.message}`,
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -212,9 +240,12 @@ describe('UserService', () => {
|
||||
describe('updateUserAvatar', () => {
|
||||
it('should construct avatar URL and update profile', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const testBaseUrl = 'http://localhost:3001';
|
||||
vi.stubEnv('FRONTEND_URL', testBaseUrl);
|
||||
|
||||
const userId = 'user-123';
|
||||
const file = { filename: 'avatar.jpg' } as Express.Multer.File;
|
||||
const expectedUrl = '/uploads/avatars/avatar.jpg';
|
||||
const expectedUrl = `${testBaseUrl}/uploads/avatars/avatar.jpg`;
|
||||
|
||||
mocks.mockUpdateUserProfile.mockResolvedValue({} as any);
|
||||
|
||||
@@ -225,6 +256,8 @@ describe('UserService', () => {
|
||||
{ avatar_url: expectedUrl },
|
||||
logger,
|
||||
);
|
||||
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
@@ -7,8 +7,10 @@ import { AddressRepository } from './db/address.db';
|
||||
import { UserRepository } from './db/user.db';
|
||||
import type { Address, Profile, UserProfile } from '../types';
|
||||
import { ValidationError, NotFoundError } from './db/errors.db';
|
||||
import { DatabaseError } from './processingErrors';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import type { TokenCleanupJobData } from '../types/job-data';
|
||||
import { getBaseUrl } from '../utils/serverUtils';
|
||||
|
||||
/**
|
||||
* Encapsulates user-related business logic that may involve multiple repository calls.
|
||||
@@ -27,27 +29,26 @@ class UserService {
|
||||
addressData: Partial<Address>,
|
||||
logger: Logger,
|
||||
): Promise<number> {
|
||||
return db.withTransaction(async (client) => {
|
||||
// Instantiate repositories with the transactional client
|
||||
const addressRepo = new AddressRepository(client);
|
||||
const userRepo = new UserRepository(client);
|
||||
|
||||
const addressId = await addressRepo.upsertAddress(
|
||||
{ ...addressData, address_id: userprofile.address_id ?? undefined },
|
||||
logger,
|
||||
);
|
||||
|
||||
// If the user didn't have an address_id before, update their profile to link it.
|
||||
if (!userprofile.address_id) {
|
||||
await userRepo.updateUserProfile(
|
||||
userprofile.user.user_id,
|
||||
{ address_id: addressId },
|
||||
return db
|
||||
.withTransaction(async (client) => {
|
||||
const addressRepo = new AddressRepository(client);
|
||||
const userRepo = new UserRepository(client);
|
||||
const addressId = await addressRepo.upsertAddress(
|
||||
{ ...addressData, address_id: userprofile.address_id ?? undefined },
|
||||
logger,
|
||||
);
|
||||
}
|
||||
|
||||
return addressId;
|
||||
});
|
||||
if (!userprofile.address_id) {
|
||||
await userRepo.updateUserProfile(userprofile.user.user_id, { address_id: addressId }, logger);
|
||||
}
|
||||
return addressId;
|
||||
})
|
||||
.catch((error) => {
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, userId: userprofile.user.user_id }, `Transaction to upsert user address failed: ${errorMessage}`);
|
||||
// Wrap the original error in a service-level DatabaseError to standardize the error contract,
|
||||
// as this is an unexpected failure within the transaction boundary.
|
||||
throw new DatabaseError(errorMessage);
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -55,27 +56,21 @@ class UserService {
|
||||
* @param job The BullMQ job object.
|
||||
* @returns An object containing the count of deleted tokens.
|
||||
*/
|
||||
async processTokenCleanupJob(
|
||||
job: Job<TokenCleanupJobData>,
|
||||
): Promise<{ deletedCount: number }> {
|
||||
async processTokenCleanupJob(job: Job<TokenCleanupJobData>): Promise<{ deletedCount: number }> {
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
});
|
||||
|
||||
logger.info('Picked up expired token cleanup job.');
|
||||
|
||||
try {
|
||||
const deletedCount = await db.userRepo.deleteExpiredResetTokens(logger);
|
||||
logger.info(`Successfully deleted ${deletedCount} expired tokens.`);
|
||||
return { deletedCount };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{ err: wrappedError, attemptsMade: job.attemptsMade },
|
||||
'Expired token cleanup job failed.',
|
||||
);
|
||||
throw wrappedError;
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, attemptsMade: job.attemptsMade }, `Expired token cleanup job failed: ${errorMessage}`);
|
||||
// This is a background job, but wrapping in a standard error type is good practice.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -87,12 +82,20 @@ class UserService {
|
||||
* @returns The updated user profile.
|
||||
*/
|
||||
async updateUserAvatar(userId: string, file: Express.Multer.File, logger: Logger): Promise<Profile> {
|
||||
const avatarUrl = `/uploads/avatars/${file.filename}`;
|
||||
return db.userRepo.updateUserProfile(
|
||||
userId,
|
||||
{ avatar_url: avatarUrl },
|
||||
logger,
|
||||
);
|
||||
try {
|
||||
const baseUrl = getBaseUrl(logger);
|
||||
const avatarUrl = `${baseUrl}/uploads/avatars/${file.filename}`;
|
||||
return await db.userRepo.updateUserProfile(userId, { avatar_url: avatarUrl }, logger);
|
||||
} catch (error) {
|
||||
// Re-throw known application errors without logging them as system errors.
|
||||
if (error instanceof NotFoundError) {
|
||||
throw error;
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, userId }, `Failed to update user avatar: ${errorMessage}`);
|
||||
// Wrap unexpected errors.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
/**
|
||||
* Updates a user's password after hashing it.
|
||||
@@ -101,9 +104,16 @@ class UserService {
|
||||
* @param logger The logger instance.
|
||||
*/
|
||||
async updateUserPassword(userId: string, newPassword: string, logger: Logger): Promise<void> {
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||
await db.userRepo.updateUserPassword(userId, hashedPassword, logger);
|
||||
try {
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||
await db.userRepo.updateUserPassword(userId, hashedPassword, logger);
|
||||
} catch (error) {
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, userId }, `Failed to update user password: ${errorMessage}`);
|
||||
// Wrap unexpected errors.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -113,19 +123,25 @@ class UserService {
|
||||
* @param logger The logger instance.
|
||||
*/
|
||||
async deleteUserAccount(userId: string, password: string, logger: Logger): Promise<void> {
|
||||
const userWithHash = await db.userRepo.findUserWithPasswordHashById(userId, logger);
|
||||
if (!userWithHash || !userWithHash.password_hash) {
|
||||
// This case should be rare for a logged-in user but is a good safeguard.
|
||||
throw new NotFoundError('User not found or password not set.');
|
||||
try {
|
||||
const userWithHash = await db.userRepo.findUserWithPasswordHashById(userId, logger);
|
||||
if (!userWithHash || !userWithHash.password_hash) {
|
||||
throw new NotFoundError('User not found or password not set.');
|
||||
}
|
||||
const isMatch = await bcrypt.compare(password, userWithHash.password_hash);
|
||||
if (!isMatch) {
|
||||
throw new ValidationError([], 'Incorrect password.');
|
||||
}
|
||||
await db.userRepo.deleteUserById(userId, logger);
|
||||
} catch (error) {
|
||||
if (error instanceof NotFoundError || error instanceof ValidationError) {
|
||||
throw error;
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, userId }, `Failed to delete user account: ${errorMessage}`);
|
||||
// Wrap unexpected errors.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
|
||||
const isMatch = await bcrypt.compare(password, userWithHash.password_hash);
|
||||
if (!isMatch) {
|
||||
// Use ValidationError for a 400-level response in the route
|
||||
throw new ValidationError([], 'Incorrect password.');
|
||||
}
|
||||
|
||||
await db.userRepo.deleteUserById(userId, logger);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -135,18 +151,21 @@ class UserService {
|
||||
* @param logger The logger instance.
|
||||
* @returns The address object.
|
||||
*/
|
||||
async getUserAddress(
|
||||
userProfile: UserProfile,
|
||||
addressId: number,
|
||||
logger: Logger,
|
||||
): Promise<Address> {
|
||||
// Security check: Ensure the requested addressId matches the one on the user's profile.
|
||||
async getUserAddress(userProfile: UserProfile, addressId: number, logger: Logger): Promise<Address> {
|
||||
if (userProfile.address_id !== addressId) {
|
||||
// Use ValidationError to trigger a 403 Forbidden response in the route handler.
|
||||
throw new ValidationError([], 'Forbidden: You can only access your own address.');
|
||||
}
|
||||
// The repo method will throw a NotFoundError if the address doesn't exist.
|
||||
return db.addressRepo.getAddressById(addressId, logger);
|
||||
try {
|
||||
return await db.addressRepo.getAddressById(addressId, logger);
|
||||
} catch (error) {
|
||||
if (error instanceof NotFoundError) {
|
||||
throw error;
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
logger.error({ err: error, userId: userProfile.user.user_id, addressId }, `Failed to get user address: ${errorMessage}`);
|
||||
// Wrap unexpected errors.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -160,7 +179,17 @@ class UserService {
|
||||
if (deleterId === userToDeleteId) {
|
||||
throw new ValidationError([], 'Admins cannot delete their own account.');
|
||||
}
|
||||
await db.userRepo.deleteUserById(userToDeleteId, log);
|
||||
try {
|
||||
await db.userRepo.deleteUserById(userToDeleteId, log);
|
||||
} catch (error) {
|
||||
if (error instanceof ValidationError) {
|
||||
throw error;
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred.';
|
||||
log.error({ err: error, deleterId, userToDeleteId }, `Admin failed to delete user account: ${errorMessage}`);
|
||||
// Wrap unexpected errors.
|
||||
throw new DatabaseError(errorMessage);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -44,7 +44,6 @@ const fsAdapter: IFileSystem = {
|
||||
const flyerProcessingService = new FlyerProcessingService(
|
||||
new FlyerFileHandler(fsAdapter, execAsync),
|
||||
new FlyerAiProcessor(aiService, db.personalizationRepo),
|
||||
db,
|
||||
fsAdapter,
|
||||
cleanupQueue,
|
||||
new FlyerDataTransformer(),
|
||||
@@ -92,6 +91,8 @@ export const flyerWorker = new Worker<FlyerJobData>(
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WORKER_CONCURRENCY || '1', 10),
|
||||
// Increase lock duration to prevent jobs from being re-processed prematurely.
|
||||
lockDuration: parseInt(process.env.WORKER_LOCK_DURATION || '30000', 10),
|
||||
},
|
||||
);
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
// src/tests/e2e/auth.e2e.test.ts
|
||||
import { describe, it, expect, afterAll, beforeAll } from 'vitest';
|
||||
import { describe, it, expect, afterAll, beforeAll } from 'vitest';
|
||||
import * as apiClient from '../../services/apiClient';
|
||||
import { cleanupDb } from '../utils/cleanup';
|
||||
import { createAndLoginUser, TEST_PASSWORD } from '../utils/testHelpers';
|
||||
@@ -13,15 +13,19 @@ describe('Authentication E2E Flow', () => {
|
||||
let testUser: UserProfile;
|
||||
const createdUserIds: string[] = [];
|
||||
|
||||
beforeAll(async () => {
|
||||
beforeAll(async () => {
|
||||
// Create a user that can be used for login-related tests in this suite.
|
||||
const { user } = await createAndLoginUser({
|
||||
email: `e2e-login-user-${Date.now()}@example.com`,
|
||||
fullName: 'E2E Login User',
|
||||
// E2E tests use apiClient which doesn't need the `request` object.
|
||||
});
|
||||
testUser = user;
|
||||
createdUserIds.push(user.user.user_id);
|
||||
try {
|
||||
const { user } = await createAndLoginUser({
|
||||
email: `e2e-login-user-${Date.now()}@example.com`,
|
||||
fullName: 'E2E Login User',
|
||||
});
|
||||
testUser = user;
|
||||
createdUserIds.push(user.user.user_id);
|
||||
} catch (error) {
|
||||
console.error('[FATAL] Setup failed. DB might be down.', error);
|
||||
throw error;
|
||||
}
|
||||
});
|
||||
|
||||
afterAll(async () => {
|
||||
@@ -70,7 +74,7 @@ describe('Authentication E2E Flow', () => {
|
||||
const firstResponse = await apiClient.registerUser(email, TEST_PASSWORD, 'Duplicate User');
|
||||
const firstData = await firstResponse.json();
|
||||
expect(firstResponse.status).toBe(201);
|
||||
createdUserIds.push(firstData.userprofile.user.user_id); // Add for cleanup
|
||||
createdUserIds.push(firstData.userprofile.user.user_id);
|
||||
|
||||
// Act 2: Attempt to register the same user again
|
||||
const secondResponse = await apiClient.registerUser(email, TEST_PASSWORD, 'Duplicate User');
|
||||
@@ -174,15 +178,35 @@ describe('Authentication E2E Flow', () => {
|
||||
expect(registerResponse.status).toBe(201);
|
||||
createdUserIds.push(registerData.userprofile.user.user_id);
|
||||
|
||||
// Act 1: Request a password reset.
|
||||
// The test environment returns the token directly in the response for E2E testing.
|
||||
// Instead of a fixed delay, poll by attempting to log in. This is more robust
|
||||
// and confirms the user record is committed and readable by subsequent transactions.
|
||||
let loginSuccess = false;
|
||||
for (let i = 0; i < 10; i++) {
|
||||
// Poll for up to 10 seconds
|
||||
const loginResponse = await apiClient.loginUser(email, TEST_PASSWORD, false);
|
||||
if (loginResponse.ok) {
|
||||
loginSuccess = true;
|
||||
break;
|
||||
}
|
||||
await new Promise((resolve) => setTimeout(resolve, 1000));
|
||||
}
|
||||
expect(loginSuccess, 'User should be able to log in after registration. DB might be lagging.').toBe(true);
|
||||
|
||||
// Act 1: Request a password reset
|
||||
const forgotResponse = await apiClient.requestPasswordReset(email);
|
||||
const forgotData = await forgotResponse.json();
|
||||
const resetToken = forgotData.token;
|
||||
|
||||
// --- DEBUG SECTION FOR FAILURE ---
|
||||
if (!resetToken) {
|
||||
console.error(' [DEBUG FAILURE] Token missing in response:', JSON.stringify(forgotData, null, 2));
|
||||
console.error(' [DEBUG FAILURE] This usually means the backend hit a DB error or is not in NODE_ENV=test mode.');
|
||||
}
|
||||
// ---------------------------------
|
||||
|
||||
// Assert 1: Check that we received a token.
|
||||
expect(forgotResponse.status).toBe(200);
|
||||
expect(resetToken).toBeDefined();
|
||||
expect(resetToken, 'Backend returned 200 but no token. Check backend logs for "Connection terminated" errors.').toBeDefined();
|
||||
expect(resetToken).toBeTypeOf('string');
|
||||
|
||||
// Act 2: Use the token to set a new password.
|
||||
@@ -194,7 +218,7 @@ describe('Authentication E2E Flow', () => {
|
||||
expect(resetResponse.status).toBe(200);
|
||||
expect(resetData.message).toBe('Password has been reset successfully.');
|
||||
|
||||
// Act 3 & Assert 3 (Verification): Log in with the NEW password to confirm the change.
|
||||
// Act 3: Log in with the NEW password
|
||||
const loginResponse = await apiClient.loginUser(email, newPassword, false);
|
||||
const loginData = await loginResponse.json();
|
||||
|
||||
|
||||
@@ -89,7 +89,7 @@ describe('E2E Flyer Upload and Processing Workflow', () => {
|
||||
|
||||
// 5. Poll for job completion
|
||||
let jobStatus;
|
||||
const maxRetries = 30; // Poll for up to 90 seconds
|
||||
const maxRetries = 60; // Poll for up to 180 seconds
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000)); // Wait 3s
|
||||
|
||||
@@ -106,5 +106,5 @@ describe('E2E Flyer Upload and Processing Workflow', () => {
|
||||
expect(jobStatus.state).toBe('completed');
|
||||
flyerId = jobStatus.returnValue?.flyerId;
|
||||
expect(flyerId).toBeTypeOf('number');
|
||||
}, 120000); // Extended timeout for AI processing
|
||||
}, 240000); // Extended timeout for AI processing
|
||||
});
|
||||
@@ -163,8 +163,8 @@ describe('Admin API Routes Integration Tests', () => {
|
||||
// Before each modification test, create a fresh flyer item and a correction for it.
|
||||
beforeEach(async () => {
|
||||
const flyerRes = await getPool().query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||
VALUES ($1, 'admin-test.jpg', 'https://example.com/flyer-images/asdmin-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum)
|
||||
VALUES ($1, 'admin-test.jpg', 'https://example.com/flyer-images/asdmin-test.jpg', 'https://example.com/flyer-images/icons/admin-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||
// The checksum must be a unique 64-character string to satisfy the DB constraint.
|
||||
// We generate a dynamic string and pad it to 64 characters.
|
||||
[testStoreId, `checksum-${Date.now()}-${Math.random()}`.padEnd(64, '0')],
|
||||
|
||||
@@ -1,48 +1,59 @@
|
||||
// src/tests/integration/db.integration.test.ts
|
||||
import { describe, it, expect } from 'vitest';
|
||||
import { describe, it, expect, beforeEach, afterEach } from 'vitest';
|
||||
import * as db from '../../services/db/index.db';
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import { getPool } from '../../services/db/connection.db';
|
||||
import { logger } from '../../services/logger.server';
|
||||
import type { UserProfile } from '../../types';
|
||||
import { cleanupDb } from '../utils/cleanup';
|
||||
|
||||
describe('Database Service Integration Tests', () => {
|
||||
it('should create a new user and be able to find them by email', async ({ onTestFinished }) => {
|
||||
let testUser: UserProfile;
|
||||
let testUserEmail: string;
|
||||
|
||||
beforeEach(async () => {
|
||||
// Arrange: Use a unique email for each test run to ensure isolation.
|
||||
const email = `test.user-${Date.now()}@example.com`;
|
||||
testUserEmail = `test.user-${Date.now()}@example.com`;
|
||||
const password = 'password123';
|
||||
const fullName = 'Test User';
|
||||
const saltRounds = 10;
|
||||
const passwordHash = await bcrypt.hash(password, saltRounds);
|
||||
|
||||
// Ensure the created user is cleaned up after this specific test finishes.
|
||||
onTestFinished(async () => {
|
||||
await getPool().query('DELETE FROM public.users WHERE email = $1', [email]);
|
||||
});
|
||||
|
||||
// Act: Call the createUser function
|
||||
const createdUser = await db.userRepo.createUser(
|
||||
email,
|
||||
testUser = await db.userRepo.createUser(
|
||||
testUserEmail,
|
||||
passwordHash,
|
||||
{ full_name: fullName },
|
||||
logger,
|
||||
);
|
||||
});
|
||||
|
||||
afterEach(async () => {
|
||||
// Ensure the created user is cleaned up after each test.
|
||||
if (testUser?.user.user_id) {
|
||||
await cleanupDb({ userIds: [testUser.user.user_id] });
|
||||
}
|
||||
});
|
||||
|
||||
it('should create a new user and have a corresponding profile', async () => {
|
||||
// Assert: Check that the user was created with the correct details
|
||||
expect(createdUser).toBeDefined();
|
||||
expect(createdUser.user.email).toBe(email); // This is correct
|
||||
expect(createdUser.user.user_id).toBeTypeOf('string');
|
||||
expect(testUser).toBeDefined();
|
||||
expect(testUser.user.email).toBe(testUserEmail);
|
||||
expect(testUser.user.user_id).toBeTypeOf('string');
|
||||
|
||||
// Also, verify the profile was created by the trigger
|
||||
const profile = await db.userRepo.findUserProfileById(testUser.user.user_id, logger);
|
||||
expect(profile).toBeDefined();
|
||||
expect(profile?.full_name).toBe('Test User');
|
||||
});
|
||||
|
||||
it('should be able to find the created user by email', async () => {
|
||||
// Act: Try to find the user we just created
|
||||
const foundUser = await db.userRepo.findUserByEmail(email, logger);
|
||||
const foundUser = await db.userRepo.findUserByEmail(testUserEmail, logger);
|
||||
|
||||
// Assert: Check that the found user matches the created user
|
||||
expect(foundUser).toBeDefined();
|
||||
expect(foundUser?.user_id).toBe(createdUser.user.user_id);
|
||||
expect(foundUser?.email).toBe(email);
|
||||
|
||||
// Also, verify the profile was created by the trigger
|
||||
const profile = await db.userRepo.findUserProfileById(createdUser.user.user_id, logger);
|
||||
expect(profile).toBeDefined();
|
||||
expect(profile?.full_name).toBe(fullName);
|
||||
expect(foundUser?.user_id).toBe(testUser.user.user_id);
|
||||
expect(foundUser?.email).toBe(testUserEmail);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -15,6 +15,7 @@ import { cleanupFiles } from '../utils/cleanupFiles';
|
||||
import piexif from 'piexifjs';
|
||||
import exifParser from 'exif-parser';
|
||||
import sharp from 'sharp';
|
||||
import { createFlyerAndItems } from '../../services/db/flyer.db';
|
||||
|
||||
|
||||
/**
|
||||
@@ -23,8 +24,30 @@ import sharp from 'sharp';
|
||||
|
||||
const request = supertest(app);
|
||||
|
||||
// Import the mocked service to control its behavior in tests.
|
||||
import { aiService } from '../../services/aiService.server';
|
||||
const { mockExtractCoreData } = vi.hoisted(() => ({
|
||||
mockExtractCoreData: vi.fn(),
|
||||
}));
|
||||
|
||||
// Mock the AI service to prevent real API calls during integration tests.
|
||||
// This is crucial for making the tests reliable and fast. We don't want to
|
||||
// depend on the external Gemini API.
|
||||
vi.mock('../../services/aiService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../../services/aiService.server')>();
|
||||
// To preserve the class instance methods of `aiService`, we must modify the
|
||||
// instance directly rather than creating a new plain object with spread syntax.
|
||||
actual.aiService.extractCoreDataFromFlyerImage = mockExtractCoreData;
|
||||
return actual;
|
||||
});
|
||||
|
||||
// Mock the database service to allow for simulating DB failures.
|
||||
// By default, it will use the real implementation.
|
||||
vi.mock('../../services/db/flyer.db', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../../services/db/flyer.db')>();
|
||||
return {
|
||||
...actual,
|
||||
createFlyerAndItems: vi.fn().mockImplementation(actual.createFlyerAndItems),
|
||||
};
|
||||
});
|
||||
|
||||
describe('Flyer Processing Background Job Integration Test', () => {
|
||||
const createdUserIds: string[] = [];
|
||||
@@ -32,23 +55,21 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
const createdFilePaths: string[] = [];
|
||||
|
||||
beforeAll(async () => {
|
||||
// This setup is now simpler as the worker handles fetching master items.
|
||||
// Setup default mock response for AI service
|
||||
const mockItems: ExtractedFlyerItem[] = [
|
||||
{
|
||||
item: 'Mocked Integration Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: 'each',
|
||||
category_name: 'Mock Category',
|
||||
},
|
||||
];
|
||||
vi.spyOn(aiService, 'extractCoreDataFromFlyerImage').mockResolvedValue({
|
||||
// Setup default mock response for the AI service's extractCoreDataFromFlyerImage method.
|
||||
mockExtractCoreData.mockResolvedValue({
|
||||
store_name: 'Mock Store',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
items: mockItems,
|
||||
items: [
|
||||
{
|
||||
item: 'Mocked Integration Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: 'each',
|
||||
category_name: 'Mock Category',
|
||||
},
|
||||
],
|
||||
});
|
||||
});
|
||||
|
||||
@@ -101,7 +122,9 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
|
||||
// Act 2: Poll for the job status until it completes.
|
||||
let jobStatus;
|
||||
const maxRetries = 60; // Poll for up to 180 seconds (60 * 3s)
|
||||
// Poll for up to 210 seconds (70 * 3s). This should be greater than the worker's
|
||||
// lockDuration (120s) to patiently wait for long-running jobs.
|
||||
const maxRetries = 70;
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
console.log(`Polling attempt ${i + 1}...`);
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000)); // Wait 3 seconds between polls
|
||||
@@ -163,11 +186,6 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
});
|
||||
createdUserIds.push(authUser.user.user_id); // Track for cleanup
|
||||
|
||||
// Use a cleanup function to delete the user even if the test fails.
|
||||
onTestFinished(async () => {
|
||||
await getPool().query('DELETE FROM public.users WHERE user_id = $1', [authUser.user.user_id]);
|
||||
});
|
||||
|
||||
// Act & Assert
|
||||
await runBackgroundProcessingTest(authUser, token);
|
||||
}, 240000); // Increase timeout to 240 seconds for this long-running test
|
||||
@@ -223,7 +241,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
|
||||
// Poll for job completion
|
||||
let jobStatus;
|
||||
const maxRetries = 30; // Poll for up to 90 seconds
|
||||
const maxRetries = 60; // Poll for up to 180 seconds
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request
|
||||
@@ -240,7 +258,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
console.error('[DEBUG] EXIF test job failed:', jobStatus.failedReason);
|
||||
}
|
||||
expect(jobStatus?.state).toBe('completed');
|
||||
const flyerId = jobStatus?.data?.flyerId;
|
||||
const flyerId = jobStatus?.returnValue?.flyerId;
|
||||
expect(flyerId).toBeTypeOf('number');
|
||||
createdFlyerIds.push(flyerId);
|
||||
|
||||
@@ -309,7 +327,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
|
||||
// Poll for job completion
|
||||
let jobStatus;
|
||||
const maxRetries = 30;
|
||||
const maxRetries = 60; // Poll for up to 180 seconds
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request
|
||||
@@ -326,7 +344,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
console.error('[DEBUG] PNG metadata test job failed:', jobStatus.failedReason);
|
||||
}
|
||||
expect(jobStatus?.state).toBe('completed');
|
||||
const flyerId = jobStatus?.data?.flyerId;
|
||||
const flyerId = jobStatus?.returnValue?.flyerId;
|
||||
expect(flyerId).toBeTypeOf('number');
|
||||
createdFlyerIds.push(flyerId);
|
||||
|
||||
@@ -345,4 +363,162 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
||||
},
|
||||
240000,
|
||||
);
|
||||
|
||||
it(
|
||||
'should handle a failure from the AI service gracefully',
|
||||
async () => {
|
||||
// Arrange: Mock the AI service to throw an error for this specific test.
|
||||
const aiError = new Error('AI model failed to extract data.');
|
||||
mockExtractCoreData.mockRejectedValueOnce(aiError);
|
||||
|
||||
// Arrange: Prepare a unique flyer file for upload.
|
||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||
const imageBuffer = await fs.readFile(imagePath);
|
||||
const uniqueContent = Buffer.concat([imageBuffer, Buffer.from(`fail-test-${Date.now()}`)]);
|
||||
const uniqueFileName = `ai-fail-test-${Date.now()}.jpg`;
|
||||
const mockImageFile = new File([uniqueContent], uniqueFileName, { type: 'image/jpeg' });
|
||||
const checksum = await generateFileChecksum(mockImageFile);
|
||||
|
||||
// Track created files for cleanup
|
||||
const uploadDir = path.resolve(__dirname, '../../../flyer-images');
|
||||
createdFilePaths.push(path.join(uploadDir, uniqueFileName));
|
||||
|
||||
// Act 1: Upload the file to start the background job.
|
||||
const uploadResponse = await request
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', checksum)
|
||||
.attach('flyerFile', uniqueContent, uniqueFileName);
|
||||
|
||||
const { jobId } = uploadResponse.body;
|
||||
expect(jobId).toBeTypeOf('string');
|
||||
|
||||
// Act 2: Poll for the job status until it completes or fails.
|
||||
let jobStatus;
|
||||
const maxRetries = 60;
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request.get(`/api/ai/jobs/${jobId}/status`);
|
||||
jobStatus = statusResponse.body;
|
||||
if (jobStatus.state === 'completed' || jobStatus.state === 'failed') {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
// Assert 1: Check that the job failed.
|
||||
expect(jobStatus?.state).toBe('failed');
|
||||
expect(jobStatus?.failedReason).toContain('AI model failed to extract data.');
|
||||
|
||||
// Assert 2: Verify the flyer was NOT saved in the database.
|
||||
const savedFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||
expect(savedFlyer).toBeUndefined();
|
||||
},
|
||||
240000,
|
||||
);
|
||||
|
||||
it(
|
||||
'should handle a database failure during flyer creation',
|
||||
async () => {
|
||||
// Arrange: Mock the database creation function to throw an error for this specific test.
|
||||
const dbError = new Error('DB transaction failed');
|
||||
vi.mocked(createFlyerAndItems).mockRejectedValueOnce(dbError);
|
||||
|
||||
// Arrange: Prepare a unique flyer file for upload.
|
||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||
const imageBuffer = await fs.readFile(imagePath);
|
||||
const uniqueContent = Buffer.concat([imageBuffer, Buffer.from(`db-fail-test-${Date.now()}`)]);
|
||||
const uniqueFileName = `db-fail-test-${Date.now()}.jpg`;
|
||||
const mockImageFile = new File([uniqueContent], uniqueFileName, { type: 'image/jpeg' });
|
||||
const checksum = await generateFileChecksum(mockImageFile);
|
||||
|
||||
// Track created files for cleanup
|
||||
const uploadDir = path.resolve(__dirname, '../../../flyer-images');
|
||||
createdFilePaths.push(path.join(uploadDir, uniqueFileName));
|
||||
|
||||
// Act 1: Upload the file to start the background job.
|
||||
const uploadResponse = await request
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', checksum)
|
||||
.attach('flyerFile', uniqueContent, uniqueFileName);
|
||||
|
||||
const { jobId } = uploadResponse.body;
|
||||
expect(jobId).toBeTypeOf('string');
|
||||
|
||||
// Act 2: Poll for the job status until it completes or fails.
|
||||
let jobStatus;
|
||||
const maxRetries = 60;
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request.get(`/api/ai/jobs/${jobId}/status`);
|
||||
jobStatus = statusResponse.body;
|
||||
if (jobStatus.state === 'completed' || jobStatus.state === 'failed') {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
// Assert 1: Check that the job failed.
|
||||
expect(jobStatus?.state).toBe('failed');
|
||||
expect(jobStatus?.failedReason).toContain('DB transaction failed');
|
||||
|
||||
// Assert 2: Verify the flyer was NOT saved in the database.
|
||||
const savedFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||
expect(savedFlyer).toBeUndefined();
|
||||
},
|
||||
240000,
|
||||
);
|
||||
|
||||
it(
|
||||
'should NOT clean up temporary files when a job fails, to allow for manual inspection',
|
||||
async () => {
|
||||
// Arrange: Mock the AI service to throw an error, causing the job to fail.
|
||||
const aiError = new Error('Simulated AI failure for cleanup test.');
|
||||
mockExtractCoreData.mockRejectedValueOnce(aiError);
|
||||
|
||||
// Arrange: Prepare a unique flyer file for upload.
|
||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||
const imageBuffer = await fs.readFile(imagePath);
|
||||
const uniqueContent = Buffer.concat([
|
||||
imageBuffer,
|
||||
Buffer.from(`cleanup-fail-test-${Date.now()}`),
|
||||
]);
|
||||
const uniqueFileName = `cleanup-fail-test-${Date.now()}.jpg`;
|
||||
const mockImageFile = new File([uniqueContent], uniqueFileName, { type: 'image/jpeg' });
|
||||
const checksum = await generateFileChecksum(mockImageFile);
|
||||
|
||||
// Track the path of the file that will be created in the uploads directory.
|
||||
const uploadDir = path.resolve(__dirname, '../../../flyer-images');
|
||||
const tempFilePath = path.join(uploadDir, uniqueFileName);
|
||||
createdFilePaths.push(tempFilePath);
|
||||
|
||||
// Act 1: Upload the file to start the background job.
|
||||
const uploadResponse = await request
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', checksum)
|
||||
.attach('flyerFile', uniqueContent, uniqueFileName);
|
||||
|
||||
const { jobId } = uploadResponse.body;
|
||||
expect(jobId).toBeTypeOf('string');
|
||||
|
||||
// Act 2: Poll for the job status until it fails.
|
||||
let jobStatus;
|
||||
const maxRetries = 60;
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request.get(`/api/ai/jobs/${jobId}/status`);
|
||||
jobStatus = statusResponse.body;
|
||||
if (jobStatus.state === 'failed') {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
// Assert 1: Check that the job actually failed.
|
||||
expect(jobStatus?.state).toBe('failed');
|
||||
expect(jobStatus?.failedReason).toContain('Simulated AI failure for cleanup test.');
|
||||
|
||||
// Assert 2: Verify the temporary file was NOT deleted.
|
||||
// We check for its existence. If it doesn't exist, fs.access will throw an error.
|
||||
await expect(fs.access(tempFilePath), 'Expected temporary file to exist after job failure, but it was deleted.');
|
||||
},
|
||||
240000,
|
||||
);
|
||||
|
||||
});
|
||||
|
||||
@@ -1,9 +1,10 @@
|
||||
// src/tests/integration/flyer.integration.test.ts
|
||||
import { describe, it, expect, beforeAll } from 'vitest';
|
||||
import { describe, it, expect, beforeAll, afterAll } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import { getPool } from '../../services/db/connection.db';
|
||||
import app from '../../../server';
|
||||
import type { Flyer, FlyerItem } from '../../types';
|
||||
import { cleanupDb } from '../utils/cleanup';
|
||||
|
||||
/**
|
||||
* @vitest-environment node
|
||||
@@ -13,6 +14,7 @@ describe('Public Flyer API Routes Integration Tests', () => {
|
||||
let flyers: Flyer[] = [];
|
||||
// Use a supertest instance for all requests in this file
|
||||
const request = supertest(app);
|
||||
let testStoreId: number;
|
||||
let createdFlyerId: number;
|
||||
|
||||
// Fetch flyers once before all tests in this suite to use in subsequent tests.
|
||||
@@ -21,12 +23,12 @@ describe('Public Flyer API Routes Integration Tests', () => {
|
||||
const storeRes = await getPool().query(
|
||||
`INSERT INTO public.stores (name) VALUES ('Integration Test Store') RETURNING store_id`,
|
||||
);
|
||||
const storeId = storeRes.rows[0].store_id;
|
||||
testStoreId = storeRes.rows[0].store_id;
|
||||
|
||||
const flyerRes = await getPool().query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||
VALUES ($1, 'integration-test.jpg', 'https://example.com/flyer-images/integration-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||
[storeId, `${Date.now().toString(16)}`.padEnd(64, '0')],
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum)
|
||||
VALUES ($1, 'integration-test.jpg', 'https://example.com/flyer-images/integration-test.jpg', 'https://example.com/flyer-images/icons/integration-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||
[testStoreId, `${Date.now().toString(16)}`.padEnd(64, '0')],
|
||||
);
|
||||
createdFlyerId = flyerRes.rows[0].flyer_id;
|
||||
|
||||
@@ -41,6 +43,14 @@ describe('Public Flyer API Routes Integration Tests', () => {
|
||||
flyers = response.body;
|
||||
});
|
||||
|
||||
afterAll(async () => {
|
||||
// Clean up the test data created in beforeAll to prevent polluting the test database.
|
||||
await cleanupDb({
|
||||
flyerIds: [createdFlyerId],
|
||||
storeIds: [testStoreId],
|
||||
});
|
||||
});
|
||||
|
||||
describe('GET /api/flyers', () => {
|
||||
it('should return a list of flyers', async () => {
|
||||
// Act: Call the API endpoint using the client function.
|
||||
|
||||
@@ -4,11 +4,13 @@ import supertest from 'supertest';
|
||||
import app from '../../../server';
|
||||
import path from 'path';
|
||||
import fs from 'node:fs/promises';
|
||||
import { getPool } from '../../services/db/connection.db';
|
||||
import { createAndLoginUser } from '../utils/testHelpers';
|
||||
import { generateFileChecksum } from '../../utils/checksum';
|
||||
import * as db from '../../services/db/index.db';
|
||||
import { cleanupDb } from '../utils/cleanup';
|
||||
import { logger } from '../../services/logger.server';
|
||||
import * as imageProcessor from '../../utils/imageProcessor';
|
||||
import type {
|
||||
UserProfile,
|
||||
UserAchievement,
|
||||
@@ -16,6 +18,7 @@ import type {
|
||||
Achievement,
|
||||
ExtractedFlyerItem,
|
||||
} from '../../types';
|
||||
import type { Flyer } from '../../types';
|
||||
import { cleanupFiles } from '../utils/cleanupFiles';
|
||||
|
||||
/**
|
||||
@@ -24,14 +27,36 @@ import { cleanupFiles } from '../utils/cleanupFiles';
|
||||
|
||||
const request = supertest(app);
|
||||
|
||||
// Import the mocked service to control its behavior in tests.
|
||||
import { aiService } from '../../services/aiService.server';
|
||||
const { mockExtractCoreData } = vi.hoisted(() => ({
|
||||
mockExtractCoreData: vi.fn(),
|
||||
}));
|
||||
|
||||
// Mock the AI service to prevent real API calls during integration tests.
|
||||
// This is crucial for making the tests reliable and fast. We don't want to
|
||||
// depend on the external Gemini API.
|
||||
vi.mock('../../services/aiService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../../services/aiService.server')>();
|
||||
// To preserve the class instance methods of `aiService`, we must modify the
|
||||
// instance directly rather than creating a new plain object with spread syntax.
|
||||
actual.aiService.extractCoreDataFromFlyerImage = mockExtractCoreData;
|
||||
return actual;
|
||||
});
|
||||
|
||||
// Mock the image processor to control icon generation for legacy uploads
|
||||
vi.mock('../../utils/imageProcessor', async () => {
|
||||
const actual = await vi.importActual<typeof imageProcessor>('../../utils/imageProcessor');
|
||||
return {
|
||||
...actual,
|
||||
generateFlyerIcon: vi.fn(),
|
||||
};
|
||||
});
|
||||
|
||||
describe('Gamification Flow Integration Test', () => {
|
||||
let testUser: UserProfile;
|
||||
let authToken: string;
|
||||
const createdFlyerIds: number[] = [];
|
||||
const createdFilePaths: string[] = [];
|
||||
const createdStoreIds: number[] = [];
|
||||
|
||||
beforeAll(async () => {
|
||||
// Create a new user specifically for this test suite to ensure a clean slate.
|
||||
@@ -41,26 +66,21 @@ describe('Gamification Flow Integration Test', () => {
|
||||
request,
|
||||
}));
|
||||
|
||||
// Mock the AI service's method to prevent actual API calls during integration tests.
|
||||
// This is crucial for making the integration test reliable. We don't want to
|
||||
// depend on the external Gemini API, which has quotas and can be slow.
|
||||
// By mocking this, we test our application's internal flow:
|
||||
// API -> Queue -> Worker -> DB -> Gamification Logic
|
||||
const mockExtractedItems: ExtractedFlyerItem[] = [
|
||||
{
|
||||
item: 'Integration Test Milk',
|
||||
price_display: '$4.99',
|
||||
price_in_cents: 499,
|
||||
quantity: '2L',
|
||||
category_name: 'Dairy',
|
||||
},
|
||||
];
|
||||
vi.spyOn(aiService, 'extractCoreDataFromFlyerImage').mockResolvedValue({
|
||||
// Setup default mock response for the AI service's extractCoreDataFromFlyerImage method.
|
||||
mockExtractCoreData.mockResolvedValue({
|
||||
store_name: 'Gamification Test Store',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
items: mockExtractedItems,
|
||||
items: [
|
||||
{
|
||||
item: 'Integration Test Milk',
|
||||
price_display: '$4.99',
|
||||
price_in_cents: 499,
|
||||
quantity: '2L',
|
||||
category_name: 'Dairy',
|
||||
},
|
||||
],
|
||||
});
|
||||
});
|
||||
|
||||
@@ -68,6 +88,7 @@ describe('Gamification Flow Integration Test', () => {
|
||||
await cleanupDb({
|
||||
userIds: testUser ? [testUser.user.user_id] : [],
|
||||
flyerIds: createdFlyerIds,
|
||||
storeIds: createdStoreIds,
|
||||
});
|
||||
await cleanupFiles(createdFilePaths);
|
||||
});
|
||||
@@ -75,6 +96,10 @@ describe('Gamification Flow Integration Test', () => {
|
||||
it(
|
||||
'should award the "First Upload" achievement after a user successfully uploads and processes their first flyer',
|
||||
async () => {
|
||||
// --- Arrange: Stub environment variables for URL generation in the background worker ---
|
||||
const testBaseUrl = 'http://localhost:3001'; // Use a fixed port for predictability
|
||||
vi.stubEnv('FRONTEND_URL', testBaseUrl);
|
||||
|
||||
// --- Arrange: Prepare a unique flyer file for upload ---
|
||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||
const imageBuffer = await fs.readFile(imagePath);
|
||||
@@ -101,7 +126,7 @@ describe('Gamification Flow Integration Test', () => {
|
||||
|
||||
// --- Act 2: Poll for job completion ---
|
||||
let jobStatus;
|
||||
const maxRetries = 30; // Poll for up to 90 seconds
|
||||
const maxRetries = 60; // Poll for up to 180 seconds
|
||||
for (let i = 0; i < maxRetries; i++) {
|
||||
await new Promise((resolve) => setTimeout(resolve, 3000));
|
||||
const statusResponse = await request
|
||||
@@ -161,7 +186,82 @@ describe('Gamification Flow Integration Test', () => {
|
||||
expect(Number(userOnLeaderboard?.points)).toBeGreaterThanOrEqual(
|
||||
firstUploadAchievement!.points_value,
|
||||
);
|
||||
|
||||
// --- Cleanup ---
|
||||
vi.unstubAllEnvs();
|
||||
},
|
||||
120000, // Increase timeout to 120 seconds for this long-running test
|
||||
240000, // Increase timeout to 240s to match other long-running processing tests
|
||||
);
|
||||
|
||||
describe('Legacy Flyer Upload', () => {
|
||||
it('should process a legacy upload and save fully qualified URLs to the database', async () => {
|
||||
// --- Arrange ---
|
||||
// 1. Stub environment variables to have a predictable base URL for the test.
|
||||
const testBaseUrl = 'https://cdn.example.com';
|
||||
vi.stubEnv('FRONTEND_URL', testBaseUrl);
|
||||
|
||||
// 2. Mock the icon generator to return a predictable filename.
|
||||
vi.mocked(imageProcessor.generateFlyerIcon).mockResolvedValue('legacy-icon.webp');
|
||||
|
||||
// 3. Prepare a unique file for upload to avoid checksum conflicts.
|
||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||
const imageBuffer = await fs.readFile(imagePath);
|
||||
const uniqueFileName = `legacy-upload-test-${Date.now()}.jpg`;
|
||||
const mockImageFile = new File([imageBuffer], uniqueFileName, { type: 'image/jpeg' });
|
||||
const checksum = await generateFileChecksum(mockImageFile);
|
||||
|
||||
// Track created files for cleanup.
|
||||
const uploadDir = path.resolve(__dirname, '../../../flyer-images');
|
||||
createdFilePaths.push(path.join(uploadDir, uniqueFileName));
|
||||
createdFilePaths.push(path.join(uploadDir, 'icons', 'legacy-icon.webp'));
|
||||
|
||||
// 4. Prepare the legacy payload (body of the request).
|
||||
const storeName = `Legacy Store - ${Date.now()}`;
|
||||
const legacyPayload = {
|
||||
checksum: checksum,
|
||||
extractedData: {
|
||||
store_name: storeName,
|
||||
items: [{ item: 'Legacy Milk', price_in_cents: 250 }],
|
||||
},
|
||||
};
|
||||
|
||||
// --- Act ---
|
||||
// 5. Make the API request.
|
||||
// Note: This assumes a legacy endpoint exists at `/api/ai/upload-legacy`.
|
||||
// This endpoint would be responsible for calling `aiService.processLegacyFlyerUpload`.
|
||||
const response = await request
|
||||
.post('/api/ai/upload-legacy')
|
||||
.set('Authorization', `Bearer ${authToken}`)
|
||||
.field('data', JSON.stringify(legacyPayload))
|
||||
.attach('flyerFile', imageBuffer, uniqueFileName);
|
||||
|
||||
// --- Assert ---
|
||||
// 6. Check for a successful response.
|
||||
expect(response.status).toBe(200);
|
||||
const newFlyer: Flyer = response.body;
|
||||
expect(newFlyer).toBeDefined();
|
||||
expect(newFlyer.flyer_id).toBeTypeOf('number');
|
||||
createdFlyerIds.push(newFlyer.flyer_id); // Add for cleanup.
|
||||
|
||||
// 7. Query the database directly to verify the saved values.
|
||||
const pool = getPool();
|
||||
const dbResult = await pool.query<Flyer>(
|
||||
'SELECT image_url, icon_url, store_id FROM public.flyers WHERE flyer_id = $1',
|
||||
[newFlyer.flyer_id],
|
||||
);
|
||||
|
||||
expect(dbResult.rowCount).toBe(1);
|
||||
const savedFlyer = dbResult.rows[0];
|
||||
// The store_id is guaranteed to exist for a saved flyer, but the generic `Flyer` type
|
||||
// might have it as optional. We use a non-null assertion `!` to satisfy TypeScript.
|
||||
createdStoreIds.push(savedFlyer.store_id!); // Add for cleanup.
|
||||
|
||||
// 8. Assert that the URLs are fully qualified.
|
||||
expect(savedFlyer.image_url).to.equal(`${testBaseUrl}/flyer-images/${uniqueFileName}`);
|
||||
expect(savedFlyer.icon_url).to.equal(`${testBaseUrl}/flyer-images/icons/legacy-icon.webp`);
|
||||
|
||||
// --- Cleanup ---
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -34,22 +34,22 @@ describe('Price History API Integration Test (/api/price-history)', () => {
|
||||
|
||||
// 3. Create two flyers with different dates
|
||||
const flyerRes1 = await pool.query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-1.jpg', 'https://example.com/flyer-images/price-test-1.jpg', 1, $2, '2025-01-01') RETURNING flyer_id`,
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-1.jpg', 'https://example.com/flyer-images/price-test-1.jpg', 'https://example.com/flyer-images/icons/price-test-1.jpg', 1, $2, '2025-01-01') RETURNING flyer_id`,
|
||||
[storeId, `${Date.now().toString(16)}1`.padEnd(64, '0')],
|
||||
);
|
||||
flyerId1 = flyerRes1.rows[0].flyer_id;
|
||||
|
||||
const flyerRes2 = await pool.query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-2.jpg', 'https://example.com/flyer-images/price-test-2.jpg', 1, $2, '2025-01-08') RETURNING flyer_id`,
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-2.jpg', 'https://example.com/flyer-images/price-test-2.jpg', 'https://example.com/flyer-images/icons/price-test-2.jpg', 1, $2, '2025-01-08') RETURNING flyer_id`,
|
||||
[storeId, `${Date.now().toString(16)}2`.padEnd(64, '0')],
|
||||
);
|
||||
flyerId2 = flyerRes2.rows[0].flyer_id; // This was a duplicate, fixed.
|
||||
|
||||
const flyerRes3 = await pool.query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-3.jpg', 'https://example.com/flyer-images/price-test-3.jpg', 1, $2, '2025-01-15') RETURNING flyer_id`,
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum, valid_from)
|
||||
VALUES ($1, 'price-test-3.jpg', 'https://example.com/flyer-images/price-test-3.jpg', 'https://example.com/flyer-images/icons/price-test-3.jpg', 1, $2, '2025-01-15') RETURNING flyer_id`,
|
||||
[storeId, `${Date.now().toString(16)}3`.padEnd(64, '0')],
|
||||
);
|
||||
flyerId3 = flyerRes3.rows[0].flyer_id;
|
||||
|
||||
@@ -77,8 +77,8 @@ describe('Public API Routes Integration Tests', () => {
|
||||
);
|
||||
testStoreId = storeRes.rows[0].store_id;
|
||||
const flyerRes = await pool.query(
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||
VALUES ($1, 'public-routes-test.jpg', 'https://example.com/flyer-images/public-routes-test.jpg', 1, $2) RETURNING *`,
|
||||
`INSERT INTO public.flyers (store_id, file_name, image_url, icon_url, item_count, checksum)
|
||||
VALUES ($1, 'public-routes-test.jpg', 'https://example.com/flyer-images/public-routes-test.jpg', 'https://example.com/flyer-images/icons/public-routes-test.jpg', 1, $2) RETURNING *`,
|
||||
[testStoreId, `${Date.now().toString(16)}`.padEnd(64, '0')],
|
||||
);
|
||||
testFlyer = flyerRes.rows[0];
|
||||
|
||||
75
src/tests/setup/mockUI.ts
Normal file
75
src/tests/setup/mockUI.ts
Normal file
@@ -0,0 +1,75 @@
|
||||
// src/tests/setup/mockUI.ts
|
||||
import { vi } from 'vitest';
|
||||
|
||||
/**
|
||||
* This setup file centralizes the mocking of common UI components for high-level tests like App.test.tsx.
|
||||
* By importing this single file into a test, all standard UI components are replaced with their mock implementations
|
||||
* from `src/tests/utils/componentMocks.tsx`, reducing boilerplate in the test files.
|
||||
*
|
||||
* Note: Mocks that require special logic (e.g., using `vi.importActual`) should remain in the test file itself.
|
||||
*/
|
||||
|
||||
vi.mock('../../components/Footer', async () => {
|
||||
const { MockFooter } = await import('../utils/componentMocks');
|
||||
return { Footer: MockFooter };
|
||||
});
|
||||
|
||||
vi.mock('../../components/Header', async () => {
|
||||
const { MockHeader } = await import('../utils/componentMocks');
|
||||
return { Header: MockHeader };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/HomePage', async () => {
|
||||
const { MockHomePage } = await import('../utils/componentMocks');
|
||||
return { HomePage: MockHomePage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/admin/AdminPage', async () => {
|
||||
const { MockAdminPage } = await import('../utils/componentMocks');
|
||||
return { AdminPage: MockAdminPage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/admin/CorrectionsPage', async () => {
|
||||
const { MockCorrectionsPage } = await import('../utils/componentMocks');
|
||||
return { CorrectionsPage: MockCorrectionsPage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/admin/AdminStatsPage', async () => {
|
||||
const { MockAdminStatsPage } = await import('../utils/componentMocks');
|
||||
return { AdminStatsPage: MockAdminStatsPage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/VoiceLabPage', async () => {
|
||||
const { MockVoiceLabPage } = await import('../utils/componentMocks');
|
||||
return { VoiceLabPage: MockVoiceLabPage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/ResetPasswordPage', async () => {
|
||||
const { MockResetPasswordPage } = await import('../utils/componentMocks');
|
||||
return { ResetPasswordPage: MockResetPasswordPage };
|
||||
});
|
||||
|
||||
vi.mock('../../pages/admin/components/ProfileManager', async () => {
|
||||
const { MockProfileManager } = await import('../utils/componentMocks');
|
||||
return { ProfileManager: MockProfileManager };
|
||||
});
|
||||
|
||||
vi.mock('../../features/voice-assistant/VoiceAssistant', async () => {
|
||||
const { MockVoiceAssistant } = await import('../utils/componentMocks');
|
||||
return { VoiceAssistant: MockVoiceAssistant };
|
||||
});
|
||||
|
||||
vi.mock('../../components/FlyerCorrectionTool', async () => {
|
||||
const { MockFlyerCorrectionTool } = await import('../utils/componentMocks');
|
||||
return { FlyerCorrectionTool: MockFlyerCorrectionTool };
|
||||
});
|
||||
|
||||
vi.mock('../../components/WhatsNewModal', async () => {
|
||||
const { MockWhatsNewModal } = await import('../utils/componentMocks');
|
||||
return { WhatsNewModal: MockWhatsNewModal };
|
||||
});
|
||||
|
||||
vi.mock('../../layouts/MainLayout', async () => {
|
||||
const { MockMainLayout } = await import('../utils/componentMocks');
|
||||
return { MainLayout: MockMainLayout };
|
||||
});
|
||||
@@ -329,6 +329,59 @@ vi.mock('react-hot-toast', () => ({
|
||||
|
||||
// --- Database Service Mocks ---
|
||||
|
||||
// Mock for db/index.db which exports repository instances used by many routes
|
||||
vi.mock('../../services/db/index.db', () => ({
|
||||
userRepo: {
|
||||
findUserProfileById: vi.fn(),
|
||||
updateUserProfile: vi.fn(),
|
||||
updateUserPreferences: vi.fn(),
|
||||
},
|
||||
personalizationRepo: {
|
||||
getWatchedItems: vi.fn(),
|
||||
removeWatchedItem: vi.fn(),
|
||||
addWatchedItem: vi.fn(),
|
||||
getUserDietaryRestrictions: vi.fn(),
|
||||
setUserDietaryRestrictions: vi.fn(),
|
||||
getUserAppliances: vi.fn(),
|
||||
setUserAppliances: vi.fn(),
|
||||
},
|
||||
shoppingRepo: {
|
||||
getShoppingLists: vi.fn(),
|
||||
createShoppingList: vi.fn(),
|
||||
deleteShoppingList: vi.fn(),
|
||||
addShoppingListItem: vi.fn(),
|
||||
updateShoppingListItem: vi.fn(),
|
||||
removeShoppingListItem: vi.fn(),
|
||||
getShoppingListById: vi.fn(),
|
||||
},
|
||||
recipeRepo: {
|
||||
deleteRecipe: vi.fn(),
|
||||
updateRecipe: vi.fn(),
|
||||
},
|
||||
addressRepo: {
|
||||
getAddressById: vi.fn(),
|
||||
upsertAddress: vi.fn(),
|
||||
},
|
||||
notificationRepo: {
|
||||
getNotificationsForUser: vi.fn(),
|
||||
markAllNotificationsAsRead: vi.fn(),
|
||||
markNotificationAsRead: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock userService used by routes
|
||||
vi.mock('../../services/userService', () => ({
|
||||
userService: {
|
||||
updateUserAvatar: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteUserAccount: vi.fn(),
|
||||
getUserAddress: vi.fn(),
|
||||
upsertUserAddress: vi.fn(),
|
||||
processTokenCleanupJob: vi.fn(),
|
||||
deleteUserAsAdmin: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('../../services/db/user.db', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../../services/db/user.db')>();
|
||||
return {
|
||||
|
||||
@@ -107,8 +107,9 @@ export const MockMainLayout: React.FC<Partial<MainLayoutProps>> = () => (
|
||||
<Outlet />
|
||||
</div>
|
||||
);
|
||||
export const MockHomePage: React.FC<Partial<HomePageProps>> = ({ onOpenCorrectionTool }) => (
|
||||
<div data-testid="home-page-mock">
|
||||
export const MockHomePage: React.FC<Partial<HomePageProps>> = ({ selectedFlyer, onOpenCorrectionTool }) => (
|
||||
<div data-testid="home-page-mock" data-selected-flyer-id={selectedFlyer?.flyer_id}>
|
||||
Mock Home Page
|
||||
<button onClick={onOpenCorrectionTool}>Open Correction Tool</button>
|
||||
</div>
|
||||
);
|
||||
|
||||
@@ -88,7 +88,10 @@ export const resetMockIds = () => {
|
||||
* @returns A complete and type-safe User object.
|
||||
*/
|
||||
export const createMockUser = (overrides: Partial<User> = {}): User => {
|
||||
const userId = overrides.user_id ?? `user-${getNextId()}`;
|
||||
// Generate a deterministic, valid UUID-like string for mock user IDs.
|
||||
// This prevents database errors in integration tests where a UUID is expected.
|
||||
const userId =
|
||||
overrides.user_id ?? `00000000-0000-0000-0000-${String(getNextId()).padStart(12, '0')}`;
|
||||
|
||||
const defaultUser: User = {
|
||||
user_id: userId,
|
||||
@@ -175,6 +178,8 @@ export const createMockFlyer = (
|
||||
store_id: overrides.store_id ?? overrides.store?.store_id,
|
||||
});
|
||||
|
||||
const baseUrl = 'http://localhost:3001'; // A reasonable default for tests
|
||||
|
||||
// Determine the final file_name to generate dependent properties from.
|
||||
const fileName = overrides.file_name ?? `flyer-${flyerId}.jpg`;
|
||||
|
||||
@@ -192,8 +197,8 @@ export const createMockFlyer = (
|
||||
const defaultFlyer: Flyer = {
|
||||
flyer_id: flyerId,
|
||||
file_name: fileName,
|
||||
image_url: `/flyer-images/${fileName}`,
|
||||
icon_url: `/flyer-images/icons/icon-${fileName.replace(/\.[^/.]+$/, '.webp')}`,
|
||||
image_url: `${baseUrl}/flyer-images/${fileName}`,
|
||||
icon_url: `${baseUrl}/flyer-images/icons/icon-${fileName.replace(/\.[^/.]+$/, '.webp')}`,
|
||||
checksum: generateMockChecksum(fileName),
|
||||
store_id: store.store_id,
|
||||
valid_from: new Date().toISOString().split('T')[0],
|
||||
|
||||
0
src/tests/utils/userService.mock.ts
Normal file
0
src/tests/utils/userService.mock.ts
Normal file
@@ -14,7 +14,7 @@ export interface Flyer {
|
||||
readonly flyer_id: number;
|
||||
file_name: string;
|
||||
image_url: string;
|
||||
icon_url?: string | null; // URL for the 64x64 icon version of the flyer
|
||||
icon_url: string; // URL for the 64x64 icon version of the flyer
|
||||
readonly checksum?: string;
|
||||
readonly store_id?: number;
|
||||
valid_from?: string | null;
|
||||
@@ -72,7 +72,7 @@ export interface FlyerItem {
|
||||
item: string;
|
||||
price_display: string;
|
||||
price_in_cents?: number | null;
|
||||
quantity?: string;
|
||||
quantity: string;
|
||||
quantity_num?: number | null;
|
||||
master_item_id?: number; // Can be updated by admin correction
|
||||
master_item_name?: string | null;
|
||||
@@ -536,7 +536,7 @@ export type ActivityLogAction =
|
||||
interface ActivityLogItemBase {
|
||||
readonly activity_log_id: number;
|
||||
readonly user_id: string | null;
|
||||
action: string;
|
||||
action: ActivityLogAction;
|
||||
display_text: string;
|
||||
icon?: string | null;
|
||||
readonly created_at: string;
|
||||
|
||||
@@ -12,11 +12,11 @@ export const requiredString = (message: string) =>
|
||||
// They are used for validation and type inference across multiple services.
|
||||
|
||||
export const ExtractedFlyerItemSchema = z.object({
|
||||
item: z.string().nullable(),
|
||||
price_display: z.string().nullable(),
|
||||
price_in_cents: z.number().nullable(),
|
||||
quantity: z.string().nullable(),
|
||||
category_name: z.string().nullable(),
|
||||
item: z.string().nullish(),
|
||||
price_display: z.string().nullish(),
|
||||
price_in_cents: z.number().nullish(),
|
||||
quantity: z.string().nullish(),
|
||||
category_name: z.string().nullish(),
|
||||
master_item_id: z.number().nullish(), // .nullish() allows null or undefined
|
||||
});
|
||||
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
// src/types/job-data.ts
|
||||
|
||||
/**
|
||||
* Defines the data structure for a flyer processing job.
|
||||
* This is the information passed to the worker when a new flyer is uploaded.
|
||||
* Defines the shape of the data payload for a flyer processing job.
|
||||
* This is the data that gets passed to the BullMQ worker.
|
||||
*/
|
||||
export interface FlyerJobData {
|
||||
filePath: string;
|
||||
@@ -11,35 +11,11 @@ export interface FlyerJobData {
|
||||
userId?: string;
|
||||
submitterIp?: string;
|
||||
userProfileAddress?: string;
|
||||
baseUrl: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for an email sending job.
|
||||
*/
|
||||
export interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a daily analytics reporting job.
|
||||
*/
|
||||
export interface AnalyticsJobData {
|
||||
reportDate: string; // e.g., '2024-10-26'
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a weekly analytics reporting job.
|
||||
*/
|
||||
export interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number; // ISO week number (1-53)
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a file cleanup job, which runs after a flyer is successfully processed.
|
||||
* Defines the shape of the data payload for a file cleanup job.
|
||||
*/
|
||||
export interface CleanupJobData {
|
||||
flyerId: number;
|
||||
@@ -47,8 +23,33 @@ export interface CleanupJobData {
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for the job that cleans up expired password reset tokens.
|
||||
* Defines the shape of the data payload for a token cleanup job.
|
||||
*/
|
||||
export interface TokenCleanupJobData {
|
||||
timestamp: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the shape of the data payload for a daily analytics report job.
|
||||
*/
|
||||
export interface AnalyticsJobData {
|
||||
reportDate: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the shape of the data payload for a weekly analytics report job.
|
||||
*/
|
||||
export interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the shape of the data payload for an email sending job.
|
||||
*/
|
||||
export interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
@@ -3,8 +3,16 @@
|
||||
* @vitest-environment jsdom
|
||||
*/
|
||||
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, Mocked } from 'vitest';
|
||||
import { convertPdfToImageFiles } from './pdfConverter';
|
||||
import { logger } from '../services/logger.client';
|
||||
|
||||
// Mock the logger before other imports to spy on its methods
|
||||
vi.mock('../services/logger.client', () => ({
|
||||
logger: {
|
||||
warn: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock the entire pdfjs-dist library
|
||||
const mockPdfPage = {
|
||||
@@ -14,7 +22,9 @@ const mockPdfPage = {
|
||||
|
||||
const mockPdfDocument = {
|
||||
numPages: 3,
|
||||
getPage: vi.fn(() => Promise.resolve(mockPdfPage)),
|
||||
// Explicitly type the mock function to accept a number and return the correct promise type.
|
||||
// This resolves the TypeScript error when using mockImplementation with arguments later.
|
||||
getPage: vi.fn<(pageNumber: number) => Promise<typeof mockPdfPage>>(() => Promise.resolve(mockPdfPage)),
|
||||
};
|
||||
|
||||
vi.mock('pdfjs-dist', () => ({
|
||||
@@ -170,15 +180,26 @@ describe('pdfConverter', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should throw an error if getContext returns null', async () => {
|
||||
it('should return partial results if one page fails to get a canvas context', async () => {
|
||||
const pdfFile = new File(['pdf-content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
// Mock getContext to fail for the first page
|
||||
mockGetContext.mockReturnValueOnce(null);
|
||||
const mockedLogger = logger as Mocked<typeof logger>;
|
||||
|
||||
await expect(convertPdfToImageFiles(pdfFile)).rejects.toThrow('Could not get canvas context'); // This was a duplicate, fixed.
|
||||
const { imageFiles, pageCount } = await convertPdfToImageFiles(pdfFile);
|
||||
|
||||
// Should still report 3 total pages
|
||||
expect(pageCount).toBe(3);
|
||||
// But only 2 images should be successfully created
|
||||
expect(imageFiles).toHaveLength(2);
|
||||
// And a warning should be logged for the failed page
|
||||
expect(mockedLogger.warn).toHaveBeenCalledWith(
|
||||
{ error: new Error('Could not get canvas context') },
|
||||
'A page failed to convert during PDF processing.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if canvas.toBlob fails', async () => {
|
||||
it('should throw an error if canvas.toBlob fails for the only page', async () => {
|
||||
const pdfFile = new File(['pdf-content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
mockPdfDocument.numPages = 1;
|
||||
|
||||
@@ -187,8 +208,9 @@ describe('pdfConverter', () => {
|
||||
callback(null);
|
||||
});
|
||||
|
||||
// The function should throw the generic "zero images" error because the only page failed.
|
||||
await expect(convertPdfToImageFiles(pdfFile)).rejects.toThrow(
|
||||
'Failed to convert page 1 of PDF to blob.',
|
||||
'PDF conversion resulted in zero images, though the PDF has pages. It might be corrupted or contain non-standard content.',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -205,19 +227,56 @@ describe('pdfConverter', () => {
|
||||
expect(getDocument).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should throw an error if conversion results in zero images for a non-empty PDF', async () => {
|
||||
it('should throw a specific error if all pages of a non-empty PDF fail to convert', async () => {
|
||||
// Arrange: Ensure the document appears to have pages
|
||||
mockPdfDocument.numPages = 1;
|
||||
const pdfFile = new File(['pdf-content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
|
||||
// Mock getPage to fail for the first page. This simulates a corrupted page
|
||||
// within an otherwise valid PDF document, which is what the function's
|
||||
// Promise.allSettled logic is designed to handle.
|
||||
// Mock getPage to fail for the only page. This simulates a scenario where
|
||||
// the PDF has pages, but none can be rendered, causing the `imageFiles` array
|
||||
// to be empty.
|
||||
vi.mocked(mockPdfDocument.getPage).mockRejectedValueOnce(new Error('Corrupted page'));
|
||||
|
||||
// Act & Assert: The function should catch the settled promise and re-throw the reason.
|
||||
// Act & Assert: The function should now catch the settled promise, find that no
|
||||
// images were generated, and throw the specific "zero images" error, covering line 133.
|
||||
await expect(convertPdfToImageFiles(pdfFile)).rejects.toThrow(
|
||||
'PDF conversion resulted in zero images, though the PDF has pages. It might be corrupted or contain non-standard content.',
|
||||
);
|
||||
});
|
||||
|
||||
await expect(convertPdfToImageFiles(pdfFile)).rejects.toThrow('Corrupted page');
|
||||
it('should successfully process a PDF even if some pages fail to convert', async () => {
|
||||
// Arrange: 3-page PDF where the 2nd page will fail
|
||||
mockPdfDocument.numPages = 3;
|
||||
const pdfFile = new File(['pdf-content'], 'partial-success.pdf', { type: 'application/pdf' });
|
||||
const onProgress = vi.fn();
|
||||
const mockedLogger = logger as Mocked<typeof logger>;
|
||||
|
||||
// Mock getPage to fail only for the second page
|
||||
vi.mocked(mockPdfDocument.getPage).mockImplementation(async (pageNumber: number) => {
|
||||
if (pageNumber === 2) {
|
||||
throw new Error('Simulated page 2 corruption');
|
||||
}
|
||||
// Return the standard mock page for other pages
|
||||
return mockPdfPage;
|
||||
});
|
||||
|
||||
// Act
|
||||
const { imageFiles, pageCount } = await convertPdfToImageFiles(pdfFile, onProgress);
|
||||
|
||||
// Assert
|
||||
// Total page count should still be 3
|
||||
expect(pageCount).toBe(3);
|
||||
// Only 2 pages should have converted successfully
|
||||
expect(imageFiles).toHaveLength(2);
|
||||
// The progress callback should have been called for the 2 successful pages
|
||||
expect(onProgress).toHaveBeenCalledTimes(2);
|
||||
expect(onProgress).toHaveBeenCalledWith(1, 3);
|
||||
expect(onProgress).toHaveBeenCalledWith(3, 3);
|
||||
// The failure of page 2 should be logged as a warning
|
||||
expect(mockedLogger.warn).toHaveBeenCalledWith(
|
||||
{ error: new Error('Simulated page 2 corruption') },
|
||||
'A page failed to convert during PDF processing.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if FileReader fails', async () => {
|
||||
|
||||
@@ -116,17 +116,18 @@ export const convertPdfToImageFiles = async (
|
||||
// Process all pages in parallel and collect the results.
|
||||
const settledResults = await Promise.allSettled(pagePromises);
|
||||
|
||||
// Check for any hard failures and re-throw the first one encountered.
|
||||
const firstRejected = settledResults.find((r) => r.status === 'rejected') as
|
||||
| PromiseRejectedResult
|
||||
| undefined;
|
||||
if (firstRejected) {
|
||||
throw firstRejected.reason;
|
||||
}
|
||||
// Filter for fulfilled promises and extract their values. This allows for partial
|
||||
// success if some pages convert and others fail.
|
||||
const imageFiles = settledResults
|
||||
.filter((result): result is PromiseFulfilledResult<File> => result.status === 'fulfilled')
|
||||
.map((result) => result.value);
|
||||
|
||||
// Collect all successfully rendered image files. Since we've already checked for rejections,
|
||||
// we know all results are fulfilled and can safely extract their values.
|
||||
const imageFiles = settledResults.map((result) => (result as PromiseFulfilledResult<File>).value);
|
||||
// Log any pages that failed to convert, without stopping the entire process.
|
||||
settledResults.forEach((result) => {
|
||||
if (result.status === 'rejected') {
|
||||
logger.warn({ error: result.reason }, 'A page failed to convert during PDF processing.');
|
||||
}
|
||||
});
|
||||
|
||||
if (imageFiles.length === 0 && pageCount > 0) {
|
||||
throw new Error(
|
||||
|
||||
@@ -69,4 +69,9 @@ describe('parsePriceToCents', () => {
|
||||
expect(parsePriceToCents(' $10.99 ')).toBe(1099);
|
||||
expect(parsePriceToCents(' 99¢ ')).toBe(99);
|
||||
});
|
||||
|
||||
it('should return null for a price string that matches the pattern but results in NaN (e.g., "$." or ".")', () => {
|
||||
expect(parsePriceToCents('$.')).toBeNull();
|
||||
expect(parsePriceToCents('.')).toBeNull();
|
||||
});
|
||||
});
|
||||
|
||||
85
src/utils/serverUtils.test.ts
Normal file
85
src/utils/serverUtils.test.ts
Normal file
@@ -0,0 +1,85 @@
|
||||
// src/utils/serverUtils.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import type { Logger } from 'pino';
|
||||
import { getBaseUrl } from './serverUtils';
|
||||
|
||||
// Create a mock logger to spy on its methods
|
||||
const createMockLogger = (): Logger =>
|
||||
({
|
||||
warn: vi.fn(),
|
||||
// Add other logger methods if they were used, but only `warn` is relevant here.
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
fatal: vi.fn(),
|
||||
trace: vi.fn(),
|
||||
silent: vi.fn(),
|
||||
child: vi.fn(() => createMockLogger()),
|
||||
level: 'info',
|
||||
}) as unknown as Logger;
|
||||
|
||||
describe('serverUtils', () => {
|
||||
describe('getBaseUrl', () => {
|
||||
const originalEnv = process.env;
|
||||
let mockLogger: Logger;
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset mocks and environment variables before each test for isolation
|
||||
vi.resetModules();
|
||||
process.env = { ...originalEnv };
|
||||
mockLogger = createMockLogger();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore original environment variables after each test
|
||||
process.env = originalEnv;
|
||||
});
|
||||
|
||||
it('should use FRONTEND_URL if it is a valid URL', () => {
|
||||
process.env.FRONTEND_URL = 'https://valid.example.com';
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('https://valid.example.com');
|
||||
expect(mockLogger.warn).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should trim a trailing slash from FRONTEND_URL', () => {
|
||||
process.env.FRONTEND_URL = 'https://valid.example.com/';
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('https://valid.example.com');
|
||||
});
|
||||
|
||||
it('should use BASE_URL if FRONTEND_URL is not set', () => {
|
||||
delete process.env.FRONTEND_URL;
|
||||
process.env.BASE_URL = 'https://base.example.com';
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('https://base.example.com');
|
||||
expect(mockLogger.warn).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should fall back to localhost with default port 3000 if no URL is provided', () => {
|
||||
delete process.env.FRONTEND_URL;
|
||||
delete process.env.BASE_URL;
|
||||
delete process.env.PORT;
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('http://localhost:3000');
|
||||
expect(mockLogger.warn).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should fall back to localhost with the specified PORT if no URL is provided', () => {
|
||||
delete process.env.FRONTEND_URL;
|
||||
delete process.env.BASE_URL;
|
||||
process.env.PORT = '8888';
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('http://localhost:8888');
|
||||
});
|
||||
|
||||
it('should log a warning and fall back if FRONTEND_URL is invalid (does not start with http)', () => {
|
||||
process.env.FRONTEND_URL = 'invalid.url.com';
|
||||
const baseUrl = getBaseUrl(mockLogger);
|
||||
expect(baseUrl).toBe('http://localhost:3000');
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
"[getBaseUrl] FRONTEND_URL/BASE_URL is invalid or incomplete ('invalid.url.com'). Falling back to default local URL: http://localhost:3000",
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
26
src/utils/serverUtils.ts
Normal file
26
src/utils/serverUtils.ts
Normal file
@@ -0,0 +1,26 @@
|
||||
// src/utils/serverUtils.ts
|
||||
import type { Logger } from 'pino';
|
||||
|
||||
/**
|
||||
* Constructs a fully qualified base URL for generating absolute URLs.
|
||||
* It prioritizes `FRONTEND_URL`, then `BASE_URL`, and falls back to a localhost URL
|
||||
* based on the `PORT` environment variable. It also logs a warning if the provided
|
||||
* URL is invalid or missing.
|
||||
*
|
||||
* @param logger - The logger instance to use for warnings.
|
||||
* @returns A validated, fully qualified base URL without a trailing slash.
|
||||
*/
|
||||
export function getBaseUrl(logger: Logger): string {
|
||||
let baseUrl = (process.env.FRONTEND_URL || process.env.BASE_URL || '').trim();
|
||||
if (!baseUrl || !baseUrl.startsWith('http')) {
|
||||
const port = process.env.PORT || 3000;
|
||||
const fallbackUrl = `http://localhost:${port}`;
|
||||
if (baseUrl) {
|
||||
logger.warn(
|
||||
`[getBaseUrl] FRONTEND_URL/BASE_URL is invalid or incomplete ('${baseUrl}'). Falling back to default local URL: ${fallbackUrl}`,
|
||||
);
|
||||
}
|
||||
baseUrl = fallbackUrl;
|
||||
}
|
||||
return baseUrl.endsWith('/') ? baseUrl.slice(0, -1) : baseUrl;
|
||||
}
|
||||
@@ -44,10 +44,17 @@ const finalConfig = mergeConfig(
|
||||
// Otherwise, the inherited `exclude` rule will prevent any integration tests from running.
|
||||
// Setting it to an empty array removes all exclusion rules for this project.
|
||||
exclude: [],
|
||||
// Fix: Set environment variables to ensure generated URLs pass validation
|
||||
env: {
|
||||
NODE_ENV: 'test',
|
||||
BASE_URL: 'http://example.com', // Use a standard domain to pass strict URL validation
|
||||
PORT: '3000',
|
||||
},
|
||||
// This setup script starts the backend server before tests run.
|
||||
globalSetup: './src/tests/setup/integration-global-setup.ts',
|
||||
// The default timeout is 5000ms (5 seconds)
|
||||
testTimeout: 60000, // Increased timeout for server startup and API calls, especially AI services.
|
||||
hookTimeout: 60000,
|
||||
// "singleThread: true" is removed in modern Vitest.
|
||||
// Use fileParallelism: false to ensure test files run one by one to prevent port conflicts.
|
||||
fileParallelism: false,
|
||||
|
||||
Reference in New Issue
Block a user