Compare commits
49 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
146d4c1351 | ||
| 88625706f4 | |||
|
|
e395faed30 | ||
| e8f8399896 | |||
|
|
ac0115af2b | ||
| f24b15f19b | |||
|
|
e64426bd84 | ||
| 0ec4cd68d2 | |||
|
|
840516d2a3 | ||
| 59355c3eef | |||
| d024935fe9 | |||
|
|
5a5470634e | ||
| 392231ad63 | |||
|
|
4b1c896621 | ||
| 720920a51c | |||
|
|
460adb9506 | ||
| 7aa1f756a9 | |||
|
|
c484a8ca9b | ||
| 28d2c9f4ec | |||
|
|
ee253e9449 | ||
| b6c15e53d0 | |||
|
|
722162c2c3 | ||
| 02a76fe996 | |||
|
|
0ebb03a7ab | ||
| 748ac9e049 | |||
|
|
495edd621c | ||
| 4ffca19db6 | |||
|
|
717427c5d7 | ||
| cc438a0e36 | |||
|
|
a32a0b62fc | ||
| 342f72b713 | |||
|
|
91254d18f3 | ||
| 40580dbf15 | |||
| 7f1d74c047 | |||
|
|
ecec686347 | ||
| 86de680080 | |||
|
|
0371947065 | ||
| 296698758c | |||
|
|
18c1161587 | ||
| 0010396780 | |||
|
|
d4557e13fb | ||
| 3e41130c69 | |||
|
|
d9034563d6 | ||
| 5836a75157 | |||
|
|
790008ae0d | ||
|
|
b5b91eb968 | ||
| 38eb810e7a | |||
|
|
458588a6e7 | ||
| 0b4113417f |
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
|||||||
{
|
{
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"version": "0.6.4",
|
"version": "0.7.21",
|
||||||
"lockfileVersion": 3,
|
"lockfileVersion": 3,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"packages": {
|
"packages": {
|
||||||
"": {
|
"": {
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"version": "0.6.4",
|
"version": "0.7.21",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@bull-board/api": "^6.14.2",
|
"@bull-board/api": "^6.14.2",
|
||||||
"@bull-board/express": "^6.14.2",
|
"@bull-board/express": "^6.14.2",
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"private": true,
|
"private": true,
|
||||||
"version": "0.6.4",
|
"version": "0.7.21",
|
||||||
"type": "module",
|
"type": "module",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||||
|
|||||||
@@ -8,16 +8,23 @@
|
|||||||
CREATE TABLE IF NOT EXISTS public.addresses (
|
CREATE TABLE IF NOT EXISTS public.addresses (
|
||||||
address_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
address_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
address_line_1 TEXT NOT NULL UNIQUE,
|
address_line_1 TEXT NOT NULL UNIQUE,
|
||||||
address_line_2 TEXT,
|
|
||||||
city TEXT NOT NULL,
|
city TEXT NOT NULL,
|
||||||
province_state TEXT NOT NULL,
|
province_state TEXT NOT NULL,
|
||||||
postal_code TEXT NOT NULL,
|
postal_code TEXT NOT NULL,
|
||||||
country TEXT NOT NULL,
|
country TEXT NOT NULL,
|
||||||
|
address_line_2 TEXT,
|
||||||
latitude NUMERIC(9, 6),
|
latitude NUMERIC(9, 6),
|
||||||
longitude NUMERIC(9, 6),
|
longitude NUMERIC(9, 6),
|
||||||
location GEOGRAPHY(Point, 4326),
|
location GEOGRAPHY(Point, 4326),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT addresses_address_line_1_check CHECK (TRIM(address_line_1) <> ''),
|
||||||
|
CONSTRAINT addresses_city_check CHECK (TRIM(city) <> ''),
|
||||||
|
CONSTRAINT addresses_province_state_check CHECK (TRIM(province_state) <> ''),
|
||||||
|
CONSTRAINT addresses_postal_code_check CHECK (TRIM(postal_code) <> ''),
|
||||||
|
CONSTRAINT addresses_country_check CHECK (TRIM(country) <> ''),
|
||||||
|
CONSTRAINT addresses_latitude_check CHECK (latitude >= -90 AND latitude <= 90),
|
||||||
|
CONSTRAINT addresses_longitude_check CHECK (longitude >= -180 AND longitude <= 180)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.addresses IS 'A centralized table for storing all physical addresses for users and stores.';
|
COMMENT ON TABLE public.addresses IS 'A centralized table for storing all physical addresses for users and stores.';
|
||||||
COMMENT ON COLUMN public.addresses.latitude IS 'The geographic latitude.';
|
COMMENT ON COLUMN public.addresses.latitude IS 'The geographic latitude.';
|
||||||
@@ -31,12 +38,14 @@ CREATE TABLE IF NOT EXISTS public.users (
|
|||||||
email TEXT NOT NULL UNIQUE,
|
email TEXT NOT NULL UNIQUE,
|
||||||
password_hash TEXT,
|
password_hash TEXT,
|
||||||
refresh_token TEXT,
|
refresh_token TEXT,
|
||||||
failed_login_attempts INTEGER DEFAULT 0,
|
failed_login_attempts INTEGER DEFAULT 0 CHECK (failed_login_attempts >= 0),
|
||||||
last_failed_login TIMESTAMPTZ,
|
last_failed_login TIMESTAMPTZ,
|
||||||
last_login_at TIMESTAMPTZ,
|
last_login_at TIMESTAMPTZ,
|
||||||
last_login_ip TEXT,
|
last_login_ip TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT users_email_check CHECK (email ~* '^[A-Za-z0-9._%+-]+@[A-Za-z0-9.-]+\.[A-Za-z]{2,}$'),
|
||||||
|
CONSTRAINT users_password_hash_check CHECK (password_hash IS NULL OR TRIM(password_hash) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.users IS 'Stores user authentication information.';
|
COMMENT ON TABLE public.users IS 'Stores user authentication information.';
|
||||||
COMMENT ON COLUMN public.users.refresh_token IS 'Stores the long-lived refresh token for re-authentication.';
|
COMMENT ON COLUMN public.users.refresh_token IS 'Stores the long-lived refresh token for re-authentication.';
|
||||||
@@ -59,10 +68,13 @@ CREATE TABLE IF NOT EXISTS public.activity_log (
|
|||||||
icon TEXT,
|
icon TEXT,
|
||||||
details JSONB,
|
details JSONB,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT activity_log_action_check CHECK (TRIM(action) <> ''),
|
||||||
|
CONSTRAINT activity_log_display_text_check CHECK (TRIM(display_text) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.activity_log IS 'Logs key user and system actions for auditing and display in an activity feed.';
|
COMMENT ON TABLE public.activity_log IS 'Logs key user and system actions for auditing and display in an activity feed.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_activity_log_user_id ON public.activity_log(user_id);
|
-- This composite index is more efficient for user-specific activity feeds ordered by date.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_activity_log_user_id_created_at ON public.activity_log(user_id, created_at DESC);
|
||||||
|
|
||||||
-- 3. for public user profiles.
|
-- 3. for public user profiles.
|
||||||
-- This table is linked to the users table and stores non-sensitive user data.
|
-- This table is linked to the users table and stores non-sensitive user data.
|
||||||
@@ -72,16 +84,20 @@ CREATE TABLE IF NOT EXISTS public.profiles (
|
|||||||
full_name TEXT,
|
full_name TEXT,
|
||||||
avatar_url TEXT,
|
avatar_url TEXT,
|
||||||
address_id BIGINT REFERENCES public.addresses(address_id) ON DELETE SET NULL,
|
address_id BIGINT REFERENCES public.addresses(address_id) ON DELETE SET NULL,
|
||||||
|
points INTEGER DEFAULT 0 NOT NULL CHECK (points >= 0),
|
||||||
preferences JSONB,
|
preferences JSONB,
|
||||||
role TEXT CHECK (role IN ('admin', 'user')),
|
role TEXT CHECK (role IN ('admin', 'user')),
|
||||||
points INTEGER DEFAULT 0 NOT NULL,
|
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT profiles_full_name_check CHECK (full_name IS NULL OR TRIM(full_name) <> ''),
|
||||||
|
CONSTRAINT profiles_avatar_url_check CHECK (avatar_url IS NULL OR avatar_url ~* '^https://?.*'),
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
updated_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
updated_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.profiles IS 'Stores public-facing user data, linked to the public.users table.';
|
COMMENT ON TABLE public.profiles IS 'Stores public-facing user data, linked to the public.users table.';
|
||||||
COMMENT ON COLUMN public.profiles.address_id IS 'A foreign key to the user''s primary address in the `addresses` table.';
|
COMMENT ON COLUMN public.profiles.address_id IS 'A foreign key to the user''s primary address in the `addresses` table.';
|
||||||
|
-- This index is crucial for the gamification leaderboard feature.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_profiles_points_leaderboard ON public.profiles (points DESC, full_name ASC);
|
||||||
COMMENT ON COLUMN public.profiles.points IS 'A simple integer column to store a user''s total accumulated points from achievements.';
|
COMMENT ON COLUMN public.profiles.points IS 'A simple integer column to store a user''s total accumulated points from achievements.';
|
||||||
|
|
||||||
-- 4. The 'stores' table for normalized store data.
|
-- 4. The 'stores' table for normalized store data.
|
||||||
@@ -91,6 +107,8 @@ CREATE TABLE IF NOT EXISTS public.stores (
|
|||||||
logo_url TEXT,
|
logo_url TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT stores_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT stores_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*'),
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.stores IS 'Stores metadata for grocery store chains (e.g., Safeway, Kroger).';
|
COMMENT ON TABLE public.stores IS 'Stores metadata for grocery store chains (e.g., Safeway, Kroger).';
|
||||||
@@ -100,7 +118,8 @@ CREATE TABLE IF NOT EXISTS public.categories (
|
|||||||
category_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
category_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT categories_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.categories IS 'Stores a predefined list of grocery item categories (e.g., ''Fruits & Vegetables'', ''Dairy & Eggs'').';
|
COMMENT ON TABLE public.categories IS 'Stores a predefined list of grocery item categories (e.g., ''Fruits & Vegetables'', ''Dairy & Eggs'').';
|
||||||
|
|
||||||
@@ -116,10 +135,15 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
|||||||
valid_to DATE,
|
valid_to DATE,
|
||||||
store_address TEXT,
|
store_address TEXT,
|
||||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
item_count INTEGER DEFAULT 0 NOT NULL CHECK (item_count >= 0),
|
||||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT flyers_valid_dates_check CHECK (valid_to >= valid_from),
|
||||||
|
CONSTRAINT flyers_file_name_check CHECK (TRIM(file_name) <> ''),
|
||||||
|
CONSTRAINT flyers_image_url_check CHECK (image_url ~* '^https://?.*'),
|
||||||
|
CONSTRAINT flyers_icon_url_check CHECK (icon_url IS NULL OR icon_url ~* '^https://?.*'),
|
||||||
|
CONSTRAINT flyers_checksum_check CHECK (checksum IS NULL OR length(checksum) = 64)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.flyers IS 'Stores metadata for each processed flyer, linking it to a store and its validity period.';
|
COMMENT ON TABLE public.flyers IS 'Stores metadata for each processed flyer, linking it to a store and its validity period.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_store_id ON public.flyers(store_id);
|
CREATE INDEX IF NOT EXISTS idx_flyers_store_id ON public.flyers(store_id);
|
||||||
@@ -135,6 +159,7 @@ COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e
|
|||||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||||
@@ -147,7 +172,8 @@ CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
|||||||
allergy_info JSONB,
|
allergy_info JSONB,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
|
CONSTRAINT master_grocery_items_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.master_grocery_items IS 'The master dictionary of canonical grocery items. Each item has a unique name and is linked to a category.';
|
COMMENT ON TABLE public.master_grocery_items IS 'The master dictionary of canonical grocery items. Each item has a unique name and is linked to a category.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_master_grocery_items_category_id ON public.master_grocery_items(category_id);
|
CREATE INDEX IF NOT EXISTS idx_master_grocery_items_category_id ON public.master_grocery_items(category_id);
|
||||||
@@ -172,7 +198,9 @@ CREATE TABLE IF NOT EXISTS public.brands (
|
|||||||
logo_url TEXT,
|
logo_url TEXT,
|
||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT brands_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT brands_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||||
@@ -187,7 +215,9 @@ CREATE TABLE IF NOT EXISTS public.products (
|
|||||||
size TEXT,
|
size TEXT,
|
||||||
upc_code TEXT UNIQUE,
|
upc_code TEXT UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT products_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT products_upc_code_check CHECK (upc_code IS NULL OR upc_code ~ '^[0-9]{8,14}$')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||||
@@ -203,18 +233,22 @@ CREATE TABLE IF NOT EXISTS public.flyer_items (
|
|||||||
flyer_id BIGINT REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
flyer_id BIGINT REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
||||||
item TEXT NOT NULL,
|
item TEXT NOT NULL,
|
||||||
price_display TEXT NOT NULL,
|
price_display TEXT NOT NULL,
|
||||||
price_in_cents INTEGER,
|
price_in_cents INTEGER CHECK (price_in_cents IS NULL OR price_in_cents >= 0),
|
||||||
quantity_num NUMERIC,
|
quantity_num NUMERIC,
|
||||||
quantity TEXT NOT NULL,
|
quantity TEXT NOT NULL,
|
||||||
category_id BIGINT REFERENCES public.categories(category_id) ON DELETE SET NULL,
|
category_id BIGINT REFERENCES public.categories(category_id) ON DELETE SET NULL,
|
||||||
category_name TEXT,
|
category_name TEXT,
|
||||||
unit_price JSONB,
|
unit_price JSONB,
|
||||||
view_count INTEGER DEFAULT 0 NOT NULL,
|
view_count INTEGER DEFAULT 0 NOT NULL CHECK (view_count >= 0),
|
||||||
click_count INTEGER DEFAULT 0 NOT NULL,
|
click_count INTEGER DEFAULT 0 NOT NULL CHECK (click_count >= 0),
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT flyer_items_item_check CHECK (TRIM(item) <> ''),
|
||||||
|
CONSTRAINT flyer_items_price_display_check CHECK (TRIM(price_display) <> ''),
|
||||||
|
CONSTRAINT flyer_items_quantity_check CHECK (TRIM(quantity) <> ''),
|
||||||
|
CONSTRAINT flyer_items_category_name_check CHECK (category_name IS NULL OR TRIM(category_name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.flyer_items IS 'Stores individual items extracted from a specific flyer.';
|
COMMENT ON TABLE public.flyer_items IS 'Stores individual items extracted from a specific flyer.';
|
||||||
COMMENT ON COLUMN public.flyer_items.flyer_id IS 'Foreign key linking this item to its parent flyer in the `flyers` table.';
|
COMMENT ON COLUMN public.flyer_items.flyer_id IS 'Foreign key linking this item to its parent flyer in the `flyers` table.';
|
||||||
@@ -233,6 +267,8 @@ CREATE INDEX IF NOT EXISTS idx_flyer_items_master_item_id ON public.flyer_items(
|
|||||||
CREATE INDEX IF NOT EXISTS idx_flyer_items_category_id ON public.flyer_items(category_id);
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_category_id ON public.flyer_items(category_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyer_items_product_id ON public.flyer_items(product_id);
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_product_id ON public.flyer_items(product_id);
|
||||||
-- Add a GIN index to the 'item' column for fast fuzzy text searching.
|
-- Add a GIN index to the 'item' column for fast fuzzy text searching.
|
||||||
|
-- This partial index is optimized for queries that find the best price for an item.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_master_item_price ON public.flyer_items (master_item_id, price_in_cents ASC) WHERE price_in_cents IS NOT NULL;
|
||||||
-- This requires the pg_trgm extension.
|
-- This requires the pg_trgm extension.
|
||||||
CREATE INDEX IF NOT EXISTS flyer_items_item_trgm_idx ON public.flyer_items USING GIN (item gin_trgm_ops);
|
CREATE INDEX IF NOT EXISTS flyer_items_item_trgm_idx ON public.flyer_items USING GIN (item gin_trgm_ops);
|
||||||
|
|
||||||
@@ -241,7 +277,7 @@ CREATE TABLE IF NOT EXISTS public.user_alerts (
|
|||||||
user_alert_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
user_alert_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_watched_item_id BIGINT NOT NULL REFERENCES public.user_watched_items(user_watched_item_id) ON DELETE CASCADE,
|
user_watched_item_id BIGINT NOT NULL REFERENCES public.user_watched_items(user_watched_item_id) ON DELETE CASCADE,
|
||||||
alert_type TEXT NOT NULL CHECK (alert_type IN ('PRICE_BELOW', 'PERCENT_OFF_AVERAGE')),
|
alert_type TEXT NOT NULL CHECK (alert_type IN ('PRICE_BELOW', 'PERCENT_OFF_AVERAGE')),
|
||||||
threshold_value NUMERIC NOT NULL,
|
threshold_value NUMERIC NOT NULL CHECK (threshold_value > 0),
|
||||||
is_active BOOLEAN DEFAULT true NOT NULL,
|
is_active BOOLEAN DEFAULT true NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
@@ -259,7 +295,8 @@ CREATE TABLE IF NOT EXISTS public.notifications (
|
|||||||
link_url TEXT,
|
link_url TEXT,
|
||||||
is_read BOOLEAN DEFAULT false NOT NULL,
|
is_read BOOLEAN DEFAULT false NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT notifications_content_check CHECK (TRIM(content) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.notifications IS 'A central log of notifications generated for users, such as price alerts.';
|
COMMENT ON TABLE public.notifications IS 'A central log of notifications generated for users, such as price alerts.';
|
||||||
COMMENT ON COLUMN public.notifications.content IS 'The notification message displayed to the user.';
|
COMMENT ON COLUMN public.notifications.content IS 'The notification message displayed to the user.';
|
||||||
@@ -272,8 +309,8 @@ CREATE TABLE IF NOT EXISTS public.store_locations (
|
|||||||
store_location_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
store_location_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
address_id BIGINT NOT NULL REFERENCES public.addresses(address_id) ON DELETE CASCADE,
|
address_id BIGINT NOT NULL REFERENCES public.addresses(address_id) ON DELETE CASCADE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
|
||||||
UNIQUE(store_id, address_id),
|
UNIQUE(store_id, address_id),
|
||||||
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.store_locations IS 'Stores physical locations of stores with geographic data for proximity searches.';
|
COMMENT ON TABLE public.store_locations IS 'Stores physical locations of stores with geographic data for proximity searches.';
|
||||||
@@ -285,13 +322,14 @@ CREATE TABLE IF NOT EXISTS public.item_price_history (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
summary_date DATE NOT NULL,
|
summary_date DATE NOT NULL,
|
||||||
store_location_id BIGINT REFERENCES public.store_locations(store_location_id) ON DELETE CASCADE,
|
store_location_id BIGINT REFERENCES public.store_locations(store_location_id) ON DELETE CASCADE,
|
||||||
min_price_in_cents INTEGER,
|
min_price_in_cents INTEGER CHECK (min_price_in_cents IS NULL OR min_price_in_cents >= 0),
|
||||||
max_price_in_cents INTEGER,
|
max_price_in_cents INTEGER CHECK (max_price_in_cents IS NULL OR max_price_in_cents >= 0),
|
||||||
avg_price_in_cents INTEGER,
|
avg_price_in_cents INTEGER CHECK (avg_price_in_cents IS NULL OR avg_price_in_cents >= 0),
|
||||||
data_points_count INTEGER DEFAULT 0 NOT NULL,
|
data_points_count INTEGER DEFAULT 0 NOT NULL CHECK (data_points_count >= 0),
|
||||||
UNIQUE(master_item_id, summary_date, store_location_id),
|
UNIQUE(master_item_id, summary_date, store_location_id),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT item_price_history_price_order_check CHECK (min_price_in_cents <= max_price_in_cents)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.item_price_history IS 'Serves as a summary table to speed up charting and analytics.';
|
COMMENT ON TABLE public.item_price_history IS 'Serves as a summary table to speed up charting and analytics.';
|
||||||
COMMENT ON COLUMN public.item_price_history.summary_date IS 'The date for which the price data is summarized.';
|
COMMENT ON COLUMN public.item_price_history.summary_date IS 'The date for which the price data is summarized.';
|
||||||
@@ -308,7 +346,8 @@ CREATE TABLE IF NOT EXISTS public.master_item_aliases (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
alias TEXT NOT NULL UNIQUE,
|
alias TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT master_item_aliases_alias_check CHECK (TRIM(alias) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.master_item_aliases IS 'Stores synonyms or alternative names for master items to improve matching.';
|
COMMENT ON TABLE public.master_item_aliases IS 'Stores synonyms or alternative names for master items to improve matching.';
|
||||||
COMMENT ON COLUMN public.master_item_aliases.alias IS 'An alternative name, e.g., "Ground Chuck" for the master item "Ground Beef".';
|
COMMENT ON COLUMN public.master_item_aliases.alias IS 'An alternative name, e.g., "Ground Chuck" for the master item "Ground Beef".';
|
||||||
@@ -320,7 +359,8 @@ CREATE TABLE IF NOT EXISTS public.shopping_lists (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT shopping_lists_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_lists IS 'Stores user-created shopping lists, e.g., "Weekly Groceries".';
|
COMMENT ON TABLE public.shopping_lists IS 'Stores user-created shopping lists, e.g., "Weekly Groceries".';
|
||||||
CREATE INDEX IF NOT EXISTS idx_shopping_lists_user_id ON public.shopping_lists(user_id);
|
CREATE INDEX IF NOT EXISTS idx_shopping_lists_user_id ON public.shopping_lists(user_id);
|
||||||
@@ -331,12 +371,13 @@ CREATE TABLE IF NOT EXISTS public.shopping_list_items (
|
|||||||
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
custom_item_name TEXT,
|
custom_item_name TEXT,
|
||||||
quantity NUMERIC DEFAULT 1 NOT NULL,
|
quantity NUMERIC DEFAULT 1 NOT NULL CHECK (quantity > 0),
|
||||||
is_purchased BOOLEAN DEFAULT false NOT NULL,
|
is_purchased BOOLEAN DEFAULT false NOT NULL,
|
||||||
notes TEXT,
|
notes TEXT,
|
||||||
added_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
added_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
CONSTRAINT must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
CONSTRAINT must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL),
|
||||||
|
CONSTRAINT shopping_list_items_custom_item_name_check CHECK (custom_item_name IS NULL OR TRIM(custom_item_name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_list_items IS 'Contains individual items for a specific shopping list.';
|
COMMENT ON TABLE public.shopping_list_items IS 'Contains individual items for a specific shopping list.';
|
||||||
COMMENT ON COLUMN public.shopping_list_items.custom_item_name IS 'For items not in the master list, e.g., "Grandma''s special spice mix".';
|
COMMENT ON COLUMN public.shopping_list_items.custom_item_name IS 'For items not in the master list, e.g., "Grandma''s special spice mix".';
|
||||||
@@ -344,7 +385,6 @@ COMMENT ON COLUMN public.shopping_list_items.is_purchased IS 'Lets users check i
|
|||||||
CREATE INDEX IF NOT EXISTS idx_shopping_list_items_shopping_list_id ON public.shopping_list_items(shopping_list_id);
|
CREATE INDEX IF NOT EXISTS idx_shopping_list_items_shopping_list_id ON public.shopping_list_items(shopping_list_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_shopping_list_items_master_item_id ON public.shopping_list_items(master_item_id);
|
CREATE INDEX IF NOT EXISTS idx_shopping_list_items_master_item_id ON public.shopping_list_items(master_item_id);
|
||||||
|
|
||||||
-- 17. Manage shared access to shopping lists.
|
|
||||||
CREATE TABLE IF NOT EXISTS public.shared_shopping_lists (
|
CREATE TABLE IF NOT EXISTS public.shared_shopping_lists (
|
||||||
shared_shopping_list_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
shared_shopping_list_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
||||||
@@ -369,6 +409,7 @@ CREATE TABLE IF NOT EXISTS public.menu_plans (
|
|||||||
end_date DATE NOT NULL,
|
end_date DATE NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT menu_plans_name_check CHECK (TRIM(name) <> ''),
|
||||||
CONSTRAINT date_range_check CHECK (end_date >= start_date)
|
CONSTRAINT date_range_check CHECK (end_date >= start_date)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.menu_plans IS 'Represents a user''s meal plan for a specific period, e.g., "Week of Oct 23".';
|
COMMENT ON TABLE public.menu_plans IS 'Represents a user''s meal plan for a specific period, e.g., "Week of Oct 23".';
|
||||||
@@ -397,11 +438,13 @@ CREATE TABLE IF NOT EXISTS public.suggested_corrections (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
correction_type TEXT NOT NULL,
|
correction_type TEXT NOT NULL,
|
||||||
suggested_value TEXT NOT NULL,
|
suggested_value TEXT NOT NULL,
|
||||||
status TEXT DEFAULT 'pending' NOT NULL,
|
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'approved', 'rejected')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
reviewed_notes TEXT,
|
reviewed_notes TEXT,
|
||||||
reviewed_at TIMESTAMPTZ,
|
reviewed_at TIMESTAMPTZ,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT suggested_corrections_correction_type_check CHECK (TRIM(correction_type) <> ''),
|
||||||
|
CONSTRAINT suggested_corrections_suggested_value_check CHECK (TRIM(suggested_value) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.suggested_corrections IS 'A queue for user-submitted data corrections, enabling crowdsourced data quality improvements.';
|
COMMENT ON TABLE public.suggested_corrections IS 'A queue for user-submitted data corrections, enabling crowdsourced data quality improvements.';
|
||||||
COMMENT ON COLUMN public.suggested_corrections.correction_type IS 'The type of error the user is reporting.';
|
COMMENT ON COLUMN public.suggested_corrections.correction_type IS 'The type of error the user is reporting.';
|
||||||
@@ -417,12 +460,13 @@ CREATE TABLE IF NOT EXISTS public.user_submitted_prices (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
store_id BIGINT NOT NULL REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT NOT NULL REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
price_in_cents INTEGER NOT NULL,
|
price_in_cents INTEGER NOT NULL CHECK (price_in_cents > 0),
|
||||||
photo_url TEXT,
|
photo_url TEXT,
|
||||||
upvotes INTEGER DEFAULT 0 NOT NULL,
|
upvotes INTEGER DEFAULT 0 NOT NULL CHECK (upvotes >= 0),
|
||||||
downvotes INTEGER DEFAULT 0 NOT NULL,
|
downvotes INTEGER DEFAULT 0 NOT NULL CHECK (downvotes >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT user_submitted_prices_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_submitted_prices IS 'Stores item prices submitted by users directly from physical stores.';
|
COMMENT ON TABLE public.user_submitted_prices IS 'Stores item prices submitted by users directly from physical stores.';
|
||||||
COMMENT ON COLUMN public.user_submitted_prices.photo_url IS 'URL to user-submitted photo evidence of the price.';
|
COMMENT ON COLUMN public.user_submitted_prices.photo_url IS 'URL to user-submitted photo evidence of the price.';
|
||||||
@@ -464,20 +508,22 @@ CREATE TABLE IF NOT EXISTS public.recipes (
|
|||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
description TEXT,
|
description TEXT,
|
||||||
instructions TEXT,
|
instructions TEXT,
|
||||||
prep_time_minutes INTEGER,
|
prep_time_minutes INTEGER CHECK (prep_time_minutes IS NULL OR prep_time_minutes >= 0),
|
||||||
cook_time_minutes INTEGER,
|
cook_time_minutes INTEGER CHECK (cook_time_minutes IS NULL OR cook_time_minutes >= 0),
|
||||||
servings INTEGER,
|
servings INTEGER CHECK (servings IS NULL OR servings > 0),
|
||||||
photo_url TEXT,
|
photo_url TEXT,
|
||||||
calories_per_serving INTEGER,
|
calories_per_serving INTEGER,
|
||||||
protein_grams NUMERIC,
|
protein_grams NUMERIC,
|
||||||
fat_grams NUMERIC,
|
fat_grams NUMERIC,
|
||||||
carb_grams NUMERIC,
|
carb_grams NUMERIC,
|
||||||
avg_rating NUMERIC(2,1) DEFAULT 0.0 NOT NULL,
|
avg_rating NUMERIC(2,1) DEFAULT 0.0 NOT NULL CHECK (avg_rating >= 0.0 AND avg_rating <= 5.0),
|
||||||
status TEXT DEFAULT 'private' NOT NULL CHECK (status IN ('private', 'pending_review', 'public', 'rejected')),
|
status TEXT DEFAULT 'private' NOT NULL CHECK (status IN ('private', 'pending_review', 'public', 'rejected')),
|
||||||
rating_count INTEGER DEFAULT 0 NOT NULL,
|
rating_count INTEGER DEFAULT 0 NOT NULL CHECK (rating_count >= 0),
|
||||||
fork_count INTEGER DEFAULT 0 NOT NULL,
|
fork_count INTEGER DEFAULT 0 NOT NULL CHECK (fork_count >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipes_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT recipes_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipes IS 'Stores recipes that can be used to generate shopping lists.';
|
COMMENT ON TABLE public.recipes IS 'Stores recipes that can be used to generate shopping lists.';
|
||||||
COMMENT ON COLUMN public.recipes.servings IS 'The number of servings this recipe yields.';
|
COMMENT ON COLUMN public.recipes.servings IS 'The number of servings this recipe yields.';
|
||||||
@@ -488,11 +534,11 @@ COMMENT ON COLUMN public.recipes.calories_per_serving IS 'Optional nutritional i
|
|||||||
COMMENT ON COLUMN public.recipes.protein_grams IS 'Optional nutritional information.';
|
COMMENT ON COLUMN public.recipes.protein_grams IS 'Optional nutritional information.';
|
||||||
COMMENT ON COLUMN public.recipes.fat_grams IS 'Optional nutritional information.';
|
COMMENT ON COLUMN public.recipes.fat_grams IS 'Optional nutritional information.';
|
||||||
COMMENT ON COLUMN public.recipes.carb_grams IS 'Optional nutritional information.';
|
COMMENT ON COLUMN public.recipes.carb_grams IS 'Optional nutritional information.';
|
||||||
COMMENT ON COLUMN public.recipes.fork_count IS 'To track how many times a public recipe has been "forked" or copied by other users.';
|
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipes_user_id ON public.recipes(user_id);
|
CREATE INDEX IF NOT EXISTS idx_recipes_user_id ON public.recipes(user_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipes_original_recipe_id ON public.recipes(original_recipe_id);
|
CREATE INDEX IF NOT EXISTS idx_recipes_original_recipe_id ON public.recipes(original_recipe_id);
|
||||||
-- Add a partial unique index to ensure system-wide recipes (user_id IS NULL) have unique names.
|
-- Add a partial unique index to ensure system-wide recipes (user_id IS NULL) have unique names.
|
||||||
-- This allows different users to have recipes with the same name.
|
-- This index helps speed up sorting for recipe recommendations.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_recipes_rating_sort ON public.recipes (avg_rating DESC, rating_count DESC);
|
||||||
CREATE UNIQUE INDEX IF NOT EXISTS idx_recipes_unique_system_recipe_name ON public.recipes(name) WHERE user_id IS NULL;
|
CREATE UNIQUE INDEX IF NOT EXISTS idx_recipes_unique_system_recipe_name ON public.recipes(name) WHERE user_id IS NULL;
|
||||||
|
|
||||||
-- 27. For ingredients required for each recipe.
|
-- 27. For ingredients required for each recipe.
|
||||||
@@ -500,10 +546,11 @@ CREATE TABLE IF NOT EXISTS public.recipe_ingredients (
|
|||||||
recipe_ingredient_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
recipe_ingredient_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
recipe_id BIGINT NOT NULL REFERENCES public.recipes(recipe_id) ON DELETE CASCADE,
|
recipe_id BIGINT NOT NULL REFERENCES public.recipes(recipe_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity > 0),
|
||||||
unit TEXT NOT NULL,
|
unit TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_ingredients_unit_check CHECK (TRIM(unit) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_ingredients IS 'Defines the ingredients and quantities needed for a recipe.';
|
COMMENT ON TABLE public.recipe_ingredients IS 'Defines the ingredients and quantities needed for a recipe.';
|
||||||
COMMENT ON COLUMN public.recipe_ingredients.unit IS 'e.g., "cups", "tbsp", "g", "each".';
|
COMMENT ON COLUMN public.recipe_ingredients.unit IS 'e.g., "cups", "tbsp", "g", "each".';
|
||||||
@@ -529,7 +576,8 @@ CREATE TABLE IF NOT EXISTS public.tags (
|
|||||||
tag_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
tag_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT tags_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.tags IS 'Stores tags for categorizing recipes, e.g., "Vegetarian", "Quick & Easy".';
|
COMMENT ON TABLE public.tags IS 'Stores tags for categorizing recipes, e.g., "Vegetarian", "Quick & Easy".';
|
||||||
|
|
||||||
@@ -543,6 +591,7 @@ CREATE TABLE IF NOT EXISTS public.recipe_tags (
|
|||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_tags IS 'A linking table to associate multiple tags with a single recipe.';
|
COMMENT ON TABLE public.recipe_tags IS 'A linking table to associate multiple tags with a single recipe.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipe_tags_recipe_id ON public.recipe_tags(recipe_id);
|
CREATE INDEX IF NOT EXISTS idx_recipe_tags_recipe_id ON public.recipe_tags(recipe_id);
|
||||||
|
-- This index is crucial for functions that find recipes based on tags.
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipe_tags_tag_id ON public.recipe_tags(tag_id);
|
CREATE INDEX IF NOT EXISTS idx_recipe_tags_tag_id ON public.recipe_tags(tag_id);
|
||||||
|
|
||||||
-- 31. Store a predefined list of kitchen appliances.
|
-- 31. Store a predefined list of kitchen appliances.
|
||||||
@@ -550,7 +599,8 @@ CREATE TABLE IF NOT EXISTS public.appliances (
|
|||||||
appliance_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
appliance_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT appliances_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.appliances IS 'A predefined list of kitchen appliances (e.g., Air Fryer, Instant Pot).';
|
COMMENT ON TABLE public.appliances IS 'A predefined list of kitchen appliances (e.g., Air Fryer, Instant Pot).';
|
||||||
|
|
||||||
@@ -590,7 +640,8 @@ CREATE TABLE IF NOT EXISTS public.recipe_comments (
|
|||||||
content TEXT NOT NULL,
|
content TEXT NOT NULL,
|
||||||
status TEXT DEFAULT 'visible' NOT NULL CHECK (status IN ('visible', 'hidden', 'reported')),
|
status TEXT DEFAULT 'visible' NOT NULL CHECK (status IN ('visible', 'hidden', 'reported')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_comments_content_check CHECK (TRIM(content) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_comments IS 'Allows for threaded discussions and comments on recipes.';
|
COMMENT ON TABLE public.recipe_comments IS 'Allows for threaded discussions and comments on recipes.';
|
||||||
COMMENT ON COLUMN public.recipe_comments.parent_comment_id IS 'For threaded comments.';
|
COMMENT ON COLUMN public.recipe_comments.parent_comment_id IS 'For threaded comments.';
|
||||||
@@ -605,6 +656,7 @@ CREATE TABLE IF NOT EXISTS public.pantry_locations (
|
|||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT pantry_locations_name_check CHECK (TRIM(name) <> ''),
|
||||||
UNIQUE(user_id, name)
|
UNIQUE(user_id, name)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.pantry_locations IS 'User-defined locations for organizing pantry items (e.g., "Fridge", "Freezer", "Spice Rack").';
|
COMMENT ON TABLE public.pantry_locations IS 'User-defined locations for organizing pantry items (e.g., "Fridge", "Freezer", "Spice Rack").';
|
||||||
@@ -618,8 +670,9 @@ CREATE TABLE IF NOT EXISTS public.planned_meals (
|
|||||||
plan_date DATE NOT NULL,
|
plan_date DATE NOT NULL,
|
||||||
meal_type TEXT NOT NULL,
|
meal_type TEXT NOT NULL,
|
||||||
servings_to_cook INTEGER,
|
servings_to_cook INTEGER,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT planned_meals_meal_type_check CHECK (TRIM(meal_type) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.planned_meals IS 'Assigns a recipe to a specific day and meal type within a user''s menu plan.';
|
COMMENT ON TABLE public.planned_meals IS 'Assigns a recipe to a specific day and meal type within a user''s menu plan.';
|
||||||
COMMENT ON COLUMN public.planned_meals.meal_type IS 'The designated meal for the recipe, e.g., ''Breakfast'', ''Lunch'', ''Dinner''.';
|
COMMENT ON COLUMN public.planned_meals.meal_type IS 'The designated meal for the recipe, e.g., ''Breakfast'', ''Lunch'', ''Dinner''.';
|
||||||
@@ -631,7 +684,7 @@ CREATE TABLE IF NOT EXISTS public.pantry_items (
|
|||||||
pantry_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
pantry_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity >= 0),
|
||||||
unit TEXT,
|
unit TEXT,
|
||||||
best_before_date DATE,
|
best_before_date DATE,
|
||||||
pantry_location_id BIGINT REFERENCES public.pantry_locations(pantry_location_id) ON DELETE SET NULL,
|
pantry_location_id BIGINT REFERENCES public.pantry_locations(pantry_location_id) ON DELETE SET NULL,
|
||||||
@@ -640,7 +693,6 @@ CREATE TABLE IF NOT EXISTS public.pantry_items (
|
|||||||
UNIQUE(user_id, master_item_id, unit)
|
UNIQUE(user_id, master_item_id, unit)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.pantry_items IS 'Tracks a user''s personal inventory of grocery items to enable smart shopping lists.';
|
COMMENT ON TABLE public.pantry_items IS 'Tracks a user''s personal inventory of grocery items to enable smart shopping lists.';
|
||||||
COMMENT ON COLUMN public.pantry_items.quantity IS 'The current amount of the item. Convention: use grams for weight, mL for volume where applicable.';
|
|
||||||
COMMENT ON COLUMN public.pantry_items.pantry_location_id IS 'Links the item to a user-defined location like "Fridge" or "Freezer".';
|
COMMENT ON COLUMN public.pantry_items.pantry_location_id IS 'Links the item to a user-defined location like "Fridge" or "Freezer".';
|
||||||
COMMENT ON COLUMN public.pantry_items.unit IS 'e.g., ''g'', ''ml'', ''items''. Should align with recipe_ingredients.unit and quantity convention.';
|
COMMENT ON COLUMN public.pantry_items.unit IS 'e.g., ''g'', ''ml'', ''items''. Should align with recipe_ingredients.unit and quantity convention.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_pantry_items_user_id ON public.pantry_items(user_id);
|
CREATE INDEX IF NOT EXISTS idx_pantry_items_user_id ON public.pantry_items(user_id);
|
||||||
@@ -654,7 +706,8 @@ CREATE TABLE IF NOT EXISTS public.password_reset_tokens (
|
|||||||
token_hash TEXT NOT NULL UNIQUE,
|
token_hash TEXT NOT NULL UNIQUE,
|
||||||
expires_at TIMESTAMPTZ NOT NULL,
|
expires_at TIMESTAMPTZ NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT password_reset_tokens_token_hash_check CHECK (TRIM(token_hash) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.password_reset_tokens IS 'Stores secure, single-use tokens for password reset requests.';
|
COMMENT ON TABLE public.password_reset_tokens IS 'Stores secure, single-use tokens for password reset requests.';
|
||||||
COMMENT ON COLUMN public.password_reset_tokens.token_hash IS 'A bcrypt hash of the reset token sent to the user.';
|
COMMENT ON COLUMN public.password_reset_tokens.token_hash IS 'A bcrypt hash of the reset token sent to the user.';
|
||||||
@@ -669,10 +722,13 @@ CREATE TABLE IF NOT EXISTS public.unit_conversions (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
from_unit TEXT NOT NULL,
|
from_unit TEXT NOT NULL,
|
||||||
to_unit TEXT NOT NULL,
|
to_unit TEXT NOT NULL,
|
||||||
factor NUMERIC NOT NULL,
|
factor NUMERIC NOT NULL CHECK (factor > 0),
|
||||||
UNIQUE(master_item_id, from_unit, to_unit),
|
UNIQUE(master_item_id, from_unit, to_unit),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT unit_conversions_from_unit_check CHECK (TRIM(from_unit) <> ''),
|
||||||
|
CONSTRAINT unit_conversions_to_unit_check CHECK (TRIM(to_unit) <> ''),
|
||||||
|
CONSTRAINT unit_conversions_units_check CHECK (from_unit <> to_unit)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.unit_conversions IS 'Stores item-specific unit conversion factors (e.g., grams of flour to cups).';
|
COMMENT ON TABLE public.unit_conversions IS 'Stores item-specific unit conversion factors (e.g., grams of flour to cups).';
|
||||||
COMMENT ON COLUMN public.unit_conversions.factor IS 'The multiplication factor to convert from_unit to to_unit.';
|
COMMENT ON COLUMN public.unit_conversions.factor IS 'The multiplication factor to convert from_unit to to_unit.';
|
||||||
@@ -686,7 +742,8 @@ CREATE TABLE IF NOT EXISTS public.user_item_aliases (
|
|||||||
alias TEXT NOT NULL,
|
alias TEXT NOT NULL,
|
||||||
UNIQUE(user_id, alias),
|
UNIQUE(user_id, alias),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT user_item_aliases_alias_check CHECK (TRIM(alias) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_item_aliases IS 'Allows users to create personal aliases for grocery items (e.g., "Dad''s Cereal").';
|
COMMENT ON TABLE public.user_item_aliases IS 'Allows users to create personal aliases for grocery items (e.g., "Dad''s Cereal").';
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_item_aliases_user_id ON public.user_item_aliases(user_id);
|
CREATE INDEX IF NOT EXISTS idx_user_item_aliases_user_id ON public.user_item_aliases(user_id);
|
||||||
@@ -723,7 +780,8 @@ CREATE TABLE IF NOT EXISTS public.recipe_collections (
|
|||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
description TEXT,
|
description TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_collections_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_collections IS 'Allows users to create personal collections of recipes (e.g., "Holiday Baking").';
|
COMMENT ON TABLE public.recipe_collections IS 'Allows users to create personal collections of recipes (e.g., "Holiday Baking").';
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipe_collections_user_id ON public.recipe_collections(user_id);
|
CREATE INDEX IF NOT EXISTS idx_recipe_collections_user_id ON public.recipe_collections(user_id);
|
||||||
@@ -748,8 +806,11 @@ CREATE TABLE IF NOT EXISTS public.shared_recipe_collections (
|
|||||||
shared_with_user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
shared_with_user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
permission_level TEXT NOT NULL CHECK (permission_level IN ('view', 'edit')),
|
permission_level TEXT NOT NULL CHECK (permission_level IN ('view', 'edit')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
UNIQUE(recipe_collection_id, shared_with_user_id)
|
UNIQUE(recipe_collection_id, shared_with_user_id)
|
||||||
);
|
);
|
||||||
|
-- This index is crucial for efficiently finding all collections shared with a specific user.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_shared_recipe_collections_shared_with ON public.shared_recipe_collections(shared_with_user_id);
|
||||||
|
|
||||||
-- 45. Log user search queries for analysis.
|
-- 45. Log user search queries for analysis.
|
||||||
CREATE TABLE IF NOT EXISTS public.search_queries (
|
CREATE TABLE IF NOT EXISTS public.search_queries (
|
||||||
@@ -759,7 +820,8 @@ CREATE TABLE IF NOT EXISTS public.search_queries (
|
|||||||
result_count INTEGER,
|
result_count INTEGER,
|
||||||
was_successful BOOLEAN,
|
was_successful BOOLEAN,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT search_queries_query_text_check CHECK (TRIM(query_text) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.search_queries IS 'Logs user search queries to analyze search effectiveness and identify gaps in data.';
|
COMMENT ON TABLE public.search_queries IS 'Logs user search queries to analyze search effectiveness and identify gaps in data.';
|
||||||
COMMENT ON COLUMN public.search_queries.was_successful IS 'Indicates if the user interacted with a search result.';
|
COMMENT ON COLUMN public.search_queries.was_successful IS 'Indicates if the user interacted with a search result.';
|
||||||
@@ -785,10 +847,11 @@ CREATE TABLE IF NOT EXISTS public.shopping_trip_items (
|
|||||||
shopping_trip_id BIGINT NOT NULL REFERENCES public.shopping_trips(shopping_trip_id) ON DELETE CASCADE,
|
shopping_trip_id BIGINT NOT NULL REFERENCES public.shopping_trips(shopping_trip_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
custom_item_name TEXT,
|
custom_item_name TEXT,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity > 0),
|
||||||
price_paid_cents INTEGER,
|
price_paid_cents INTEGER,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT shopping_trip_items_custom_item_name_check CHECK (custom_item_name IS NULL OR TRIM(custom_item_name) <> ''),
|
||||||
CONSTRAINT trip_must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
CONSTRAINT trip_must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_trip_items IS 'A historical log of items purchased during a shopping trip.';
|
COMMENT ON TABLE public.shopping_trip_items IS 'A historical log of items purchased during a shopping trip.';
|
||||||
@@ -802,7 +865,8 @@ CREATE TABLE IF NOT EXISTS public.dietary_restrictions (
|
|||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
type TEXT NOT NULL CHECK (type IN ('diet', 'allergy')),
|
type TEXT NOT NULL CHECK (type IN ('diet', 'allergy')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT dietary_restrictions_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.dietary_restrictions IS 'A predefined list of common diets (e.g., Vegan) and allergies (e.g., Nut Allergy).';
|
COMMENT ON TABLE public.dietary_restrictions IS 'A predefined list of common diets (e.g., Vegan) and allergies (e.g., Nut Allergy).';
|
||||||
|
|
||||||
@@ -815,6 +879,7 @@ CREATE TABLE IF NOT EXISTS public.user_dietary_restrictions (
|
|||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_dietary_restrictions IS 'Connects users to their selected dietary needs and allergies.';
|
COMMENT ON TABLE public.user_dietary_restrictions IS 'Connects users to their selected dietary needs and allergies.';
|
||||||
|
-- This index is crucial for functions that filter recipes based on user diets/allergies.
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_dietary_restrictions_user_id ON public.user_dietary_restrictions(user_id);
|
CREATE INDEX IF NOT EXISTS idx_user_dietary_restrictions_user_id ON public.user_dietary_restrictions(user_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_dietary_restrictions_restriction_id ON public.user_dietary_restrictions(restriction_id);
|
CREATE INDEX IF NOT EXISTS idx_user_dietary_restrictions_restriction_id ON public.user_dietary_restrictions(restriction_id);
|
||||||
|
|
||||||
@@ -840,6 +905,7 @@ CREATE TABLE IF NOT EXISTS public.user_follows (
|
|||||||
CONSTRAINT cant_follow_self CHECK (follower_id <> following_id)
|
CONSTRAINT cant_follow_self CHECK (follower_id <> following_id)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_follows IS 'Stores user following relationships to build a social graph.';
|
COMMENT ON TABLE public.user_follows IS 'Stores user following relationships to build a social graph.';
|
||||||
|
-- This index is crucial for efficiently generating a user's activity feed.
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_follows_follower_id ON public.user_follows(follower_id);
|
CREATE INDEX IF NOT EXISTS idx_user_follows_follower_id ON public.user_follows(follower_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_follows_following_id ON public.user_follows(following_id);
|
CREATE INDEX IF NOT EXISTS idx_user_follows_following_id ON public.user_follows(following_id);
|
||||||
|
|
||||||
@@ -850,12 +916,13 @@ CREATE TABLE IF NOT EXISTS public.receipts (
|
|||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
receipt_image_url TEXT NOT NULL,
|
receipt_image_url TEXT NOT NULL,
|
||||||
transaction_date TIMESTAMPTZ,
|
transaction_date TIMESTAMPTZ,
|
||||||
total_amount_cents INTEGER,
|
total_amount_cents INTEGER CHECK (total_amount_cents IS NULL OR total_amount_cents >= 0),
|
||||||
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'processing', 'completed', 'failed')),
|
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'processing', 'completed', 'failed')),
|
||||||
raw_text TEXT,
|
raw_text TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
processed_at TIMESTAMPTZ,
|
processed_at TIMESTAMPTZ,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT receipts_receipt_image_url_check CHECK (receipt_image_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.receipts IS 'Stores uploaded user receipts for purchase tracking and analysis.';
|
COMMENT ON TABLE public.receipts IS 'Stores uploaded user receipts for purchase tracking and analysis.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_receipts_user_id ON public.receipts(user_id);
|
CREATE INDEX IF NOT EXISTS idx_receipts_user_id ON public.receipts(user_id);
|
||||||
@@ -866,13 +933,14 @@ CREATE TABLE IF NOT EXISTS public.receipt_items (
|
|||||||
receipt_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
receipt_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
receipt_id BIGINT NOT NULL REFERENCES public.receipts(receipt_id) ON DELETE CASCADE,
|
receipt_id BIGINT NOT NULL REFERENCES public.receipts(receipt_id) ON DELETE CASCADE,
|
||||||
raw_item_description TEXT NOT NULL,
|
raw_item_description TEXT NOT NULL,
|
||||||
quantity NUMERIC DEFAULT 1 NOT NULL,
|
quantity NUMERIC DEFAULT 1 NOT NULL CHECK (quantity > 0),
|
||||||
price_paid_cents INTEGER NOT NULL,
|
price_paid_cents INTEGER NOT NULL CHECK (price_paid_cents >= 0),
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
||||||
status TEXT DEFAULT 'unmatched' NOT NULL CHECK (status IN ('unmatched', 'matched', 'needs_review', 'ignored')),
|
status TEXT DEFAULT 'unmatched' NOT NULL CHECK (status IN ('unmatched', 'matched', 'needs_review', 'ignored')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT receipt_items_raw_item_description_check CHECK (TRIM(raw_item_description) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.receipt_items IS 'Stores individual line items extracted from a user receipt.';
|
COMMENT ON TABLE public.receipt_items IS 'Stores individual line items extracted from a user receipt.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_receipt_items_receipt_id ON public.receipt_items(receipt_id);
|
CREATE INDEX IF NOT EXISTS idx_receipt_items_receipt_id ON public.receipt_items(receipt_id);
|
||||||
@@ -885,7 +953,6 @@ CREATE TABLE IF NOT EXISTS public.schema_info (
|
|||||||
deployed_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
deployed_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.schema_info IS 'Stores metadata about the deployed schema, such as a hash of the schema file, to detect changes.';
|
COMMENT ON TABLE public.schema_info IS 'Stores metadata about the deployed schema, such as a hash of the schema file, to detect changes.';
|
||||||
COMMENT ON COLUMN public.schema_info.environment IS 'The deployment environment (e.g., ''development'', ''test'', ''production'').';
|
|
||||||
COMMENT ON COLUMN public.schema_info.schema_hash IS 'A SHA-256 hash of the master_schema_rollup.sql file at the time of deployment.';
|
COMMENT ON COLUMN public.schema_info.schema_hash IS 'A SHA-256 hash of the master_schema_rollup.sql file at the time of deployment.';
|
||||||
|
|
||||||
-- 55. Store user reactions to various entities (e.g., recipes, comments).
|
-- 55. Store user reactions to various entities (e.g., recipes, comments).
|
||||||
@@ -912,8 +979,10 @@ CREATE TABLE IF NOT EXISTS public.achievements (
|
|||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
description TEXT NOT NULL,
|
description TEXT NOT NULL,
|
||||||
icon TEXT,
|
icon TEXT,
|
||||||
points_value INTEGER NOT NULL DEFAULT 0,
|
points_value INTEGER NOT NULL DEFAULT 0 CHECK (points_value >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT achievements_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT achievements_description_check CHECK (TRIM(description) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.achievements IS 'A static table defining the available achievements users can earn.';
|
COMMENT ON TABLE public.achievements IS 'A static table defining the available achievements users can earn.';
|
||||||
|
|
||||||
@@ -934,11 +1003,12 @@ CREATE TABLE IF NOT EXISTS public.budgets (
|
|||||||
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
amount_cents INTEGER NOT NULL,
|
amount_cents INTEGER NOT NULL CHECK (amount_cents > 0),
|
||||||
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
||||||
start_date DATE NOT NULL,
|
start_date DATE NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT budgets_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
||||||
|
|||||||
@@ -23,16 +23,23 @@
|
|||||||
CREATE TABLE IF NOT EXISTS public.addresses (
|
CREATE TABLE IF NOT EXISTS public.addresses (
|
||||||
address_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
address_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
address_line_1 TEXT NOT NULL UNIQUE,
|
address_line_1 TEXT NOT NULL UNIQUE,
|
||||||
address_line_2 TEXT,
|
|
||||||
city TEXT NOT NULL,
|
city TEXT NOT NULL,
|
||||||
province_state TEXT NOT NULL,
|
province_state TEXT NOT NULL,
|
||||||
postal_code TEXT NOT NULL,
|
postal_code TEXT NOT NULL,
|
||||||
country TEXT NOT NULL,
|
country TEXT NOT NULL,
|
||||||
|
address_line_2 TEXT,
|
||||||
latitude NUMERIC(9, 6),
|
latitude NUMERIC(9, 6),
|
||||||
longitude NUMERIC(9, 6),
|
longitude NUMERIC(9, 6),
|
||||||
location GEOGRAPHY(Point, 4326),
|
location GEOGRAPHY(Point, 4326),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT addresses_address_line_1_check CHECK (TRIM(address_line_1) <> ''),
|
||||||
|
CONSTRAINT addresses_city_check CHECK (TRIM(city) <> ''),
|
||||||
|
CONSTRAINT addresses_province_state_check CHECK (TRIM(province_state) <> ''),
|
||||||
|
CONSTRAINT addresses_postal_code_check CHECK (TRIM(postal_code) <> ''),
|
||||||
|
CONSTRAINT addresses_country_check CHECK (TRIM(country) <> ''),
|
||||||
|
CONSTRAINT addresses_latitude_check CHECK (latitude >= -90 AND latitude <= 90),
|
||||||
|
CONSTRAINT addresses_longitude_check CHECK (longitude >= -180 AND longitude <= 180)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.addresses IS 'A centralized table for storing all physical addresses for users and stores.';
|
COMMENT ON TABLE public.addresses IS 'A centralized table for storing all physical addresses for users and stores.';
|
||||||
COMMENT ON COLUMN public.addresses.latitude IS 'The geographic latitude.';
|
COMMENT ON COLUMN public.addresses.latitude IS 'The geographic latitude.';
|
||||||
@@ -45,14 +52,16 @@ CREATE INDEX IF NOT EXISTS addresses_location_idx ON public.addresses USING GIST
|
|||||||
CREATE TABLE IF NOT EXISTS public.users (
|
CREATE TABLE IF NOT EXISTS public.users (
|
||||||
user_id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
|
user_id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
|
||||||
email TEXT NOT NULL UNIQUE,
|
email TEXT NOT NULL UNIQUE,
|
||||||
password_hash TEXT,
|
password_hash TEXT,
|
||||||
refresh_token TEXT,
|
refresh_token TEXT,
|
||||||
failed_login_attempts INTEGER DEFAULT 0,
|
failed_login_attempts INTEGER DEFAULT 0 CHECK (failed_login_attempts >= 0),
|
||||||
last_failed_login TIMESTAMPTZ,
|
last_failed_login TIMESTAMPTZ,
|
||||||
last_login_at TIMESTAMPTZ,
|
last_login_at TIMESTAMPTZ,
|
||||||
last_login_ip TEXT,
|
last_login_ip TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT users_email_check CHECK (email ~* '^[A-Za-z0-9._%+-]+@[A-Za-z0-9.-]+\.[A-Za-z]{2,}$'),
|
||||||
|
CONSTRAINT users_password_hash_check CHECK (password_hash IS NULL OR TRIM(password_hash) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.users IS 'Stores user authentication information.';
|
COMMENT ON TABLE public.users IS 'Stores user authentication information.';
|
||||||
COMMENT ON COLUMN public.users.refresh_token IS 'Stores the long-lived refresh token for re-authentication.';
|
COMMENT ON COLUMN public.users.refresh_token IS 'Stores the long-lived refresh token for re-authentication.';
|
||||||
@@ -74,11 +83,14 @@ CREATE TABLE IF NOT EXISTS public.activity_log (
|
|||||||
display_text TEXT NOT NULL,
|
display_text TEXT NOT NULL,
|
||||||
icon TEXT,
|
icon TEXT,
|
||||||
details JSONB,
|
details JSONB,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT activity_log_action_check CHECK (TRIM(action) <> ''),
|
||||||
|
CONSTRAINT activity_log_display_text_check CHECK (TRIM(display_text) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.activity_log IS 'Logs key user and system actions for auditing and display in an activity feed.';
|
COMMENT ON TABLE public.activity_log IS 'Logs key user and system actions for auditing and display in an activity feed.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_activity_log_user_id ON public.activity_log(user_id);
|
-- This composite index is more efficient for user-specific activity feeds ordered by date.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_activity_log_user_id_created_at ON public.activity_log(user_id, created_at DESC);
|
||||||
|
|
||||||
-- 3. for public user profiles.
|
-- 3. for public user profiles.
|
||||||
-- This table is linked to the users table and stores non-sensitive user data.
|
-- This table is linked to the users table and stores non-sensitive user data.
|
||||||
@@ -88,16 +100,20 @@ CREATE TABLE IF NOT EXISTS public.profiles (
|
|||||||
full_name TEXT,
|
full_name TEXT,
|
||||||
avatar_url TEXT,
|
avatar_url TEXT,
|
||||||
address_id BIGINT REFERENCES public.addresses(address_id) ON DELETE SET NULL,
|
address_id BIGINT REFERENCES public.addresses(address_id) ON DELETE SET NULL,
|
||||||
points INTEGER DEFAULT 0 NOT NULL,
|
points INTEGER DEFAULT 0 NOT NULL CHECK (points >= 0),
|
||||||
preferences JSONB,
|
preferences JSONB,
|
||||||
role TEXT CHECK (role IN ('admin', 'user')),
|
role TEXT CHECK (role IN ('admin', 'user')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
CONSTRAINT profiles_full_name_check CHECK (full_name IS NULL OR TRIM(full_name) <> ''),
|
||||||
|
CONSTRAINT profiles_avatar_url_check CHECK (avatar_url IS NULL OR avatar_url ~* '^https://?.*'),
|
||||||
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
updated_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
updated_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.profiles IS 'Stores public-facing user data, linked to the public.users table.';
|
COMMENT ON TABLE public.profiles IS 'Stores public-facing user data, linked to the public.users table.';
|
||||||
COMMENT ON COLUMN public.profiles.address_id IS 'A foreign key to the user''s primary address in the `addresses` table.';
|
COMMENT ON COLUMN public.profiles.address_id IS 'A foreign key to the user''s primary address in the `addresses` table.';
|
||||||
|
-- This index is crucial for the gamification leaderboard feature.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_profiles_points_leaderboard ON public.profiles (points DESC, full_name ASC);
|
||||||
COMMENT ON COLUMN public.profiles.points IS 'A simple integer column to store a user''s total accumulated points from achievements.';
|
COMMENT ON COLUMN public.profiles.points IS 'A simple integer column to store a user''s total accumulated points from achievements.';
|
||||||
|
|
||||||
-- 4. The 'stores' table for normalized store data.
|
-- 4. The 'stores' table for normalized store data.
|
||||||
@@ -107,7 +123,9 @@ CREATE TABLE IF NOT EXISTS public.stores (
|
|||||||
logo_url TEXT,
|
logo_url TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
CONSTRAINT stores_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT stores_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*'),
|
||||||
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.stores IS 'Stores metadata for grocery store chains (e.g., Safeway, Kroger).';
|
COMMENT ON TABLE public.stores IS 'Stores metadata for grocery store chains (e.g., Safeway, Kroger).';
|
||||||
|
|
||||||
@@ -116,7 +134,8 @@ CREATE TABLE IF NOT EXISTS public.categories (
|
|||||||
category_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
category_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT categories_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.categories IS 'Stores a predefined list of grocery item categories (e.g., ''Fruits & Vegetables'', ''Dairy & Eggs'').';
|
COMMENT ON TABLE public.categories IS 'Stores a predefined list of grocery item categories (e.g., ''Fruits & Vegetables'', ''Dairy & Eggs'').';
|
||||||
|
|
||||||
@@ -126,16 +145,21 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
|||||||
file_name TEXT NOT NULL,
|
file_name TEXT NOT NULL,
|
||||||
image_url TEXT NOT NULL,
|
image_url TEXT NOT NULL,
|
||||||
icon_url TEXT,
|
icon_url TEXT,
|
||||||
checksum TEXT UNIQUE,
|
checksum TEXT UNIQUE,
|
||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
valid_from DATE,
|
valid_from DATE,
|
||||||
valid_to DATE,
|
valid_to DATE,
|
||||||
store_address TEXT,
|
store_address TEXT,
|
||||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
item_count INTEGER DEFAULT 0 NOT NULL CHECK (item_count >= 0),
|
||||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT flyers_valid_dates_check CHECK (valid_to >= valid_from),
|
||||||
|
CONSTRAINT flyers_file_name_check CHECK (TRIM(file_name) <> ''),
|
||||||
|
CONSTRAINT flyers_image_url_check CHECK (image_url ~* '^https://?.*'),
|
||||||
|
CONSTRAINT flyers_icon_url_check CHECK (icon_url IS NULL OR icon_url ~* '^https://?.*'),
|
||||||
|
CONSTRAINT flyers_checksum_check CHECK (checksum IS NULL OR length(checksum) = 64)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.flyers IS 'Stores metadata for each processed flyer, linking it to a store and its validity period.';
|
COMMENT ON TABLE public.flyers IS 'Stores metadata for each processed flyer, linking it to a store and its validity period.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_store_id ON public.flyers(store_id);
|
CREATE INDEX IF NOT EXISTS idx_flyers_store_id ON public.flyers(store_id);
|
||||||
@@ -151,9 +175,9 @@ COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e
|
|||||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
|
||||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
@@ -163,7 +187,8 @@ CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
|||||||
allergy_info JSONB,
|
allergy_info JSONB,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL
|
created_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||||
|
CONSTRAINT master_grocery_items_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.master_grocery_items IS 'The master dictionary of canonical grocery items. Each item has a unique name and is linked to a category.';
|
COMMENT ON TABLE public.master_grocery_items IS 'The master dictionary of canonical grocery items. Each item has a unique name and is linked to a category.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_master_grocery_items_category_id ON public.master_grocery_items(category_id);
|
CREATE INDEX IF NOT EXISTS idx_master_grocery_items_category_id ON public.master_grocery_items(category_id);
|
||||||
@@ -188,7 +213,9 @@ CREATE TABLE IF NOT EXISTS public.brands (
|
|||||||
logo_url TEXT,
|
logo_url TEXT,
|
||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT brands_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT brands_logo_url_check CHECK (logo_url IS NULL OR logo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||||
@@ -203,7 +230,9 @@ CREATE TABLE IF NOT EXISTS public.products (
|
|||||||
size TEXT,
|
size TEXT,
|
||||||
upc_code TEXT UNIQUE,
|
upc_code TEXT UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT products_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT products_upc_code_check CHECK (upc_code IS NULL OR upc_code ~ '^[0-9]{8,14}$')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||||
@@ -219,18 +248,22 @@ CREATE TABLE IF NOT EXISTS public.flyer_items (
|
|||||||
flyer_id BIGINT REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
flyer_id BIGINT REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
||||||
item TEXT NOT NULL,
|
item TEXT NOT NULL,
|
||||||
price_display TEXT NOT NULL,
|
price_display TEXT NOT NULL,
|
||||||
price_in_cents INTEGER,
|
price_in_cents INTEGER CHECK (price_in_cents IS NULL OR price_in_cents >= 0),
|
||||||
quantity_num NUMERIC,
|
quantity_num NUMERIC,
|
||||||
quantity TEXT NOT NULL,
|
quantity TEXT NOT NULL,
|
||||||
category_id BIGINT REFERENCES public.categories(category_id) ON DELETE SET NULL,
|
category_id BIGINT REFERENCES public.categories(category_id) ON DELETE SET NULL,
|
||||||
category_name TEXT,
|
category_name TEXT,
|
||||||
unit_price JSONB,
|
unit_price JSONB,
|
||||||
view_count INTEGER DEFAULT 0 NOT NULL,
|
view_count INTEGER DEFAULT 0 NOT NULL CHECK (view_count >= 0),
|
||||||
click_count INTEGER DEFAULT 0 NOT NULL,
|
click_count INTEGER DEFAULT 0 NOT NULL CHECK (click_count >= 0),
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT flyer_items_item_check CHECK (TRIM(item) <> ''),
|
||||||
|
CONSTRAINT flyer_items_price_display_check CHECK (TRIM(price_display) <> ''),
|
||||||
|
CONSTRAINT flyer_items_quantity_check CHECK (TRIM(quantity) <> ''),
|
||||||
|
CONSTRAINT flyer_items_category_name_check CHECK (category_name IS NULL OR TRIM(category_name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.flyer_items IS 'Stores individual items extracted from a specific flyer.';
|
COMMENT ON TABLE public.flyer_items IS 'Stores individual items extracted from a specific flyer.';
|
||||||
COMMENT ON COLUMN public.flyer_items.flyer_id IS 'Foreign key linking this item to its parent flyer in the `flyers` table.';
|
COMMENT ON COLUMN public.flyer_items.flyer_id IS 'Foreign key linking this item to its parent flyer in the `flyers` table.';
|
||||||
@@ -249,6 +282,8 @@ CREATE INDEX IF NOT EXISTS idx_flyer_items_master_item_id ON public.flyer_items(
|
|||||||
CREATE INDEX IF NOT EXISTS idx_flyer_items_category_id ON public.flyer_items(category_id);
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_category_id ON public.flyer_items(category_id);
|
||||||
CREATE INDEX IF NOT EXISTS idx_flyer_items_product_id ON public.flyer_items(product_id);
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_product_id ON public.flyer_items(product_id);
|
||||||
-- Add a GIN index to the 'item' column for fast fuzzy text searching.
|
-- Add a GIN index to the 'item' column for fast fuzzy text searching.
|
||||||
|
-- This partial index is optimized for queries that find the best price for an item.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_flyer_items_master_item_price ON public.flyer_items (master_item_id, price_in_cents ASC) WHERE price_in_cents IS NOT NULL;
|
||||||
-- This requires the pg_trgm extension.
|
-- This requires the pg_trgm extension.
|
||||||
CREATE INDEX IF NOT EXISTS flyer_items_item_trgm_idx ON public.flyer_items USING GIN (item gin_trgm_ops);
|
CREATE INDEX IF NOT EXISTS flyer_items_item_trgm_idx ON public.flyer_items USING GIN (item gin_trgm_ops);
|
||||||
|
|
||||||
@@ -257,7 +292,7 @@ CREATE TABLE IF NOT EXISTS public.user_alerts (
|
|||||||
user_alert_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
user_alert_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_watched_item_id BIGINT NOT NULL REFERENCES public.user_watched_items(user_watched_item_id) ON DELETE CASCADE,
|
user_watched_item_id BIGINT NOT NULL REFERENCES public.user_watched_items(user_watched_item_id) ON DELETE CASCADE,
|
||||||
alert_type TEXT NOT NULL CHECK (alert_type IN ('PRICE_BELOW', 'PERCENT_OFF_AVERAGE')),
|
alert_type TEXT NOT NULL CHECK (alert_type IN ('PRICE_BELOW', 'PERCENT_OFF_AVERAGE')),
|
||||||
threshold_value NUMERIC NOT NULL,
|
threshold_value NUMERIC NOT NULL CHECK (threshold_value > 0),
|
||||||
is_active BOOLEAN DEFAULT true NOT NULL,
|
is_active BOOLEAN DEFAULT true NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
@@ -275,7 +310,8 @@ CREATE TABLE IF NOT EXISTS public.notifications (
|
|||||||
link_url TEXT,
|
link_url TEXT,
|
||||||
is_read BOOLEAN DEFAULT false NOT NULL,
|
is_read BOOLEAN DEFAULT false NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT notifications_content_check CHECK (TRIM(content) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.notifications IS 'A central log of notifications generated for users, such as price alerts.';
|
COMMENT ON TABLE public.notifications IS 'A central log of notifications generated for users, such as price alerts.';
|
||||||
COMMENT ON COLUMN public.notifications.content IS 'The notification message displayed to the user.';
|
COMMENT ON COLUMN public.notifications.content IS 'The notification message displayed to the user.';
|
||||||
@@ -301,13 +337,14 @@ CREATE TABLE IF NOT EXISTS public.item_price_history (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
summary_date DATE NOT NULL,
|
summary_date DATE NOT NULL,
|
||||||
store_location_id BIGINT REFERENCES public.store_locations(store_location_id) ON DELETE CASCADE,
|
store_location_id BIGINT REFERENCES public.store_locations(store_location_id) ON DELETE CASCADE,
|
||||||
min_price_in_cents INTEGER,
|
min_price_in_cents INTEGER CHECK (min_price_in_cents IS NULL OR min_price_in_cents >= 0),
|
||||||
max_price_in_cents INTEGER,
|
max_price_in_cents INTEGER CHECK (max_price_in_cents IS NULL OR max_price_in_cents >= 0),
|
||||||
avg_price_in_cents INTEGER,
|
avg_price_in_cents INTEGER CHECK (avg_price_in_cents IS NULL OR avg_price_in_cents >= 0),
|
||||||
data_points_count INTEGER DEFAULT 0 NOT NULL,
|
data_points_count INTEGER DEFAULT 0 NOT NULL CHECK (data_points_count >= 0),
|
||||||
UNIQUE(master_item_id, summary_date, store_location_id),
|
UNIQUE(master_item_id, summary_date, store_location_id),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT item_price_history_price_order_check CHECK (min_price_in_cents <= max_price_in_cents)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.item_price_history IS 'Serves as a summary table to speed up charting and analytics.';
|
COMMENT ON TABLE public.item_price_history IS 'Serves as a summary table to speed up charting and analytics.';
|
||||||
COMMENT ON COLUMN public.item_price_history.summary_date IS 'The date for which the price data is summarized.';
|
COMMENT ON COLUMN public.item_price_history.summary_date IS 'The date for which the price data is summarized.';
|
||||||
@@ -324,7 +361,8 @@ CREATE TABLE IF NOT EXISTS public.master_item_aliases (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
alias TEXT NOT NULL UNIQUE,
|
alias TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT master_item_aliases_alias_check CHECK (TRIM(alias) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.master_item_aliases IS 'Stores synonyms or alternative names for master items to improve matching.';
|
COMMENT ON TABLE public.master_item_aliases IS 'Stores synonyms or alternative names for master items to improve matching.';
|
||||||
COMMENT ON COLUMN public.master_item_aliases.alias IS 'An alternative name, e.g., "Ground Chuck" for the master item "Ground Beef".';
|
COMMENT ON COLUMN public.master_item_aliases.alias IS 'An alternative name, e.g., "Ground Chuck" for the master item "Ground Beef".';
|
||||||
@@ -336,7 +374,8 @@ CREATE TABLE IF NOT EXISTS public.shopping_lists (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT shopping_lists_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_lists IS 'Stores user-created shopping lists, e.g., "Weekly Groceries".';
|
COMMENT ON TABLE public.shopping_lists IS 'Stores user-created shopping lists, e.g., "Weekly Groceries".';
|
||||||
CREATE INDEX IF NOT EXISTS idx_shopping_lists_user_id ON public.shopping_lists(user_id);
|
CREATE INDEX IF NOT EXISTS idx_shopping_lists_user_id ON public.shopping_lists(user_id);
|
||||||
@@ -347,12 +386,13 @@ CREATE TABLE IF NOT EXISTS public.shopping_list_items (
|
|||||||
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
shopping_list_id BIGINT NOT NULL REFERENCES public.shopping_lists(shopping_list_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
custom_item_name TEXT,
|
custom_item_name TEXT,
|
||||||
quantity NUMERIC DEFAULT 1 NOT NULL,
|
quantity NUMERIC DEFAULT 1 NOT NULL CHECK (quantity > 0),
|
||||||
is_purchased BOOLEAN DEFAULT false NOT NULL,
|
is_purchased BOOLEAN DEFAULT false NOT NULL,
|
||||||
notes TEXT,
|
notes TEXT,
|
||||||
added_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
added_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
CONSTRAINT must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
CONSTRAINT must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL),
|
||||||
|
CONSTRAINT shopping_list_items_custom_item_name_check CHECK (custom_item_name IS NULL OR TRIM(custom_item_name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_list_items IS 'Contains individual items for a specific shopping list.';
|
COMMENT ON TABLE public.shopping_list_items IS 'Contains individual items for a specific shopping list.';
|
||||||
COMMENT ON COLUMN public.shopping_list_items.custom_item_name IS 'For items not in the master list, e.g., "Grandma''s special spice mix".';
|
COMMENT ON COLUMN public.shopping_list_items.custom_item_name IS 'For items not in the master list, e.g., "Grandma''s special spice mix".';
|
||||||
@@ -384,7 +424,8 @@ CREATE TABLE IF NOT EXISTS public.menu_plans (
|
|||||||
start_date DATE NOT NULL,
|
start_date DATE NOT NULL,
|
||||||
end_date DATE NOT NULL,
|
end_date DATE NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT menu_plans_name_check CHECK (TRIM(name) <> ''),
|
||||||
CONSTRAINT date_range_check CHECK (end_date >= start_date)
|
CONSTRAINT date_range_check CHECK (end_date >= start_date)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.menu_plans IS 'Represents a user''s meal plan for a specific period, e.g., "Week of Oct 23".';
|
COMMENT ON TABLE public.menu_plans IS 'Represents a user''s meal plan for a specific period, e.g., "Week of Oct 23".';
|
||||||
@@ -413,11 +454,13 @@ CREATE TABLE IF NOT EXISTS public.suggested_corrections (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
correction_type TEXT NOT NULL,
|
correction_type TEXT NOT NULL,
|
||||||
suggested_value TEXT NOT NULL,
|
suggested_value TEXT NOT NULL,
|
||||||
status TEXT DEFAULT 'pending' NOT NULL,
|
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'approved', 'rejected')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
reviewed_notes TEXT,
|
reviewed_notes TEXT,
|
||||||
reviewed_at TIMESTAMPTZ,
|
reviewed_at TIMESTAMPTZ,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT suggested_corrections_correction_type_check CHECK (TRIM(correction_type) <> ''),
|
||||||
|
CONSTRAINT suggested_corrections_suggested_value_check CHECK (TRIM(suggested_value) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.suggested_corrections IS 'A queue for user-submitted data corrections, enabling crowdsourced data quality improvements.';
|
COMMENT ON TABLE public.suggested_corrections IS 'A queue for user-submitted data corrections, enabling crowdsourced data quality improvements.';
|
||||||
COMMENT ON COLUMN public.suggested_corrections.correction_type IS 'The type of error the user is reporting.';
|
COMMENT ON COLUMN public.suggested_corrections.correction_type IS 'The type of error the user is reporting.';
|
||||||
@@ -433,12 +476,13 @@ CREATE TABLE IF NOT EXISTS public.user_submitted_prices (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
store_id BIGINT NOT NULL REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT NOT NULL REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
price_in_cents INTEGER NOT NULL,
|
price_in_cents INTEGER NOT NULL CHECK (price_in_cents > 0),
|
||||||
photo_url TEXT,
|
photo_url TEXT,
|
||||||
upvotes INTEGER DEFAULT 0 NOT NULL,
|
upvotes INTEGER DEFAULT 0 NOT NULL CHECK (upvotes >= 0),
|
||||||
downvotes INTEGER DEFAULT 0 NOT NULL,
|
downvotes INTEGER DEFAULT 0 NOT NULL CHECK (downvotes >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT user_submitted_prices_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_submitted_prices IS 'Stores item prices submitted by users directly from physical stores.';
|
COMMENT ON TABLE public.user_submitted_prices IS 'Stores item prices submitted by users directly from physical stores.';
|
||||||
COMMENT ON COLUMN public.user_submitted_prices.photo_url IS 'URL to user-submitted photo evidence of the price.';
|
COMMENT ON COLUMN public.user_submitted_prices.photo_url IS 'URL to user-submitted photo evidence of the price.';
|
||||||
@@ -449,7 +493,8 @@ CREATE INDEX IF NOT EXISTS idx_user_submitted_prices_master_item_id ON public.us
|
|||||||
-- 22. Log flyer items that could not be automatically matched to a master item.
|
-- 22. Log flyer items that could not be automatically matched to a master item.
|
||||||
CREATE TABLE IF NOT EXISTS public.unmatched_flyer_items (
|
CREATE TABLE IF NOT EXISTS public.unmatched_flyer_items (
|
||||||
unmatched_flyer_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
unmatched_flyer_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
flyer_item_id BIGINT NOT NULL REFERENCES public.flyer_items(flyer_item_id) ON DELETE CASCADE, status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'resolved', 'ignored')),
|
flyer_item_id BIGINT NOT NULL REFERENCES public.flyer_items(flyer_item_id) ON DELETE CASCADE,
|
||||||
|
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'resolved', 'ignored')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
reviewed_at TIMESTAMPTZ,
|
reviewed_at TIMESTAMPTZ,
|
||||||
UNIQUE(flyer_item_id),
|
UNIQUE(flyer_item_id),
|
||||||
@@ -479,20 +524,22 @@ CREATE TABLE IF NOT EXISTS public.recipes (
|
|||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
description TEXT,
|
description TEXT,
|
||||||
instructions TEXT,
|
instructions TEXT,
|
||||||
prep_time_minutes INTEGER,
|
prep_time_minutes INTEGER CHECK (prep_time_minutes IS NULL OR prep_time_minutes >= 0),
|
||||||
cook_time_minutes INTEGER,
|
cook_time_minutes INTEGER CHECK (cook_time_minutes IS NULL OR cook_time_minutes >= 0),
|
||||||
servings INTEGER,
|
servings INTEGER CHECK (servings IS NULL OR servings > 0),
|
||||||
photo_url TEXT,
|
photo_url TEXT,
|
||||||
calories_per_serving INTEGER,
|
calories_per_serving INTEGER,
|
||||||
protein_grams NUMERIC,
|
protein_grams NUMERIC,
|
||||||
fat_grams NUMERIC,
|
fat_grams NUMERIC,
|
||||||
carb_grams NUMERIC,
|
carb_grams NUMERIC,
|
||||||
avg_rating NUMERIC(2,1) DEFAULT 0.0 NOT NULL,
|
avg_rating NUMERIC(2,1) DEFAULT 0.0 NOT NULL CHECK (avg_rating >= 0.0 AND avg_rating <= 5.0),
|
||||||
status TEXT DEFAULT 'private' NOT NULL CHECK (status IN ('private', 'pending_review', 'public', 'rejected')),
|
status TEXT DEFAULT 'private' NOT NULL CHECK (status IN ('private', 'pending_review', 'public', 'rejected')),
|
||||||
rating_count INTEGER DEFAULT 0 NOT NULL,
|
rating_count INTEGER DEFAULT 0 NOT NULL CHECK (rating_count >= 0),
|
||||||
fork_count INTEGER DEFAULT 0 NOT NULL,
|
fork_count INTEGER DEFAULT 0 NOT NULL CHECK (fork_count >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipes_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT recipes_photo_url_check CHECK (photo_url IS NULL OR photo_url ~* '^https://?.*')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipes IS 'Stores recipes that can be used to generate shopping lists.';
|
COMMENT ON TABLE public.recipes IS 'Stores recipes that can be used to generate shopping lists.';
|
||||||
COMMENT ON COLUMN public.recipes.servings IS 'The number of servings this recipe yields.';
|
COMMENT ON COLUMN public.recipes.servings IS 'The number of servings this recipe yields.';
|
||||||
@@ -507,6 +554,8 @@ CREATE INDEX IF NOT EXISTS idx_recipes_user_id ON public.recipes(user_id);
|
|||||||
CREATE INDEX IF NOT EXISTS idx_recipes_original_recipe_id ON public.recipes(original_recipe_id);
|
CREATE INDEX IF NOT EXISTS idx_recipes_original_recipe_id ON public.recipes(original_recipe_id);
|
||||||
-- Add a partial unique index to ensure system-wide recipes (user_id IS NULL) have unique names.
|
-- Add a partial unique index to ensure system-wide recipes (user_id IS NULL) have unique names.
|
||||||
-- This allows different users to have recipes with the same name.
|
-- This allows different users to have recipes with the same name.
|
||||||
|
-- This index helps speed up sorting for recipe recommendations.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_recipes_rating_sort ON public.recipes (avg_rating DESC, rating_count DESC);
|
||||||
CREATE UNIQUE INDEX IF NOT EXISTS idx_recipes_unique_system_recipe_name ON public.recipes(name) WHERE user_id IS NULL;
|
CREATE UNIQUE INDEX IF NOT EXISTS idx_recipes_unique_system_recipe_name ON public.recipes(name) WHERE user_id IS NULL;
|
||||||
|
|
||||||
-- 27. For ingredients required for each recipe.
|
-- 27. For ingredients required for each recipe.
|
||||||
@@ -514,10 +563,11 @@ CREATE TABLE IF NOT EXISTS public.recipe_ingredients (
|
|||||||
recipe_ingredient_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
recipe_ingredient_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
recipe_id BIGINT NOT NULL REFERENCES public.recipes(recipe_id) ON DELETE CASCADE,
|
recipe_id BIGINT NOT NULL REFERENCES public.recipes(recipe_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity > 0),
|
||||||
unit TEXT NOT NULL,
|
unit TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_ingredients_unit_check CHECK (TRIM(unit) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_ingredients IS 'Defines the ingredients and quantities needed for a recipe.';
|
COMMENT ON TABLE public.recipe_ingredients IS 'Defines the ingredients and quantities needed for a recipe.';
|
||||||
COMMENT ON COLUMN public.recipe_ingredients.unit IS 'e.g., "cups", "tbsp", "g", "each".';
|
COMMENT ON COLUMN public.recipe_ingredients.unit IS 'e.g., "cups", "tbsp", "g", "each".';
|
||||||
@@ -544,7 +594,8 @@ CREATE TABLE IF NOT EXISTS public.tags (
|
|||||||
tag_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
tag_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT tags_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.tags IS 'Stores tags for categorizing recipes, e.g., "Vegetarian", "Quick & Easy".';
|
COMMENT ON TABLE public.tags IS 'Stores tags for categorizing recipes, e.g., "Vegetarian", "Quick & Easy".';
|
||||||
|
|
||||||
@@ -566,7 +617,8 @@ CREATE TABLE IF NOT EXISTS public.appliances (
|
|||||||
appliance_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
appliance_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT appliances_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.appliances IS 'A predefined list of kitchen appliances (e.g., Air Fryer, Instant Pot).';
|
COMMENT ON TABLE public.appliances IS 'A predefined list of kitchen appliances (e.g., Air Fryer, Instant Pot).';
|
||||||
|
|
||||||
@@ -606,7 +658,8 @@ CREATE TABLE IF NOT EXISTS public.recipe_comments (
|
|||||||
content TEXT NOT NULL,
|
content TEXT NOT NULL,
|
||||||
status TEXT DEFAULT 'visible' NOT NULL CHECK (status IN ('visible', 'hidden', 'reported')),
|
status TEXT DEFAULT 'visible' NOT NULL CHECK (status IN ('visible', 'hidden', 'reported')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_comments_content_check CHECK (TRIM(content) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_comments IS 'Allows for threaded discussions and comments on recipes.';
|
COMMENT ON TABLE public.recipe_comments IS 'Allows for threaded discussions and comments on recipes.';
|
||||||
COMMENT ON COLUMN public.recipe_comments.parent_comment_id IS 'For threaded comments.';
|
COMMENT ON COLUMN public.recipe_comments.parent_comment_id IS 'For threaded comments.';
|
||||||
@@ -620,7 +673,8 @@ CREATE TABLE IF NOT EXISTS public.pantry_locations (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT pantry_locations_name_check CHECK (TRIM(name) <> ''),
|
||||||
UNIQUE(user_id, name)
|
UNIQUE(user_id, name)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.pantry_locations IS 'User-defined locations for organizing pantry items (e.g., "Fridge", "Freezer", "Spice Rack").';
|
COMMENT ON TABLE public.pantry_locations IS 'User-defined locations for organizing pantry items (e.g., "Fridge", "Freezer", "Spice Rack").';
|
||||||
@@ -634,7 +688,8 @@ CREATE TABLE IF NOT EXISTS public.planned_meals (
|
|||||||
plan_date DATE NOT NULL,
|
plan_date DATE NOT NULL,
|
||||||
meal_type TEXT NOT NULL,
|
meal_type TEXT NOT NULL,
|
||||||
servings_to_cook INTEGER,
|
servings_to_cook INTEGER,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT planned_meals_meal_type_check CHECK (TRIM(meal_type) <> ''),
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.planned_meals IS 'Assigns a recipe to a specific day and meal type within a user''s menu plan.';
|
COMMENT ON TABLE public.planned_meals IS 'Assigns a recipe to a specific day and meal type within a user''s menu plan.';
|
||||||
@@ -647,7 +702,7 @@ CREATE TABLE IF NOT EXISTS public.pantry_items (
|
|||||||
pantry_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
pantry_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity >= 0),
|
||||||
unit TEXT,
|
unit TEXT,
|
||||||
best_before_date DATE,
|
best_before_date DATE,
|
||||||
pantry_location_id BIGINT REFERENCES public.pantry_locations(pantry_location_id) ON DELETE SET NULL,
|
pantry_location_id BIGINT REFERENCES public.pantry_locations(pantry_location_id) ON DELETE SET NULL,
|
||||||
@@ -670,7 +725,8 @@ CREATE TABLE IF NOT EXISTS public.password_reset_tokens (
|
|||||||
token_hash TEXT NOT NULL UNIQUE,
|
token_hash TEXT NOT NULL UNIQUE,
|
||||||
expires_at TIMESTAMPTZ NOT NULL,
|
expires_at TIMESTAMPTZ NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT password_reset_tokens_token_hash_check CHECK (TRIM(token_hash) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.password_reset_tokens IS 'Stores secure, single-use tokens for password reset requests.';
|
COMMENT ON TABLE public.password_reset_tokens IS 'Stores secure, single-use tokens for password reset requests.';
|
||||||
COMMENT ON COLUMN public.password_reset_tokens.token_hash IS 'A bcrypt hash of the reset token sent to the user.';
|
COMMENT ON COLUMN public.password_reset_tokens.token_hash IS 'A bcrypt hash of the reset token sent to the user.';
|
||||||
@@ -685,10 +741,13 @@ CREATE TABLE IF NOT EXISTS public.unit_conversions (
|
|||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
from_unit TEXT NOT NULL,
|
from_unit TEXT NOT NULL,
|
||||||
to_unit TEXT NOT NULL,
|
to_unit TEXT NOT NULL,
|
||||||
factor NUMERIC NOT NULL,
|
factor NUMERIC NOT NULL CHECK (factor > 0),
|
||||||
UNIQUE(master_item_id, from_unit, to_unit),
|
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
UNIQUE(master_item_id, from_unit, to_unit),
|
||||||
|
CONSTRAINT unit_conversions_from_unit_check CHECK (TRIM(from_unit) <> ''),
|
||||||
|
CONSTRAINT unit_conversions_to_unit_check CHECK (TRIM(to_unit) <> ''),
|
||||||
|
CONSTRAINT unit_conversions_units_check CHECK (from_unit <> to_unit)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.unit_conversions IS 'Stores item-specific unit conversion factors (e.g., grams of flour to cups).';
|
COMMENT ON TABLE public.unit_conversions IS 'Stores item-specific unit conversion factors (e.g., grams of flour to cups).';
|
||||||
COMMENT ON COLUMN public.unit_conversions.factor IS 'The multiplication factor to convert from_unit to to_unit.';
|
COMMENT ON COLUMN public.unit_conversions.factor IS 'The multiplication factor to convert from_unit to to_unit.';
|
||||||
@@ -700,9 +759,10 @@ CREATE TABLE IF NOT EXISTS public.user_item_aliases (
|
|||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||||
alias TEXT NOT NULL,
|
alias TEXT NOT NULL,
|
||||||
UNIQUE(user_id, alias),
|
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
UNIQUE(user_id, alias),
|
||||||
|
CONSTRAINT user_item_aliases_alias_check CHECK (TRIM(alias) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.user_item_aliases IS 'Allows users to create personal aliases for grocery items (e.g., "Dad''s Cereal").';
|
COMMENT ON TABLE public.user_item_aliases IS 'Allows users to create personal aliases for grocery items (e.g., "Dad''s Cereal").';
|
||||||
CREATE INDEX IF NOT EXISTS idx_user_item_aliases_user_id ON public.user_item_aliases(user_id);
|
CREATE INDEX IF NOT EXISTS idx_user_item_aliases_user_id ON public.user_item_aliases(user_id);
|
||||||
@@ -739,7 +799,8 @@ CREATE TABLE IF NOT EXISTS public.recipe_collections (
|
|||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
description TEXT,
|
description TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT recipe_collections_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.recipe_collections IS 'Allows users to create personal collections of recipes (e.g., "Holiday Baking").';
|
COMMENT ON TABLE public.recipe_collections IS 'Allows users to create personal collections of recipes (e.g., "Holiday Baking").';
|
||||||
CREATE INDEX IF NOT EXISTS idx_recipe_collections_user_id ON public.recipe_collections(user_id);
|
CREATE INDEX IF NOT EXISTS idx_recipe_collections_user_id ON public.recipe_collections(user_id);
|
||||||
@@ -764,8 +825,11 @@ CREATE TABLE IF NOT EXISTS public.shared_recipe_collections (
|
|||||||
shared_with_user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
shared_with_user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
permission_level TEXT NOT NULL CHECK (permission_level IN ('view', 'edit')),
|
permission_level TEXT NOT NULL CHECK (permission_level IN ('view', 'edit')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
UNIQUE(recipe_collection_id, shared_with_user_id)
|
UNIQUE(recipe_collection_id, shared_with_user_id)
|
||||||
);
|
);
|
||||||
|
-- This index is crucial for efficiently finding all collections shared with a specific user.
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_shared_recipe_collections_shared_with ON public.shared_recipe_collections(shared_with_user_id);
|
||||||
|
|
||||||
-- 45. Log user search queries for analysis.
|
-- 45. Log user search queries for analysis.
|
||||||
CREATE TABLE IF NOT EXISTS public.search_queries (
|
CREATE TABLE IF NOT EXISTS public.search_queries (
|
||||||
@@ -775,7 +839,8 @@ CREATE TABLE IF NOT EXISTS public.search_queries (
|
|||||||
result_count INTEGER,
|
result_count INTEGER,
|
||||||
was_successful BOOLEAN,
|
was_successful BOOLEAN,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT search_queries_query_text_check CHECK (TRIM(query_text) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.search_queries IS 'Logs user search queries to analyze search effectiveness and identify gaps in data.';
|
COMMENT ON TABLE public.search_queries IS 'Logs user search queries to analyze search effectiveness and identify gaps in data.';
|
||||||
COMMENT ON COLUMN public.search_queries.was_successful IS 'Indicates if the user interacted with a search result.';
|
COMMENT ON COLUMN public.search_queries.was_successful IS 'Indicates if the user interacted with a search result.';
|
||||||
@@ -801,10 +866,11 @@ CREATE TABLE IF NOT EXISTS public.shopping_trip_items (
|
|||||||
shopping_trip_id BIGINT NOT NULL REFERENCES public.shopping_trips(shopping_trip_id) ON DELETE CASCADE,
|
shopping_trip_id BIGINT NOT NULL REFERENCES public.shopping_trips(shopping_trip_id) ON DELETE CASCADE,
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
custom_item_name TEXT,
|
custom_item_name TEXT,
|
||||||
quantity NUMERIC NOT NULL,
|
quantity NUMERIC NOT NULL CHECK (quantity > 0),
|
||||||
price_paid_cents INTEGER,
|
price_paid_cents INTEGER,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT shopping_trip_items_custom_item_name_check CHECK (custom_item_name IS NULL OR TRIM(custom_item_name) <> ''),
|
||||||
CONSTRAINT trip_must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
CONSTRAINT trip_must_have_item_identifier CHECK (master_item_id IS NOT NULL OR custom_item_name IS NOT NULL)
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.shopping_trip_items IS 'A historical log of items purchased during a shopping trip.';
|
COMMENT ON TABLE public.shopping_trip_items IS 'A historical log of items purchased during a shopping trip.';
|
||||||
@@ -818,7 +884,8 @@ CREATE TABLE IF NOT EXISTS public.dietary_restrictions (
|
|||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
type TEXT NOT NULL CHECK (type IN ('diet', 'allergy')),
|
type TEXT NOT NULL CHECK (type IN ('diet', 'allergy')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT dietary_restrictions_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.dietary_restrictions IS 'A predefined list of common diets (e.g., Vegan) and allergies (e.g., Nut Allergy).';
|
COMMENT ON TABLE public.dietary_restrictions IS 'A predefined list of common diets (e.g., Vegan) and allergies (e.g., Nut Allergy).';
|
||||||
|
|
||||||
@@ -868,11 +935,12 @@ CREATE TABLE IF NOT EXISTS public.receipts (
|
|||||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE CASCADE,
|
||||||
receipt_image_url TEXT NOT NULL,
|
receipt_image_url TEXT NOT NULL,
|
||||||
transaction_date TIMESTAMPTZ,
|
transaction_date TIMESTAMPTZ,
|
||||||
total_amount_cents INTEGER,
|
total_amount_cents INTEGER CHECK (total_amount_cents IS NULL OR total_amount_cents >= 0),
|
||||||
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'processing', 'completed', 'failed')),
|
status TEXT DEFAULT 'pending' NOT NULL CHECK (status IN ('pending', 'processing', 'completed', 'failed')),
|
||||||
raw_text TEXT,
|
raw_text TEXT,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
processed_at TIMESTAMPTZ,
|
processed_at TIMESTAMPTZ,
|
||||||
|
CONSTRAINT receipts_receipt_image_url_check CHECK (receipt_image_url ~* '^https://?.*'),
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.receipts IS 'Stores uploaded user receipts for purchase tracking and analysis.';
|
COMMENT ON TABLE public.receipts IS 'Stores uploaded user receipts for purchase tracking and analysis.';
|
||||||
@@ -884,13 +952,14 @@ CREATE TABLE IF NOT EXISTS public.receipt_items (
|
|||||||
receipt_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
receipt_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
receipt_id BIGINT NOT NULL REFERENCES public.receipts(receipt_id) ON DELETE CASCADE,
|
receipt_id BIGINT NOT NULL REFERENCES public.receipts(receipt_id) ON DELETE CASCADE,
|
||||||
raw_item_description TEXT NOT NULL,
|
raw_item_description TEXT NOT NULL,
|
||||||
quantity NUMERIC DEFAULT 1 NOT NULL,
|
quantity NUMERIC DEFAULT 1 NOT NULL CHECK (quantity > 0),
|
||||||
price_paid_cents INTEGER NOT NULL,
|
price_paid_cents INTEGER NOT NULL CHECK (price_paid_cents >= 0),
|
||||||
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
master_item_id BIGINT REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE SET NULL,
|
||||||
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
product_id BIGINT REFERENCES public.products(product_id) ON DELETE SET NULL,
|
||||||
status TEXT DEFAULT 'unmatched' NOT NULL CHECK (status IN ('unmatched', 'matched', 'needs_review', 'ignored')),
|
status TEXT DEFAULT 'unmatched' NOT NULL CHECK (status IN ('unmatched', 'matched', 'needs_review', 'ignored')),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT receipt_items_raw_item_description_check CHECK (TRIM(raw_item_description) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.receipt_items IS 'Stores individual line items extracted from a user receipt.';
|
COMMENT ON TABLE public.receipt_items IS 'Stores individual line items extracted from a user receipt.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_receipt_items_receipt_id ON public.receipt_items(receipt_id);
|
CREATE INDEX IF NOT EXISTS idx_receipt_items_receipt_id ON public.receipt_items(receipt_id);
|
||||||
@@ -929,11 +998,12 @@ CREATE TABLE IF NOT EXISTS public.budgets (
|
|||||||
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
budget_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||||
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
user_id UUID NOT NULL REFERENCES public.users(user_id) ON DELETE CASCADE,
|
||||||
name TEXT NOT NULL,
|
name TEXT NOT NULL,
|
||||||
amount_cents INTEGER NOT NULL,
|
amount_cents INTEGER NOT NULL CHECK (amount_cents > 0),
|
||||||
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
period TEXT NOT NULL CHECK (period IN ('weekly', 'monthly')),
|
||||||
start_date DATE NOT NULL,
|
start_date DATE NOT NULL,
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT budgets_name_check CHECK (TRIM(name) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
COMMENT ON TABLE public.budgets IS 'Allows users to set weekly or monthly grocery budgets for spending tracking.';
|
||||||
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
CREATE INDEX IF NOT EXISTS idx_budgets_user_id ON public.budgets(user_id);
|
||||||
@@ -944,8 +1014,10 @@ CREATE TABLE IF NOT EXISTS public.achievements (
|
|||||||
name TEXT NOT NULL UNIQUE,
|
name TEXT NOT NULL UNIQUE,
|
||||||
description TEXT NOT NULL,
|
description TEXT NOT NULL,
|
||||||
icon TEXT,
|
icon TEXT,
|
||||||
points_value INTEGER NOT NULL DEFAULT 0,
|
points_value INTEGER NOT NULL DEFAULT 0 CHECK (points_value >= 0),
|
||||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||||
|
CONSTRAINT achievements_name_check CHECK (TRIM(name) <> ''),
|
||||||
|
CONSTRAINT achievements_description_check CHECK (TRIM(description) <> '')
|
||||||
);
|
);
|
||||||
COMMENT ON TABLE public.achievements IS 'A static table defining the available achievements users can earn.';
|
COMMENT ON TABLE public.achievements IS 'A static table defining the available achievements users can earn.';
|
||||||
|
|
||||||
@@ -1176,7 +1248,8 @@ INSERT INTO public.achievements (name, description, icon, points_value) VALUES
|
|||||||
('List Sharer', 'Share a shopping list with another user for the first time.', 'list', 20),
|
('List Sharer', 'Share a shopping list with another user for the first time.', 'list', 20),
|
||||||
('First Favorite', 'Mark a recipe as one of your favorites.', 'heart', 5),
|
('First Favorite', 'Mark a recipe as one of your favorites.', 'heart', 5),
|
||||||
('First Fork', 'Make a personal copy of a public recipe.', 'git-fork', 10),
|
('First Fork', 'Make a personal copy of a public recipe.', 'git-fork', 10),
|
||||||
('First Budget Created', 'Create your first budget to track spending.', 'piggy-bank', 15)
|
('First Budget Created', 'Create your first budget to track spending.', 'piggy-bank', 15),
|
||||||
|
('First-Upload', 'Upload your first flyer.', 'upload-cloud', 25)
|
||||||
ON CONFLICT (name) DO NOTHING;
|
ON CONFLICT (name) DO NOTHING;
|
||||||
|
|
||||||
-- ============================================================================
|
-- ============================================================================
|
||||||
@@ -2485,16 +2558,21 @@ DROP FUNCTION IF EXISTS public.log_new_flyer();
|
|||||||
CREATE OR REPLACE FUNCTION public.log_new_flyer()
|
CREATE OR REPLACE FUNCTION public.log_new_flyer()
|
||||||
RETURNS TRIGGER AS $$
|
RETURNS TRIGGER AS $$
|
||||||
BEGIN
|
BEGIN
|
||||||
INSERT INTO public.activity_log (action, display_text, icon, details)
|
-- If the flyer was uploaded by a registered user, award the 'First-Upload' achievement.
|
||||||
|
-- The award_achievement function handles checking if the user already has it.
|
||||||
|
IF NEW.uploaded_by IS NOT NULL THEN
|
||||||
|
PERFORM public.award_achievement(NEW.uploaded_by, 'First-Upload');
|
||||||
|
END IF;
|
||||||
|
|
||||||
|
INSERT INTO public.activity_log (user_id, action, display_text, icon, details)
|
||||||
VALUES (
|
VALUES (
|
||||||
|
NEW.uploaded_by, -- Log the user who uploaded it
|
||||||
'flyer_uploaded',
|
'flyer_uploaded',
|
||||||
'A new flyer for ' || (SELECT name FROM public.stores WHERE store_id = NEW.store_id) || ' has been uploaded.',
|
'A new flyer for ' || (SELECT name FROM public.stores WHERE store_id = NEW.store_id) || ' has been uploaded.',
|
||||||
'file-text',
|
'file-text',
|
||||||
jsonb_build_object(
|
jsonb_build_object(
|
||||||
'flyer_id', NEW.flyer_id,
|
'flyer_id', NEW.flyer_id,
|
||||||
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id),
|
'store_name', (SELECT name FROM public.stores WHERE store_id = NEW.store_id)
|
||||||
'valid_from', to_char(NEW.valid_from, 'YYYY-MM-DD'),
|
|
||||||
'valid_to', to_char(NEW.valid_to, 'YYYY-MM-DD')
|
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
RETURN NEW;
|
RETURN NEW;
|
||||||
@@ -2601,6 +2679,7 @@ CREATE TRIGGER on_new_recipe_collection_share
|
|||||||
CREATE OR REPLACE FUNCTION public.get_best_sale_prices_for_all_users()
|
CREATE OR REPLACE FUNCTION public.get_best_sale_prices_for_all_users()
|
||||||
RETURNS TABLE(
|
RETURNS TABLE(
|
||||||
user_id uuid,
|
user_id uuid,
|
||||||
|
|
||||||
email text,
|
email text,
|
||||||
full_name text,
|
full_name text,
|
||||||
master_item_id integer,
|
master_item_id integer,
|
||||||
@@ -2615,6 +2694,7 @@ BEGIN
|
|||||||
WITH
|
WITH
|
||||||
-- Step 1: Find all flyer items that are currently on sale and have a valid price.
|
-- Step 1: Find all flyer items that are currently on sale and have a valid price.
|
||||||
current_sales AS (
|
current_sales AS (
|
||||||
|
|
||||||
SELECT
|
SELECT
|
||||||
fi.master_item_id,
|
fi.master_item_id,
|
||||||
fi.price_in_cents,
|
fi.price_in_cents,
|
||||||
@@ -2623,14 +2703,18 @@ BEGIN
|
|||||||
f.valid_to
|
f.valid_to
|
||||||
FROM public.flyer_items fi
|
FROM public.flyer_items fi
|
||||||
JOIN public.flyers f ON fi.flyer_id = f.flyer_id
|
JOIN public.flyers f ON fi.flyer_id = f.flyer_id
|
||||||
|
JOIN public.stores s ON f.store_id = s.store_id
|
||||||
WHERE
|
WHERE
|
||||||
|
|
||||||
fi.master_item_id IS NOT NULL
|
fi.master_item_id IS NOT NULL
|
||||||
AND fi.price_in_cents IS NOT NULL
|
AND fi.price_in_cents IS NOT NULL
|
||||||
AND f.valid_to >= CURRENT_DATE
|
AND f.valid_to >= CURRENT_DATE
|
||||||
),
|
),
|
||||||
-- Step 2: For each master item, find its absolute best (lowest) price across all current sales.
|
-- Step 2: For each master item, find its absolute best (lowest) price across all current sales.
|
||||||
-- We use a window function to rank the sales for each item by price.
|
-- We use a window function to rank the sales for each item by price.
|
||||||
|
|
||||||
best_prices AS (
|
best_prices AS (
|
||||||
|
|
||||||
SELECT
|
SELECT
|
||||||
cs.master_item_id,
|
cs.master_item_id,
|
||||||
cs.price_in_cents AS best_price_in_cents,
|
cs.price_in_cents AS best_price_in_cents,
|
||||||
@@ -2643,6 +2727,7 @@ BEGIN
|
|||||||
)
|
)
|
||||||
-- Step 3: Join the best-priced items with the user watchlist and user details.
|
-- Step 3: Join the best-priced items with the user watchlist and user details.
|
||||||
SELECT
|
SELECT
|
||||||
|
|
||||||
u.user_id,
|
u.user_id,
|
||||||
u.email,
|
u.email,
|
||||||
p.full_name,
|
p.full_name,
|
||||||
@@ -2662,6 +2747,7 @@ BEGIN
|
|||||||
JOIN public.master_grocery_items mgi ON bp.master_item_id = mgi.master_grocery_item_id
|
JOIN public.master_grocery_items mgi ON bp.master_item_id = mgi.master_grocery_item_id
|
||||||
WHERE
|
WHERE
|
||||||
-- Only include the items that are at their absolute best price (rank = 1).
|
-- Only include the items that are at their absolute best price (rank = 1).
|
||||||
|
|
||||||
bp.price_rank = 1;
|
bp.price_rank = 1;
|
||||||
END;
|
END;
|
||||||
$$ LANGUAGE plpgsql;
|
$$ LANGUAGE plpgsql;
|
||||||
|
|||||||
33
src/components/Dashboard.tsx
Normal file
33
src/components/Dashboard.tsx
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import { RecipeSuggester } from '../components/RecipeSuggester';
|
||||||
|
import { FlyerCountDisplay } from '../components/FlyerCountDisplay';
|
||||||
|
import { Leaderboard } from '../components/Leaderboard';
|
||||||
|
|
||||||
|
export const Dashboard: React.FC = () => {
|
||||||
|
return (
|
||||||
|
<div className="max-w-7xl mx-auto px-4 sm:px-6 lg:px-8 py-8">
|
||||||
|
<h1 className="text-2xl font-bold text-gray-900 dark:text-white mb-6">Dashboard</h1>
|
||||||
|
|
||||||
|
<div className="grid grid-cols-1 lg:grid-cols-3 gap-6">
|
||||||
|
{/* Main Content Area */}
|
||||||
|
<div className="lg:col-span-2 space-y-6">
|
||||||
|
{/* Recipe Suggester Section */}
|
||||||
|
<RecipeSuggester />
|
||||||
|
|
||||||
|
{/* Other Dashboard Widgets */}
|
||||||
|
<div className="bg-white dark:bg-gray-800 shadow rounded-lg p-6">
|
||||||
|
<h2 className="text-lg font-medium text-gray-900 dark:text-white mb-4">Your Flyers</h2>
|
||||||
|
<FlyerCountDisplay />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{/* Sidebar Area */}
|
||||||
|
<div className="space-y-6">
|
||||||
|
<Leaderboard />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
|
export default Dashboard;
|
||||||
80
src/components/RecipeSuggester.tsx
Normal file
80
src/components/RecipeSuggester.tsx
Normal file
@@ -0,0 +1,80 @@
|
|||||||
|
// src/components/RecipeSuggester.tsx
|
||||||
|
import React, { useState, useCallback } from 'react';
|
||||||
|
import { suggestRecipe } from '../services/apiClient';
|
||||||
|
import { logger } from '../services/logger.client';
|
||||||
|
|
||||||
|
export const RecipeSuggester: React.FC = () => {
|
||||||
|
const [ingredients, setIngredients] = useState<string>('');
|
||||||
|
const [suggestion, setSuggestion] = useState<string | null>(null);
|
||||||
|
const [isLoading, setIsLoading] = useState<boolean>(false);
|
||||||
|
const [error, setError] = useState<string | null>(null);
|
||||||
|
|
||||||
|
const handleSubmit = useCallback(async (event: React.FormEvent<HTMLFormElement>) => {
|
||||||
|
event.preventDefault();
|
||||||
|
setIsLoading(true);
|
||||||
|
setError(null);
|
||||||
|
setSuggestion(null);
|
||||||
|
|
||||||
|
const ingredientList = ingredients.split(',').map(item => item.trim()).filter(Boolean);
|
||||||
|
|
||||||
|
if (ingredientList.length === 0) {
|
||||||
|
setError('Please enter at least one ingredient.');
|
||||||
|
setIsLoading(false);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
const response = await suggestRecipe(ingredientList);
|
||||||
|
const data = await response.json();
|
||||||
|
|
||||||
|
if (!response.ok) {
|
||||||
|
throw new Error(data.message || 'Failed to get suggestion.');
|
||||||
|
}
|
||||||
|
|
||||||
|
setSuggestion(data.suggestion);
|
||||||
|
} catch (err) {
|
||||||
|
const errorMessage = err instanceof Error ? err.message : 'An unknown error occurred.';
|
||||||
|
logger.error({ error: err }, 'Failed to fetch recipe suggestion.');
|
||||||
|
setError(errorMessage);
|
||||||
|
} finally {
|
||||||
|
setIsLoading(false);
|
||||||
|
}
|
||||||
|
}, [ingredients]);
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className="bg-white dark:bg-gray-800 shadow rounded-lg p-6">
|
||||||
|
<h2 className="text-xl font-semibold text-gray-900 dark:text-white mb-2">Get a Recipe Suggestion</h2>
|
||||||
|
<p className="text-gray-600 dark:text-gray-400 mb-4">Enter some ingredients you have, separated by commas.</p>
|
||||||
|
<form onSubmit={handleSubmit}>
|
||||||
|
<div className="mb-4">
|
||||||
|
<label htmlFor="ingredients-input" className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">Ingredients:</label>
|
||||||
|
<input
|
||||||
|
id="ingredients-input"
|
||||||
|
type="text"
|
||||||
|
value={ingredients}
|
||||||
|
onChange={(e) => setIngredients(e.target.value)}
|
||||||
|
placeholder="e.g., chicken, rice, broccoli"
|
||||||
|
disabled={isLoading}
|
||||||
|
className="block w-full rounded-md border-gray-300 shadow-sm focus:border-blue-500 focus:ring-blue-500 dark:bg-gray-700 dark:border-gray-600 dark:text-white sm:text-sm p-2 border"
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
<button type="submit" disabled={isLoading} className="w-full flex justify-center py-2 px-4 border border-transparent rounded-md shadow-sm text-sm font-medium text-white bg-blue-600 hover:bg-blue-700 focus:outline-none focus:ring-2 focus:ring-offset-2 focus:ring-blue-500 disabled:opacity-50 transition-colors">
|
||||||
|
{isLoading ? 'Getting suggestion...' : 'Suggest a Recipe'}
|
||||||
|
</button>
|
||||||
|
</form>
|
||||||
|
|
||||||
|
{error && (
|
||||||
|
<div className="mt-4 p-4 bg-red-50 dark:bg-red-900/50 text-red-700 dark:text-red-200 rounded-md text-sm">{error}</div>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{suggestion && (
|
||||||
|
<div className="mt-6 bg-gray-50 dark:bg-gray-700/50 rounded-lg p-4 border border-gray-200 dark:border-gray-600">
|
||||||
|
<div className="prose dark:prose-invert max-w-none">
|
||||||
|
<h5 className="text-lg font-medium text-gray-900 dark:text-white mb-2">Recipe Suggestion</h5>
|
||||||
|
<p className="text-gray-700 dark:text-gray-300 whitespace-pre-wrap">{suggestion}</p>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
};
|
||||||
@@ -111,7 +111,7 @@ async function main() {
|
|||||||
|
|
||||||
const flyerQuery = `
|
const flyerQuery = `
|
||||||
INSERT INTO public.flyers (file_name, image_url, checksum, store_id, valid_from, valid_to)
|
INSERT INTO public.flyers (file_name, image_url, checksum, store_id, valid_from, valid_to)
|
||||||
VALUES ('safeway-flyer.jpg', '/sample-assets/safeway-flyer.jpg', 'sample-checksum-123', ${storeMap.get('Safeway')}, $1, $2)
|
VALUES ('safeway-flyer.jpg', 'https://example.com/flyer-images/safeway-flyer.jpg', 'a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0a0', ${storeMap.get('Safeway')}, $1, $2)
|
||||||
RETURNING flyer_id;
|
RETURNING flyer_id;
|
||||||
`;
|
`;
|
||||||
const flyerRes = await client.query<{ flyer_id: number }>(flyerQuery, [
|
const flyerRes = await client.query<{ flyer_id: number }>(flyerQuery, [
|
||||||
|
|||||||
@@ -15,7 +15,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
|||||||
// FIX: Stabilize the apiFunction passed to useApi.
|
// FIX: Stabilize the apiFunction passed to useApi.
|
||||||
// By wrapping this in useCallback, we ensure the same function instance is passed to
|
// By wrapping this in useCallback, we ensure the same function instance is passed to
|
||||||
// useApi on every render. This prevents the `execute` function returned by `useApi`
|
// useApi on every render. This prevents the `execute` function returned by `useApi`
|
||||||
// from being recreated, which in turn breaks the infinite re-render loop in the useEffect below.
|
// from being recreated, which in turn breaks the infinite re-render loop in the useEffect.
|
||||||
const getProfileCallback = useCallback(() => apiClient.getAuthenticatedUserProfile(), []);
|
const getProfileCallback = useCallback(() => apiClient.getAuthenticatedUserProfile(), []);
|
||||||
|
|
||||||
const { execute: checkTokenApi } = useApi<UserProfile, []>(getProfileCallback);
|
const { execute: checkTokenApi } = useApi<UserProfile, []>(getProfileCallback);
|
||||||
|
|||||||
@@ -4,17 +4,21 @@ import { FlyersContext, FlyersContextType } from '../contexts/FlyersContext';
|
|||||||
import type { Flyer } from '../types';
|
import type { Flyer } from '../types';
|
||||||
import * as apiClient from '../services/apiClient';
|
import * as apiClient from '../services/apiClient';
|
||||||
import { useInfiniteQuery } from '../hooks/useInfiniteQuery';
|
import { useInfiniteQuery } from '../hooks/useInfiniteQuery';
|
||||||
|
import { useCallback } from 'react';
|
||||||
|
|
||||||
export const FlyersProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
export const FlyersProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
||||||
|
// Memoize the fetch function to ensure stability for the useInfiniteQuery hook.
|
||||||
|
const fetchFlyersFn = useCallback(apiClient.fetchFlyers, []);
|
||||||
|
|
||||||
const {
|
const {
|
||||||
data: flyers,
|
data: flyers,
|
||||||
isLoading: isLoadingFlyers,
|
isLoading: isLoadingFlyers,
|
||||||
error: flyersError,
|
error: flyersError,
|
||||||
fetchNextPage: fetchNextFlyersPage,
|
fetchNextPage: fetchNextFlyersPage,
|
||||||
hasNextPage: hasNextFlyersPage,
|
hasNextPage: hasNextFlyersPage,
|
||||||
refetch: refetchFlyers,
|
refetch: refetchFlyers,
|
||||||
isRefetching: isRefetchingFlyers,
|
isRefetching: isRefetchingFlyers,
|
||||||
} = useInfiniteQuery<Flyer>(apiClient.fetchFlyers);
|
} = useInfiniteQuery<Flyer>(fetchFlyersFn);
|
||||||
|
|
||||||
const value: FlyersContextType = {
|
const value: FlyersContextType = {
|
||||||
flyers: flyers || [],
|
flyers: flyers || [],
|
||||||
@@ -26,5 +30,5 @@ export const FlyersProvider: React.FC<{ children: ReactNode }> = ({ children })
|
|||||||
refetchFlyers,
|
refetchFlyers,
|
||||||
};
|
};
|
||||||
|
|
||||||
return <FlyersContext.Provider value={value}>{children}</FlyersContext.Provider>;
|
return <FlyersContext.Provider value={value}>{children}</FlyersContext.Provider>;
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,14 +1,22 @@
|
|||||||
// src/providers/MasterItemsProvider.tsx
|
// src/providers/MasterItemsProvider.tsx
|
||||||
import React, { ReactNode, useMemo } from 'react';
|
import React, { ReactNode, useMemo, useEffect, useCallback } from 'react';
|
||||||
import { MasterItemsContext } from '../contexts/MasterItemsContext';
|
import { MasterItemsContext } from '../contexts/MasterItemsContext';
|
||||||
import type { MasterGroceryItem } from '../types';
|
import type { MasterGroceryItem } from '../types';
|
||||||
import * as apiClient from '../services/apiClient';
|
import * as apiClient from '../services/apiClient';
|
||||||
import { useApiOnMount } from '../hooks/useApiOnMount';
|
import { useApiOnMount } from '../hooks/useApiOnMount';
|
||||||
|
import { logger } from '../services/logger.client';
|
||||||
|
|
||||||
export const MasterItemsProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
export const MasterItemsProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
||||||
const { data, loading, error } = useApiOnMount<MasterGroceryItem[], []>(() =>
|
// LOGGING: Check if the provider is unmounting/remounting repeatedly
|
||||||
apiClient.fetchMasterItems(),
|
useEffect(() => {
|
||||||
);
|
logger.debug('MasterItemsProvider: MOUNTED');
|
||||||
|
return () => logger.debug('MasterItemsProvider: UNMOUNTED');
|
||||||
|
}, []);
|
||||||
|
|
||||||
|
// Memoize the fetch function to ensure stability for the useApiOnMount hook.
|
||||||
|
const fetchFn = useCallback(() => apiClient.fetchMasterItems(), []);
|
||||||
|
|
||||||
|
const { data, loading, error } = useApiOnMount<MasterGroceryItem[], []>(fetchFn);
|
||||||
|
|
||||||
const value = useMemo(
|
const value = useMemo(
|
||||||
() => ({
|
() => ({
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
// src/providers/UserDataProvider.tsx
|
// src/providers/UserDataProvider.tsx
|
||||||
import React, { useState, useEffect, useMemo, ReactNode } from 'react';
|
import { logger } from '../services/logger.client';
|
||||||
|
import React, { useState, useEffect, useMemo, ReactNode, useCallback } from 'react';
|
||||||
import { UserDataContext } from '../contexts/UserDataContext';
|
import { UserDataContext } from '../contexts/UserDataContext';
|
||||||
import type { MasterGroceryItem, ShoppingList } from '../types';
|
import type { MasterGroceryItem, ShoppingList } from '../types';
|
||||||
import * as apiClient from '../services/apiClient';
|
import * as apiClient from '../services/apiClient';
|
||||||
@@ -9,18 +10,25 @@ import { useAuth } from '../hooks/useAuth';
|
|||||||
export const UserDataProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
export const UserDataProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
||||||
const { userProfile } = useAuth();
|
const { userProfile } = useAuth();
|
||||||
|
|
||||||
|
// Wrap the API calls in useCallback to prevent unnecessary re-renders.
|
||||||
|
const fetchWatchedItemsFn = useCallback(
|
||||||
|
() => apiClient.fetchWatchedItems(),
|
||||||
|
[],
|
||||||
|
);
|
||||||
|
const fetchShoppingListsFn = useCallback(() => apiClient.fetchShoppingLists(), []);
|
||||||
|
|
||||||
const {
|
const {
|
||||||
data: watchedItemsData,
|
data: watchedItemsData,
|
||||||
loading: isLoadingWatched,
|
loading: isLoadingWatched,
|
||||||
error: watchedItemsError,
|
error: watchedItemsError,
|
||||||
} = useApiOnMount<MasterGroceryItem[], []>(() => apiClient.fetchWatchedItems(), [userProfile], {
|
} = useApiOnMount<MasterGroceryItem[], []>(fetchWatchedItemsFn, [userProfile], {
|
||||||
enabled: !!userProfile,
|
enabled: !!userProfile,
|
||||||
});
|
});
|
||||||
const {
|
const {
|
||||||
data: shoppingListsData,
|
data: shoppingListsData,
|
||||||
loading: isLoadingShoppingLists,
|
loading: isLoadingShoppingLists,
|
||||||
error: shoppingListsError,
|
error: shoppingListsError,
|
||||||
} = useApiOnMount<ShoppingList[], []>(() => apiClient.fetchShoppingLists(), [userProfile], {
|
} = useApiOnMount<ShoppingList[], []>(fetchShoppingListsFn, [userProfile], {
|
||||||
enabled: !!userProfile,
|
enabled: !!userProfile,
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -32,7 +40,7 @@ export const UserDataProvider: React.FC<{ children: ReactNode }> = ({ children }
|
|||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
// When the user logs out (user becomes null), immediately clear all user-specific data.
|
// When the user logs out (user becomes null), immediately clear all user-specific data.
|
||||||
// This also serves to clear out old data when a new user logs in, before their new data arrives.
|
// This also serves to clear out old data when a new user logs in, before their new data arrives.
|
||||||
if (!userProfile) {
|
if (!userProfile) {
|
||||||
setWatchedItems([]);
|
setWatchedItems([]);
|
||||||
setShoppingLists([]);
|
setShoppingLists([]);
|
||||||
return;
|
return;
|
||||||
@@ -60,7 +68,7 @@ export const UserDataProvider: React.FC<{ children: ReactNode }> = ({ children }
|
|||||||
watchedItemsError,
|
watchedItemsError,
|
||||||
shoppingListsError,
|
shoppingListsError,
|
||||||
],
|
],
|
||||||
);
|
);
|
||||||
|
|
||||||
return <UserDataContext.Provider value={value}>{children}</UserDataContext.Provider>;
|
return <UserDataContext.Provider value={value}>{children}</UserDataContext.Provider>;
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -165,6 +165,38 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('should allow registration with an empty string for avatar_url', async () => {
|
||||||
|
// Arrange
|
||||||
|
const email = 'avatar-user@test.com';
|
||||||
|
const mockNewUser = createMockUserProfile({
|
||||||
|
user: { user_id: 'avatar-user-id', email },
|
||||||
|
});
|
||||||
|
mockedAuthService.registerAndLoginUser.mockResolvedValue({
|
||||||
|
newUserProfile: mockNewUser,
|
||||||
|
accessToken: 'avatar-access-token',
|
||||||
|
refreshToken: 'avatar-refresh-token',
|
||||||
|
});
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const response = await supertest(app).post('/api/auth/register').send({
|
||||||
|
email,
|
||||||
|
password: strongPassword,
|
||||||
|
full_name: 'Avatar User',
|
||||||
|
avatar_url: '', // Send an empty string
|
||||||
|
});
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(201);
|
||||||
|
expect(response.body.message).toBe('User registered successfully!');
|
||||||
|
expect(mockedAuthService.registerAndLoginUser).toHaveBeenCalledWith(
|
||||||
|
email,
|
||||||
|
strongPassword,
|
||||||
|
'Avatar User',
|
||||||
|
undefined, // The preprocess step in the Zod schema should convert '' to undefined
|
||||||
|
mockLogger,
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
it('should set a refresh token cookie on successful registration', async () => {
|
it('should set a refresh token cookie on successful registration', async () => {
|
||||||
const mockNewUser = createMockUserProfile({
|
const mockNewUser = createMockUserProfile({
|
||||||
user: { user_id: 'new-user-id', email: 'cookie@test.com' },
|
user: { user_id: 'new-user-id', email: 'cookie@test.com' },
|
||||||
|
|||||||
@@ -51,7 +51,11 @@ const registerSchema = z.object({
|
|||||||
}),
|
}),
|
||||||
// Sanitize optional string inputs.
|
// Sanitize optional string inputs.
|
||||||
full_name: z.string().trim().optional(),
|
full_name: z.string().trim().optional(),
|
||||||
avatar_url: z.string().trim().url().optional(),
|
// Allow empty string or valid URL. If empty string is received, convert to undefined.
|
||||||
|
avatar_url: z.preprocess(
|
||||||
|
(val) => (val === '' ? undefined : val),
|
||||||
|
z.string().trim().url().optional(),
|
||||||
|
),
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -482,8 +482,8 @@ describe('Passport Configuration', () => {
|
|||||||
const mockReq: Partial<Request> = {
|
const mockReq: Partial<Request> = {
|
||||||
// An object that is not a valid UserProfile (e.g., missing 'role')
|
// An object that is not a valid UserProfile (e.g., missing 'role')
|
||||||
user: {
|
user: {
|
||||||
user_id: 'invalid-user-id',
|
user: { user_id: 'invalid-user-id' }, // Missing 'role' property
|
||||||
} as any,
|
} as unknown as UserProfile, // Cast to UserProfile to satisfy req.user type, but it's intentionally malformed
|
||||||
};
|
};
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
|
|||||||
@@ -19,6 +19,12 @@ router.get(
|
|||||||
validateRequest(emptySchema),
|
validateRequest(emptySchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
|
// LOGGING: Track how often this heavy DB call is actually made vs served from cache
|
||||||
|
req.log.info('Fetching master items list from database...');
|
||||||
|
|
||||||
|
// Optimization: This list changes rarely. Instruct clients to cache it for 1 hour (3600s).
|
||||||
|
res.set('Cache-Control', 'public, max-age=3600');
|
||||||
|
|
||||||
const masterItems = await db.personalizationRepo.getAllMasterItems(req.log);
|
const masterItems = await db.personalizationRepo.getAllMasterItems(req.log);
|
||||||
res.json(masterItems);
|
res.json(masterItems);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
|||||||
109
src/routes/reactions.routes.ts
Normal file
109
src/routes/reactions.routes.ts
Normal file
@@ -0,0 +1,109 @@
|
|||||||
|
import { Router, Request, Response, NextFunction } from 'express';
|
||||||
|
import { z } from 'zod';
|
||||||
|
import { reactionRepo } from '../services/db/index.db';
|
||||||
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
|
import passport from './passport.routes';
|
||||||
|
import { requiredString } from '../utils/zodUtils';
|
||||||
|
import { UserProfile } from '../types';
|
||||||
|
|
||||||
|
const router = Router();
|
||||||
|
|
||||||
|
// --- Zod Schemas for Reaction Routes ---
|
||||||
|
|
||||||
|
const getReactionsSchema = z.object({
|
||||||
|
query: z.object({
|
||||||
|
userId: z.string().uuid().optional(),
|
||||||
|
entityType: z.string().optional(),
|
||||||
|
entityId: z.string().optional(),
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
const toggleReactionSchema = z.object({
|
||||||
|
body: z.object({
|
||||||
|
entity_type: requiredString('entity_type is required.'),
|
||||||
|
entity_id: requiredString('entity_id is required.'),
|
||||||
|
reaction_type: requiredString('reaction_type is required.'),
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
const getReactionSummarySchema = z.object({
|
||||||
|
query: z.object({
|
||||||
|
entityType: requiredString('entityType is required.'),
|
||||||
|
entityId: requiredString('entityId is required.'),
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
// --- Routes ---
|
||||||
|
|
||||||
|
/**
|
||||||
|
* GET /api/reactions - Fetches user reactions based on query filters.
|
||||||
|
* Supports filtering by userId, entityType, and entityId.
|
||||||
|
* This is a public endpoint.
|
||||||
|
*/
|
||||||
|
router.get(
|
||||||
|
'/',
|
||||||
|
validateRequest(getReactionsSchema),
|
||||||
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
|
try {
|
||||||
|
const { query } = getReactionsSchema.parse({ query: req.query });
|
||||||
|
const reactions = await reactionRepo.getReactions(query, req.log);
|
||||||
|
res.json(reactions);
|
||||||
|
} catch (error) {
|
||||||
|
req.log.error({ error }, 'Error fetching user reactions');
|
||||||
|
next(error);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* GET /api/reactions/summary - Fetches a summary of reactions for a specific entity.
|
||||||
|
* Example: /api/reactions/summary?entityType=recipe&entityId=123
|
||||||
|
* This is a public endpoint.
|
||||||
|
*/
|
||||||
|
router.get(
|
||||||
|
'/summary',
|
||||||
|
validateRequest(getReactionSummarySchema),
|
||||||
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
|
try {
|
||||||
|
const { query } = getReactionSummarySchema.parse({ query: req.query });
|
||||||
|
const summary = await reactionRepo.getReactionSummary(query.entityType, query.entityId, req.log);
|
||||||
|
res.json(summary);
|
||||||
|
} catch (error) {
|
||||||
|
req.log.error({ error }, 'Error fetching reaction summary');
|
||||||
|
next(error);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /api/reactions/toggle - Toggles a user's reaction to an entity.
|
||||||
|
* This is a protected endpoint.
|
||||||
|
*/
|
||||||
|
router.post(
|
||||||
|
'/toggle',
|
||||||
|
passport.authenticate('jwt', { session: false }),
|
||||||
|
validateRequest(toggleReactionSchema),
|
||||||
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
|
const userProfile = req.user as UserProfile;
|
||||||
|
type ToggleReactionRequest = z.infer<typeof toggleReactionSchema>;
|
||||||
|
const { body } = req as unknown as ToggleReactionRequest;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const reactionData = {
|
||||||
|
user_id: userProfile.user.user_id,
|
||||||
|
...body,
|
||||||
|
};
|
||||||
|
const result = await reactionRepo.toggleReaction(reactionData, req.log);
|
||||||
|
if (result) {
|
||||||
|
res.status(201).json({ message: 'Reaction added.', reaction: result });
|
||||||
|
} else {
|
||||||
|
res.status(200).json({ message: 'Reaction removed.' });
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
req.log.error({ error, body }, 'Error toggling user reaction');
|
||||||
|
next(error);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
export default router;
|
||||||
@@ -2,6 +2,8 @@
|
|||||||
import { Router } from 'express';
|
import { Router } from 'express';
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import * as db from '../services/db/index.db';
|
import * as db from '../services/db/index.db';
|
||||||
|
import { aiService } from '../services/aiService.server';
|
||||||
|
import passport from './passport.routes';
|
||||||
import { validateRequest } from '../middleware/validation.middleware';
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
import { requiredString, numericIdParam, optionalNumeric } from '../utils/zodUtils';
|
import { requiredString, numericIdParam, optionalNumeric } from '../utils/zodUtils';
|
||||||
|
|
||||||
@@ -28,6 +30,12 @@ const byIngredientAndTagSchema = z.object({
|
|||||||
|
|
||||||
const recipeIdParamsSchema = numericIdParam('recipeId');
|
const recipeIdParamsSchema = numericIdParam('recipeId');
|
||||||
|
|
||||||
|
const suggestRecipeSchema = z.object({
|
||||||
|
body: z.object({
|
||||||
|
ingredients: z.array(z.string().min(1)).nonempty('At least one ingredient is required.'),
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* GET /api/recipes/by-sale-percentage - Get recipes based on the percentage of their ingredients on sale.
|
* GET /api/recipes/by-sale-percentage - Get recipes based on the percentage of their ingredients on sale.
|
||||||
*/
|
*/
|
||||||
@@ -121,4 +129,31 @@ router.get('/:recipeId', validateRequest(recipeIdParamsSchema), async (req, res,
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /api/recipes/suggest - Generates a simple recipe suggestion from a list of ingredients.
|
||||||
|
* This is a protected endpoint.
|
||||||
|
*/
|
||||||
|
router.post(
|
||||||
|
'/suggest',
|
||||||
|
passport.authenticate('jwt', { session: false }),
|
||||||
|
validateRequest(suggestRecipeSchema),
|
||||||
|
async (req, res, next) => {
|
||||||
|
try {
|
||||||
|
const { body } = req as unknown as z.infer<typeof suggestRecipeSchema>;
|
||||||
|
const suggestion = await aiService.generateRecipeSuggestion(body.ingredients, req.log);
|
||||||
|
|
||||||
|
if (!suggestion) {
|
||||||
|
return res
|
||||||
|
.status(503)
|
||||||
|
.json({ message: 'AI service is currently unavailable or failed to generate a suggestion.' });
|
||||||
|
}
|
||||||
|
|
||||||
|
res.json({ suggestion });
|
||||||
|
} catch (error) {
|
||||||
|
req.log.error({ error }, 'Error generating recipe suggestion');
|
||||||
|
next(error);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
export default router;
|
export default router;
|
||||||
|
|||||||
@@ -482,6 +482,12 @@ describe('User Routes (/api/users)', () => {
|
|||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(response.body).toEqual(mockAddedItem);
|
expect(response.body).toEqual(mockAddedItem);
|
||||||
|
expect(db.shoppingRepo.addShoppingListItem).toHaveBeenCalledWith(
|
||||||
|
listId,
|
||||||
|
mockUserProfile.user.user_id,
|
||||||
|
itemData,
|
||||||
|
expectLogger,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 400 on foreign key error when adding an item', async () => {
|
it('should return 400 on foreign key error when adding an item', async () => {
|
||||||
@@ -519,6 +525,12 @@ describe('User Routes (/api/users)', () => {
|
|||||||
|
|
||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(response.body).toEqual(mockUpdatedItem);
|
expect(response.body).toEqual(mockUpdatedItem);
|
||||||
|
expect(db.shoppingRepo.updateShoppingListItem).toHaveBeenCalledWith(
|
||||||
|
itemId,
|
||||||
|
mockUserProfile.user.user_id,
|
||||||
|
updates,
|
||||||
|
expectLogger,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 404 if item to update is not found', async () => {
|
it('should return 404 if item to update is not found', async () => {
|
||||||
@@ -554,6 +566,11 @@ describe('User Routes (/api/users)', () => {
|
|||||||
vi.mocked(db.shoppingRepo.removeShoppingListItem).mockResolvedValue(undefined);
|
vi.mocked(db.shoppingRepo.removeShoppingListItem).mockResolvedValue(undefined);
|
||||||
const response = await supertest(app).delete('/api/users/shopping-lists/items/101');
|
const response = await supertest(app).delete('/api/users/shopping-lists/items/101');
|
||||||
expect(response.status).toBe(204);
|
expect(response.status).toBe(204);
|
||||||
|
expect(db.shoppingRepo.removeShoppingListItem).toHaveBeenCalledWith(
|
||||||
|
101,
|
||||||
|
mockUserProfile.user.user_id,
|
||||||
|
expectLogger,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 404 if item to delete is not found', async () => {
|
it('should return 404 if item to delete is not found', async () => {
|
||||||
@@ -585,6 +602,27 @@ describe('User Routes (/api/users)', () => {
|
|||||||
expect(response.body).toEqual(updatedProfile);
|
expect(response.body).toEqual(updatedProfile);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('should allow updating the profile with an empty string for avatar_url', async () => {
|
||||||
|
// Arrange
|
||||||
|
const profileUpdates = { avatar_url: '' };
|
||||||
|
// The service should receive `undefined` after Zod preprocessing
|
||||||
|
const updatedProfile = createMockUserProfile({ ...mockUserProfile, avatar_url: undefined });
|
||||||
|
vi.mocked(db.userRepo.updateUserProfile).mockResolvedValue(updatedProfile);
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const response = await supertest(app).put('/api/users/profile').send(profileUpdates);
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(response.body).toEqual(updatedProfile);
|
||||||
|
// Verify that the Zod schema preprocessed the empty string to undefined
|
||||||
|
expect(db.userRepo.updateUserProfile).toHaveBeenCalledWith(
|
||||||
|
mockUserProfile.user.user_id,
|
||||||
|
{ avatar_url: undefined },
|
||||||
|
expectLogger,
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
it('should return 500 on a generic database error', async () => {
|
it('should return 500 on a generic database error', async () => {
|
||||||
const dbError = new Error('DB Connection Failed');
|
const dbError = new Error('DB Connection Failed');
|
||||||
vi.mocked(db.userRepo.updateUserProfile).mockRejectedValue(dbError);
|
vi.mocked(db.userRepo.updateUserProfile).mockRejectedValue(dbError);
|
||||||
|
|||||||
@@ -26,7 +26,13 @@ const router = express.Router();
|
|||||||
|
|
||||||
const updateProfileSchema = z.object({
|
const updateProfileSchema = z.object({
|
||||||
body: z
|
body: z
|
||||||
.object({ full_name: z.string().optional(), avatar_url: z.string().url().optional() })
|
.object({
|
||||||
|
full_name: z.string().optional(),
|
||||||
|
avatar_url: z.preprocess(
|
||||||
|
(val) => (val === '' ? undefined : val),
|
||||||
|
z.string().trim().url().optional(),
|
||||||
|
),
|
||||||
|
})
|
||||||
.refine((data) => Object.keys(data).length > 0, {
|
.refine((data) => Object.keys(data).length > 0, {
|
||||||
message: 'At least one field to update must be provided.',
|
message: 'At least one field to update must be provided.',
|
||||||
}),
|
}),
|
||||||
@@ -472,10 +478,16 @@ router.post(
|
|||||||
validateRequest(addShoppingListItemSchema),
|
validateRequest(addShoppingListItemSchema),
|
||||||
async (req, res, next: NextFunction) => {
|
async (req, res, next: NextFunction) => {
|
||||||
logger.debug(`[ROUTE] POST /api/users/shopping-lists/:listId/items - ENTER`);
|
logger.debug(`[ROUTE] POST /api/users/shopping-lists/:listId/items - ENTER`);
|
||||||
|
const userProfile = req.user as UserProfile;
|
||||||
// Apply ADR-003 pattern for type safety
|
// Apply ADR-003 pattern for type safety
|
||||||
const { params, body } = req as unknown as AddShoppingListItemRequest;
|
const { params, body } = req as unknown as AddShoppingListItemRequest;
|
||||||
try {
|
try {
|
||||||
const newItem = await db.shoppingRepo.addShoppingListItem(params.listId, body, req.log);
|
const newItem = await db.shoppingRepo.addShoppingListItem(
|
||||||
|
params.listId,
|
||||||
|
userProfile.user.user_id,
|
||||||
|
body,
|
||||||
|
req.log,
|
||||||
|
);
|
||||||
res.status(201).json(newItem);
|
res.status(201).json(newItem);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof ForeignKeyConstraintError) {
|
if (error instanceof ForeignKeyConstraintError) {
|
||||||
@@ -506,11 +518,13 @@ router.put(
|
|||||||
validateRequest(updateShoppingListItemSchema),
|
validateRequest(updateShoppingListItemSchema),
|
||||||
async (req, res, next: NextFunction) => {
|
async (req, res, next: NextFunction) => {
|
||||||
logger.debug(`[ROUTE] PUT /api/users/shopping-lists/items/:itemId - ENTER`);
|
logger.debug(`[ROUTE] PUT /api/users/shopping-lists/items/:itemId - ENTER`);
|
||||||
|
const userProfile = req.user as UserProfile;
|
||||||
// Apply ADR-003 pattern for type safety
|
// Apply ADR-003 pattern for type safety
|
||||||
const { params, body } = req as unknown as UpdateShoppingListItemRequest;
|
const { params, body } = req as unknown as UpdateShoppingListItemRequest;
|
||||||
try {
|
try {
|
||||||
const updatedItem = await db.shoppingRepo.updateShoppingListItem(
|
const updatedItem = await db.shoppingRepo.updateShoppingListItem(
|
||||||
params.itemId,
|
params.itemId,
|
||||||
|
userProfile.user.user_id,
|
||||||
body,
|
body,
|
||||||
req.log,
|
req.log,
|
||||||
);
|
);
|
||||||
@@ -535,10 +549,11 @@ router.delete(
|
|||||||
validateRequest(shoppingListItemIdSchema),
|
validateRequest(shoppingListItemIdSchema),
|
||||||
async (req, res, next: NextFunction) => {
|
async (req, res, next: NextFunction) => {
|
||||||
logger.debug(`[ROUTE] DELETE /api/users/shopping-lists/items/:itemId - ENTER`);
|
logger.debug(`[ROUTE] DELETE /api/users/shopping-lists/items/:itemId - ENTER`);
|
||||||
|
const userProfile = req.user as UserProfile;
|
||||||
// Apply ADR-003 pattern for type safety
|
// Apply ADR-003 pattern for type safety
|
||||||
const { params } = req as unknown as DeleteShoppingListItemRequest;
|
const { params } = req as unknown as DeleteShoppingListItemRequest;
|
||||||
try {
|
try {
|
||||||
await db.shoppingRepo.removeShoppingListItem(params.itemId, req.log);
|
await db.shoppingRepo.removeShoppingListItem(params.itemId, userProfile.user.user_id, req.log);
|
||||||
res.status(204).send();
|
res.status(204).send();
|
||||||
} catch (error: unknown) {
|
} catch (error: unknown) {
|
||||||
logger.error(
|
logger.error(
|
||||||
|
|||||||
@@ -1,17 +1,19 @@
|
|||||||
// src/services/aiService.server.test.ts
|
// src/services/aiService.server.test.ts
|
||||||
import { describe, it, expect, vi, beforeEach, afterEach, type Mock } from 'vitest';
|
import { describe, it, expect, vi, beforeEach, afterEach, type Mock } from 'vitest';
|
||||||
|
import type { Job } from 'bullmq';
|
||||||
import { createMockLogger } from '../tests/utils/mockLogger';
|
import { createMockLogger } from '../tests/utils/mockLogger';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { FlyerStatus, MasterGroceryItem, UserProfile } from '../types';
|
import type { FlyerStatus, MasterGroceryItem, UserProfile } from '../types';
|
||||||
// Import the class, not the singleton instance, so we can instantiate it with mocks.
|
// Import the class, not the singleton instance, so we can instantiate it with mocks.
|
||||||
import {
|
import {
|
||||||
AIService,
|
AIService,
|
||||||
AiFlyerDataSchema,
|
|
||||||
aiService as aiServiceSingleton,
|
aiService as aiServiceSingleton,
|
||||||
DuplicateFlyerError,
|
DuplicateFlyerError,
|
||||||
|
type RawFlyerItem,
|
||||||
} from './aiService.server';
|
} from './aiService.server';
|
||||||
import { createMockMasterGroceryItem } from '../tests/utils/mockFactories';
|
import { createMockMasterGroceryItem, createMockFlyer } from '../tests/utils/mockFactories';
|
||||||
import { ValidationError } from './db/errors.db';
|
import { ValidationError } from './db/errors.db';
|
||||||
|
import { AiFlyerDataSchema } from '../types/ai';
|
||||||
|
|
||||||
// Mock the logger to prevent the real pino instance from being created, which causes issues with 'pino-pretty' in tests.
|
// Mock the logger to prevent the real pino instance from being created, which causes issues with 'pino-pretty' in tests.
|
||||||
vi.mock('./logger.server', () => ({
|
vi.mock('./logger.server', () => ({
|
||||||
@@ -128,14 +130,7 @@ describe('AI Service (Server)', () => {
|
|||||||
const resultEmpty = AiFlyerDataSchema.safeParse(dataWithEmpty);
|
const resultEmpty = AiFlyerDataSchema.safeParse(dataWithEmpty);
|
||||||
|
|
||||||
expect(resultNull.success).toBe(false);
|
expect(resultNull.success).toBe(false);
|
||||||
if (!resultNull.success) {
|
// Null checks fail with a generic type error, which is acceptable.
|
||||||
expect(resultNull.error.issues[0].message).toBe('Store name cannot be empty');
|
|
||||||
}
|
|
||||||
|
|
||||||
expect(resultEmpty.success).toBe(false);
|
|
||||||
if (!resultEmpty.success) {
|
|
||||||
expect(resultEmpty.error.issues[0].message).toBe('Store name cannot be empty');
|
|
||||||
}
|
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -789,7 +784,7 @@ describe('AI Service (Server)', () => {
|
|||||||
} as UserProfile;
|
} as UserProfile;
|
||||||
|
|
||||||
it('should throw DuplicateFlyerError if flyer already exists', async () => {
|
it('should throw DuplicateFlyerError if flyer already exists', async () => {
|
||||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 99 } as any);
|
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 99, checksum: 'checksum123', file_name: 'test.pdf', image_url: '/flyer-images/test.pdf', icon_url: '/flyer-images/icons/test.webp', store_id: 1, status: 'processed', item_count: 0, created_at: new Date().toISOString(), updated_at: new Date().toISOString() });
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
aiServiceInstance.enqueueFlyerProcessing(
|
aiServiceInstance.enqueueFlyerProcessing(
|
||||||
@@ -804,7 +799,7 @@ describe('AI Service (Server)', () => {
|
|||||||
|
|
||||||
it('should enqueue job with user address if profile exists', async () => {
|
it('should enqueue job with user address if profile exists', async () => {
|
||||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job123' } as any);
|
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job123' } as unknown as Job);
|
||||||
|
|
||||||
const result = await aiServiceInstance.enqueueFlyerProcessing(
|
const result = await aiServiceInstance.enqueueFlyerProcessing(
|
||||||
mockFile,
|
mockFile,
|
||||||
@@ -827,7 +822,7 @@ describe('AI Service (Server)', () => {
|
|||||||
|
|
||||||
it('should enqueue job without address if profile is missing', async () => {
|
it('should enqueue job without address if profile is missing', async () => {
|
||||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job456' } as any);
|
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job456' } as unknown as Job);
|
||||||
|
|
||||||
await aiServiceInstance.enqueueFlyerProcessing(
|
await aiServiceInstance.enqueueFlyerProcessing(
|
||||||
mockFile,
|
mockFile,
|
||||||
@@ -856,7 +851,7 @@ describe('AI Service (Server)', () => {
|
|||||||
const mockProfile = { user: { user_id: 'u1' } } as UserProfile;
|
const mockProfile = { user: { user_id: 'u1' } } as UserProfile;
|
||||||
|
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
// Default success mocks
|
// Default success mocks. Use createMockFlyer for a more complete mock.
|
||||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||||
vi.mocked(generateFlyerIcon).mockResolvedValue('icon.jpg');
|
vi.mocked(generateFlyerIcon).mockResolvedValue('icon.jpg');
|
||||||
vi.mocked(createFlyerAndItems).mockResolvedValue({
|
vi.mocked(createFlyerAndItems).mockResolvedValue({
|
||||||
@@ -893,7 +888,7 @@ describe('AI Service (Server)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should throw DuplicateFlyerError if checksum exists', async () => {
|
it('should throw DuplicateFlyerError if checksum exists', async () => {
|
||||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 55 } as any);
|
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(createMockFlyer({ flyer_id: 55 }));
|
||||||
const body = { checksum: 'dup-sum' };
|
const body = { checksum: 'dup-sum' };
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
@@ -1058,4 +1053,56 @@ describe('AI Service (Server)', () => {
|
|||||||
expect(aiServiceSingleton).toBeInstanceOf(AIService);
|
expect(aiServiceSingleton).toBeInstanceOf(AIService);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('_normalizeExtractedItems (private method)', () => {
|
||||||
|
it('should correctly normalize items with null or undefined price_in_cents', () => {
|
||||||
|
const rawItems: RawFlyerItem[] = [
|
||||||
|
{
|
||||||
|
item: 'Valid Item',
|
||||||
|
price_display: '$1.99',
|
||||||
|
price_in_cents: 199,
|
||||||
|
quantity: '1',
|
||||||
|
category_name: 'Category A',
|
||||||
|
master_item_id: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
item: 'Item with Null Price',
|
||||||
|
price_display: null,
|
||||||
|
price_in_cents: null, // Test case for null
|
||||||
|
quantity: '1',
|
||||||
|
category_name: 'Category B',
|
||||||
|
master_item_id: 2,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
item: 'Item with Undefined Price',
|
||||||
|
price_display: '$2.99',
|
||||||
|
price_in_cents: undefined, // Test case for undefined
|
||||||
|
quantity: '1',
|
||||||
|
category_name: 'Category C',
|
||||||
|
master_item_id: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
item: null, // Test null item name
|
||||||
|
price_display: undefined, // Test undefined display price
|
||||||
|
price_in_cents: 50,
|
||||||
|
quantity: null, // Test null quantity
|
||||||
|
category_name: undefined, // Test undefined category
|
||||||
|
master_item_id: null, // Test null master_item_id
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
// Access the private method for testing
|
||||||
|
const normalized = (aiServiceInstance as any)._normalizeExtractedItems(rawItems);
|
||||||
|
|
||||||
|
expect(normalized).toHaveLength(4);
|
||||||
|
expect(normalized[0].price_in_cents).toBe(199);
|
||||||
|
expect(normalized[1].price_in_cents).toBe(null); // null should remain null
|
||||||
|
expect(normalized[2].price_in_cents).toBe(null); // undefined should become null
|
||||||
|
expect(normalized[3].item).toBe('Unknown Item');
|
||||||
|
expect(normalized[3].quantity).toBe('');
|
||||||
|
expect(normalized[3].category_name).toBe('Other/Miscellaneous');
|
||||||
|
expect(normalized[3].master_item_id).toBeUndefined(); // nullish coalescing to undefined
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -4,7 +4,6 @@
|
|||||||
* It is intended to be used only by the backend (e.g., server.ts) and should never be imported into client-side code.
|
* It is intended to be used only by the backend (e.g., server.ts) and should never be imported into client-side code.
|
||||||
* The `.server.ts` naming convention helps enforce this separation.
|
* The `.server.ts` naming convention helps enforce this separation.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { GoogleGenAI, type GenerateContentResponse, type Content, type Tool } from '@google/genai';
|
import { GoogleGenAI, type GenerateContentResponse, type Content, type Tool } from '@google/genai';
|
||||||
import fsPromises from 'node:fs/promises';
|
import fsPromises from 'node:fs/promises';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
@@ -26,29 +25,11 @@ import type { Job } from 'bullmq';
|
|||||||
import { createFlyerAndItems } from './db/flyer.db';
|
import { createFlyerAndItems } from './db/flyer.db';
|
||||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||||
import path from 'path';
|
import path from 'path';
|
||||||
import { ValidationError } from './db/errors.db';
|
import { ValidationError } from './db/errors.db'; // Keep this import for ValidationError
|
||||||
|
import {
|
||||||
// Helper for consistent required string validation (handles missing/null/empty)
|
AiFlyerDataSchema,
|
||||||
const requiredString = (message: string) =>
|
ExtractedFlyerItemSchema,
|
||||||
z.preprocess((val) => val ?? '', z.string().min(1, message));
|
} from '../types/ai'; // Import consolidated schemas
|
||||||
|
|
||||||
// --- Zod Schemas for AI Response Validation (exported for the transformer) ---
|
|
||||||
const ExtractedFlyerItemSchema = z.object({
|
|
||||||
item: z.string(),
|
|
||||||
price_display: z.string(),
|
|
||||||
price_in_cents: z.number().nullable(),
|
|
||||||
quantity: z.string(),
|
|
||||||
category_name: z.string(),
|
|
||||||
master_item_id: z.number().nullish(), // .nullish() allows null or undefined
|
|
||||||
});
|
|
||||||
|
|
||||||
export const AiFlyerDataSchema = z.object({
|
|
||||||
store_name: requiredString('Store name cannot be empty'),
|
|
||||||
valid_from: z.string().nullable(),
|
|
||||||
valid_to: z.string().nullable(),
|
|
||||||
store_address: z.string().nullable(),
|
|
||||||
items: z.array(ExtractedFlyerItemSchema),
|
|
||||||
});
|
|
||||||
|
|
||||||
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
||||||
checksum?: string;
|
checksum?: string;
|
||||||
@@ -81,6 +62,7 @@ interface IAiClient {
|
|||||||
generateContent(request: {
|
generateContent(request: {
|
||||||
contents: Content[];
|
contents: Content[];
|
||||||
tools?: Tool[];
|
tools?: Tool[];
|
||||||
|
useLiteModels?: boolean;
|
||||||
}): Promise<GenerateContentResponse>;
|
}): Promise<GenerateContentResponse>;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -89,10 +71,10 @@ interface IAiClient {
|
|||||||
* This type is intentionally loose to accommodate potential null/undefined values
|
* This type is intentionally loose to accommodate potential null/undefined values
|
||||||
* from the AI before they are cleaned and normalized.
|
* from the AI before they are cleaned and normalized.
|
||||||
*/
|
*/
|
||||||
type RawFlyerItem = {
|
export type RawFlyerItem = {
|
||||||
item: string;
|
item: string | null;
|
||||||
price_display: string | null | undefined;
|
price_display: string | null | undefined;
|
||||||
price_in_cents: number | null;
|
price_in_cents: number | null | undefined;
|
||||||
quantity: string | null | undefined;
|
quantity: string | null | undefined;
|
||||||
category_name: string | null | undefined;
|
category_name: string | null | undefined;
|
||||||
master_item_id?: number | null | undefined;
|
master_item_id?: number | null | undefined;
|
||||||
@@ -112,7 +94,8 @@ export class AIService {
|
|||||||
// The fallback list is ordered by preference (speed/cost vs. power).
|
// The fallback list is ordered by preference (speed/cost vs. power).
|
||||||
// We try the fastest models first, then the more powerful 'pro' model as a high-quality fallback,
|
// We try the fastest models first, then the more powerful 'pro' model as a high-quality fallback,
|
||||||
// and finally the 'lite' model as a last resort.
|
// and finally the 'lite' model as a last resort.
|
||||||
private readonly models = [ 'gemini-3-flash-preview', 'gemini-2.5-flash', 'gemini-2.5-flash-lite'];
|
private readonly models = [ 'gemini-3-flash-preview', 'gemini-2.5-flash', 'gemini-2.5-flash-lite', 'gemma-3-27b', 'gemma-3-12b'];
|
||||||
|
private readonly models_lite = ["gemma-3-4b", "gemma-3-2b", "gemma-3-1b"];
|
||||||
|
|
||||||
constructor(logger: Logger, aiClient?: IAiClient, fs?: IFileSystem) {
|
constructor(logger: Logger, aiClient?: IAiClient, fs?: IFileSystem) {
|
||||||
this.logger = logger;
|
this.logger = logger;
|
||||||
@@ -175,7 +158,9 @@ export class AIService {
|
|||||||
throw new Error('AIService.generateContent requires at least one content element.');
|
throw new Error('AIService.generateContent requires at least one content element.');
|
||||||
}
|
}
|
||||||
|
|
||||||
return this._generateWithFallback(genAI, request);
|
const { useLiteModels, ...apiReq } = request;
|
||||||
|
const models = useLiteModels ? this.models_lite : this.models;
|
||||||
|
return this._generateWithFallback(genAI, apiReq, models);
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
: {
|
: {
|
||||||
@@ -213,10 +198,11 @@ export class AIService {
|
|||||||
private async _generateWithFallback(
|
private async _generateWithFallback(
|
||||||
genAI: GoogleGenAI,
|
genAI: GoogleGenAI,
|
||||||
request: { contents: Content[]; tools?: Tool[] },
|
request: { contents: Content[]; tools?: Tool[] },
|
||||||
|
models: string[] = this.models,
|
||||||
): Promise<GenerateContentResponse> {
|
): Promise<GenerateContentResponse> {
|
||||||
let lastError: Error | null = null;
|
let lastError: Error | null = null;
|
||||||
|
|
||||||
for (const modelName of this.models) {
|
for (const modelName of models) {
|
||||||
try {
|
try {
|
||||||
this.logger.info(
|
this.logger.info(
|
||||||
`[AIService Adapter] Attempting to generate content with model: ${modelName}`,
|
`[AIService Adapter] Attempting to generate content with model: ${modelName}`,
|
||||||
@@ -507,7 +493,7 @@ export class AIService {
|
|||||||
userProfileAddress?: string,
|
userProfileAddress?: string,
|
||||||
logger: Logger = this.logger,
|
logger: Logger = this.logger,
|
||||||
): Promise<{
|
): Promise<{
|
||||||
store_name: string;
|
store_name: string | null;
|
||||||
valid_from: string | null;
|
valid_from: string | null;
|
||||||
valid_to: string | null;
|
valid_to: string | null;
|
||||||
store_address: string | null;
|
store_address: string | null;
|
||||||
@@ -606,6 +592,8 @@ export class AIService {
|
|||||||
item.category_name === null || item.category_name === undefined
|
item.category_name === null || item.category_name === undefined
|
||||||
? 'Other/Miscellaneous'
|
? 'Other/Miscellaneous'
|
||||||
: String(item.category_name),
|
: String(item.category_name),
|
||||||
|
// Ensure undefined is converted to null to match the Zod schema.
|
||||||
|
price_in_cents: item.price_in_cents ?? null,
|
||||||
master_item_id: item.master_item_id ?? undefined,
|
master_item_id: item.master_item_id ?? undefined,
|
||||||
}));
|
}));
|
||||||
}
|
}
|
||||||
@@ -685,6 +673,33 @@ export class AIService {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generates a simple recipe suggestion based on a list of ingredients.
|
||||||
|
* Uses the 'lite' models for faster/cheaper generation.
|
||||||
|
* @param ingredients List of available ingredients.
|
||||||
|
* @param logger Logger instance.
|
||||||
|
* @returns The recipe suggestion text.
|
||||||
|
*/
|
||||||
|
async generateRecipeSuggestion(
|
||||||
|
ingredients: string[],
|
||||||
|
logger: Logger = this.logger,
|
||||||
|
): Promise<string | null> {
|
||||||
|
const prompt = `Suggest a simple recipe using these ingredients: ${ingredients.join(', ')}. Keep it brief.`;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const result = await this.rateLimiter(() =>
|
||||||
|
this.aiClient.generateContent({
|
||||||
|
contents: [{ parts: [{ text: prompt }] }],
|
||||||
|
useLiteModels: true,
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
return result.text || null;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error({ err: error }, 'Failed to generate recipe suggestion');
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* SERVER-SIDE FUNCTION
|
* SERVER-SIDE FUNCTION
|
||||||
* Uses Google Maps grounding to find nearby stores and plan a shopping trip.
|
* Uses Google Maps grounding to find nearby stores and plan a shopping trip.
|
||||||
|
|||||||
@@ -543,6 +543,13 @@ describe('API Client', () => {
|
|||||||
await apiClient.deleteRecipe(recipeId);
|
await apiClient.deleteRecipe(recipeId);
|
||||||
expect(capturedUrl?.pathname).toBe(`/api/recipes/${recipeId}`);
|
expect(capturedUrl?.pathname).toBe(`/api/recipes/${recipeId}`);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('suggestRecipe should send a POST request with ingredients', async () => {
|
||||||
|
const ingredients = ['chicken', 'rice'];
|
||||||
|
await apiClient.suggestRecipe(ingredients);
|
||||||
|
expect(capturedUrl?.pathname).toBe('/api/recipes/suggest');
|
||||||
|
expect(capturedBody).toEqual({ ingredients });
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('User Profile and Settings API Functions', () => {
|
describe('User Profile and Settings API Functions', () => {
|
||||||
@@ -933,7 +940,7 @@ describe('API Client', () => {
|
|||||||
|
|
||||||
it('logSearchQuery should send a POST request with query data', async () => {
|
it('logSearchQuery should send a POST request with query data', async () => {
|
||||||
const queryData = createMockSearchQueryPayload({ query_text: 'apples', result_count: 10, was_successful: true });
|
const queryData = createMockSearchQueryPayload({ query_text: 'apples', result_count: 10, was_successful: true });
|
||||||
await apiClient.logSearchQuery(queryData);
|
await apiClient.logSearchQuery(queryData as any);
|
||||||
expect(capturedUrl?.pathname).toBe('/api/search/log');
|
expect(capturedUrl?.pathname).toBe('/api/search/log');
|
||||||
expect(capturedBody).toEqual(queryData);
|
expect(capturedBody).toEqual(queryData);
|
||||||
});
|
});
|
||||||
@@ -960,7 +967,7 @@ describe('API Client', () => {
|
|||||||
result_count: 0,
|
result_count: 0,
|
||||||
was_successful: false,
|
was_successful: false,
|
||||||
});
|
});
|
||||||
await apiClient.logSearchQuery(queryData);
|
await apiClient.logSearchQuery(queryData as any);
|
||||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -283,7 +283,10 @@ export const fetchFlyerById = (flyerId: number): Promise<Response> =>
|
|||||||
* Fetches all master grocery items from the backend.
|
* Fetches all master grocery items from the backend.
|
||||||
* @returns A promise that resolves to an array of MasterGroceryItem objects.
|
* @returns A promise that resolves to an array of MasterGroceryItem objects.
|
||||||
*/
|
*/
|
||||||
export const fetchMasterItems = (): Promise<Response> => publicGet('/personalization/master-items');
|
export const fetchMasterItems = (): Promise<Response> => {
|
||||||
|
logger.debug('apiClient: fetchMasterItems called');
|
||||||
|
return publicGet('/personalization/master-items');
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Fetches all categories from the backend.
|
* Fetches all categories from the backend.
|
||||||
@@ -633,6 +636,20 @@ export const addRecipeComment = (
|
|||||||
): Promise<Response> =>
|
): Promise<Response> =>
|
||||||
authedPost(`/recipes/${recipeId}/comments`, { content, parentCommentId }, { tokenOverride });
|
authedPost(`/recipes/${recipeId}/comments`, { content, parentCommentId }, { tokenOverride });
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Requests a simple recipe suggestion from the AI based on a list of ingredients.
|
||||||
|
* @param ingredients An array of ingredient strings.
|
||||||
|
* @param tokenOverride Optional token for testing.
|
||||||
|
* @returns A promise that resolves to the API response containing the suggestion.
|
||||||
|
*/
|
||||||
|
export const suggestRecipe = (
|
||||||
|
ingredients: string[],
|
||||||
|
tokenOverride?: string,
|
||||||
|
): Promise<Response> => {
|
||||||
|
// This is a protected endpoint, so we use authedPost.
|
||||||
|
return authedPost('/recipes/suggest', { ingredients }, { tokenOverride });
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Deletes a recipe.
|
* Deletes a recipe.
|
||||||
* @param recipeId The ID of the recipe to delete.
|
* @param recipeId The ID of the recipe to delete.
|
||||||
|
|||||||
@@ -17,6 +17,11 @@ describe('AuthService', () => {
|
|||||||
user_id: 'user-123',
|
user_id: 'user-123',
|
||||||
email: 'test@example.com',
|
email: 'test@example.com',
|
||||||
password_hash: 'hashed-password',
|
password_hash: 'hashed-password',
|
||||||
|
failed_login_attempts: 0,
|
||||||
|
last_failed_login: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
refresh_token: null,
|
||||||
};
|
};
|
||||||
const mockUserProfile: UserProfile = {
|
const mockUserProfile: UserProfile = {
|
||||||
user: mockUser,
|
user: mockUser,
|
||||||
@@ -205,7 +210,7 @@ describe('AuthService', () => {
|
|||||||
|
|
||||||
describe('resetPassword', () => {
|
describe('resetPassword', () => {
|
||||||
it('should process password reset for existing user', async () => {
|
it('should process password reset for existing user', async () => {
|
||||||
vi.mocked(userRepo.findUserByEmail).mockResolvedValue(mockUser as any);
|
vi.mocked(userRepo.findUserByEmail).mockResolvedValue(mockUser);
|
||||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-token');
|
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-token');
|
||||||
|
|
||||||
const result = await authService.resetPassword('test@example.com', reqLog);
|
const result = await authService.resetPassword('test@example.com', reqLog);
|
||||||
@@ -284,7 +289,7 @@ describe('AuthService', () => {
|
|||||||
|
|
||||||
describe('getUserByRefreshToken', () => {
|
describe('getUserByRefreshToken', () => {
|
||||||
it('should return user profile if token exists', async () => {
|
it('should return user profile if token exists', async () => {
|
||||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123' } as any);
|
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123', email: 'test@example.com', created_at: new Date().toISOString(), updated_at: new Date().toISOString() });
|
||||||
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
||||||
|
|
||||||
const result = await authService.getUserByRefreshToken('valid-token', reqLog);
|
const result = await authService.getUserByRefreshToken('valid-token', reqLog);
|
||||||
@@ -318,7 +323,7 @@ describe('AuthService', () => {
|
|||||||
|
|
||||||
describe('refreshAccessToken', () => {
|
describe('refreshAccessToken', () => {
|
||||||
it('should return new access token if user found', async () => {
|
it('should return new access token if user found', async () => {
|
||||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123' } as any);
|
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123', email: 'test@example.com', created_at: new Date().toISOString(), updated_at: new Date().toISOString() });
|
||||||
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
||||||
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
||||||
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { UniqueConstraintError, NotFoundError } from './errors.db';
|
import { UniqueConstraintError, NotFoundError, handleDbError } from './errors.db';
|
||||||
import { Address } from '../../types';
|
import { Address } from '../../types';
|
||||||
|
|
||||||
export class AddressRepository {
|
export class AddressRepository {
|
||||||
@@ -30,11 +30,9 @@ export class AddressRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) {
|
handleDbError(error, logger, 'Database error in getAddressById', { addressId }, {
|
||||||
throw error;
|
defaultMessage: 'Failed to retrieve address.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, addressId }, 'Database error in getAddressById');
|
|
||||||
throw new Error('Failed to retrieve address.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -78,10 +76,10 @@ export class AddressRepository {
|
|||||||
const res = await this.db.query<{ address_id: number }>(query, values);
|
const res = await this.db.query<{ address_id: number }>(query, values);
|
||||||
return res.rows[0].address_id;
|
return res.rows[0].address_id;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, address }, 'Database error in upsertAddress');
|
handleDbError(error, logger, 'Database error in upsertAddress', { address }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23505')
|
uniqueMessage: 'An identical address already exists.',
|
||||||
throw new UniqueConstraintError('An identical address already exists.');
|
defaultMessage: 'Failed to upsert address.',
|
||||||
throw new Error('Failed to upsert address.');
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -203,7 +203,11 @@ describe('Admin DB Service', () => {
|
|||||||
.mockRejectedValueOnce(new Error('DB Read Error'));
|
.mockRejectedValueOnce(new Error('DB Read Error'));
|
||||||
|
|
||||||
// The Promise.all should reject, and the function should re-throw the error
|
// The Promise.all should reject, and the function should re-throw the error
|
||||||
await expect(adminRepo.getApplicationStats(mockLogger)).rejects.toThrow('DB Read Error');
|
// The handleDbError function wraps the original error in a new one with a default message,
|
||||||
|
// so we should test for that specific message.
|
||||||
|
await expect(adminRepo.getApplicationStats(mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to retrieve application statistics.',
|
||||||
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: expect.any(Error) },
|
{ err: expect.any(Error) },
|
||||||
'Database error in getApplicationStats',
|
'Database error in getApplicationStats',
|
||||||
@@ -277,7 +281,7 @@ describe('Admin DB Service', () => {
|
|||||||
'Failed to get most frequent sale items.',
|
'Failed to get most frequent sale items.',
|
||||||
);
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError },
|
{ err: dbError, days: 30, limit: 10 },
|
||||||
'Database error in getMostFrequentSaleItems',
|
'Database error in getMostFrequentSaleItems',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -688,7 +692,9 @@ describe('Admin DB Service', () => {
|
|||||||
it('should re-throw a generic error if the database query fails for other reasons', async () => {
|
it('should re-throw a generic error if the database query fails for other reasons', async () => {
|
||||||
const dbError = new Error('DB Error');
|
const dbError = new Error('DB Error');
|
||||||
mockDb.query.mockRejectedValue(dbError);
|
mockDb.query.mockRejectedValue(dbError);
|
||||||
await expect(adminRepo.updateUserRole('1', 'admin', mockLogger)).rejects.toThrow('DB Error');
|
await expect(adminRepo.updateUserRole('1', 'admin', mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to update user role.',
|
||||||
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, userId: '1', role: 'admin' },
|
{ err: dbError, userId: '1', role: 'admin' },
|
||||||
'Database error in updateUserRole',
|
'Database error in updateUserRole',
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/admin.db.ts
|
// src/services/db/admin.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool, withTransaction } from './connection.db';
|
import { getPool, withTransaction } from './connection.db';
|
||||||
import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { ForeignKeyConstraintError, NotFoundError, CheckConstraintError, handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import {
|
import {
|
||||||
SuggestedCorrection,
|
SuggestedCorrection,
|
||||||
@@ -41,6 +41,7 @@ export class AdminRepository {
|
|||||||
sc.correction_type,
|
sc.correction_type,
|
||||||
sc.suggested_value,
|
sc.suggested_value,
|
||||||
sc.status,
|
sc.status,
|
||||||
|
sc.updated_at,
|
||||||
sc.created_at,
|
sc.created_at,
|
||||||
fi.item as flyer_item_name,
|
fi.item as flyer_item_name,
|
||||||
fi.price_display as flyer_item_price_display,
|
fi.price_display as flyer_item_price_display,
|
||||||
@@ -54,8 +55,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<SuggestedCorrection>(query);
|
const res = await this.db.query<SuggestedCorrection>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getSuggestedCorrections');
|
handleDbError(error, logger, 'Database error in getSuggestedCorrections', {}, {
|
||||||
throw new Error('Failed to retrieve suggested corrections.');
|
defaultMessage: 'Failed to retrieve suggested corrections.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -73,8 +75,10 @@ export class AdminRepository {
|
|||||||
await this.db.query('SELECT public.approve_correction($1)', [correctionId]);
|
await this.db.query('SELECT public.approve_correction($1)', [correctionId]);
|
||||||
logger.info(`Successfully approved and applied correction ID: ${correctionId}`);
|
logger.info(`Successfully approved and applied correction ID: ${correctionId}`);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, correctionId }, 'Database transaction error in approveCorrection');
|
handleDbError(error, logger, 'Database transaction error in approveCorrection', { correctionId }, {
|
||||||
throw new Error('Failed to approve correction.');
|
fkMessage: 'The suggested master item ID does not exist.',
|
||||||
|
defaultMessage: 'Failed to approve correction.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -95,8 +99,9 @@ export class AdminRepository {
|
|||||||
logger.info(`Successfully rejected correction ID: ${correctionId}`);
|
logger.info(`Successfully rejected correction ID: ${correctionId}`);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, correctionId }, 'Database error in rejectCorrection');
|
handleDbError(error, logger, 'Database error in rejectCorrection', { correctionId }, {
|
||||||
throw new Error('Failed to reject correction.');
|
defaultMessage: 'Failed to reject correction.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -121,8 +126,9 @@ export class AdminRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error({ err: error, correctionId }, 'Database error in updateSuggestedCorrection');
|
handleDbError(error, logger, 'Database error in updateSuggestedCorrection', { correctionId }, {
|
||||||
throw new Error('Failed to update suggested correction.');
|
defaultMessage: 'Failed to update suggested correction.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -168,8 +174,9 @@ export class AdminRepository {
|
|||||||
recipeCount: parseInt(recipeCountRes.rows[0].count, 10),
|
recipeCount: parseInt(recipeCountRes.rows[0].count, 10),
|
||||||
};
|
};
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getApplicationStats');
|
handleDbError(error, logger, 'Database error in getApplicationStats', {}, {
|
||||||
throw error; // Re-throw the original error to be handled by the caller
|
defaultMessage: 'Failed to retrieve application statistics.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -212,8 +219,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query(query);
|
const res = await this.db.query(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getDailyStatsForLast30Days');
|
handleDbError(error, logger, 'Database error in getDailyStatsForLast30Days', {}, {
|
||||||
throw new Error('Failed to retrieve daily statistics.');
|
defaultMessage: 'Failed to retrieve daily statistics.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -254,8 +262,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<MostFrequentSaleItem>(query, [days, limit]);
|
const res = await this.db.query<MostFrequentSaleItem>(query, [days, limit]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getMostFrequentSaleItems');
|
handleDbError(error, logger, 'Database error in getMostFrequentSaleItems', { days, limit }, {
|
||||||
throw new Error('Failed to get most frequent sale items.');
|
defaultMessage: 'Failed to get most frequent sale items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -283,11 +292,10 @@ export class AdminRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in updateRecipeCommentStatus', { commentId, status }, {
|
||||||
{ err: error, commentId, status },
|
checkMessage: 'Invalid status provided for recipe comment.',
|
||||||
'Database error in updateRecipeCommentStatus',
|
defaultMessage: 'Failed to update recipe comment status.',
|
||||||
);
|
});
|
||||||
throw new Error('Failed to update recipe comment status.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -301,6 +309,7 @@ export class AdminRepository {
|
|||||||
SELECT
|
SELECT
|
||||||
ufi.unmatched_flyer_item_id,
|
ufi.unmatched_flyer_item_id,
|
||||||
ufi.status,
|
ufi.status,
|
||||||
|
ufi.updated_at,
|
||||||
ufi.created_at,
|
ufi.created_at,
|
||||||
fi.flyer_item_id as flyer_item_id,
|
fi.flyer_item_id as flyer_item_id,
|
||||||
fi.item as flyer_item_name,
|
fi.item as flyer_item_name,
|
||||||
@@ -317,8 +326,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<UnmatchedFlyerItem>(query);
|
const res = await this.db.query<UnmatchedFlyerItem>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getUnmatchedFlyerItems');
|
handleDbError(error, logger, 'Database error in getUnmatchedFlyerItems', {}, {
|
||||||
throw new Error('Failed to retrieve unmatched flyer items.');
|
defaultMessage: 'Failed to retrieve unmatched flyer items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -344,8 +354,10 @@ export class AdminRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error({ err: error, recipeId, status }, 'Database error in updateRecipeStatus');
|
handleDbError(error, logger, 'Database error in updateRecipeStatus', { recipeId, status }, {
|
||||||
throw new Error('Failed to update recipe status.'); // Keep generic for other DB errors
|
checkMessage: 'Invalid status provided for recipe.',
|
||||||
|
defaultMessage: 'Failed to update recipe status.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -397,11 +409,13 @@ export class AdminRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, unmatchedFlyerItemId, masterItemId },
|
error,
|
||||||
|
logger,
|
||||||
'Database transaction error in resolveUnmatchedFlyerItem',
|
'Database transaction error in resolveUnmatchedFlyerItem',
|
||||||
|
{ unmatchedFlyerItemId, masterItemId },
|
||||||
|
{ fkMessage: 'The specified master item ID does not exist.', defaultMessage: 'Failed to resolve unmatched flyer item.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to resolve unmatched flyer item.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -422,11 +436,13 @@ export class AdminRepository {
|
|||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, unmatchedFlyerItemId },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in ignoreUnmatchedFlyerItem',
|
'Database error in ignoreUnmatchedFlyerItem',
|
||||||
|
{ unmatchedFlyerItemId },
|
||||||
|
{ defaultMessage: 'Failed to ignore unmatched flyer item.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to ignore unmatched flyer item.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -442,8 +458,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<ActivityLogItem>('SELECT * FROM public.get_activity_log($1, $2)', [limit, offset]);
|
const res = await this.db.query<ActivityLogItem>('SELECT * FROM public.get_activity_log($1, $2)', [limit, offset]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, limit, offset }, 'Database error in getActivityLog');
|
handleDbError(error, logger, 'Database error in getActivityLog', { limit, offset }, {
|
||||||
throw new Error('Failed to retrieve activity log.');
|
defaultMessage: 'Failed to retrieve activity log.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -544,8 +561,9 @@ export class AdminRepository {
|
|||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, brandId }, 'Database error in updateBrandLogo');
|
handleDbError(error, logger, 'Database error in updateBrandLogo', { brandId }, {
|
||||||
throw new Error('Failed to update brand logo in database.');
|
defaultMessage: 'Failed to update brand logo in database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -569,8 +587,10 @@ export class AdminRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, receiptId, status }, 'Database error in updateReceiptStatus');
|
handleDbError(error, logger, 'Database error in updateReceiptStatus', { receiptId, status }, {
|
||||||
throw new Error('Failed to update receipt status.');
|
checkMessage: 'Invalid status provided for receipt.',
|
||||||
|
defaultMessage: 'Failed to update receipt status.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -583,8 +603,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<AdminUserView>(query);
|
const res = await this.db.query<AdminUserView>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getAllUsers');
|
handleDbError(error, logger, 'Database error in getAllUsers', {}, {
|
||||||
throw new Error('Failed to retrieve all users.');
|
defaultMessage: 'Failed to retrieve all users.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -605,14 +626,14 @@ export class AdminRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, role }, 'Database error in updateUserRole');
|
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
|
||||||
throw new ForeignKeyConstraintError('The specified user does not exist.');
|
|
||||||
}
|
|
||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
throw error; // Re-throw to be handled by the route
|
handleDbError(error, logger, 'Database error in updateUserRole', { userId, role }, {
|
||||||
|
fkMessage: 'The specified user does not exist.',
|
||||||
|
checkMessage: 'Invalid role provided for user.',
|
||||||
|
defaultMessage: 'Failed to update user role.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -639,8 +660,9 @@ export class AdminRepository {
|
|||||||
const res = await this.db.query<Flyer>(query);
|
const res = await this.db.query<Flyer>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getFlyersForReview');
|
handleDbError(error, logger, 'Database error in getFlyersForReview', {}, {
|
||||||
throw new Error('Failed to retrieve flyers for review.');
|
defaultMessage: 'Failed to retrieve flyers for review.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -249,6 +249,17 @@ describe('Budget DB Service', () => {
|
|||||||
expect(result).toEqual(mockUpdatedBudget);
|
expect(result).toEqual(mockUpdatedBudget);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('should prevent a user from updating a budget they do not own', async () => {
|
||||||
|
// Arrange: Mock the query to return 0 rows affected
|
||||||
|
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||||
|
|
||||||
|
// Act & Assert: Attempt to update with a different user ID should throw an error.
|
||||||
|
await expect(
|
||||||
|
budgetRepo.updateBudget(1, 'another-user', { name: 'Updated Groceries' }, mockLogger),
|
||||||
|
).rejects.toThrow('Budget not found or user does not have permission to update.');
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
it('should throw an error if no rows are updated', async () => {
|
it('should throw an error if no rows are updated', async () => {
|
||||||
// Arrange: Mock the query to return 0 rows affected
|
// Arrange: Mock the query to return 0 rows affected
|
||||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/budget.db.ts
|
// src/services/db/budget.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool, withTransaction } from './connection.db';
|
import { getPool, withTransaction } from './connection.db';
|
||||||
import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { NotFoundError, handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { Budget, SpendingByCategory } from '../../types';
|
import type { Budget, SpendingByCategory } from '../../types';
|
||||||
import { GamificationRepository } from './gamification.db';
|
import { GamificationRepository } from './gamification.db';
|
||||||
@@ -28,8 +28,9 @@ export class BudgetRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getBudgetsForUser');
|
handleDbError(error, logger, 'Database error in getBudgetsForUser', { userId }, {
|
||||||
throw new Error('Failed to retrieve budgets.');
|
defaultMessage: 'Failed to retrieve budgets.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -59,14 +60,12 @@ export class BudgetRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The patch requested this specific error handling.
|
handleDbError(error, logger, 'Database error in createBudget', { budgetData, userId }, {
|
||||||
// Type-safe check for a PostgreSQL error code.
|
fkMessage: 'The specified user does not exist.',
|
||||||
// This ensures 'error' is an object with a 'code' property before we access it.
|
notNullMessage: 'One or more required budget fields are missing.',
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
checkMessage: 'Invalid value provided for budget period.',
|
||||||
throw new ForeignKeyConstraintError('The specified user does not exist.');
|
defaultMessage: 'Failed to create budget.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, budgetData, userId }, 'Database error in createBudget');
|
|
||||||
throw new Error('Failed to create budget.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -99,8 +98,9 @@ export class BudgetRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, budgetId, userId }, 'Database error in updateBudget');
|
handleDbError(error, logger, 'Database error in updateBudget', { budgetId, userId }, {
|
||||||
throw new Error('Failed to update budget.');
|
defaultMessage: 'Failed to update budget.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -120,8 +120,9 @@ export class BudgetRepository {
|
|||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, budgetId, userId }, 'Database error in deleteBudget');
|
handleDbError(error, logger, 'Database error in deleteBudget', { budgetId, userId }, {
|
||||||
throw new Error('Failed to delete budget.');
|
defaultMessage: 'Failed to delete budget.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -145,11 +146,13 @@ export class BudgetRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, userId, startDate, endDate },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in getSpendingByCategory',
|
'Database error in getSpendingByCategory',
|
||||||
|
{ userId, startDate, endDate },
|
||||||
|
{ defaultMessage: 'Failed to get spending analysis.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to get spending analysis.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -6,6 +6,7 @@
|
|||||||
// src/services/db/connection.db.ts
|
// src/services/db/connection.db.ts
|
||||||
import { Pool, PoolConfig, PoolClient, types } from 'pg';
|
import { Pool, PoolConfig, PoolClient, types } from 'pg';
|
||||||
import { logger } from '../logger.server';
|
import { logger } from '../logger.server';
|
||||||
|
import { handleDbError } from './errors.db';
|
||||||
|
|
||||||
// --- Singleton Pool Instance ---
|
// --- Singleton Pool Instance ---
|
||||||
// This variable will hold the single, shared connection pool for the entire application.
|
// This variable will hold the single, shared connection pool for the entire application.
|
||||||
@@ -105,8 +106,9 @@ export async function checkTablesExist(tableNames: string[]): Promise<string[]>
|
|||||||
|
|
||||||
return missingTables;
|
return missingTables;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in checkTablesExist');
|
handleDbError(error, logger, 'Database error in checkTablesExist', {}, {
|
||||||
throw new Error('Failed to check for tables in database.');
|
defaultMessage: 'Failed to check for tables in database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
160
src/services/db/conversion.db.test.ts
Normal file
160
src/services/db/conversion.db.test.ts
Normal file
@@ -0,0 +1,160 @@
|
|||||||
|
// src/services/db/conversion.db.test.ts
|
||||||
|
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||||
|
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||||
|
import { getPool } from './connection.db';
|
||||||
|
import { conversionRepo } from './conversion.db';
|
||||||
|
import { NotFoundError } from './errors.db';
|
||||||
|
import type { UnitConversion } from '../../types';
|
||||||
|
|
||||||
|
// Un-mock the module we are testing
|
||||||
|
vi.unmock('./conversion.db');
|
||||||
|
|
||||||
|
// Mock dependencies
|
||||||
|
vi.mock('./connection.db', () => ({
|
||||||
|
getPool: vi.fn(),
|
||||||
|
}));
|
||||||
|
|
||||||
|
vi.mock('../logger.server', () => ({
|
||||||
|
logger: {
|
||||||
|
info: vi.fn(),
|
||||||
|
warn: vi.fn(),
|
||||||
|
error: vi.fn(),
|
||||||
|
debug: vi.fn(),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
import { logger as mockLogger } from '../logger.server';
|
||||||
|
|
||||||
|
describe('Conversion DB Service', () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
// Make getPool return our mock instance for each test
|
||||||
|
vi.mocked(getPool).mockReturnValue(mockPoolInstance as any);
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('getConversions', () => {
|
||||||
|
it('should return all conversions if no filters are provided', async () => {
|
||||||
|
const mockConversions: UnitConversion[] = [
|
||||||
|
{
|
||||||
|
unit_conversion_id: 1,
|
||||||
|
master_item_id: 1,
|
||||||
|
from_unit: 'g',
|
||||||
|
to_unit: 'kg',
|
||||||
|
factor: 0.001,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
},
|
||||||
|
];
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: mockConversions });
|
||||||
|
|
||||||
|
const result = await conversionRepo.getConversions({}, mockLogger);
|
||||||
|
|
||||||
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
|
expect.stringContaining('SELECT * FROM public.unit_conversions'),
|
||||||
|
expect.any(Array),
|
||||||
|
);
|
||||||
|
// Check that WHERE clause is not present for master_item_id
|
||||||
|
expect(mockPoolInstance.query.mock.calls[0][0]).not.toContain('WHERE master_item_id');
|
||||||
|
expect(result).toEqual(mockConversions);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should filter by masterItemId', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||||
|
|
||||||
|
await conversionRepo.getConversions({ masterItemId: 123 }, mockLogger);
|
||||||
|
|
||||||
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
|
expect.stringContaining('WHERE master_item_id = $1'),
|
||||||
|
[123],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw an error if the database query fails', async () => {
|
||||||
|
const dbError = new Error('DB Error');
|
||||||
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
|
|
||||||
|
await expect(conversionRepo.getConversions({}, mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to retrieve unit conversions.',
|
||||||
|
);
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, filters: {} },
|
||||||
|
'Database error in getConversions',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('createConversion', () => {
|
||||||
|
const newConversion = {
|
||||||
|
master_item_id: 1,
|
||||||
|
from_unit: 'cup',
|
||||||
|
to_unit: 'ml',
|
||||||
|
factor: 236.588,
|
||||||
|
};
|
||||||
|
|
||||||
|
it('should insert a new conversion and return it', async () => {
|
||||||
|
const mockCreatedConversion: UnitConversion = {
|
||||||
|
unit_conversion_id: 1,
|
||||||
|
...newConversion,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
};
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockCreatedConversion] });
|
||||||
|
|
||||||
|
const result = await conversionRepo.createConversion(newConversion, mockLogger);
|
||||||
|
|
||||||
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
|
expect.stringContaining('INSERT INTO public.unit_conversions'),
|
||||||
|
[1, 'cup', 'ml', 236.588],
|
||||||
|
);
|
||||||
|
expect(result).toEqual(mockCreatedConversion);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw an error if the database query fails', async () => {
|
||||||
|
const dbError = new Error('DB Error');
|
||||||
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
|
|
||||||
|
await expect(conversionRepo.createConversion(newConversion, mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to create unit conversion.',
|
||||||
|
);
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, conversionData: newConversion },
|
||||||
|
'Database error in createConversion',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('deleteConversion', () => {
|
||||||
|
it('should delete a conversion if found', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 1 });
|
||||||
|
|
||||||
|
await conversionRepo.deleteConversion(1, mockLogger);
|
||||||
|
|
||||||
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
|
'DELETE FROM public.unit_conversions WHERE unit_conversion_id = $1',
|
||||||
|
[1],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw NotFoundError if conversion is not found', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||||
|
|
||||||
|
await expect(conversionRepo.deleteConversion(999, mockLogger)).rejects.toThrow(NotFoundError);
|
||||||
|
await expect(conversionRepo.deleteConversion(999, mockLogger)).rejects.toThrow(
|
||||||
|
'Unit conversion with ID 999 not found.',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw an error if the database query fails', async () => {
|
||||||
|
const dbError = new Error('DB Error');
|
||||||
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
|
|
||||||
|
await expect(conversionRepo.deleteConversion(1, mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to delete unit conversion.',
|
||||||
|
);
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, conversionId: 1 },
|
||||||
|
'Database error in deleteConversion',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
78
src/services/db/conversion.db.ts
Normal file
78
src/services/db/conversion.db.ts
Normal file
@@ -0,0 +1,78 @@
|
|||||||
|
// src/services/db/conversion.db.ts
|
||||||
|
import type { Logger } from 'pino';
|
||||||
|
import { getPool } from './connection.db';
|
||||||
|
import { handleDbError, NotFoundError } from './errors.db';
|
||||||
|
import type { UnitConversion } from '../../types';
|
||||||
|
|
||||||
|
export const conversionRepo = {
|
||||||
|
/**
|
||||||
|
* Fetches unit conversions, optionally filtered by master_item_id.
|
||||||
|
*/
|
||||||
|
async getConversions(
|
||||||
|
filters: { masterItemId?: number },
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<UnitConversion[]> {
|
||||||
|
const { masterItemId } = filters;
|
||||||
|
try {
|
||||||
|
let query = 'SELECT * FROM public.unit_conversions';
|
||||||
|
const params: any[] = [];
|
||||||
|
|
||||||
|
if (masterItemId) {
|
||||||
|
query += ' WHERE master_item_id = $1';
|
||||||
|
params.push(masterItemId);
|
||||||
|
}
|
||||||
|
|
||||||
|
query += ' ORDER BY master_item_id, from_unit, to_unit';
|
||||||
|
|
||||||
|
const result = await getPool().query<UnitConversion>(query, params);
|
||||||
|
return result.rows;
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in getConversions', { filters }, {
|
||||||
|
defaultMessage: 'Failed to retrieve unit conversions.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a new unit conversion rule.
|
||||||
|
*/
|
||||||
|
async createConversion(
|
||||||
|
conversionData: Omit<UnitConversion, 'unit_conversion_id' | 'created_at' | 'updated_at'>,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<UnitConversion> {
|
||||||
|
const { master_item_id, from_unit, to_unit, factor } = conversionData;
|
||||||
|
try {
|
||||||
|
const res = await getPool().query<UnitConversion>(
|
||||||
|
'INSERT INTO public.unit_conversions (master_item_id, from_unit, to_unit, factor) VALUES ($1, $2, $3, $4) RETURNING *',
|
||||||
|
[master_item_id, from_unit, to_unit, factor],
|
||||||
|
);
|
||||||
|
return res.rows[0];
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in createConversion', { conversionData }, {
|
||||||
|
fkMessage: 'The specified master item does not exist.',
|
||||||
|
uniqueMessage: 'This conversion rule already exists for this item.',
|
||||||
|
checkMessage: 'Invalid unit conversion data provided (e.g., factor must be > 0, units cannot be the same).',
|
||||||
|
defaultMessage: 'Failed to create unit conversion.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Deletes a unit conversion rule.
|
||||||
|
*/
|
||||||
|
async deleteConversion(conversionId: number, logger: Logger): Promise<void> {
|
||||||
|
try {
|
||||||
|
const res = await getPool().query(
|
||||||
|
'DELETE FROM public.unit_conversions WHERE unit_conversion_id = $1',
|
||||||
|
[conversionId],
|
||||||
|
);
|
||||||
|
if (res.rowCount === 0) {
|
||||||
|
throw new NotFoundError(`Unit conversion with ID ${conversionId} not found.`);
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in deleteConversion', { conversionId }, {
|
||||||
|
defaultMessage: 'Failed to delete unit conversion.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
},
|
||||||
|
};
|
||||||
@@ -82,15 +82,15 @@ describe('Deals DB Service', () => {
|
|||||||
expect(result).toEqual([]);
|
expect(result).toEqual([]);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should re-throw the error if the database query fails', async () => {
|
it('should throw a generic error if the database query fails', async () => {
|
||||||
const dbError = new Error('DB Connection Error');
|
const dbError = new Error('DB Connection Error');
|
||||||
mockDb.query.mockRejectedValue(dbError);
|
mockDb.query.mockRejectedValue(dbError);
|
||||||
|
|
||||||
await expect(dealsRepo.findBestPricesForWatchedItems('user-1', mockLogger)).rejects.toThrow(
|
await expect(dealsRepo.findBestPricesForWatchedItems('user-1', mockLogger)).rejects.toThrow(
|
||||||
dbError,
|
'Failed to find best prices for watched items.',
|
||||||
);
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError },
|
{ err: dbError, userId: 'user-1' },
|
||||||
'Database error in findBestPricesForWatchedItems',
|
'Database error in findBestPricesForWatchedItems',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ import { WatchedItemDeal } from '../../types';
|
|||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { logger as globalLogger } from '../logger.server';
|
import { logger as globalLogger } from '../logger.server';
|
||||||
|
import { handleDbError } from './errors.db';
|
||||||
|
|
||||||
export class DealsRepository {
|
export class DealsRepository {
|
||||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||||
@@ -69,8 +70,9 @@ export class DealsRepository {
|
|||||||
const { rows } = await this.db.query<WatchedItemDeal>(query, [userId]);
|
const { rows } = await this.db.query<WatchedItemDeal>(query, [userId]);
|
||||||
return rows;
|
return rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in findBestPricesForWatchedItems');
|
handleDbError(error, logger, 'Database error in findBestPricesForWatchedItems', { userId }, {
|
||||||
throw error; // Re-throw the original error to be handled by the global error handler
|
defaultMessage: 'Failed to find best prices for watched items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
// src/services/db/errors.db.ts
|
// src/services/db/errors.db.ts
|
||||||
|
import type { Logger } from 'pino';
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Base class for custom database errors to ensure they have a status property.
|
* Base class for custom database errors to ensure they have a status property.
|
||||||
@@ -35,6 +36,46 @@ export class ForeignKeyConstraintError extends DatabaseError {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Thrown when a 'not null' constraint is violated.
|
||||||
|
* Corresponds to PostgreSQL error code '23502'.
|
||||||
|
*/
|
||||||
|
export class NotNullConstraintError extends DatabaseError {
|
||||||
|
constructor(message = 'A required field was left null.') {
|
||||||
|
super(message, 400); // 400 Bad Request
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Thrown when a 'check' constraint is violated.
|
||||||
|
* Corresponds to PostgreSQL error code '23514'.
|
||||||
|
*/
|
||||||
|
export class CheckConstraintError extends DatabaseError {
|
||||||
|
constructor(message = 'A check constraint was violated.') {
|
||||||
|
super(message, 400); // 400 Bad Request
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Thrown when a value has an invalid text representation for its data type (e.g., 'abc' for an integer).
|
||||||
|
* Corresponds to PostgreSQL error code '22P02'.
|
||||||
|
*/
|
||||||
|
export class InvalidTextRepresentationError extends DatabaseError {
|
||||||
|
constructor(message = 'A value has an invalid format for its data type.') {
|
||||||
|
super(message, 400); // 400 Bad Request
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Thrown when a numeric value is out of range for its data type (e.g., too large for an integer).
|
||||||
|
* Corresponds to PostgreSQL error code '22003'.
|
||||||
|
*/
|
||||||
|
export class NumericValueOutOfRangeError extends DatabaseError {
|
||||||
|
constructor(message = 'A numeric value is out of the allowed range.') {
|
||||||
|
super(message, 400); // 400 Bad Request
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Thrown when a specific record is not found in the database.
|
* Thrown when a specific record is not found in the database.
|
||||||
*/
|
*/
|
||||||
@@ -73,3 +114,50 @@ export class FileUploadError extends Error {
|
|||||||
this.name = 'FileUploadError';
|
this.name = 'FileUploadError';
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface HandleDbErrorOptions {
|
||||||
|
entityName?: string;
|
||||||
|
uniqueMessage?: string;
|
||||||
|
fkMessage?: string;
|
||||||
|
notNullMessage?: string;
|
||||||
|
checkMessage?: string;
|
||||||
|
invalidTextMessage?: string;
|
||||||
|
numericOutOfRangeMessage?: string;
|
||||||
|
defaultMessage?: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Centralized error handler for database repositories.
|
||||||
|
* Logs the error and throws appropriate custom errors based on PostgreSQL error codes.
|
||||||
|
*/
|
||||||
|
export function handleDbError(
|
||||||
|
error: unknown,
|
||||||
|
logger: Logger,
|
||||||
|
logMessage: string,
|
||||||
|
logContext: Record<string, unknown>,
|
||||||
|
options: HandleDbErrorOptions = {},
|
||||||
|
): never {
|
||||||
|
// If it's already a known domain error (like NotFoundError thrown manually), rethrow it.
|
||||||
|
if (error instanceof DatabaseError) {
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Log the raw error
|
||||||
|
logger.error({ err: error, ...logContext }, logMessage);
|
||||||
|
|
||||||
|
if (error instanceof Error && 'code' in error) {
|
||||||
|
const code = (error as any).code;
|
||||||
|
|
||||||
|
if (code === '23505') throw new UniqueConstraintError(options.uniqueMessage);
|
||||||
|
if (code === '23503') throw new ForeignKeyConstraintError(options.fkMessage);
|
||||||
|
if (code === '23502') throw new NotNullConstraintError(options.notNullMessage);
|
||||||
|
if (code === '23514') throw new CheckConstraintError(options.checkMessage);
|
||||||
|
if (code === '22P02') throw new InvalidTextRepresentationError(options.invalidTextMessage);
|
||||||
|
if (code === '22003') throw new NumericValueOutOfRangeError(options.numericOutOfRangeMessage);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fallback generic error
|
||||||
|
throw new Error(
|
||||||
|
options.defaultMessage || `Failed to perform operation on ${options.entityName || 'database'}.`,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|||||||
@@ -274,7 +274,7 @@ describe('Flyer DB Service', () => {
|
|||||||
ForeignKeyConstraintError,
|
ForeignKeyConstraintError,
|
||||||
);
|
);
|
||||||
await expect(flyerRepo.insertFlyerItems(999, itemsData, mockLogger)).rejects.toThrow(
|
await expect(flyerRepo.insertFlyerItems(999, itemsData, mockLogger)).rejects.toThrow(
|
||||||
'The specified flyer does not exist.',
|
'The specified flyer, category, master item, or product does not exist.',
|
||||||
);
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, flyerId: 999 },
|
{ err: dbError, flyerId: 999 },
|
||||||
@@ -285,10 +285,10 @@ describe('Flyer DB Service', () => {
|
|||||||
it('should throw a generic error if the database query fails', async () => {
|
it('should throw a generic error if the database query fails', async () => {
|
||||||
const dbError = new Error('DB Connection Error');
|
const dbError = new Error('DB Connection Error');
|
||||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
// The implementation now re-throws the original error, so we should expect that.
|
// The implementation wraps the error using handleDbError
|
||||||
await expect(
|
await expect(
|
||||||
flyerRepo.insertFlyerItems(1, [{ item: 'Test' } as FlyerItemInsert], mockLogger),
|
flyerRepo.insertFlyerItems(1, [{ item: 'Test' } as FlyerItemInsert], mockLogger),
|
||||||
).rejects.toThrow(dbError);
|
).rejects.toThrow('An unknown error occurred while inserting flyer items.');
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, flyerId: 1 },
|
{ err: dbError, flyerId: 1 },
|
||||||
'Database error in insertFlyerItems',
|
'Database error in insertFlyerItems',
|
||||||
@@ -691,11 +691,7 @@ describe('Flyer DB Service', () => {
|
|||||||
);
|
);
|
||||||
|
|
||||||
await expect(flyerRepo.deleteFlyer(999, mockLogger)).rejects.toThrow(
|
await expect(flyerRepo.deleteFlyer(999, mockLogger)).rejects.toThrow(
|
||||||
'Failed to delete flyer.',
|
'Flyer with ID 999 not found.',
|
||||||
);
|
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
|
||||||
{ err: expect.any(NotFoundError), flyerId: 999 },
|
|
||||||
'Database transaction error in deleteFlyer',
|
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool, withTransaction } from './connection.db';
|
import { getPool, withTransaction } from './connection.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { UniqueConstraintError, NotFoundError, handleDbError } from './errors.db';
|
||||||
import type {
|
import type {
|
||||||
Flyer,
|
Flyer,
|
||||||
FlyerItem,
|
FlyerItem,
|
||||||
@@ -97,18 +97,25 @@ export class FlyerRepository {
|
|||||||
flyerData.store_address, // $8
|
flyerData.store_address, // $8
|
||||||
flyerData.status, // $9
|
flyerData.status, // $9
|
||||||
flyerData.item_count, // $10
|
flyerData.item_count, // $10
|
||||||
flyerData.uploaded_by, // $11
|
flyerData.uploaded_by ?? null, // $11
|
||||||
];
|
];
|
||||||
|
|
||||||
const result = await this.db.query<Flyer>(query, values);
|
const result = await this.db.query<Flyer>(query, values);
|
||||||
return result.rows[0];
|
return result.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerData }, 'Database error in insertFlyer');
|
const isChecksumError =
|
||||||
// Check for a unique constraint violation on the 'checksum' column.
|
error instanceof Error && error.message.includes('flyers_checksum_check');
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23505') {
|
|
||||||
throw new UniqueConstraintError('A flyer with this checksum already exists.');
|
handleDbError(error, logger, 'Database error in insertFlyer', { flyerData }, {
|
||||||
}
|
uniqueMessage: 'A flyer with this checksum already exists.',
|
||||||
throw new Error('Failed to insert flyer into database.');
|
fkMessage: 'The specified user or store for this flyer does not exist.',
|
||||||
|
// Provide a more specific message for the checksum constraint violation,
|
||||||
|
// which is a common issue during seeding or testing with placeholder data.
|
||||||
|
checkMessage: isChecksumError
|
||||||
|
? 'The provided checksum is invalid or does not meet format requirements (e.g., must be a 64-character SHA-256 hash).'
|
||||||
|
: 'Invalid status provided for flyer.',
|
||||||
|
defaultMessage: 'Failed to insert flyer into database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -159,16 +166,10 @@ export class FlyerRepository {
|
|||||||
const result = await this.db.query<FlyerItem>(query, values);
|
const result = await this.db.query<FlyerItem>(query, values);
|
||||||
return result.rows;
|
return result.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerId }, 'Database error in insertFlyerItems');
|
handleDbError(error, logger, 'Database error in insertFlyerItems', { flyerId }, {
|
||||||
// Check for a foreign key violation, which would mean the flyerId is invalid.
|
fkMessage: 'The specified flyer, category, master item, or product does not exist.',
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
defaultMessage: 'An unknown error occurred while inserting flyer items.',
|
||||||
throw new ForeignKeyConstraintError('The specified flyer does not exist.');
|
});
|
||||||
}
|
|
||||||
// Preserve the original error if it's not a foreign key violation,
|
|
||||||
// allowing transactional functions to catch and identify the specific failure.
|
|
||||||
// This is a higher-level fix for the test failure in `createFlyerAndItems`.
|
|
||||||
if (error instanceof Error) throw error;
|
|
||||||
throw new Error('An unknown error occurred while inserting flyer items.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -179,15 +180,16 @@ export class FlyerRepository {
|
|||||||
async getAllBrands(logger: Logger): Promise<Brand[]> {
|
async getAllBrands(logger: Logger): Promise<Brand[]> {
|
||||||
try {
|
try {
|
||||||
const query = `
|
const query = `
|
||||||
SELECT s.store_id as brand_id, s.name, s.logo_url
|
SELECT s.store_id as brand_id, s.name, s.logo_url, s.created_at, s.updated_at
|
||||||
FROM public.stores s
|
FROM public.stores s
|
||||||
ORDER BY s.name;
|
ORDER BY s.name;
|
||||||
`;
|
`;
|
||||||
const res = await this.db.query<Brand>(query);
|
const res = await this.db.query<Brand>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getAllBrands');
|
handleDbError(error, logger, 'Database error in getAllBrands', {}, {
|
||||||
throw new Error('Failed to retrieve brands from database.');
|
defaultMessage: 'Failed to retrieve brands from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -226,8 +228,9 @@ export class FlyerRepository {
|
|||||||
const res = await this.db.query<Flyer>(query, [limit, offset]);
|
const res = await this.db.query<Flyer>(query, [limit, offset]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, limit, offset }, 'Database error in getFlyers');
|
handleDbError(error, logger, 'Database error in getFlyers', { limit, offset }, {
|
||||||
throw new Error('Failed to retrieve flyers from database.');
|
defaultMessage: 'Failed to retrieve flyers from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -244,8 +247,9 @@ export class FlyerRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerId }, 'Database error in getFlyerItems');
|
handleDbError(error, logger, 'Database error in getFlyerItems', { flyerId }, {
|
||||||
throw new Error('Failed to retrieve flyer items from database.');
|
defaultMessage: 'Failed to retrieve flyer items from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -262,8 +266,9 @@ export class FlyerRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerIds }, 'Database error in getFlyerItemsForFlyers');
|
handleDbError(error, logger, 'Database error in getFlyerItemsForFlyers', { flyerIds }, {
|
||||||
throw new Error('Failed to retrieve flyer items in batch from database.');
|
defaultMessage: 'Failed to retrieve flyer items in batch from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -283,8 +288,9 @@ export class FlyerRepository {
|
|||||||
);
|
);
|
||||||
return parseInt(res.rows[0].count, 10);
|
return parseInt(res.rows[0].count, 10);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerIds }, 'Database error in countFlyerItemsForFlyers');
|
handleDbError(error, logger, 'Database error in countFlyerItemsForFlyers', { flyerIds }, {
|
||||||
throw new Error('Failed to count flyer items in batch from database.');
|
defaultMessage: 'Failed to count flyer items in batch from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -300,8 +306,9 @@ export class FlyerRepository {
|
|||||||
]);
|
]);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, checksum }, 'Database error in findFlyerByChecksum');
|
handleDbError(error, logger, 'Database error in findFlyerByChecksum', { checksum }, {
|
||||||
throw new Error('Failed to find flyer by checksum in database.');
|
defaultMessage: 'Failed to find flyer by checksum in database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -353,8 +360,9 @@ export class FlyerRepository {
|
|||||||
logger.info(`Successfully deleted flyer with ID: ${flyerId}`);
|
logger.info(`Successfully deleted flyer with ID: ${flyerId}`);
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, flyerId }, 'Database transaction error in deleteFlyer');
|
handleDbError(error, logger, 'Database transaction error in deleteFlyer', { flyerId }, {
|
||||||
throw new Error('Failed to delete flyer.');
|
defaultMessage: 'Failed to delete flyer.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/gamification.db.ts
|
// src/services/db/gamification.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
import { ForeignKeyConstraintError } from './errors.db';
|
import { handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { Achievement, UserAchievement, LeaderboardUser } from '../../types';
|
import { Achievement, UserAchievement, LeaderboardUser } from '../../types';
|
||||||
|
|
||||||
@@ -25,8 +25,9 @@ export class GamificationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getAllAchievements');
|
handleDbError(error, logger, 'Database error in getAllAchievements', {}, {
|
||||||
throw new Error('Failed to retrieve achievements.');
|
defaultMessage: 'Failed to retrieve achievements.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -49,7 +50,8 @@ export class GamificationRepository {
|
|||||||
a.name,
|
a.name,
|
||||||
a.description,
|
a.description,
|
||||||
a.icon,
|
a.icon,
|
||||||
a.points_value
|
a.points_value,
|
||||||
|
a.created_at
|
||||||
FROM public.user_achievements ua
|
FROM public.user_achievements ua
|
||||||
JOIN public.achievements a ON ua.achievement_id = a.achievement_id
|
JOIN public.achievements a ON ua.achievement_id = a.achievement_id
|
||||||
WHERE ua.user_id = $1
|
WHERE ua.user_id = $1
|
||||||
@@ -58,8 +60,9 @@ export class GamificationRepository {
|
|||||||
const res = await this.db.query<UserAchievement & Achievement>(query, [userId]);
|
const res = await this.db.query<UserAchievement & Achievement>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getUserAchievements');
|
handleDbError(error, logger, 'Database error in getUserAchievements', { userId }, {
|
||||||
throw new Error('Failed to retrieve user achievements.');
|
defaultMessage: 'Failed to retrieve user achievements.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -75,12 +78,10 @@ export class GamificationRepository {
|
|||||||
try {
|
try {
|
||||||
await this.db.query('SELECT public.award_achievement($1, $2)', [userId, achievementName]); // This was a duplicate, fixed.
|
await this.db.query('SELECT public.award_achievement($1, $2)', [userId, achievementName]); // This was a duplicate, fixed.
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, achievementName }, 'Database error in awardAchievement');
|
handleDbError(error, logger, 'Database error in awardAchievement', { userId, achievementName }, {
|
||||||
// Check for a foreign key violation, which would mean the user or achievement name is invalid.
|
fkMessage: 'The specified user or achievement does not exist.',
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
defaultMessage: 'Failed to award achievement.',
|
||||||
throw new ForeignKeyConstraintError('The specified user or achievement does not exist.');
|
});
|
||||||
}
|
|
||||||
throw new Error('Failed to award achievement.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -105,8 +106,9 @@ export class GamificationRepository {
|
|||||||
const res = await this.db.query<LeaderboardUser>(query, [limit]);
|
const res = await this.db.query<LeaderboardUser>(query, [limit]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, limit }, 'Database error in getLeaderboard');
|
handleDbError(error, logger, 'Database error in getLeaderboard', { limit }, {
|
||||||
throw new Error('Failed to retrieve leaderboard.');
|
defaultMessage: 'Failed to retrieve leaderboard.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,6 +10,8 @@ import { NotificationRepository } from './notification.db';
|
|||||||
import { BudgetRepository } from './budget.db';
|
import { BudgetRepository } from './budget.db';
|
||||||
import { GamificationRepository } from './gamification.db';
|
import { GamificationRepository } from './gamification.db';
|
||||||
import { AdminRepository } from './admin.db';
|
import { AdminRepository } from './admin.db';
|
||||||
|
import { reactionRepo } from './reaction.db';
|
||||||
|
import { conversionRepo } from './conversion.db';
|
||||||
|
|
||||||
const userRepo = new UserRepository();
|
const userRepo = new UserRepository();
|
||||||
const flyerRepo = new FlyerRepository();
|
const flyerRepo = new FlyerRepository();
|
||||||
@@ -33,5 +35,7 @@ export {
|
|||||||
budgetRepo,
|
budgetRepo,
|
||||||
gamificationRepo,
|
gamificationRepo,
|
||||||
adminRepo,
|
adminRepo,
|
||||||
|
reactionRepo,
|
||||||
|
conversionRepo,
|
||||||
withTransaction,
|
withTransaction,
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -195,7 +195,7 @@ describe('Notification DB Service', () => {
|
|||||||
notificationRepo.createBulkNotifications(notificationsToCreate, mockLogger),
|
notificationRepo.createBulkNotifications(notificationsToCreate, mockLogger),
|
||||||
).rejects.toThrow(ForeignKeyConstraintError);
|
).rejects.toThrow(ForeignKeyConstraintError);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError },
|
{ err: dbError, notifications: notificationsToCreate },
|
||||||
'Database error in createBulkNotifications',
|
'Database error in createBulkNotifications',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -208,7 +208,7 @@ describe('Notification DB Service', () => {
|
|||||||
notificationRepo.createBulkNotifications(notificationsToCreate, mockLogger),
|
notificationRepo.createBulkNotifications(notificationsToCreate, mockLogger),
|
||||||
).rejects.toThrow('Failed to create bulk notifications.');
|
).rejects.toThrow('Failed to create bulk notifications.');
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError },
|
{ err: dbError, notifications: notificationsToCreate },
|
||||||
'Database error in createBulkNotifications',
|
'Database error in createBulkNotifications',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -264,6 +264,16 @@ describe('Notification DB Service', () => {
|
|||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('markNotificationAsRead - Ownership Check', () => {
|
||||||
|
it('should not mark a notification as read if the user does not own it', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||||
|
|
||||||
|
await expect(notificationRepo.markNotificationAsRead(1, 'wrong-user', mockLogger)).rejects.toThrow(
|
||||||
|
'Notification not found or user does not have permission.',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
describe('markAllNotificationsAsRead', () => {
|
describe('markAllNotificationsAsRead', () => {
|
||||||
it('should execute an UPDATE query to mark all notifications as read for a user', async () => {
|
it('should execute an UPDATE query to mark all notifications as read for a user', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 3 });
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 3 });
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/notification.db.ts
|
// src/services/db/notification.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { NotFoundError, handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { Notification } from '../../types';
|
import type { Notification } from '../../types';
|
||||||
|
|
||||||
@@ -34,14 +34,10 @@ export class NotificationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in createNotification', { userId, content, linkUrl }, {
|
||||||
{ err: error, userId, content, linkUrl },
|
fkMessage: 'The specified user does not exist.',
|
||||||
'Database error in createNotification',
|
defaultMessage: 'Failed to create notification.',
|
||||||
);
|
});
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
|
||||||
throw new ForeignKeyConstraintError('The specified user does not exist.');
|
|
||||||
}
|
|
||||||
throw new Error('Failed to create notification.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -78,11 +74,10 @@ export class NotificationRepository {
|
|||||||
|
|
||||||
await this.db.query(query, [userIds, contents, linkUrls]);
|
await this.db.query(query, [userIds, contents, linkUrls]);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in createBulkNotifications');
|
handleDbError(error, logger, 'Database error in createBulkNotifications', { notifications }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'One or more of the specified users do not exist.',
|
||||||
throw new ForeignKeyConstraintError('One or more of the specified users do not exist.');
|
defaultMessage: 'Failed to create bulk notifications.',
|
||||||
}
|
});
|
||||||
throw new Error('Failed to create bulk notifications.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -113,11 +108,13 @@ export class NotificationRepository {
|
|||||||
const res = await this.db.query<Notification>(query, params);
|
const res = await this.db.query<Notification>(query, params);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, userId, limit, offset, includeRead },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in getNotificationsForUser',
|
'Database error in getNotificationsForUser',
|
||||||
|
{ userId, limit, offset, includeRead },
|
||||||
|
{ defaultMessage: 'Failed to retrieve notifications.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to retrieve notifications.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -133,8 +130,9 @@ export class NotificationRepository {
|
|||||||
[userId],
|
[userId],
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in markAllNotificationsAsRead');
|
handleDbError(error, logger, 'Database error in markAllNotificationsAsRead', { userId }, {
|
||||||
throw new Error('Failed to mark notifications as read.');
|
defaultMessage: 'Failed to mark notifications as read.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -161,12 +159,13 @@ export class NotificationRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
handleDbError(
|
||||||
logger.error(
|
error,
|
||||||
{ err: error, notificationId, userId },
|
logger,
|
||||||
'Database error in markNotificationAsRead',
|
'Database error in markNotificationAsRead',
|
||||||
|
{ notificationId, userId },
|
||||||
|
{ defaultMessage: 'Failed to mark notification as read.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to mark notification as read.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -184,8 +183,9 @@ export class NotificationRepository {
|
|||||||
);
|
);
|
||||||
return res.rowCount ?? 0;
|
return res.rowCount ?? 0;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, daysOld }, 'Database error in deleteOldNotifications');
|
handleDbError(error, logger, 'Database error in deleteOldNotifications', { daysOld }, {
|
||||||
throw new Error('Failed to delete old notifications.');
|
defaultMessage: 'Failed to delete old notifications.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ import type { Pool, PoolClient } from 'pg';
|
|||||||
import { withTransaction } from './connection.db';
|
import { withTransaction } from './connection.db';
|
||||||
import { PersonalizationRepository } from './personalization.db';
|
import { PersonalizationRepository } from './personalization.db';
|
||||||
import type { MasterGroceryItem, UserAppliance, DietaryRestriction, Appliance } from '../../types';
|
import type { MasterGroceryItem, UserAppliance, DietaryRestriction, Appliance } from '../../types';
|
||||||
import { createMockMasterGroceryItem } from '../../tests/utils/mockFactories';
|
import { createMockMasterGroceryItem, createMockUserAppliance } from '../../tests/utils/mockFactories';
|
||||||
|
|
||||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||||
vi.unmock('./personalization.db');
|
vi.unmock('./personalization.db');
|
||||||
@@ -46,9 +46,6 @@ describe('Personalization DB Service', () => {
|
|||||||
|
|
||||||
describe('getAllMasterItems', () => {
|
describe('getAllMasterItems', () => {
|
||||||
it('should execute the correct query and return master items', async () => {
|
it('should execute the correct query and return master items', async () => {
|
||||||
console.log(
|
|
||||||
'[TEST DEBUG] Running test: getAllMasterItems > should execute the correct query',
|
|
||||||
);
|
|
||||||
const mockItems: MasterGroceryItem[] = [
|
const mockItems: MasterGroceryItem[] = [
|
||||||
createMockMasterGroceryItem({ master_grocery_item_id: 1, name: 'Apples' }),
|
createMockMasterGroceryItem({ master_grocery_item_id: 1, name: 'Apples' }),
|
||||||
];
|
];
|
||||||
@@ -64,8 +61,6 @@ describe('Personalization DB Service', () => {
|
|||||||
LEFT JOIN public.categories c ON mgi.category_id = c.category_id
|
LEFT JOIN public.categories c ON mgi.category_id = c.category_id
|
||||||
ORDER BY mgi.name ASC`;
|
ORDER BY mgi.name ASC`;
|
||||||
|
|
||||||
console.log('[TEST DEBUG] mockQuery calls:', JSON.stringify(mockQuery.mock.calls, null, 2));
|
|
||||||
|
|
||||||
// The query string in the implementation has a lot of whitespace from the template literal.
|
// The query string in the implementation has a lot of whitespace from the template literal.
|
||||||
// This updated expectation matches the new query exactly.
|
// This updated expectation matches the new query exactly.
|
||||||
expect(mockQuery).toHaveBeenCalledWith(expectedQuery);
|
expect(mockQuery).toHaveBeenCalledWith(expectedQuery);
|
||||||
@@ -649,8 +644,8 @@ describe('Personalization DB Service', () => {
|
|||||||
describe('setUserAppliances', () => {
|
describe('setUserAppliances', () => {
|
||||||
it('should execute a transaction to set appliances', async () => {
|
it('should execute a transaction to set appliances', async () => {
|
||||||
const mockNewAppliances: UserAppliance[] = [
|
const mockNewAppliances: UserAppliance[] = [
|
||||||
{ user_id: 'user-123', appliance_id: 1 },
|
createMockUserAppliance({ user_id: 'user-123', appliance_id: 1 }),
|
||||||
{ user_id: 'user-123', appliance_id: 2 },
|
createMockUserAppliance({ user_id: 'user-123', appliance_id: 2 }),
|
||||||
];
|
];
|
||||||
const mockClientQuery = vi.fn();
|
const mockClientQuery = vi.fn();
|
||||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/personalization.db.ts
|
// src/services/db/personalization.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool, withTransaction } from './connection.db';
|
import { getPool, withTransaction } from './connection.db';
|
||||||
import { ForeignKeyConstraintError } from './errors.db';
|
import { handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import {
|
import {
|
||||||
MasterGroceryItem,
|
MasterGroceryItem,
|
||||||
@@ -40,8 +40,9 @@ export class PersonalizationRepository {
|
|||||||
const res = await this.db.query<MasterGroceryItem>(query);
|
const res = await this.db.query<MasterGroceryItem>(query);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getAllMasterItems');
|
handleDbError(error, logger, 'Database error in getAllMasterItems', {}, {
|
||||||
throw new Error('Failed to retrieve master grocery items.');
|
defaultMessage: 'Failed to retrieve master grocery items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -62,8 +63,9 @@ export class PersonalizationRepository {
|
|||||||
const res = await this.db.query<MasterGroceryItem>(query, [userId]);
|
const res = await this.db.query<MasterGroceryItem>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getWatchedItems');
|
handleDbError(error, logger, 'Database error in getWatchedItems', { userId }, {
|
||||||
throw new Error('Failed to retrieve watched items.');
|
defaultMessage: 'Failed to retrieve watched items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -79,8 +81,9 @@ export class PersonalizationRepository {
|
|||||||
[userId, masterItemId],
|
[userId, masterItemId],
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, masterItemId }, 'Database error in removeWatchedItem');
|
handleDbError(error, logger, 'Database error in removeWatchedItem', { userId, masterItemId }, {
|
||||||
throw new Error('Failed to remove item from watchlist.');
|
defaultMessage: 'Failed to remove item from watchlist.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -100,8 +103,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, pantryItemId }, 'Database error in findPantryItemOwner');
|
handleDbError(error, logger, 'Database error in findPantryItemOwner', { pantryItemId }, {
|
||||||
throw new Error('Failed to retrieve pantry item owner from database.');
|
defaultMessage: 'Failed to retrieve pantry item owner from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -156,18 +160,17 @@ export class PersonalizationRepository {
|
|||||||
return masterItem;
|
return masterItem;
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The withTransaction helper will handle rollback. We just need to handle specific errors.
|
handleDbError(
|
||||||
if (error instanceof Error && 'code' in error) {
|
error,
|
||||||
if (error.code === '23503') {
|
logger,
|
||||||
// foreign_key_violation
|
|
||||||
throw new ForeignKeyConstraintError('The specified user or category does not exist.');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
logger.error(
|
|
||||||
{ err: error, userId, itemName, categoryName },
|
|
||||||
'Transaction error in addWatchedItem',
|
'Transaction error in addWatchedItem',
|
||||||
|
{ userId, itemName, categoryName },
|
||||||
|
{
|
||||||
|
fkMessage: 'The specified user or category does not exist.',
|
||||||
|
uniqueMessage: 'A master grocery item with this name was created by another process.',
|
||||||
|
defaultMessage: 'Failed to add item to watchlist.',
|
||||||
|
},
|
||||||
);
|
);
|
||||||
throw new Error('Failed to add item to watchlist.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -186,8 +189,9 @@ export class PersonalizationRepository {
|
|||||||
>('SELECT * FROM public.get_best_sale_prices_for_all_users()');
|
>('SELECT * FROM public.get_best_sale_prices_for_all_users()');
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getBestSalePricesForAllUsers');
|
handleDbError(error, logger, 'Database error in getBestSalePricesForAllUsers', {}, {
|
||||||
throw new Error('Failed to get best sale prices for all users.');
|
defaultMessage: 'Failed to get best sale prices for all users.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -200,8 +204,9 @@ export class PersonalizationRepository {
|
|||||||
const res = await this.db.query<Appliance>('SELECT * FROM public.appliances ORDER BY name');
|
const res = await this.db.query<Appliance>('SELECT * FROM public.appliances ORDER BY name');
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getAppliances');
|
handleDbError(error, logger, 'Database error in getAppliances', {}, {
|
||||||
throw new Error('Failed to get appliances.');
|
defaultMessage: 'Failed to get appliances.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -216,8 +221,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in getDietaryRestrictions');
|
handleDbError(error, logger, 'Database error in getDietaryRestrictions', {}, {
|
||||||
throw new Error('Failed to get dietary restrictions.');
|
defaultMessage: 'Failed to get dietary restrictions.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -236,8 +242,9 @@ export class PersonalizationRepository {
|
|||||||
const res = await this.db.query<DietaryRestriction>(query, [userId]);
|
const res = await this.db.query<DietaryRestriction>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getUserDietaryRestrictions');
|
handleDbError(error, logger, 'Database error in getUserDietaryRestrictions', { userId }, {
|
||||||
throw new Error('Failed to get user dietary restrictions.');
|
defaultMessage: 'Failed to get user dietary restrictions.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -266,17 +273,13 @@ export class PersonalizationRepository {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// Check for a foreign key violation, which would mean an invalid ID was provided.
|
handleDbError(
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
error,
|
||||||
throw new ForeignKeyConstraintError(
|
logger,
|
||||||
'One or more of the specified restriction IDs are invalid.',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logger.error(
|
|
||||||
{ err: error, userId, restrictionIds },
|
|
||||||
'Database error in setUserDietaryRestrictions',
|
'Database error in setUserDietaryRestrictions',
|
||||||
|
{ userId, restrictionIds },
|
||||||
|
{ fkMessage: 'One or more of the specified restriction IDs are invalid.', defaultMessage: 'Failed to set user dietary restrictions.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to set user dietary restrictions.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -306,12 +309,10 @@ export class PersonalizationRepository {
|
|||||||
return newAppliances;
|
return newAppliances;
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// Check for a foreign key violation, which would mean an invalid ID was provided.
|
handleDbError(error, logger, 'Database error in setUserAppliances', { userId, applianceIds }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'Invalid appliance ID',
|
||||||
throw new ForeignKeyConstraintError('Invalid appliance ID');
|
defaultMessage: 'Failed to set user appliances.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, userId, applianceIds }, 'Database error in setUserAppliances');
|
|
||||||
throw new Error('Failed to set user appliances.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -330,8 +331,9 @@ export class PersonalizationRepository {
|
|||||||
const res = await this.db.query<Appliance>(query, [userId]);
|
const res = await this.db.query<Appliance>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getUserAppliances');
|
handleDbError(error, logger, 'Database error in getUserAppliances', { userId }, {
|
||||||
throw new Error('Failed to get user appliances.');
|
defaultMessage: 'Failed to get user appliances.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -348,8 +350,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in findRecipesFromPantry');
|
handleDbError(error, logger, 'Database error in findRecipesFromPantry', { userId }, {
|
||||||
throw new Error('Failed to find recipes from pantry.');
|
defaultMessage: 'Failed to find recipes from pantry.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -371,8 +374,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, limit }, 'Database error in recommendRecipesForUser');
|
handleDbError(error, logger, 'Database error in recommendRecipesForUser', { userId, limit }, {
|
||||||
throw new Error('Failed to recommend recipes.');
|
defaultMessage: 'Failed to recommend recipes.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -389,8 +393,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getBestSalePricesForUser');
|
handleDbError(error, logger, 'Database error in getBestSalePricesForUser', { userId }, {
|
||||||
throw new Error('Failed to get best sale prices.');
|
defaultMessage: 'Failed to get best sale prices.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -410,8 +415,9 @@ export class PersonalizationRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, pantryItemId }, 'Database error in suggestPantryItemConversions');
|
handleDbError(error, logger, 'Database error in suggestPantryItemConversions', { pantryItemId }, {
|
||||||
throw new Error('Failed to suggest pantry item conversions.');
|
defaultMessage: 'Failed to suggest pantry item conversions.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -428,8 +434,9 @@ export class PersonalizationRepository {
|
|||||||
); // This is a standalone function, no change needed here.
|
); // This is a standalone function, no change needed here.
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getRecipesForUserDiets');
|
handleDbError(error, logger, 'Database error in getRecipesForUserDiets', { userId }, {
|
||||||
throw new Error('Failed to get recipes compatible with user diet.');
|
defaultMessage: 'Failed to get recipes compatible with user diet.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,6 +2,7 @@
|
|||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { PriceHistoryData } from '../../types';
|
import type { PriceHistoryData } from '../../types';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
|
import { handleDbError } from './errors.db';
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Repository for fetching price-related data.
|
* Repository for fetching price-related data.
|
||||||
@@ -51,11 +52,13 @@ export const priceRepo = {
|
|||||||
);
|
);
|
||||||
return result.rows;
|
return result.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, masterItemIds, limit, offset },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in getPriceHistory',
|
'Database error in getPriceHistory',
|
||||||
|
{ masterItemIds, limit, offset },
|
||||||
|
{ defaultMessage: 'Failed to retrieve price history.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to retrieve price history.');
|
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
225
src/services/db/reaction.db.test.ts
Normal file
225
src/services/db/reaction.db.test.ts
Normal file
@@ -0,0 +1,225 @@
|
|||||||
|
// src/services/db/reaction.db.test.ts
|
||||||
|
import { describe, it, expect, vi, beforeEach, Mock } from 'vitest';
|
||||||
|
import type { Pool, PoolClient } from 'pg';
|
||||||
|
import { ReactionRepository } from './reaction.db';
|
||||||
|
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||||
|
import { withTransaction } from './connection.db';
|
||||||
|
import { ForeignKeyConstraintError } from './errors.db';
|
||||||
|
import type { UserReaction } from '../../types';
|
||||||
|
|
||||||
|
// Un-mock the module we are testing
|
||||||
|
vi.unmock('./reaction.db');
|
||||||
|
|
||||||
|
// Mock dependencies
|
||||||
|
vi.mock('../logger.server', () => ({
|
||||||
|
logger: {
|
||||||
|
info: vi.fn(),
|
||||||
|
warn: vi.fn(),
|
||||||
|
error: vi.fn(),
|
||||||
|
debug: vi.fn(),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
import { logger as mockLogger } from '../logger.server';
|
||||||
|
|
||||||
|
vi.mock('./connection.db', async (importOriginal) => {
|
||||||
|
const actual = await importOriginal<typeof import('./connection.db')>();
|
||||||
|
return { ...actual, withTransaction: vi.fn() };
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('Reaction DB Service', () => {
|
||||||
|
let reactionRepo: ReactionRepository;
|
||||||
|
const mockDb = {
|
||||||
|
query: vi.fn(),
|
||||||
|
};
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
vi.clearAllMocks();
|
||||||
|
reactionRepo = new ReactionRepository(mockDb);
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('getReactions', () => {
|
||||||
|
it('should build a query with no filters', async () => {
|
||||||
|
mockDb.query.mockResolvedValue({ rows: [] });
|
||||||
|
await reactionRepo.getReactions({}, mockLogger);
|
||||||
|
expect(mockDb.query).toHaveBeenCalledWith(
|
||||||
|
'SELECT * FROM public.user_reactions WHERE 1=1 ORDER BY created_at DESC',
|
||||||
|
[],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should build a query with a userId filter', async () => {
|
||||||
|
mockDb.query.mockResolvedValue({ rows: [] });
|
||||||
|
await reactionRepo.getReactions({ userId: 'user-1' }, mockLogger);
|
||||||
|
expect(mockDb.query).toHaveBeenCalledWith(
|
||||||
|
'SELECT * FROM public.user_reactions WHERE 1=1 AND user_id = $1 ORDER BY created_at DESC',
|
||||||
|
['user-1'],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should build a query with all filters', async () => {
|
||||||
|
mockDb.query.mockResolvedValue({ rows: [] });
|
||||||
|
await reactionRepo.getReactions(
|
||||||
|
{ userId: 'user-1', entityType: 'recipe', entityId: '123' },
|
||||||
|
mockLogger,
|
||||||
|
);
|
||||||
|
expect(mockDb.query).toHaveBeenCalledWith(
|
||||||
|
'SELECT * FROM public.user_reactions WHERE 1=1 AND user_id = $1 AND entity_type = $2 AND entity_id = $3 ORDER BY created_at DESC',
|
||||||
|
['user-1', 'recipe', '123'],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should return an array of reactions on success', async () => {
|
||||||
|
const mockReactions: UserReaction[] = [
|
||||||
|
{
|
||||||
|
reaction_id: 1,
|
||||||
|
user_id: 'user-1',
|
||||||
|
entity_type: 'recipe',
|
||||||
|
entity_id: '123',
|
||||||
|
reaction_type: 'like',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
},
|
||||||
|
];
|
||||||
|
mockDb.query.mockResolvedValue({ rows: mockReactions });
|
||||||
|
const result = await reactionRepo.getReactions({}, mockLogger);
|
||||||
|
expect(result).toEqual(mockReactions);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw an error if the database query fails', async () => {
|
||||||
|
const dbError = new Error('DB Error');
|
||||||
|
mockDb.query.mockRejectedValue(dbError);
|
||||||
|
await expect(reactionRepo.getReactions({}, mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to retrieve user reactions.',
|
||||||
|
);
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, filters: {} },
|
||||||
|
'Database error in getReactions',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('toggleReaction', () => {
|
||||||
|
const reactionData = {
|
||||||
|
user_id: 'user-1',
|
||||||
|
entity_type: 'recipe',
|
||||||
|
entity_id: '123',
|
||||||
|
reaction_type: 'like',
|
||||||
|
};
|
||||||
|
|
||||||
|
it('should remove an existing reaction and return null', async () => {
|
||||||
|
const mockClient = { query: vi.fn() };
|
||||||
|
// Mock DELETE returning 1 row, indicating a reaction was deleted
|
||||||
|
(mockClient.query as Mock).mockResolvedValueOnce({ rowCount: 1 });
|
||||||
|
|
||||||
|
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||||
|
return callback(mockClient as unknown as PoolClient);
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await reactionRepo.toggleReaction(reactionData, mockLogger);
|
||||||
|
|
||||||
|
expect(result).toBeNull();
|
||||||
|
expect(mockClient.query).toHaveBeenCalledWith(
|
||||||
|
'DELETE FROM public.user_reactions WHERE user_id = $1 AND entity_type = $2 AND entity_id = $3 AND reaction_type = $4',
|
||||||
|
['user-1', 'recipe', '123', 'like'],
|
||||||
|
);
|
||||||
|
// Ensure INSERT was not called
|
||||||
|
expect(mockClient.query).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should add a new reaction and return it if it does not exist', async () => {
|
||||||
|
const mockClient = { query: vi.fn() };
|
||||||
|
const mockCreatedReaction: UserReaction = {
|
||||||
|
reaction_id: 1,
|
||||||
|
...reactionData,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
};
|
||||||
|
|
||||||
|
// Mock DELETE returning 0 rows, then mock INSERT returning the new reaction
|
||||||
|
(mockClient.query as Mock)
|
||||||
|
.mockResolvedValueOnce({ rowCount: 0 }) // DELETE
|
||||||
|
.mockResolvedValueOnce({ rows: [mockCreatedReaction] }); // INSERT
|
||||||
|
|
||||||
|
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||||
|
return callback(mockClient as unknown as PoolClient);
|
||||||
|
});
|
||||||
|
|
||||||
|
const result = await reactionRepo.toggleReaction(reactionData, mockLogger);
|
||||||
|
|
||||||
|
expect(result).toEqual(mockCreatedReaction);
|
||||||
|
expect(mockClient.query).toHaveBeenCalledTimes(2);
|
||||||
|
expect(mockClient.query).toHaveBeenCalledWith(
|
||||||
|
'INSERT INTO public.user_reactions (user_id, entity_type, entity_id, reaction_type) VALUES ($1, $2, $3, $4) RETURNING *',
|
||||||
|
['user-1', 'recipe', '123', 'like'],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw ForeignKeyConstraintError if user or entity does not exist', async () => {
|
||||||
|
const dbError = new Error('violates foreign key constraint');
|
||||||
|
(dbError as Error & { code: string }).code = '23503';
|
||||||
|
|
||||||
|
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
||||||
|
const mockClient = { query: vi.fn().mockRejectedValue(dbError) };
|
||||||
|
await expect(callback(mockClient as unknown as PoolClient)).rejects.toThrow(dbError);
|
||||||
|
throw dbError;
|
||||||
|
});
|
||||||
|
|
||||||
|
await expect(reactionRepo.toggleReaction(reactionData, mockLogger)).rejects.toThrow(
|
||||||
|
ForeignKeyConstraintError,
|
||||||
|
);
|
||||||
|
await expect(reactionRepo.toggleReaction(reactionData, mockLogger)).rejects.toThrow(
|
||||||
|
'The specified user or entity does not exist.',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw a generic error if the transaction fails', async () => {
|
||||||
|
const dbError = new Error('Transaction failed');
|
||||||
|
vi.mocked(withTransaction).mockRejectedValue(dbError);
|
||||||
|
|
||||||
|
await expect(reactionRepo.toggleReaction(reactionData, mockLogger)).rejects.toThrow(
|
||||||
|
'Failed to toggle user reaction.',
|
||||||
|
);
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, reactionData },
|
||||||
|
'Database error in toggleReaction',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('getReactionSummary', () => {
|
||||||
|
it('should return a summary of reactions for an entity', async () => {
|
||||||
|
const mockSummary = [
|
||||||
|
{ reaction_type: 'like', count: 5 },
|
||||||
|
{ reaction_type: 'heart', count: 2 },
|
||||||
|
];
|
||||||
|
// This method uses getPool() directly, so we mock the main instance
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: mockSummary });
|
||||||
|
|
||||||
|
const result = await reactionRepo.getReactionSummary('recipe', '123', mockLogger);
|
||||||
|
|
||||||
|
expect(result).toEqual(mockSummary);
|
||||||
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
|
expect.stringContaining('GROUP BY reaction_type'),
|
||||||
|
['recipe', '123'],
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should return an empty array if there are no reactions', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||||
|
const result = await reactionRepo.getReactionSummary('recipe', '456', mockLogger);
|
||||||
|
expect(result).toEqual([]);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should throw an error if the database query fails', async () => {
|
||||||
|
const dbError = new Error('DB Error');
|
||||||
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
|
await expect(
|
||||||
|
reactionRepo.getReactionSummary('recipe', '123', mockLogger),
|
||||||
|
).rejects.toThrow('Failed to retrieve reaction summary.');
|
||||||
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
|
{ err: dbError, entityType: 'recipe', entityId: '123' },
|
||||||
|
'Database error in getReactionSummary',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
131
src/services/db/reaction.db.ts
Normal file
131
src/services/db/reaction.db.ts
Normal file
@@ -0,0 +1,131 @@
|
|||||||
|
// src/services/db/reaction.db.ts
|
||||||
|
import type { Pool, PoolClient } from 'pg';
|
||||||
|
import type { Logger } from 'pino';
|
||||||
|
import { getPool, withTransaction } from './connection.db';
|
||||||
|
import { handleDbError } from './errors.db';
|
||||||
|
import type { UserReaction } from '../../types';
|
||||||
|
|
||||||
|
export class ReactionRepository {
|
||||||
|
private db: Pick<Pool | PoolClient, 'query'>;
|
||||||
|
|
||||||
|
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||||
|
this.db = db;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Fetches user reactions based on query filters.
|
||||||
|
* Supports filtering by user_id, entity_type, and entity_id.
|
||||||
|
*/
|
||||||
|
async getReactions(
|
||||||
|
filters: {
|
||||||
|
userId?: string;
|
||||||
|
entityType?: string;
|
||||||
|
entityId?: string;
|
||||||
|
},
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<UserReaction[]> {
|
||||||
|
const { userId, entityType, entityId } = filters;
|
||||||
|
try {
|
||||||
|
let query = 'SELECT * FROM public.user_reactions WHERE 1=1';
|
||||||
|
const params: any[] = [];
|
||||||
|
let paramCount = 1;
|
||||||
|
|
||||||
|
if (userId) {
|
||||||
|
query += ` AND user_id = $${paramCount++}`;
|
||||||
|
params.push(userId);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (entityType) {
|
||||||
|
query += ` AND entity_type = $${paramCount++}`;
|
||||||
|
params.push(entityType);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (entityId) {
|
||||||
|
query += ` AND entity_id = $${paramCount++}`;
|
||||||
|
params.push(entityId);
|
||||||
|
}
|
||||||
|
|
||||||
|
query += ' ORDER BY created_at DESC';
|
||||||
|
|
||||||
|
const result = await this.db.query<UserReaction>(query, params);
|
||||||
|
return result.rows;
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in getReactions', { filters }, {
|
||||||
|
defaultMessage: 'Failed to retrieve user reactions.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Toggles a user's reaction to an entity.
|
||||||
|
* If the reaction exists, it's deleted. If it doesn't, it's created.
|
||||||
|
* @returns The created UserReaction if a reaction was added, or null if it was removed.
|
||||||
|
*/
|
||||||
|
async toggleReaction(
|
||||||
|
reactionData: Omit<UserReaction, 'reaction_id' | 'created_at' | 'updated_at'>,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<UserReaction | null> {
|
||||||
|
const { user_id, entity_type, entity_id, reaction_type } = reactionData;
|
||||||
|
|
||||||
|
try {
|
||||||
|
return await withTransaction(async (client) => {
|
||||||
|
const deleteRes = await client.query(
|
||||||
|
'DELETE FROM public.user_reactions WHERE user_id = $1 AND entity_type = $2 AND entity_id = $3 AND reaction_type = $4',
|
||||||
|
[user_id, entity_type, entity_id, reaction_type],
|
||||||
|
);
|
||||||
|
|
||||||
|
if ((deleteRes.rowCount ?? 0) > 0) {
|
||||||
|
logger.debug({ reactionData }, 'Reaction removed.');
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
const insertRes = await client.query<UserReaction>(
|
||||||
|
'INSERT INTO public.user_reactions (user_id, entity_type, entity_id, reaction_type) VALUES ($1, $2, $3, $4) RETURNING *',
|
||||||
|
[user_id, entity_type, entity_id, reaction_type],
|
||||||
|
);
|
||||||
|
|
||||||
|
logger.debug({ reaction: insertRes.rows[0] }, 'Reaction added.');
|
||||||
|
return insertRes.rows[0];
|
||||||
|
});
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in toggleReaction', { reactionData }, {
|
||||||
|
fkMessage: 'The specified user or entity does not exist.',
|
||||||
|
defaultMessage: 'Failed to toggle user reaction.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets a summary of reactions for a specific entity.
|
||||||
|
* Counts the number of each reaction_type.
|
||||||
|
* @param entityType The type of the entity (e.g., 'recipe').
|
||||||
|
* @param entityId The ID of the entity.
|
||||||
|
* @param logger The pino logger instance.
|
||||||
|
* @returns A promise that resolves to an array of reaction summaries.
|
||||||
|
*/
|
||||||
|
async getReactionSummary(
|
||||||
|
entityType: string,
|
||||||
|
entityId: string,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<{ reaction_type: string; count: number }[]> {
|
||||||
|
try {
|
||||||
|
const query = `
|
||||||
|
SELECT
|
||||||
|
reaction_type,
|
||||||
|
COUNT(*)::int as count
|
||||||
|
FROM public.user_reactions
|
||||||
|
WHERE entity_type = $1 AND entity_id = $2
|
||||||
|
GROUP BY reaction_type
|
||||||
|
ORDER BY count DESC;
|
||||||
|
`;
|
||||||
|
const result = await getPool().query<{ reaction_type: string; count: number }>(query, [entityType, entityId]);
|
||||||
|
return result.rows;
|
||||||
|
} catch (error) {
|
||||||
|
handleDbError(error, logger, 'Database error in getReactionSummary', { entityType, entityId }, {
|
||||||
|
defaultMessage: 'Failed to retrieve reaction summary.',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const reactionRepo = new ReactionRepository();
|
||||||
@@ -268,6 +268,17 @@ describe('Recipe DB Service', () => {
|
|||||||
);
|
);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
describe('deleteRecipe - Ownership Check', () => {
|
||||||
|
it('should not delete recipe if the user does not own it and is not an admin', async () => {
|
||||||
|
mockQuery.mockResolvedValue({ rowCount: 0 });
|
||||||
|
|
||||||
|
await expect(recipeRepo.deleteRecipe(1, 'wrong-user', false, mockLogger)).rejects.toThrow(
|
||||||
|
'Recipe not found or user does not have permission to delete.',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
describe('updateRecipe', () => {
|
describe('updateRecipe', () => {
|
||||||
it('should execute an UPDATE query with the correct fields', async () => {
|
it('should execute an UPDATE query with the correct fields', async () => {
|
||||||
@@ -382,6 +393,7 @@ describe('Recipe DB Service', () => {
|
|||||||
content: 'Great!',
|
content: 'Great!',
|
||||||
status: 'visible',
|
status: 'visible',
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
mockQuery.mockResolvedValue({ rows: [mockComment] });
|
mockQuery.mockResolvedValue({ rows: [mockComment] });
|
||||||
|
|
||||||
@@ -441,10 +453,6 @@ describe('Recipe DB Service', () => {
|
|||||||
await expect(recipeRepo.forkRecipe('user-123', 1, mockLogger)).rejects.toThrow(
|
await expect(recipeRepo.forkRecipe('user-123', 1, mockLogger)).rejects.toThrow(
|
||||||
'Recipe is not public and cannot be forked.',
|
'Recipe is not public and cannot be forked.',
|
||||||
);
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
|
||||||
{ err: dbError, userId: 'user-123', originalRecipeId: 1 },
|
|
||||||
'Database error in forkRecipe',
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw a generic error if the database query fails', async () => {
|
it('should throw a generic error if the database query fails', async () => {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/recipe.db.ts
|
// src/services/db/recipe.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
import { ForeignKeyConstraintError, NotFoundError, UniqueConstraintError } from './errors.db';
|
import { NotFoundError, UniqueConstraintError, handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { Recipe, FavoriteRecipe, RecipeComment } from '../../types';
|
import type { Recipe, FavoriteRecipe, RecipeComment } from '../../types';
|
||||||
|
|
||||||
@@ -25,8 +25,9 @@ export class RecipeRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, minPercentage }, 'Database error in getRecipesBySalePercentage');
|
handleDbError(error, logger, 'Database error in getRecipesBySalePercentage', { minPercentage }, {
|
||||||
throw new Error('Failed to get recipes by sale percentage.');
|
defaultMessage: 'Failed to get recipes by sale percentage.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -43,11 +44,13 @@ export class RecipeRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, minIngredients },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in getRecipesByMinSaleIngredients',
|
'Database error in getRecipesByMinSaleIngredients',
|
||||||
|
{ minIngredients },
|
||||||
|
{ defaultMessage: 'Failed to get recipes by minimum sale ingredients.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to get recipes by minimum sale ingredients.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -69,11 +72,13 @@ export class RecipeRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, ingredient, tag },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in findRecipesByIngredientAndTag',
|
'Database error in findRecipesByIngredientAndTag',
|
||||||
|
{ ingredient, tag },
|
||||||
|
{ defaultMessage: 'Failed to find recipes by ingredient and tag.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to find recipes by ingredient and tag.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -90,8 +95,9 @@ export class RecipeRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getUserFavoriteRecipes');
|
handleDbError(error, logger, 'Database error in getUserFavoriteRecipes', { userId }, {
|
||||||
throw new Error('Failed to get favorite recipes.');
|
defaultMessage: 'Failed to get favorite recipes.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -118,14 +124,10 @@ export class RecipeRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof UniqueConstraintError) {
|
handleDbError(error, logger, 'Database error in addFavoriteRecipe', { userId, recipeId }, {
|
||||||
throw error;
|
fkMessage: 'The specified user or recipe does not exist.',
|
||||||
}
|
defaultMessage: 'Failed to add favorite recipe.',
|
||||||
logger.error({ err: error, userId, recipeId }, 'Database error in addFavoriteRecipe');
|
});
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
|
||||||
throw new ForeignKeyConstraintError('The specified user or recipe does not exist.');
|
|
||||||
}
|
|
||||||
throw new Error('Failed to add favorite recipe.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -144,11 +146,9 @@ export class RecipeRepository {
|
|||||||
throw new NotFoundError('Favorite recipe not found for this user.');
|
throw new NotFoundError('Favorite recipe not found for this user.');
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) {
|
handleDbError(error, logger, 'Database error in removeFavoriteRecipe', { userId, recipeId }, {
|
||||||
throw error;
|
defaultMessage: 'Failed to remove favorite recipe.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, userId, recipeId }, 'Database error in removeFavoriteRecipe');
|
|
||||||
throw new Error('Failed to remove favorite recipe.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -178,9 +178,9 @@ export class RecipeRepository {
|
|||||||
throw new NotFoundError('Recipe not found or user does not have permission to delete.');
|
throw new NotFoundError('Recipe not found or user does not have permission to delete.');
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
handleDbError(error, logger, 'Database error in deleteRecipe', { recipeId, userId, isAdmin }, {
|
||||||
logger.error({ err: error, recipeId, userId, isAdmin }, 'Database error in deleteRecipe');
|
defaultMessage: 'Failed to delete recipe.',
|
||||||
throw new Error('Failed to delete recipe.');
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -239,15 +239,13 @@ export class RecipeRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// Re-throw specific, known errors to allow for more precise error handling in the calling code.
|
// Explicitly re-throw the "No fields" error before it gets caught by the generic handler.
|
||||||
if (
|
if (error instanceof Error && error.message === 'No fields provided to update.') {
|
||||||
error instanceof NotFoundError ||
|
|
||||||
(error instanceof Error && error.message.includes('No fields provided'))
|
|
||||||
) {
|
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error({ err: error, recipeId, userId, updates }, 'Database error in updateRecipe');
|
handleDbError(error, logger, 'Database error in updateRecipe', { recipeId, userId, updates }, {
|
||||||
throw new Error('Failed to update recipe.');
|
defaultMessage: 'Failed to update recipe.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -261,8 +259,20 @@ export class RecipeRepository {
|
|||||||
const query = `
|
const query = `
|
||||||
SELECT
|
SELECT
|
||||||
r.*,
|
r.*,
|
||||||
COALESCE(json_agg(DISTINCT jsonb_build_object('recipe_ingredient_id', ri.recipe_ingredient_id, 'master_item_name', mgi.name, 'quantity', ri.quantity, 'unit', ri.unit)) FILTER (WHERE ri.recipe_ingredient_id IS NOT NULL), '[]') AS ingredients,
|
COALESCE(json_agg(DISTINCT jsonb_build_object(
|
||||||
COALESCE(json_agg(DISTINCT jsonb_build_object('tag_id', t.tag_id, 'name', t.name)) FILTER (WHERE t.tag_id IS NOT NULL), '[]') AS tags
|
'recipe_ingredient_id', ri.recipe_ingredient_id,
|
||||||
|
'master_item_name', mgi.name,
|
||||||
|
'quantity', ri.quantity,
|
||||||
|
'unit', ri.unit,
|
||||||
|
'created_at', ri.created_at,
|
||||||
|
'updated_at', ri.updated_at
|
||||||
|
)) FILTER (WHERE ri.recipe_ingredient_id IS NOT NULL), '[]') AS ingredients,
|
||||||
|
COALESCE(json_agg(DISTINCT jsonb_build_object(
|
||||||
|
'tag_id', t.tag_id,
|
||||||
|
'name', t.name,
|
||||||
|
'created_at', t.created_at,
|
||||||
|
'updated_at', t.updated_at
|
||||||
|
)) FILTER (WHERE t.tag_id IS NOT NULL), '[]') AS tags
|
||||||
FROM public.recipes r
|
FROM public.recipes r
|
||||||
LEFT JOIN public.recipe_ingredients ri ON r.recipe_id = ri.recipe_id
|
LEFT JOIN public.recipe_ingredients ri ON r.recipe_id = ri.recipe_id
|
||||||
LEFT JOIN public.master_grocery_items mgi ON ri.master_item_id = mgi.master_grocery_item_id
|
LEFT JOIN public.master_grocery_items mgi ON ri.master_item_id = mgi.master_grocery_item_id
|
||||||
@@ -277,11 +287,9 @@ export class RecipeRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) {
|
handleDbError(error, logger, 'Database error in getRecipeById', { recipeId }, {
|
||||||
throw error;
|
defaultMessage: 'Failed to retrieve recipe.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, recipeId }, 'Database error in getRecipeById');
|
|
||||||
throw new Error('Failed to retrieve recipe.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -305,8 +313,9 @@ export class RecipeRepository {
|
|||||||
const res = await this.db.query<RecipeComment>(query, [recipeId]);
|
const res = await this.db.query<RecipeComment>(query, [recipeId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, recipeId }, 'Database error in getRecipeComments');
|
handleDbError(error, logger, 'Database error in getRecipeComments', { recipeId }, {
|
||||||
throw new Error('Failed to get recipe comments.');
|
defaultMessage: 'Failed to get recipe comments.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -332,18 +341,13 @@ export class RecipeRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, recipeId, userId, parentCommentId },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in addRecipeComment',
|
'Database error in addRecipeComment',
|
||||||
|
{ recipeId, userId, parentCommentId },
|
||||||
|
{ fkMessage: 'The specified recipe, user, or parent comment does not exist.', defaultMessage: 'Failed to add recipe comment.' },
|
||||||
);
|
);
|
||||||
// Check for specific PostgreSQL error codes
|
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
|
||||||
// foreign_key_violation
|
|
||||||
throw new ForeignKeyConstraintError(
|
|
||||||
'The specified recipe, user, or parent comment does not exist.',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
throw new Error('Failed to add recipe comment.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -361,13 +365,15 @@ export class RecipeRepository {
|
|||||||
]);
|
]);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, originalRecipeId }, 'Database error in forkRecipe');
|
|
||||||
// The fork_recipe function could fail if the original recipe doesn't exist or isn't public.
|
// The fork_recipe function could fail if the original recipe doesn't exist or isn't public.
|
||||||
if (error instanceof Error && 'code' in error && error.code === 'P0001') {
|
if (error instanceof Error && 'code' in error && error.code === 'P0001') {
|
||||||
// raise_exception
|
// raise_exception
|
||||||
throw new Error(error.message); // Re-throw the user-friendly message from the DB function.
|
throw new Error(error.message); // Re-throw the user-friendly message from the DB function.
|
||||||
}
|
}
|
||||||
throw new Error('Failed to fork recipe.');
|
handleDbError(error, logger, 'Database error in forkRecipe', { userId, originalRecipeId }, {
|
||||||
|
fkMessage: 'The specified user or original recipe does not exist.',
|
||||||
|
defaultMessage: 'Failed to fork recipe.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -166,7 +166,7 @@ describe('Shopping DB Service', () => {
|
|||||||
it('should throw an error if no rows are deleted (list not found or wrong user)', async () => {
|
it('should throw an error if no rows are deleted (list not found or wrong user)', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'DELETE' });
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'DELETE' });
|
||||||
await expect(shoppingRepo.deleteShoppingList(999, 'user-1', mockLogger)).rejects.toThrow(
|
await expect(shoppingRepo.deleteShoppingList(999, 'user-1', mockLogger)).rejects.toThrow(
|
||||||
'Failed to delete shopping list.',
|
'Shopping list not found or user does not have permission to delete.',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -190,13 +190,14 @@ describe('Shopping DB Service', () => {
|
|||||||
|
|
||||||
const result = await shoppingRepo.addShoppingListItem(
|
const result = await shoppingRepo.addShoppingListItem(
|
||||||
1,
|
1,
|
||||||
|
'user-1',
|
||||||
{ customItemName: 'Custom Item' },
|
{ customItemName: 'Custom Item' },
|
||||||
mockLogger,
|
mockLogger,
|
||||||
);
|
);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
||||||
[1, null, 'Custom Item'],
|
[1, null, 'Custom Item', 'user-1'],
|
||||||
);
|
);
|
||||||
expect(result).toEqual(mockItem);
|
expect(result).toEqual(mockItem);
|
||||||
});
|
});
|
||||||
@@ -205,11 +206,11 @@ describe('Shopping DB Service', () => {
|
|||||||
const mockItem = createMockShoppingListItem({ master_item_id: 123 });
|
const mockItem = createMockShoppingListItem({ master_item_id: 123 });
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockItem] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockItem] });
|
||||||
|
|
||||||
const result = await shoppingRepo.addShoppingListItem(1, { masterItemId: 123 }, mockLogger);
|
const result = await shoppingRepo.addShoppingListItem(1, 'user-1', { masterItemId: 123 }, mockLogger);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
||||||
[1, 123, null],
|
[1, 123, null, 'user-1'],
|
||||||
);
|
);
|
||||||
expect(result).toEqual(mockItem);
|
expect(result).toEqual(mockItem);
|
||||||
});
|
});
|
||||||
@@ -223,19 +224,20 @@ describe('Shopping DB Service', () => {
|
|||||||
|
|
||||||
const result = await shoppingRepo.addShoppingListItem(
|
const result = await shoppingRepo.addShoppingListItem(
|
||||||
1,
|
1,
|
||||||
|
'user-1',
|
||||||
{ masterItemId: 123, customItemName: 'Organic Apples' },
|
{ masterItemId: 123, customItemName: 'Organic Apples' },
|
||||||
mockLogger,
|
mockLogger,
|
||||||
);
|
);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
expect.stringContaining('INSERT INTO public.shopping_list_items'),
|
||||||
[1, 123, 'Organic Apples'],
|
[1, 123, 'Organic Apples', 'user-1'],
|
||||||
);
|
);
|
||||||
expect(result).toEqual(mockItem);
|
expect(result).toEqual(mockItem);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw an error if both masterItemId and customItemName are missing', async () => {
|
it('should throw an error if both masterItemId and customItemName are missing', async () => {
|
||||||
await expect(shoppingRepo.addShoppingListItem(1, {}, mockLogger)).rejects.toThrow(
|
await expect(shoppingRepo.addShoppingListItem(1, 'user-1', {}, mockLogger)).rejects.toThrow(
|
||||||
'Either masterItemId or customItemName must be provided.',
|
'Either masterItemId or customItemName must be provided.',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -244,19 +246,19 @@ describe('Shopping DB Service', () => {
|
|||||||
const dbError = new Error('violates foreign key constraint');
|
const dbError = new Error('violates foreign key constraint');
|
||||||
(dbError as Error & { code: string }).code = '23503';
|
(dbError as Error & { code: string }).code = '23503';
|
||||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
await expect(
|
await expect(shoppingRepo.addShoppingListItem(999, 'user-1', { masterItemId: 999 }, mockLogger)).rejects.toThrow(
|
||||||
shoppingRepo.addShoppingListItem(999, { masterItemId: 999 }, mockLogger),
|
'Referenced list or item does not exist.',
|
||||||
).rejects.toThrow('Referenced list or item does not exist.');
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw a generic error if the database query fails', async () => {
|
it('should throw a generic error if the database query fails', async () => {
|
||||||
const dbError = new Error('DB Connection Error');
|
const dbError = new Error('DB Connection Error');
|
||||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
await expect(
|
await expect(
|
||||||
shoppingRepo.addShoppingListItem(1, { customItemName: 'Test' }, mockLogger),
|
shoppingRepo.addShoppingListItem(1, 'user-1', { customItemName: 'Test' }, mockLogger),
|
||||||
).rejects.toThrow('Failed to add item to shopping list.');
|
).rejects.toThrow('Failed to add item to shopping list.');
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, listId: 1, item: { customItemName: 'Test' } },
|
{ err: dbError, listId: 1, userId: 'user-1', item: { customItemName: 'Test' } },
|
||||||
'Database error in addShoppingListItem',
|
'Database error in addShoppingListItem',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -269,13 +271,14 @@ describe('Shopping DB Service', () => {
|
|||||||
|
|
||||||
const result = await shoppingRepo.updateShoppingListItem(
|
const result = await shoppingRepo.updateShoppingListItem(
|
||||||
1,
|
1,
|
||||||
|
'user-1',
|
||||||
{ is_purchased: true },
|
{ is_purchased: true },
|
||||||
mockLogger,
|
mockLogger,
|
||||||
);
|
);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
'UPDATE public.shopping_list_items SET is_purchased = $1 WHERE shopping_list_item_id = $2 RETURNING *',
|
expect.stringContaining('UPDATE public.shopping_list_items sli'),
|
||||||
[true, 1],
|
[true, 1, 'user-1'],
|
||||||
);
|
);
|
||||||
expect(result).toEqual(mockItem);
|
expect(result).toEqual(mockItem);
|
||||||
});
|
});
|
||||||
@@ -285,11 +288,11 @@ describe('Shopping DB Service', () => {
|
|||||||
const mockItem = createMockShoppingListItem({ shopping_list_item_id: 1, ...updates });
|
const mockItem = createMockShoppingListItem({ shopping_list_item_id: 1, ...updates });
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockItem], rowCount: 1 });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockItem], rowCount: 1 });
|
||||||
|
|
||||||
const result = await shoppingRepo.updateShoppingListItem(1, updates, mockLogger);
|
const result = await shoppingRepo.updateShoppingListItem(1, 'user-1', updates, mockLogger);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
'UPDATE public.shopping_list_items SET quantity = $1, is_purchased = $2, notes = $3 WHERE shopping_list_item_id = $4 RETURNING *',
|
expect.stringContaining('UPDATE public.shopping_list_items sli'),
|
||||||
[updates.quantity, updates.is_purchased, updates.notes, 1],
|
[updates.quantity, updates.is_purchased, updates.notes, 1, 'user-1'],
|
||||||
);
|
);
|
||||||
expect(result).toEqual(mockItem);
|
expect(result).toEqual(mockItem);
|
||||||
});
|
});
|
||||||
@@ -297,13 +300,13 @@ describe('Shopping DB Service', () => {
|
|||||||
it('should throw an error if the item to update is not found', async () => {
|
it('should throw an error if the item to update is not found', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'UPDATE' });
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'UPDATE' });
|
||||||
await expect(
|
await expect(
|
||||||
shoppingRepo.updateShoppingListItem(999, { quantity: 5 }, mockLogger),
|
shoppingRepo.updateShoppingListItem(999, 'user-1', { quantity: 5 }, mockLogger),
|
||||||
).rejects.toThrow('Shopping list item not found.');
|
).rejects.toThrow('Shopping list item not found.');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw an error if no valid fields are provided to update', async () => {
|
it('should throw an error if no valid fields are provided to update', async () => {
|
||||||
// The function should throw before even querying the database.
|
// The function should throw before even querying the database.
|
||||||
await expect(shoppingRepo.updateShoppingListItem(1, {}, mockLogger)).rejects.toThrow(
|
await expect(shoppingRepo.updateShoppingListItem(1, 'user-1', {}, mockLogger)).rejects.toThrow(
|
||||||
'No valid fields to update.',
|
'No valid fields to update.',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -312,44 +315,65 @@ describe('Shopping DB Service', () => {
|
|||||||
const dbError = new Error('DB Connection Error');
|
const dbError = new Error('DB Connection Error');
|
||||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
await expect(
|
await expect(
|
||||||
shoppingRepo.updateShoppingListItem(1, { is_purchased: true }, mockLogger),
|
shoppingRepo.updateShoppingListItem(1, 'user-1', { is_purchased: true }, mockLogger),
|
||||||
).rejects.toThrow('Failed to update shopping list item.');
|
).rejects.toThrow('Failed to update shopping list item.');
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, itemId: 1, updates: { is_purchased: true } },
|
{ err: dbError, itemId: 1, userId: 'user-1', updates: { is_purchased: true } },
|
||||||
'Database error in updateShoppingListItem',
|
'Database error in updateShoppingListItem',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('updateShoppingListItem - Ownership Check', () => {
|
||||||
|
it('should not update an item if the user does not own the shopping list', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||||
|
|
||||||
|
await expect(
|
||||||
|
shoppingRepo.updateShoppingListItem(1, 'wrong-user', { is_purchased: true }, mockLogger),
|
||||||
|
).rejects.toThrow('Shopping list item not found.');
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
describe('removeShoppingListItem', () => {
|
describe('removeShoppingListItem', () => {
|
||||||
it('should delete an item if rowCount is 1', async () => {
|
it('should delete an item if rowCount is 1', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1, rows: [], command: 'DELETE' });
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 1, rows: [], command: 'DELETE' });
|
||||||
await expect(shoppingRepo.removeShoppingListItem(1, mockLogger)).resolves.toBeUndefined();
|
await expect(shoppingRepo.removeShoppingListItem(1, 'user-1', mockLogger)).resolves.toBeUndefined();
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
'DELETE FROM public.shopping_list_items WHERE shopping_list_item_id = $1',
|
expect.stringContaining('DELETE FROM public.shopping_list_items sli'),
|
||||||
[1],
|
[1, 'user-1'],
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw an error if no rows are deleted (item not found)', async () => {
|
it('should throw an error if no rows are deleted (item not found)', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'DELETE' });
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [], command: 'DELETE' });
|
||||||
await expect(shoppingRepo.removeShoppingListItem(999, mockLogger)).rejects.toThrow(
|
await expect(shoppingRepo.removeShoppingListItem(999, 'user-1', mockLogger)).rejects.toThrow(
|
||||||
'Shopping list item not found.',
|
'Shopping list item not found or user does not have permission.',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw a generic error if the database query fails', async () => {
|
it('should throw a generic error if the database query fails', async () => {
|
||||||
const dbError = new Error('DB Connection Error');
|
const dbError = new Error('DB Connection Error');
|
||||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||||
await expect(shoppingRepo.removeShoppingListItem(1, mockLogger)).rejects.toThrow(
|
await expect(shoppingRepo.removeShoppingListItem(1, 'user-1', mockLogger)).rejects.toThrow(
|
||||||
'Failed to remove item from shopping list.',
|
'Failed to remove item from shopping list.',
|
||||||
);
|
);
|
||||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||||
{ err: dbError, itemId: 1 },
|
{ err: dbError, itemId: 1, userId: 'user-1' },
|
||||||
'Database error in removeShoppingListItem',
|
'Database error in removeShoppingListItem',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
describe('removeShoppingListItem - Ownership Check', () => {
|
||||||
|
it('should not remove an item if the user does not own the shopping list', async () => {
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||||
|
|
||||||
|
await expect(shoppingRepo.removeShoppingListItem(1, 'wrong-user', mockLogger)).rejects.toThrow(
|
||||||
|
'Shopping list item not found or user does not have permission.',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
describe('completeShoppingList', () => {
|
describe('completeShoppingList', () => {
|
||||||
it('should call the complete_shopping_list database function', async () => {
|
it('should call the complete_shopping_list database function', async () => {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
// src/services/db/shopping.db.ts
|
// src/services/db/shopping.db.ts
|
||||||
import type { Pool, PoolClient } from 'pg';
|
import type { Pool, PoolClient } from 'pg';
|
||||||
import { getPool, withTransaction } from './connection.db';
|
import { getPool, withTransaction } from './connection.db';
|
||||||
import { ForeignKeyConstraintError, UniqueConstraintError, NotFoundError } from './errors.db';
|
import { NotFoundError, handleDbError } from './errors.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import {
|
import {
|
||||||
ShoppingList,
|
ShoppingList,
|
||||||
@@ -29,8 +29,7 @@ export class ShoppingRepository {
|
|||||||
async getShoppingLists(userId: string, logger: Logger): Promise<ShoppingList[]> {
|
async getShoppingLists(userId: string, logger: Logger): Promise<ShoppingList[]> {
|
||||||
try {
|
try {
|
||||||
const query = `
|
const query = `
|
||||||
SELECT
|
SELECT sl.shopping_list_id, sl.name, sl.created_at, sl.updated_at,
|
||||||
sl.shopping_list_id, sl.name, sl.created_at,
|
|
||||||
COALESCE(json_agg(
|
COALESCE(json_agg(
|
||||||
json_build_object(
|
json_build_object(
|
||||||
'shopping_list_item_id', sli.shopping_list_item_id,
|
'shopping_list_item_id', sli.shopping_list_item_id,
|
||||||
@@ -40,6 +39,7 @@ export class ShoppingRepository {
|
|||||||
'quantity', sli.quantity,
|
'quantity', sli.quantity,
|
||||||
'is_purchased', sli.is_purchased,
|
'is_purchased', sli.is_purchased,
|
||||||
'added_at', sli.added_at,
|
'added_at', sli.added_at,
|
||||||
|
'updated_at', sli.updated_at,
|
||||||
'master_item', json_build_object('name', mgi.name)
|
'master_item', json_build_object('name', mgi.name)
|
||||||
)
|
)
|
||||||
) FILTER (WHERE sli.shopping_list_item_id IS NOT NULL), '[]'::json) as items
|
) FILTER (WHERE sli.shopping_list_item_id IS NOT NULL), '[]'::json) as items
|
||||||
@@ -53,8 +53,9 @@ export class ShoppingRepository {
|
|||||||
const res = await this.db.query<ShoppingList>(query, [userId]);
|
const res = await this.db.query<ShoppingList>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getShoppingLists');
|
handleDbError(error, logger, 'Database error in getShoppingLists', { userId }, {
|
||||||
throw new Error('Failed to retrieve shopping lists.');
|
defaultMessage: 'Failed to retrieve shopping lists.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -67,18 +68,15 @@ export class ShoppingRepository {
|
|||||||
async createShoppingList(userId: string, name: string, logger: Logger): Promise<ShoppingList> {
|
async createShoppingList(userId: string, name: string, logger: Logger): Promise<ShoppingList> {
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<ShoppingList>(
|
const res = await this.db.query<ShoppingList>(
|
||||||
'INSERT INTO public.shopping_lists (user_id, name) VALUES ($1, $2) RETURNING shopping_list_id, user_id, name, created_at',
|
'INSERT INTO public.shopping_lists (user_id, name) VALUES ($1, $2) RETURNING shopping_list_id, user_id, name, created_at, updated_at',
|
||||||
[userId, name],
|
[userId, name],
|
||||||
);
|
);
|
||||||
return { ...res.rows[0], items: [] };
|
return { ...res.rows[0], items: [] };
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The patch requested this specific error handling.
|
handleDbError(error, logger, 'Database error in createShoppingList', { userId, name }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'The specified user does not exist.',
|
||||||
throw new ForeignKeyConstraintError('The specified user does not exist.');
|
defaultMessage: 'Failed to create shopping list.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, userId, name }, 'Database error in createShoppingList');
|
|
||||||
// The patch requested this specific error handling.
|
|
||||||
throw new Error('Failed to create shopping list.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -91,8 +89,7 @@ export class ShoppingRepository {
|
|||||||
async getShoppingListById(listId: number, userId: string, logger: Logger): Promise<ShoppingList> {
|
async getShoppingListById(listId: number, userId: string, logger: Logger): Promise<ShoppingList> {
|
||||||
try {
|
try {
|
||||||
const query = `
|
const query = `
|
||||||
SELECT
|
SELECT sl.shopping_list_id, sl.name, sl.created_at, sl.updated_at,
|
||||||
sl.shopping_list_id, sl.name, sl.created_at,
|
|
||||||
COALESCE(json_agg(
|
COALESCE(json_agg(
|
||||||
json_build_object(
|
json_build_object(
|
||||||
'shopping_list_item_id', sli.shopping_list_item_id,
|
'shopping_list_item_id', sli.shopping_list_item_id,
|
||||||
@@ -102,6 +99,7 @@ export class ShoppingRepository {
|
|||||||
'quantity', sli.quantity,
|
'quantity', sli.quantity,
|
||||||
'is_purchased', sli.is_purchased,
|
'is_purchased', sli.is_purchased,
|
||||||
'added_at', sli.added_at,
|
'added_at', sli.added_at,
|
||||||
|
'updated_at', sli.updated_at,
|
||||||
'master_item', json_build_object('name', mgi.name)
|
'master_item', json_build_object('name', mgi.name)
|
||||||
)
|
)
|
||||||
) FILTER (WHERE sli.shopping_list_item_id IS NOT NULL), '[]'::json) as items
|
) FILTER (WHERE sli.shopping_list_item_id IS NOT NULL), '[]'::json) as items
|
||||||
@@ -120,8 +118,9 @@ export class ShoppingRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, listId, userId }, 'Database error in getShoppingListById');
|
handleDbError(error, logger, 'Database error in getShoppingListById', { listId, userId }, {
|
||||||
throw new Error('Failed to retrieve shopping list.');
|
defaultMessage: 'Failed to retrieve shopping list.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -143,8 +142,9 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, listId, userId }, 'Database error in deleteShoppingList');
|
handleDbError(error, logger, 'Database error in deleteShoppingList', { listId, userId }, {
|
||||||
throw new Error('Failed to delete shopping list.');
|
defaultMessage: 'Failed to delete shopping list.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -156,6 +156,7 @@ export class ShoppingRepository {
|
|||||||
*/
|
*/
|
||||||
async addShoppingListItem(
|
async addShoppingListItem(
|
||||||
listId: number,
|
listId: number,
|
||||||
|
userId: string,
|
||||||
item: { masterItemId?: number; customItemName?: string },
|
item: { masterItemId?: number; customItemName?: string },
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<ShoppingListItem> {
|
): Promise<ShoppingListItem> {
|
||||||
@@ -165,18 +166,33 @@ export class ShoppingRepository {
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<ShoppingListItem>(
|
const query = `
|
||||||
'INSERT INTO public.shopping_list_items (shopping_list_id, master_item_id, custom_item_name) VALUES ($1, $2, $3) RETURNING *',
|
INSERT INTO public.shopping_list_items (shopping_list_id, master_item_id, custom_item_name)
|
||||||
[listId, item.masterItemId ?? null, item.customItemName ?? null],
|
SELECT $1, $2, $3
|
||||||
);
|
WHERE EXISTS (
|
||||||
|
SELECT 1 FROM public.shopping_lists WHERE shopping_list_id = $1 AND user_id = $4
|
||||||
|
)
|
||||||
|
RETURNING *;
|
||||||
|
`;
|
||||||
|
const res = await this.db.query<ShoppingListItem>(query, [
|
||||||
|
listId,
|
||||||
|
item.masterItemId ?? null,
|
||||||
|
item.customItemName ?? null,
|
||||||
|
userId,
|
||||||
|
]);
|
||||||
|
|
||||||
|
if (res.rowCount === 0) {
|
||||||
|
throw new NotFoundError('Shopping list not found or user does not have permission.');
|
||||||
|
}
|
||||||
|
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The patch requested this specific error handling.
|
if (error instanceof NotFoundError) throw error;
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
handleDbError(error, logger, 'Database error in addShoppingListItem', { listId, userId, item }, {
|
||||||
throw new ForeignKeyConstraintError('Referenced list or item does not exist.');
|
fkMessage: 'Referenced list or item does not exist.',
|
||||||
}
|
checkMessage: 'Shopping list item must have a master item or a custom name.',
|
||||||
logger.error({ err: error, listId, item }, 'Database error in addShoppingListItem');
|
defaultMessage: 'Failed to add item to shopping list.',
|
||||||
throw new Error('Failed to add item to shopping list.');
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -184,20 +200,25 @@ export class ShoppingRepository {
|
|||||||
* Removes an item from a shopping list.
|
* Removes an item from a shopping list.
|
||||||
* @param itemId The ID of the shopping list item to remove.
|
* @param itemId The ID of the shopping list item to remove.
|
||||||
*/
|
*/
|
||||||
async removeShoppingListItem(itemId: number, logger: Logger): Promise<void> {
|
async removeShoppingListItem(itemId: number, userId: string, logger: Logger): Promise<void> {
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query(
|
const query = `
|
||||||
'DELETE FROM public.shopping_list_items WHERE shopping_list_item_id = $1',
|
DELETE FROM public.shopping_list_items sli
|
||||||
[itemId],
|
WHERE sli.shopping_list_item_id = $1
|
||||||
);
|
AND EXISTS (
|
||||||
// The patch requested this specific error handling.
|
SELECT 1 FROM public.shopping_lists sl
|
||||||
|
WHERE sl.shopping_list_id = sli.shopping_list_id AND sl.user_id = $2
|
||||||
|
);
|
||||||
|
`;
|
||||||
|
const res = await this.db.query(query, [itemId, userId]);
|
||||||
if (res.rowCount === 0) {
|
if (res.rowCount === 0) {
|
||||||
throw new NotFoundError('Shopping list item not found.');
|
throw new NotFoundError('Shopping list item not found or user does not have permission.');
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error({ err: error, itemId }, 'Database error in removeShoppingListItem');
|
handleDbError(error, logger, 'Database error in removeShoppingListItem', { itemId, userId }, {
|
||||||
throw new Error('Failed to remove item from shopping list.');
|
defaultMessage: 'Failed to remove item from shopping list.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
@@ -218,11 +239,13 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, menuPlanId, userId },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in generateShoppingListForMenuPlan',
|
'Database error in generateShoppingListForMenuPlan',
|
||||||
|
{ menuPlanId, userId },
|
||||||
|
{ defaultMessage: 'Failed to generate shopping list for menu plan.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to generate shopping list for menu plan.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -246,11 +269,13 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(
|
||||||
{ err: error, menuPlanId, shoppingListId, userId },
|
error,
|
||||||
|
logger,
|
||||||
'Database error in addMenuPlanToShoppingList',
|
'Database error in addMenuPlanToShoppingList',
|
||||||
|
{ menuPlanId, shoppingListId, userId },
|
||||||
|
{ fkMessage: 'The specified menu plan, shopping list, or an item within the plan does not exist.', defaultMessage: 'Failed to add menu plan to shopping list.' },
|
||||||
);
|
);
|
||||||
throw new Error('Failed to add menu plan to shopping list.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -267,8 +292,9 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getPantryLocations');
|
handleDbError(error, logger, 'Database error in getPantryLocations', { userId }, {
|
||||||
throw new Error('Failed to get pantry locations.');
|
defaultMessage: 'Failed to get pantry locations.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -290,13 +316,12 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23505') {
|
handleDbError(error, logger, 'Database error in createPantryLocation', { userId, name }, {
|
||||||
throw new UniqueConstraintError('A pantry location with this name already exists.');
|
uniqueMessage: 'A pantry location with this name already exists.',
|
||||||
} else if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'User not found',
|
||||||
throw new ForeignKeyConstraintError('User not found');
|
notNullMessage: 'Pantry location name cannot be null.',
|
||||||
}
|
defaultMessage: 'Failed to create pantry location.',
|
||||||
logger.error({ err: error, userId, name }, 'Database error in createPantryLocation');
|
});
|
||||||
throw new Error('Failed to create pantry location.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -308,6 +333,7 @@ export class ShoppingRepository {
|
|||||||
*/
|
*/
|
||||||
async updateShoppingListItem(
|
async updateShoppingListItem(
|
||||||
itemId: number,
|
itemId: number,
|
||||||
|
userId: string,
|
||||||
updates: Partial<ShoppingListItem>,
|
updates: Partial<ShoppingListItem>,
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<ShoppingListItem> {
|
): Promise<ShoppingListItem> {
|
||||||
@@ -337,10 +363,19 @@ export class ShoppingRepository {
|
|||||||
}
|
}
|
||||||
|
|
||||||
values.push(itemId);
|
values.push(itemId);
|
||||||
const query = `UPDATE public.shopping_list_items SET ${setClauses.join(', ')} WHERE shopping_list_item_id = $${valueIndex} RETURNING *`;
|
values.push(userId);
|
||||||
|
|
||||||
|
const query = `
|
||||||
|
UPDATE public.shopping_list_items sli
|
||||||
|
SET ${setClauses.join(', ')}
|
||||||
|
FROM public.shopping_lists sl
|
||||||
|
WHERE sli.shopping_list_item_id = $${valueIndex}
|
||||||
|
AND sli.shopping_list_id = sl.shopping_list_id
|
||||||
|
AND sl.user_id = $${valueIndex + 1}
|
||||||
|
RETURNING sli.*;
|
||||||
|
`;
|
||||||
|
|
||||||
const res = await this.db.query<ShoppingListItem>(query, values);
|
const res = await this.db.query<ShoppingListItem>(query, values);
|
||||||
// The patch requested this specific error handling.
|
|
||||||
if (res.rowCount === 0) {
|
if (res.rowCount === 0) {
|
||||||
throw new NotFoundError('Shopping list item not found.');
|
throw new NotFoundError('Shopping list item not found.');
|
||||||
}
|
}
|
||||||
@@ -353,8 +388,9 @@ export class ShoppingRepository {
|
|||||||
) {
|
) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error({ err: error, itemId, updates }, 'Database error in updateShoppingListItem');
|
handleDbError(error, logger, 'Database error in updateShoppingListItem', { itemId, userId, updates }, {
|
||||||
throw new Error('Failed to update shopping list item.');
|
defaultMessage: 'Failed to update shopping list item.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -378,15 +414,10 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0].complete_shopping_list;
|
return res.rows[0].complete_shopping_list;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The patch requested this specific error handling.
|
handleDbError(error, logger, 'Database error in completeShoppingList', { shoppingListId, userId }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'The specified shopping list does not exist.',
|
||||||
throw new ForeignKeyConstraintError('The specified shopping list does not exist.');
|
defaultMessage: 'Failed to complete shopping list.',
|
||||||
}
|
});
|
||||||
logger.error(
|
|
||||||
{ err: error, shoppingListId, userId },
|
|
||||||
'Database error in completeShoppingList',
|
|
||||||
);
|
|
||||||
throw new Error('Failed to complete shopping list.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -399,13 +430,15 @@ export class ShoppingRepository {
|
|||||||
try {
|
try {
|
||||||
const query = `
|
const query = `
|
||||||
SELECT
|
SELECT
|
||||||
st.shopping_trip_id, st.user_id, st.shopping_list_id, st.completed_at, st.total_spent_cents,
|
st.shopping_trip_id, st.user_id, st.shopping_list_id, st.completed_at, st.total_spent_cents, st.updated_at,
|
||||||
COALESCE(
|
COALESCE(
|
||||||
json_agg(
|
json_agg(
|
||||||
json_build_object(
|
json_build_object(
|
||||||
'shopping_trip_item_id', sti.shopping_trip_item_id,
|
'shopping_trip_item_id', sti.shopping_trip_item_id,
|
||||||
'master_item_id', sti.master_item_id,
|
'master_item_id', sti.master_item_id,
|
||||||
'custom_item_name', sti.custom_item_name,
|
'custom_item_name', sti.custom_item_name,
|
||||||
|
'created_at', sti.created_at,
|
||||||
|
'updated_at', sti.updated_at,
|
||||||
'quantity', sti.quantity,
|
'quantity', sti.quantity,
|
||||||
'price_paid_cents', sti.price_paid_cents,
|
'price_paid_cents', sti.price_paid_cents,
|
||||||
'master_item_name', mgi.name
|
'master_item_name', mgi.name
|
||||||
@@ -423,8 +456,9 @@ export class ShoppingRepository {
|
|||||||
const res = await this.db.query<ShoppingTrip>(query, [userId]);
|
const res = await this.db.query<ShoppingTrip>(query, [userId]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId }, 'Database error in getShoppingTripHistory');
|
handleDbError(error, logger, 'Database error in getShoppingTripHistory', { userId }, {
|
||||||
throw new Error('Failed to retrieve shopping trip history.');
|
defaultMessage: 'Failed to retrieve shopping trip history.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -444,12 +478,10 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// The patch requested this specific error handling.
|
handleDbError(error, logger, 'Database error in createReceipt', { userId, receiptImageUrl }, {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
fkMessage: 'User not found',
|
||||||
throw new ForeignKeyConstraintError('User not found');
|
defaultMessage: 'Failed to create receipt record.',
|
||||||
}
|
});
|
||||||
logger.error({ err: error, userId, receiptImageUrl }, 'Database error in createReceipt');
|
|
||||||
throw new Error('Failed to create receipt record.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -463,7 +495,14 @@ export class ShoppingRepository {
|
|||||||
receiptId: number,
|
receiptId: number,
|
||||||
items: Omit<
|
items: Omit<
|
||||||
ReceiptItem,
|
ReceiptItem,
|
||||||
'receipt_item_id' | 'receipt_id' | 'status' | 'master_item_id' | 'product_id' | 'quantity'
|
| 'receipt_item_id'
|
||||||
|
| 'receipt_id'
|
||||||
|
| 'status'
|
||||||
|
| 'master_item_id'
|
||||||
|
| 'product_id'
|
||||||
|
| 'quantity'
|
||||||
|
| 'created_at'
|
||||||
|
| 'updated_at'
|
||||||
>[],
|
>[],
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
@@ -479,7 +518,6 @@ export class ShoppingRepository {
|
|||||||
logger.info(`Successfully processed items for receipt ID: ${receiptId}`);
|
logger.info(`Successfully processed items for receipt ID: ${receiptId}`);
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, receiptId }, 'Database transaction error in processReceiptItems');
|
|
||||||
// After the transaction fails and is rolled back by withTransaction,
|
// After the transaction fails and is rolled back by withTransaction,
|
||||||
// update the receipt status in a separate, non-transactional query.
|
// update the receipt status in a separate, non-transactional query.
|
||||||
try {
|
try {
|
||||||
@@ -492,7 +530,10 @@ export class ShoppingRepository {
|
|||||||
'Failed to update receipt status to "failed" after transaction rollback.',
|
'Failed to update receipt status to "failed" after transaction rollback.',
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
throw new Error('Failed to process and save receipt items.');
|
handleDbError(error, logger, 'Database transaction error in processReceiptItems', { receiptId }, {
|
||||||
|
fkMessage: 'The specified receipt or an item within it does not exist.',
|
||||||
|
defaultMessage: 'Failed to process and save receipt items.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -509,8 +550,9 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, receiptId }, 'Database error in findDealsForReceipt');
|
handleDbError(error, logger, 'Database error in findDealsForReceipt', { receiptId }, {
|
||||||
throw new Error('Failed to find deals for receipt.');
|
defaultMessage: 'Failed to find deals for receipt.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -530,8 +572,9 @@ export class ShoppingRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, receiptId }, 'Database error in findReceiptOwner');
|
handleDbError(error, logger, 'Database error in findReceiptOwner', { receiptId }, {
|
||||||
throw new Error('Failed to retrieve receipt owner from database.');
|
defaultMessage: 'Failed to retrieve receipt owner from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -25,9 +25,9 @@ import { withTransaction } from './connection.db';
|
|||||||
import { UserRepository, exportUserData } from './user.db';
|
import { UserRepository, exportUserData } from './user.db';
|
||||||
|
|
||||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||||
import { createMockUserProfile } from '../../tests/utils/mockFactories';
|
import { createMockUserProfile, createMockUser } from '../../tests/utils/mockFactories';
|
||||||
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||||
import type { Profile, ActivityLogItem, SearchQuery, UserProfile } from '../../types';
|
import type { Profile, ActivityLogItem, SearchQuery, UserProfile, User } from '../../types';
|
||||||
|
|
||||||
// Mock other db services that are used by functions in user.db.ts
|
// Mock other db services that are used by functions in user.db.ts
|
||||||
// Update mocks to put methods on prototype so spyOn works in exportUserData tests
|
// Update mocks to put methods on prototype so spyOn works in exportUserData tests
|
||||||
@@ -70,7 +70,16 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserByEmail', () => {
|
describe('findUserByEmail', () => {
|
||||||
it('should execute the correct query and return a user', async () => {
|
it('should execute the correct query and return a user', async () => {
|
||||||
const mockUser = { user_id: '123', email: 'test@example.com' };
|
const mockUser = {
|
||||||
|
user_id: '123',
|
||||||
|
email: 'test@example.com',
|
||||||
|
password_hash: 'some-hash',
|
||||||
|
failed_login_attempts: 0,
|
||||||
|
last_failed_login: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
refresh_token: null,
|
||||||
|
};
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockUser] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockUser] });
|
||||||
|
|
||||||
const result = await userRepo.findUserByEmail('test@example.com', mockLogger);
|
const result = await userRepo.findUserByEmail('test@example.com', mockLogger);
|
||||||
@@ -107,8 +116,12 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('createUser', () => {
|
describe('createUser', () => {
|
||||||
it('should execute a transaction to create a user and profile', async () => {
|
it('should execute a transaction to create a user and profile', async () => {
|
||||||
const mockUser = { user_id: 'new-user-id', email: 'new@example.com' };
|
const mockUser = {
|
||||||
const now = new Date().toISOString();
|
user_id: 'new-user-id',
|
||||||
|
email: 'new@example.com',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
};
|
||||||
// This is the flat structure returned by the DB query inside createUser
|
// This is the flat structure returned by the DB query inside createUser
|
||||||
const mockDbProfile = {
|
const mockDbProfile = {
|
||||||
user_id: 'new-user-id',
|
user_id: 'new-user-id',
|
||||||
@@ -118,24 +131,31 @@ describe('User DB Service', () => {
|
|||||||
avatar_url: null,
|
avatar_url: null,
|
||||||
points: 0,
|
points: 0,
|
||||||
preferences: null,
|
preferences: null,
|
||||||
created_at: now,
|
created_at: new Date().toISOString(),
|
||||||
updated_at: now,
|
updated_at: new Date().toISOString(),
|
||||||
|
user_created_at: new Date().toISOString(),
|
||||||
|
user_updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
// This is the nested structure the function is expected to return
|
// This is the nested structure the function is expected to return
|
||||||
const expectedProfile: UserProfile = {
|
const expectedProfile: UserProfile = {
|
||||||
user: { user_id: 'new-user-id', email: 'new@example.com' },
|
user: {
|
||||||
|
user_id: mockDbProfile.user_id,
|
||||||
|
email: mockDbProfile.email,
|
||||||
|
created_at: mockDbProfile.user_created_at,
|
||||||
|
updated_at: mockDbProfile.user_updated_at,
|
||||||
|
},
|
||||||
full_name: 'New User',
|
full_name: 'New User',
|
||||||
avatar_url: null,
|
avatar_url: null,
|
||||||
role: 'user',
|
role: 'user',
|
||||||
points: 0,
|
points: 0,
|
||||||
preferences: null,
|
preferences: null,
|
||||||
created_at: now,
|
created_at: mockDbProfile.created_at,
|
||||||
updated_at: now,
|
updated_at: mockDbProfile.updated_at,
|
||||||
};
|
};
|
||||||
|
|
||||||
vi.mocked(withTransaction).mockImplementation(async (callback) => {
|
vi.mocked(withTransaction).mockImplementation(async (callback: any) => {
|
||||||
const mockClient = { query: vi.fn() };
|
const mockClient = { query: vi.fn(), release: vi.fn() };
|
||||||
mockClient.query
|
(mockClient.query as Mock)
|
||||||
.mockResolvedValueOnce({ rows: [] }) // set_config
|
.mockResolvedValueOnce({ rows: [] }) // set_config
|
||||||
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
.mockResolvedValueOnce({ rows: [mockUser] }) // INSERT user
|
||||||
.mockResolvedValueOnce({ rows: [mockDbProfile] }); // SELECT profile
|
.mockResolvedValueOnce({ rows: [mockDbProfile] }); // SELECT profile
|
||||||
@@ -149,16 +169,11 @@ describe('User DB Service', () => {
|
|||||||
mockLogger,
|
mockLogger,
|
||||||
);
|
);
|
||||||
|
|
||||||
console.log(
|
|
||||||
'[TEST DEBUG] createUser - Result from function:',
|
|
||||||
JSON.stringify(result, null, 2),
|
|
||||||
);
|
|
||||||
console.log(
|
|
||||||
'[TEST DEBUG] createUser - Expected result:',
|
|
||||||
JSON.stringify(expectedProfile, null, 2),
|
|
||||||
);
|
|
||||||
|
|
||||||
// Use objectContaining because the real implementation might have other DB-generated fields.
|
// Use objectContaining because the real implementation might have other DB-generated fields.
|
||||||
|
// We can't do a deep equality check on the user object because the mock factory will generate different timestamps.
|
||||||
|
expect(result.user.user_id).toEqual(expectedProfile.user.user_id);
|
||||||
|
expect(result.full_name).toEqual(expectedProfile.full_name);
|
||||||
|
// eslint-disable-next-line @typescript-eslint/no-unused-vars
|
||||||
expect(result).toEqual(expect.objectContaining(expectedProfile));
|
expect(result).toEqual(expect.objectContaining(expectedProfile));
|
||||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||||
});
|
});
|
||||||
@@ -222,9 +237,7 @@ describe('User DB Service', () => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
expect(mockLogger.warn).toHaveBeenCalledWith(`Attempted to create a user with an existing email: exists@example.com`);
|
||||||
`Attempted to create a user with an existing email: exists@example.com`,
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should throw an error if profile is not found after user creation', async () => {
|
it('should throw an error if profile is not found after user creation', async () => {
|
||||||
@@ -255,8 +268,7 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserWithProfileByEmail', () => {
|
describe('findUserWithProfileByEmail', () => {
|
||||||
it('should query for a user and their profile by email', async () => {
|
it('should query for a user and their profile by email', async () => {
|
||||||
const now = new Date().toISOString();
|
const mockDbResult: any = {
|
||||||
const mockDbResult = {
|
|
||||||
user_id: '123',
|
user_id: '123',
|
||||||
email: 'test@example.com',
|
email: 'test@example.com',
|
||||||
password_hash: 'hash',
|
password_hash: 'hash',
|
||||||
@@ -268,9 +280,11 @@ describe('User DB Service', () => {
|
|||||||
role: 'user' as const,
|
role: 'user' as const,
|
||||||
points: 0,
|
points: 0,
|
||||||
preferences: null,
|
preferences: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
user_created_at: new Date().toISOString(),
|
||||||
|
user_updated_at: new Date().toISOString(),
|
||||||
address_id: null,
|
address_id: null,
|
||||||
created_at: now,
|
|
||||||
updated_at: now,
|
|
||||||
};
|
};
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockDbResult] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockDbResult] });
|
||||||
|
|
||||||
@@ -281,9 +295,12 @@ describe('User DB Service', () => {
|
|||||||
points: 0,
|
points: 0,
|
||||||
preferences: null,
|
preferences: null,
|
||||||
address_id: null,
|
address_id: null,
|
||||||
created_at: now,
|
user: {
|
||||||
updated_at: now,
|
user_id: '123',
|
||||||
user: { user_id: '123', email: 'test@example.com' },
|
email: 'test@example.com',
|
||||||
|
created_at: expect.any(String),
|
||||||
|
updated_at: expect.any(String),
|
||||||
|
},
|
||||||
password_hash: 'hash',
|
password_hash: 'hash',
|
||||||
failed_login_attempts: 0,
|
failed_login_attempts: 0,
|
||||||
last_failed_login: null,
|
last_failed_login: null,
|
||||||
@@ -292,15 +309,6 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
const result = await userRepo.findUserWithProfileByEmail('test@example.com', mockLogger);
|
const result = await userRepo.findUserWithProfileByEmail('test@example.com', mockLogger);
|
||||||
|
|
||||||
console.log(
|
|
||||||
'[TEST DEBUG] findUserWithProfileByEmail - Result from function:',
|
|
||||||
JSON.stringify(result, null, 2),
|
|
||||||
);
|
|
||||||
console.log(
|
|
||||||
'[TEST DEBUG] findUserWithProfileByEmail - Expected result:',
|
|
||||||
JSON.stringify(expectedResult, null, 2),
|
|
||||||
);
|
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('JOIN public.profiles'),
|
expect.stringContaining('JOIN public.profiles'),
|
||||||
['test@example.com'],
|
['test@example.com'],
|
||||||
@@ -329,7 +337,11 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserById', () => {
|
describe('findUserById', () => {
|
||||||
it('should query for a user by their ID', async () => {
|
it('should query for a user by their ID', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ user_id: '123' }], rowCount: 1 });
|
const mockUser = createMockUser({ user_id: '123' });
|
||||||
|
mockPoolInstance.query.mockResolvedValue({
|
||||||
|
rows: [mockUser],
|
||||||
|
rowCount: 1,
|
||||||
|
});
|
||||||
await userRepo.findUserById('123', mockLogger);
|
await userRepo.findUserById('123', mockLogger);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('FROM public.users WHERE user_id = $1'),
|
expect.stringContaining('FROM public.users WHERE user_id = $1'),
|
||||||
@@ -359,13 +371,16 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserWithPasswordHashById', () => {
|
describe('findUserWithPasswordHashById', () => {
|
||||||
it('should query for a user and their password hash by ID', async () => {
|
it('should query for a user and their password hash by ID', async () => {
|
||||||
|
const mockUser = createMockUser({ user_id: '123' });
|
||||||
|
const mockUserWithHash = { ...mockUser, password_hash: 'hash' };
|
||||||
|
|
||||||
mockPoolInstance.query.mockResolvedValue({
|
mockPoolInstance.query.mockResolvedValue({
|
||||||
rows: [{ user_id: '123', password_hash: 'hash' }],
|
rows: [mockUserWithHash],
|
||||||
rowCount: 1,
|
rowCount: 1,
|
||||||
});
|
});
|
||||||
await userRepo.findUserWithPasswordHashById('123', mockLogger);
|
await userRepo.findUserWithPasswordHashById('123', mockLogger);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('SELECT user_id, email, password_hash'),
|
expect.stringContaining('SELECT user_id, email, password_hash, created_at, updated_at'),
|
||||||
['123'],
|
['123'],
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -395,7 +410,11 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserProfileById', () => {
|
describe('findUserProfileById', () => {
|
||||||
it('should query for a user profile by user ID', async () => {
|
it('should query for a user profile by user ID', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ user_id: '123' }] });
|
const mockProfile = createMockUserProfile({
|
||||||
|
user: createMockUser({ user_id: '123' }),
|
||||||
|
});
|
||||||
|
// The query returns a user object inside, so we need to mock that structure.
|
||||||
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockProfile] });
|
||||||
await userRepo.findUserProfileById('123', mockLogger);
|
await userRepo.findUserProfileById('123', mockLogger);
|
||||||
// The actual query uses 'p.user_id' due to the join alias
|
// The actual query uses 'p.user_id' due to the join alias
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
@@ -426,7 +445,7 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('updateUserProfile', () => {
|
describe('updateUserProfile', () => {
|
||||||
it('should execute an UPDATE query for the user profile', async () => {
|
it('should execute an UPDATE query for the user profile', async () => {
|
||||||
const mockProfile: Profile = {
|
const mockProfile: any = {
|
||||||
full_name: 'Updated Name',
|
full_name: 'Updated Name',
|
||||||
role: 'user',
|
role: 'user',
|
||||||
points: 0,
|
points: 0,
|
||||||
@@ -444,7 +463,7 @@ describe('User DB Service', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should execute an UPDATE query for avatar_url', async () => {
|
it('should execute an UPDATE query for avatar_url', async () => {
|
||||||
const mockProfile: Profile = {
|
const mockProfile: any = {
|
||||||
avatar_url: 'new-avatar.png',
|
avatar_url: 'new-avatar.png',
|
||||||
role: 'user',
|
role: 'user',
|
||||||
points: 0,
|
points: 0,
|
||||||
@@ -462,7 +481,7 @@ describe('User DB Service', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should execute an UPDATE query for address_id', async () => {
|
it('should execute an UPDATE query for address_id', async () => {
|
||||||
const mockProfile: Profile = {
|
const mockProfile: any = {
|
||||||
address_id: 99,
|
address_id: 99,
|
||||||
role: 'user',
|
role: 'user',
|
||||||
points: 0,
|
points: 0,
|
||||||
@@ -480,8 +499,8 @@ describe('User DB Service', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should fetch the current profile if no update fields are provided', async () => {
|
it('should fetch the current profile if no update fields are provided', async () => {
|
||||||
const mockProfile: Profile = createMockUserProfile({
|
const mockProfile: UserProfile = createMockUserProfile({
|
||||||
user: { user_id: '123', email: '123@example.com' },
|
user: createMockUser({ user_id: '123', email: '123@example.com' }),
|
||||||
full_name: 'Current Name',
|
full_name: 'Current Name',
|
||||||
});
|
});
|
||||||
// FIX: Instead of mocking `mockResolvedValue` on the instance method which might fail if not spied correctly,
|
// FIX: Instead of mocking `mockResolvedValue` on the instance method which might fail if not spied correctly,
|
||||||
@@ -520,7 +539,7 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('updateUserPreferences', () => {
|
describe('updateUserPreferences', () => {
|
||||||
it('should execute an UPDATE query for user preferences', async () => {
|
it('should execute an UPDATE query for user preferences', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [{}] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [createMockUserProfile()] });
|
||||||
await userRepo.updateUserPreferences('123', { darkMode: true }, mockLogger);
|
await userRepo.updateUserPreferences('123', { darkMode: true }, mockLogger);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining("SET preferences = COALESCE(preferences, '{}'::jsonb) || $1"),
|
expect.stringContaining("SET preferences = COALESCE(preferences, '{}'::jsonb) || $1"),
|
||||||
@@ -616,7 +635,11 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('findUserByRefreshToken', () => {
|
describe('findUserByRefreshToken', () => {
|
||||||
it('should query for a user by their refresh token', async () => {
|
it('should query for a user by their refresh token', async () => {
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ user_id: '123' }], rowCount: 1 });
|
const mockUser = createMockUser({ user_id: '123' });
|
||||||
|
mockPoolInstance.query.mockResolvedValue({
|
||||||
|
rows: [mockUser],
|
||||||
|
rowCount: 1,
|
||||||
|
});
|
||||||
await userRepo.findUserByRefreshToken('a-token', mockLogger);
|
await userRepo.findUserByRefreshToken('a-token', mockLogger);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('WHERE refresh_token = $1'),
|
expect.stringContaining('WHERE refresh_token = $1'),
|
||||||
@@ -788,7 +811,7 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
const findProfileSpy = vi.spyOn(UserRepository.prototype, 'findUserProfileById');
|
const findProfileSpy = vi.spyOn(UserRepository.prototype, 'findUserProfileById');
|
||||||
findProfileSpy.mockResolvedValue(
|
findProfileSpy.mockResolvedValue(
|
||||||
createMockUserProfile({ user: { user_id: '123', email: '123@example.com' } }),
|
createMockUserProfile({ user: createMockUser({ user_id: '123', email: '123@example.com' }) }),
|
||||||
);
|
);
|
||||||
const getWatchedItemsSpy = vi.spyOn(PersonalizationRepository.prototype, 'getWatchedItems');
|
const getWatchedItemsSpy = vi.spyOn(PersonalizationRepository.prototype, 'getWatchedItems');
|
||||||
getWatchedItemsSpy.mockResolvedValue([]);
|
getWatchedItemsSpy.mockResolvedValue([]);
|
||||||
@@ -815,9 +838,7 @@ describe('User DB Service', () => {
|
|||||||
);
|
);
|
||||||
|
|
||||||
// Act & Assert: The outer function catches the NotFoundError and re-throws it.
|
// Act & Assert: The outer function catches the NotFoundError and re-throws it.
|
||||||
await expect(exportUserData('123', mockLogger)).rejects.toThrow(
|
await expect(exportUserData('123', mockLogger)).rejects.toThrow('Profile not found');
|
||||||
'Failed to export user data.',
|
|
||||||
);
|
|
||||||
expect(withTransaction).toHaveBeenCalledTimes(1);
|
expect(withTransaction).toHaveBeenCalledTimes(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -898,8 +919,8 @@ describe('User DB Service', () => {
|
|||||||
user_id: 'following-1',
|
user_id: 'following-1',
|
||||||
action: 'recipe_created',
|
action: 'recipe_created',
|
||||||
display_text: 'Created a new recipe',
|
display_text: 'Created a new recipe',
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
details: { recipe_id: 1, recipe_name: 'Test Recipe' },
|
details: { recipe_id: 1, recipe_name: 'Test Recipe' },
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
updated_at: new Date().toISOString(),
|
updated_at: new Date().toISOString(),
|
||||||
},
|
},
|
||||||
];
|
];
|
||||||
@@ -935,16 +956,17 @@ describe('User DB Service', () => {
|
|||||||
|
|
||||||
describe('logSearchQuery', () => {
|
describe('logSearchQuery', () => {
|
||||||
it('should execute an INSERT query and return the new search query log', async () => {
|
it('should execute an INSERT query and return the new search query log', async () => {
|
||||||
const queryData: Omit<SearchQuery, 'search_query_id' | 'created_at'> = {
|
const queryData: Omit<SearchQuery, 'search_query_id' | 'created_at' | 'updated_at'> = {
|
||||||
user_id: 'user-123',
|
user_id: 'user-123',
|
||||||
query_text: 'best chicken recipes',
|
query_text: 'best chicken recipes',
|
||||||
result_count: 5,
|
result_count: 5,
|
||||||
was_successful: true,
|
was_successful: true,
|
||||||
};
|
};
|
||||||
const mockLoggedQuery: SearchQuery = {
|
const mockLoggedQuery: any = {
|
||||||
search_query_id: 1,
|
search_query_id: 1,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
...queryData,
|
...queryData,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockLoggedQuery] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockLoggedQuery] });
|
||||||
|
|
||||||
@@ -966,8 +988,9 @@ describe('User DB Service', () => {
|
|||||||
};
|
};
|
||||||
const mockLoggedQuery: SearchQuery = {
|
const mockLoggedQuery: SearchQuery = {
|
||||||
search_query_id: 2,
|
search_query_id: 2,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
...queryData,
|
...queryData,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockLoggedQuery] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [mockLoggedQuery] });
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
import { Pool, PoolClient } from 'pg';
|
import { Pool, PoolClient } from 'pg';
|
||||||
import { getPool } from './connection.db';
|
import { getPool } from './connection.db';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { UniqueConstraintError, ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
import { NotFoundError, handleDbError, UniqueConstraintError } from './errors.db';
|
||||||
import {
|
import {
|
||||||
Profile,
|
Profile,
|
||||||
MasterGroceryItem,
|
MasterGroceryItem,
|
||||||
@@ -10,6 +10,7 @@ import {
|
|||||||
ActivityLogItem,
|
ActivityLogItem,
|
||||||
UserProfile,
|
UserProfile,
|
||||||
SearchQuery,
|
SearchQuery,
|
||||||
|
User,
|
||||||
} from '../../types';
|
} from '../../types';
|
||||||
import { ShoppingRepository } from './shopping.db';
|
import { ShoppingRepository } from './shopping.db';
|
||||||
import { PersonalizationRepository } from './personalization.db';
|
import { PersonalizationRepository } from './personalization.db';
|
||||||
@@ -26,6 +27,8 @@ interface DbUser {
|
|||||||
refresh_token?: string | null;
|
refresh_token?: string | null;
|
||||||
failed_login_attempts: number;
|
failed_login_attempts: number;
|
||||||
last_failed_login: string | null; // This will be a date string from the DB
|
last_failed_login: string | null; // This will be a date string from the DB
|
||||||
|
created_at: string;
|
||||||
|
updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export class UserRepository {
|
export class UserRepository {
|
||||||
@@ -43,7 +46,7 @@ export class UserRepository {
|
|||||||
logger.debug({ email }, `[DB findUserByEmail] Searching for user.`);
|
logger.debug({ email }, `[DB findUserByEmail] Searching for user.`);
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<DbUser>(
|
const res = await this.db.query<DbUser>(
|
||||||
'SELECT user_id, email, password_hash, refresh_token, failed_login_attempts, last_failed_login FROM public.users WHERE email = $1',
|
'SELECT user_id, email, password_hash, refresh_token, failed_login_attempts, last_failed_login, created_at, updated_at FROM public.users WHERE email = $1',
|
||||||
[email],
|
[email],
|
||||||
);
|
);
|
||||||
const userFound = res.rows[0];
|
const userFound = res.rows[0];
|
||||||
@@ -52,8 +55,9 @@ export class UserRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, email }, 'Database error in findUserByEmail');
|
handleDbError(error, logger, 'Database error in findUserByEmail', { email }, {
|
||||||
throw new Error('Failed to retrieve user from database.');
|
defaultMessage: 'Failed to retrieve user from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -90,7 +94,7 @@ export class UserRepository {
|
|||||||
|
|
||||||
// After the trigger has run, fetch the complete profile data.
|
// After the trigger has run, fetch the complete profile data.
|
||||||
const profileQuery = `
|
const profileQuery = `
|
||||||
SELECT u.user_id, u.email, p.full_name, p.avatar_url, p.role, p.points, p.preferences, p.created_at, p.updated_at
|
SELECT u.user_id, u.email, u.created_at as user_created_at, u.updated_at as user_updated_at, p.full_name, p.avatar_url, p.role, p.points, p.preferences, p.created_at, p.updated_at
|
||||||
FROM public.users u
|
FROM public.users u
|
||||||
JOIN public.profiles p ON u.user_id = p.user_id
|
JOIN public.profiles p ON u.user_id = p.user_id
|
||||||
WHERE u.user_id = $1;
|
WHERE u.user_id = $1;
|
||||||
@@ -108,6 +112,8 @@ export class UserRepository {
|
|||||||
user: {
|
user: {
|
||||||
user_id: flatProfile.user_id,
|
user_id: flatProfile.user_id,
|
||||||
email: flatProfile.email,
|
email: flatProfile.email,
|
||||||
|
created_at: flatProfile.user_created_at,
|
||||||
|
updated_at: flatProfile.user_updated_at,
|
||||||
},
|
},
|
||||||
full_name: flatProfile.full_name,
|
full_name: flatProfile.full_name,
|
||||||
avatar_url: flatProfile.avatar_url,
|
avatar_url: flatProfile.avatar_url,
|
||||||
@@ -121,14 +127,16 @@ export class UserRepository {
|
|||||||
logger.debug({ user: fullUserProfile }, `[DB createUser] Fetched full profile for new user:`);
|
logger.debug({ user: fullUserProfile }, `[DB createUser] Fetched full profile for new user:`);
|
||||||
return fullUserProfile;
|
return fullUserProfile;
|
||||||
}).catch((error) => {
|
}).catch((error) => {
|
||||||
// Check for specific PostgreSQL error codes
|
// Specific handling for unique constraint violation on user creation
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23505') {
|
if (error instanceof Error && 'code' in error && (error as any).code === '23505') {
|
||||||
logger.warn(`Attempted to create a user with an existing email: ${email}`);
|
logger.warn(`Attempted to create a user with an existing email: ${email}`);
|
||||||
throw new UniqueConstraintError('A user with this email address already exists.');
|
throw new UniqueConstraintError('A user with this email address already exists.');
|
||||||
}
|
}
|
||||||
// The withTransaction helper logs the rollback, so we just log the context here.
|
// Fallback to generic handler for all other errors
|
||||||
logger.error({ err: error, email }, 'Error during createUser transaction');
|
handleDbError(error, logger, 'Error during createUser transaction', { email }, {
|
||||||
throw new Error('Failed to create user in database.');
|
uniqueMessage: 'A user with this email address already exists.',
|
||||||
|
defaultMessage: 'Failed to create user in database.',
|
||||||
|
});
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -145,15 +153,17 @@ export class UserRepository {
|
|||||||
logger.debug({ email }, `[DB findUserWithProfileByEmail] Searching for user.`);
|
logger.debug({ email }, `[DB findUserWithProfileByEmail] Searching for user.`);
|
||||||
try {
|
try {
|
||||||
const query = `
|
const query = `
|
||||||
SELECT
|
SELECT
|
||||||
u.user_id, u.email, u.password_hash, u.refresh_token, u.failed_login_attempts, u.last_failed_login,
|
u.user_id, u.email, u.created_at as user_created_at, u.updated_at as user_updated_at, u.password_hash, u.refresh_token, u.failed_login_attempts, u.last_failed_login,
|
||||||
p.full_name, p.avatar_url, p.role, p.points, p.preferences, p.address_id,
|
p.full_name, p.avatar_url, p.role, p.points, p.preferences, p.address_id,
|
||||||
p.created_at, p.updated_at
|
p.created_at, p.updated_at
|
||||||
FROM public.users u
|
FROM public.users u
|
||||||
JOIN public.profiles p ON u.user_id = p.user_id
|
JOIN public.profiles p ON u.user_id = p.user_id
|
||||||
WHERE u.email = $1;
|
WHERE u.email = $1;
|
||||||
`;
|
`;
|
||||||
const res = await this.db.query<DbUser & Profile>(query, [email]);
|
const res = await this.db.query<
|
||||||
|
DbUser & Profile & { user_created_at: string; user_updated_at: string }
|
||||||
|
>(query, [email]);
|
||||||
const flatUser = res.rows[0];
|
const flatUser = res.rows[0];
|
||||||
|
|
||||||
if (!flatUser) {
|
if (!flatUser) {
|
||||||
@@ -173,6 +183,8 @@ export class UserRepository {
|
|||||||
user: {
|
user: {
|
||||||
user_id: flatUser.user_id,
|
user_id: flatUser.user_id,
|
||||||
email: flatUser.email,
|
email: flatUser.email,
|
||||||
|
created_at: flatUser.user_created_at,
|
||||||
|
updated_at: flatUser.user_updated_at,
|
||||||
},
|
},
|
||||||
password_hash: flatUser.password_hash,
|
password_hash: flatUser.password_hash,
|
||||||
failed_login_attempts: flatUser.failed_login_attempts,
|
failed_login_attempts: flatUser.failed_login_attempts,
|
||||||
@@ -182,8 +194,9 @@ export class UserRepository {
|
|||||||
|
|
||||||
return authableProfile;
|
return authableProfile;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, email }, 'Database error in findUserWithProfileByEmail');
|
handleDbError(error, logger, 'Database error in findUserWithProfileByEmail', { email }, {
|
||||||
throw new Error('Failed to retrieve user with profile from database.');
|
defaultMessage: 'Failed to retrieve user with profile from database.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -193,10 +206,10 @@ export class UserRepository {
|
|||||||
* @returns A promise that resolves to the user object (id, email) or undefined if not found.
|
* @returns A promise that resolves to the user object (id, email) or undefined if not found.
|
||||||
*/
|
*/
|
||||||
// prettier-ignore
|
// prettier-ignore
|
||||||
async findUserById(userId: string, logger: Logger): Promise<{ user_id: string; email: string; }> {
|
async findUserById(userId: string, logger: Logger): Promise<User> {
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<{ user_id: string; email: string }>(
|
const res = await this.db.query<User>(
|
||||||
'SELECT user_id, email FROM public.users WHERE user_id = $1',
|
'SELECT user_id, email, created_at, updated_at FROM public.users WHERE user_id = $1',
|
||||||
[userId]
|
[userId]
|
||||||
);
|
);
|
||||||
if (res.rowCount === 0) {
|
if (res.rowCount === 0) {
|
||||||
@@ -205,11 +218,9 @@ export class UserRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in findUserById', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to retrieve user by ID from database.',
|
||||||
'Database error in findUserById',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to retrieve user by ID from database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -220,10 +231,10 @@ export class UserRepository {
|
|||||||
* @returns A promise that resolves to the user object (id, email, password_hash) or undefined if not found.
|
* @returns A promise that resolves to the user object (id, email, password_hash) or undefined if not found.
|
||||||
*/
|
*/
|
||||||
// prettier-ignore
|
// prettier-ignore
|
||||||
async findUserWithPasswordHashById(userId: string, logger: Logger): Promise<{ user_id: string; email: string; password_hash: string | null }> {
|
async findUserWithPasswordHashById(userId: string, logger: Logger): Promise<User & { password_hash: string | null }> {
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<{ user_id: string; email: string; password_hash: string | null }>(
|
const res = await this.db.query<User & { password_hash: string | null }>(
|
||||||
'SELECT user_id, email, password_hash FROM public.users WHERE user_id = $1',
|
'SELECT user_id, email, password_hash, created_at, updated_at FROM public.users WHERE user_id = $1',
|
||||||
[userId]
|
[userId]
|
||||||
);
|
);
|
||||||
if ((res.rowCount ?? 0) === 0) {
|
if ((res.rowCount ?? 0) === 0) {
|
||||||
@@ -232,11 +243,9 @@ export class UserRepository {
|
|||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof NotFoundError) throw error;
|
if (error instanceof NotFoundError) throw error;
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in findUserWithPasswordHashById', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to retrieve user with sensitive data by ID from database.',
|
||||||
'Database error in findUserWithPasswordHashById',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to retrieve user with sensitive data by ID from database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -253,7 +262,9 @@ export class UserRepository {
|
|||||||
p.created_at, p.updated_at,
|
p.created_at, p.updated_at,
|
||||||
json_build_object(
|
json_build_object(
|
||||||
'user_id', u.user_id,
|
'user_id', u.user_id,
|
||||||
'email', u.email
|
'email', u.email,
|
||||||
|
'created_at', u.created_at,
|
||||||
|
'updated_at', u.updated_at
|
||||||
) as user,
|
) as user,
|
||||||
CASE
|
CASE
|
||||||
WHEN a.address_id IS NOT NULL THEN json_build_object(
|
WHEN a.address_id IS NOT NULL THEN json_build_object(
|
||||||
@@ -281,11 +292,9 @@ export class UserRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in findUserProfileById', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to retrieve user profile from database.',
|
||||||
'Database error in findUserProfileById',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to retrieve user profile from database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -330,11 +339,10 @@ export class UserRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in updateUserProfile', { userId, profileData }, {
|
||||||
{ err: error, userId, profileData },
|
fkMessage: 'The specified address does not exist.',
|
||||||
'Database error in updateUserProfile',
|
defaultMessage: 'Failed to update user profile in database.',
|
||||||
);
|
});
|
||||||
throw new Error('Failed to update user profile in database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -362,11 +370,9 @@ export class UserRepository {
|
|||||||
if (error instanceof NotFoundError) {
|
if (error instanceof NotFoundError) {
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in updateUserPreferences', { userId, preferences }, {
|
||||||
{ err: error, userId, preferences },
|
defaultMessage: 'Failed to update user preferences in database.',
|
||||||
'Database error in updateUserPreferences',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to update user preferences in database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -383,11 +389,9 @@ export class UserRepository {
|
|||||||
[passwordHash, userId]
|
[passwordHash, userId]
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in updateUserPassword', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to update user password in database.',
|
||||||
'Database error in updateUserPassword',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to update user password in database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -400,11 +404,9 @@ export class UserRepository {
|
|||||||
try {
|
try {
|
||||||
await this.db.query('DELETE FROM public.users WHERE user_id = $1', [userId]);
|
await this.db.query('DELETE FROM public.users WHERE user_id = $1', [userId]);
|
||||||
} catch (error) { // This was a duplicate, fixed.
|
} catch (error) { // This was a duplicate, fixed.
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in deleteUserById', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to delete user from database.',
|
||||||
'Database error in deleteUserById',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to delete user from database.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -421,11 +423,9 @@ export class UserRepository {
|
|||||||
[refreshToken, userId]
|
[refreshToken, userId]
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in saveRefreshToken', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to save refresh token.',
|
||||||
'Database error in saveRefreshToken',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to save refresh token.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -437,10 +437,10 @@ export class UserRepository {
|
|||||||
async findUserByRefreshToken(
|
async findUserByRefreshToken(
|
||||||
refreshToken: string,
|
refreshToken: string,
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<{ user_id: string; email: string } | undefined> {
|
): Promise<User | undefined> {
|
||||||
try {
|
try {
|
||||||
const res = await this.db.query<{ user_id: string; email: string }>(
|
const res = await this.db.query<User>(
|
||||||
'SELECT user_id, email FROM public.users WHERE refresh_token = $1',
|
'SELECT user_id, email, created_at, updated_at FROM public.users WHERE refresh_token = $1',
|
||||||
[refreshToken],
|
[refreshToken],
|
||||||
);
|
);
|
||||||
if ((res.rowCount ?? 0) === 0) {
|
if ((res.rowCount ?? 0) === 0) {
|
||||||
@@ -448,8 +448,9 @@ export class UserRepository {
|
|||||||
}
|
}
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in findUserByRefreshToken');
|
handleDbError(error, logger, 'Database error in findUserByRefreshToken', {}, {
|
||||||
throw new Error('Failed to find user by refresh token.'); // Generic error for other failures
|
defaultMessage: 'Failed to find user by refresh token.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -483,14 +484,11 @@ export class UserRepository {
|
|||||||
[userId, tokenHash, expiresAt]
|
[userId, tokenHash, expiresAt]
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
handleDbError(error, logger, 'Database error in createPasswordResetToken', { userId }, {
|
||||||
throw new ForeignKeyConstraintError('The specified user does not exist.');
|
fkMessage: 'The specified user does not exist.',
|
||||||
}
|
uniqueMessage: 'A password reset token with this hash already exists.',
|
||||||
logger.error(
|
defaultMessage: 'Failed to create password reset token.',
|
||||||
{ err: error, userId },
|
});
|
||||||
'Database error in createPasswordResetToken',
|
|
||||||
);
|
|
||||||
throw new Error('Failed to create password reset token.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -506,11 +504,9 @@ export class UserRepository {
|
|||||||
);
|
);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in getValidResetTokens', {}, {
|
||||||
{ err: error },
|
defaultMessage: 'Failed to retrieve valid reset tokens.',
|
||||||
'Database error in getValidResetTokens',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to retrieve valid reset tokens.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -545,8 +541,9 @@ export class UserRepository {
|
|||||||
);
|
);
|
||||||
return res.rowCount ?? 0;
|
return res.rowCount ?? 0;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error }, 'Database error in deleteExpiredResetTokens');
|
handleDbError(error, logger, 'Database error in deleteExpiredResetTokens', {}, {
|
||||||
throw new Error('Failed to delete expired password reset tokens.');
|
defaultMessage: 'Failed to delete expired password reset tokens.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
@@ -561,11 +558,11 @@ export class UserRepository {
|
|||||||
[followerId, followingId],
|
[followerId, followingId],
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof Error && 'code' in error && error.code === '23503') {
|
handleDbError(error, logger, 'Database error in followUser', { followerId, followingId }, {
|
||||||
throw new ForeignKeyConstraintError('One or both users do not exist.');
|
fkMessage: 'One or both users do not exist.',
|
||||||
}
|
checkMessage: 'A user cannot follow themselves.',
|
||||||
logger.error({ err: error, followerId, followingId }, 'Database error in followUser');
|
defaultMessage: 'Failed to follow user.',
|
||||||
throw new Error('Failed to follow user.');
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -581,8 +578,9 @@ export class UserRepository {
|
|||||||
[followerId, followingId],
|
[followerId, followingId],
|
||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, followerId, followingId }, 'Database error in unfollowUser');
|
handleDbError(error, logger, 'Database error in unfollowUser', { followerId, followingId }, {
|
||||||
throw new Error('Failed to unfollow user.');
|
defaultMessage: 'Failed to unfollow user.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -612,8 +610,9 @@ export class UserRepository {
|
|||||||
const res = await this.db.query<ActivityLogItem>(query, [userId, limit, offset]);
|
const res = await this.db.query<ActivityLogItem>(query, [userId, limit, offset]);
|
||||||
return res.rows;
|
return res.rows;
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, userId, limit, offset }, 'Database error in getUserFeed');
|
handleDbError(error, logger, 'Database error in getUserFeed', { userId, limit, offset }, {
|
||||||
throw new Error('Failed to retrieve user feed.');
|
defaultMessage: 'Failed to retrieve user feed.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -623,7 +622,7 @@ export class UserRepository {
|
|||||||
* @returns A promise that resolves to the created SearchQuery object.
|
* @returns A promise that resolves to the created SearchQuery object.
|
||||||
*/
|
*/
|
||||||
async logSearchQuery(
|
async logSearchQuery(
|
||||||
queryData: Omit<SearchQuery, 'search_query_id' | 'created_at'>,
|
queryData: Omit<SearchQuery, 'search_query_id' | 'created_at' | 'updated_at'>,
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<SearchQuery> {
|
): Promise<SearchQuery> {
|
||||||
const { user_id, query_text, result_count, was_successful } = queryData;
|
const { user_id, query_text, result_count, was_successful } = queryData;
|
||||||
@@ -634,8 +633,10 @@ export class UserRepository {
|
|||||||
);
|
);
|
||||||
return res.rows[0];
|
return res.rows[0];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ err: error, queryData }, 'Database error in logSearchQuery');
|
handleDbError(error, logger, 'Database error in logSearchQuery', { queryData }, {
|
||||||
throw new Error('Failed to log search query.');
|
fkMessage: 'The specified user does not exist.',
|
||||||
|
defaultMessage: 'Failed to log search query.',
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -668,10 +669,8 @@ export async function exportUserData(userId: string, logger: Logger): Promise<{
|
|||||||
return { profile, watchedItems, shoppingLists };
|
return { profile, watchedItems, shoppingLists };
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error(
|
handleDbError(error, logger, 'Database error in exportUserData', { userId }, {
|
||||||
{ err: error, userId },
|
defaultMessage: 'Failed to export user data.',
|
||||||
'Database error in exportUserData',
|
});
|
||||||
);
|
|
||||||
throw new Error('Failed to export user data.');
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||||
import { AiDataValidationError } from './processingErrors';
|
import { AiDataValidationError } from './processingErrors';
|
||||||
import { logger } from './logger.server';
|
import { logger } from './logger.server'; // Keep this import for the logger instance
|
||||||
import type { AIService } from './aiService.server';
|
import type { AIService } from './aiService.server';
|
||||||
import type { PersonalizationRepository } from './db/personalization.db';
|
import type { PersonalizationRepository } from './db/personalization.db';
|
||||||
import type { FlyerJobData } from '../types/job-data';
|
import type { FlyerJobData } from '../types/job-data';
|
||||||
@@ -63,7 +63,8 @@ describe('FlyerAiProcessor', () => {
|
|||||||
};
|
};
|
||||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||||
|
|
||||||
const result = await service.extractAndValidateData([], jobData, logger);
|
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||||
|
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||||
|
|
||||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(1);
|
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(1);
|
||||||
expect(mockPersonalizationRepo.getAllMasterItems).toHaveBeenCalledTimes(1);
|
expect(mockPersonalizationRepo.getAllMasterItems).toHaveBeenCalledTimes(1);
|
||||||
@@ -83,7 +84,8 @@ describe('FlyerAiProcessor', () => {
|
|||||||
};
|
};
|
||||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(invalidResponse as any);
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(invalidResponse as any);
|
||||||
|
|
||||||
await expect(service.extractAndValidateData([], jobData, logger)).rejects.toThrow(
|
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||||
|
await expect(service.extractAndValidateData(imagePaths, jobData, logger)).rejects.toThrow(
|
||||||
AiDataValidationError,
|
AiDataValidationError,
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
@@ -98,10 +100,11 @@ describe('FlyerAiProcessor', () => {
|
|||||||
valid_to: null,
|
valid_to: null,
|
||||||
store_address: null,
|
store_address: null,
|
||||||
};
|
};
|
||||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse as any);
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||||
const { logger } = await import('./logger.server');
|
const { logger } = await import('./logger.server');
|
||||||
|
|
||||||
const result = await service.extractAndValidateData([], jobData, logger);
|
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||||
|
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||||
|
|
||||||
// It should not throw, but return the data and log a warning.
|
// It should not throw, but return the data and log a warning.
|
||||||
expect(result.data).toEqual(mockAiResponse);
|
expect(result.data).toEqual(mockAiResponse);
|
||||||
@@ -122,9 +125,104 @@ describe('FlyerAiProcessor', () => {
|
|||||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||||
const { logger } = await import('./logger.server');
|
const { logger } = await import('./logger.server');
|
||||||
|
|
||||||
const result = await service.extractAndValidateData([], jobData, logger);
|
const imagePaths = [{ path: 'page1.jpg', mimetype: 'image/jpeg' }];
|
||||||
|
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||||
expect(result.data).toEqual(mockAiResponse);
|
expect(result.data).toEqual(mockAiResponse);
|
||||||
expect(result.needsReview).toBe(true);
|
expect(result.needsReview).toBe(true);
|
||||||
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('contains no items. The flyer will be saved with an item_count of 0. Flagging for review.'));
|
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('contains no items. The flyer will be saved with an item_count of 0. Flagging for review.'));
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('Batching Logic', () => {
|
||||||
|
it('should process images in batches and merge the results correctly', async () => {
|
||||||
|
// Arrange
|
||||||
|
const jobData = createMockJobData({});
|
||||||
|
// 5 images, with BATCH_SIZE = 4, should result in 2 batches.
|
||||||
|
const imagePaths = [
|
||||||
|
{ path: 'page1.jpg', mimetype: 'image/jpeg' },
|
||||||
|
{ path: 'page2.jpg', mimetype: 'image/jpeg' },
|
||||||
|
{ path: 'page3.jpg', mimetype: 'image/jpeg' },
|
||||||
|
{ path: 'page4.jpg', mimetype: 'image/jpeg' },
|
||||||
|
{ path: 'page5.jpg', mimetype: 'image/jpeg' },
|
||||||
|
];
|
||||||
|
|
||||||
|
const mockAiResponseBatch1 = {
|
||||||
|
store_name: 'Batch 1 Store',
|
||||||
|
valid_from: '2025-01-01',
|
||||||
|
valid_to: '2025-01-07',
|
||||||
|
store_address: '123 Batch St',
|
||||||
|
items: [
|
||||||
|
{ item: 'Item A', price_display: '$1', price_in_cents: 100, quantity: '1', category_name: 'Cat A', master_item_id: 1 },
|
||||||
|
{ item: 'Item B', price_display: '$2', price_in_cents: 200, quantity: '1', category_name: 'Cat B', master_item_id: 2 },
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
const mockAiResponseBatch2 = {
|
||||||
|
store_name: 'Batch 2 Store', // This should be ignored in the merge
|
||||||
|
valid_from: null,
|
||||||
|
valid_to: null,
|
||||||
|
store_address: null,
|
||||||
|
items: [
|
||||||
|
{ item: 'Item C', price_display: '$3', price_in_cents: 300, quantity: '1', category_name: 'Cat C', master_item_id: 3 },
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
// Mock the AI service to return different results for each batch call
|
||||||
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage)
|
||||||
|
.mockResolvedValueOnce(mockAiResponseBatch1)
|
||||||
|
.mockResolvedValueOnce(mockAiResponseBatch2);
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
// 1. AI service was called twice (for 2 batches)
|
||||||
|
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(2);
|
||||||
|
|
||||||
|
// 2. Check the arguments for each call
|
||||||
|
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenNthCalledWith(1, imagePaths.slice(0, 4), [], undefined, undefined, logger);
|
||||||
|
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenNthCalledWith(2, imagePaths.slice(4, 5), [], undefined, undefined, logger);
|
||||||
|
|
||||||
|
// 3. Check the merged data
|
||||||
|
expect(result.data.store_name).toBe('Batch 1 Store'); // Metadata from the first batch
|
||||||
|
expect(result.data.valid_from).toBe('2025-01-01');
|
||||||
|
expect(result.data.valid_to).toBe('2025-01-07');
|
||||||
|
expect(result.data.store_address).toBe('123 Batch St');
|
||||||
|
|
||||||
|
// 4. Check that items from both batches are merged
|
||||||
|
expect(result.data.items).toHaveLength(3);
|
||||||
|
expect(result.data.items).toEqual(expect.arrayContaining([
|
||||||
|
expect.objectContaining({ item: 'Item A' }),
|
||||||
|
expect.objectContaining({ item: 'Item B' }),
|
||||||
|
expect.objectContaining({ item: 'Item C' }),
|
||||||
|
]));
|
||||||
|
|
||||||
|
// 5. Check that the job is not flagged for review
|
||||||
|
expect(result.needsReview).toBe(false);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fill in missing metadata from subsequent batches', async () => {
|
||||||
|
// Arrange
|
||||||
|
const jobData = createMockJobData({});
|
||||||
|
const imagePaths = [
|
||||||
|
{ path: 'page1.jpg', mimetype: 'image/jpeg' }, { path: 'page2.jpg', mimetype: 'image/jpeg' }, { path: 'page3.jpg', mimetype: 'image/jpeg' }, { path: 'page4.jpg', mimetype: 'image/jpeg' }, { path: 'page5.jpg', mimetype: 'image/jpeg' },
|
||||||
|
];
|
||||||
|
|
||||||
|
const mockAiResponseBatch1 = { store_name: null, valid_from: '2025-01-01', valid_to: '2025-01-07', store_address: null, items: [{ item: 'Item A', price_display: '$1', price_in_cents: 100, quantity: '1', category_name: 'Cat A', master_item_id: 1 }] };
|
||||||
|
const mockAiResponseBatch2 = { store_name: 'Batch 2 Store', valid_from: '2025-01-02', valid_to: null, store_address: '456 Subsequent St', items: [{ item: 'Item C', price_display: '$3', price_in_cents: 300, quantity: '1', category_name: 'Cat C', master_item_id: 3 }] };
|
||||||
|
|
||||||
|
vi.mocked(mockAiService.extractCoreDataFromFlyerImage)
|
||||||
|
.mockResolvedValueOnce(mockAiResponseBatch1)
|
||||||
|
.mockResolvedValueOnce(mockAiResponseBatch2);
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const result = await service.extractAndValidateData(imagePaths, jobData, logger);
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(result.data.store_name).toBe('Batch 2 Store'); // Filled from batch 2
|
||||||
|
expect(result.data.valid_from).toBe('2025-01-01'); // Kept from batch 1
|
||||||
|
expect(result.data.valid_to).toBe('2025-01-07'); // Kept from batch 1
|
||||||
|
expect(result.data.store_address).toBe('456 Subsequent St'); // Filled from batch 2
|
||||||
|
expect(result.data.items).toHaveLength(2);
|
||||||
|
});
|
||||||
|
});
|
||||||
});
|
});
|
||||||
@@ -5,28 +5,11 @@ import type { AIService } from './aiService.server';
|
|||||||
import type { PersonalizationRepository } from './db/personalization.db';
|
import type { PersonalizationRepository } from './db/personalization.db';
|
||||||
import { AiDataValidationError } from './processingErrors';
|
import { AiDataValidationError } from './processingErrors';
|
||||||
import type { FlyerJobData } from '../types/job-data';
|
import type { FlyerJobData } from '../types/job-data';
|
||||||
|
import {
|
||||||
// Helper for consistent required string validation (handles missing/null/empty)
|
AiFlyerDataSchema,
|
||||||
const requiredString = (message: string) =>
|
ExtractedFlyerItemSchema,
|
||||||
z.preprocess((val) => val ?? '', z.string().min(1, message));
|
requiredString,
|
||||||
|
} from '../types/ai'; // Import consolidated schemas and helper
|
||||||
// --- Zod Schemas for AI Response Validation ---
|
|
||||||
const ExtractedFlyerItemSchema = z.object({
|
|
||||||
item: z.string().nullable(),
|
|
||||||
price_display: z.string().nullable(),
|
|
||||||
price_in_cents: z.number().nullable(),
|
|
||||||
quantity: z.string().nullable(),
|
|
||||||
category_name: z.string().nullable(),
|
|
||||||
master_item_id: z.number().nullish(),
|
|
||||||
});
|
|
||||||
|
|
||||||
export const AiFlyerDataSchema = z.object({
|
|
||||||
store_name: z.string().nullable(),
|
|
||||||
valid_from: z.string().nullable(),
|
|
||||||
valid_to: z.string().nullable(),
|
|
||||||
store_address: z.string().nullable(),
|
|
||||||
items: z.array(ExtractedFlyerItemSchema),
|
|
||||||
});
|
|
||||||
|
|
||||||
export type ValidatedAiDataType = z.infer<typeof AiFlyerDataSchema>;
|
export type ValidatedAiDataType = z.infer<typeof AiFlyerDataSchema>;
|
||||||
|
|
||||||
@@ -94,19 +77,64 @@ export class FlyerAiProcessor {
|
|||||||
jobData: FlyerJobData,
|
jobData: FlyerJobData,
|
||||||
logger: Logger,
|
logger: Logger,
|
||||||
): Promise<AiProcessorResult> {
|
): Promise<AiProcessorResult> {
|
||||||
logger.info(`Starting AI data extraction.`);
|
logger.info(`Starting AI data extraction for ${imagePaths.length} pages.`);
|
||||||
const { submitterIp, userProfileAddress } = jobData;
|
const { submitterIp, userProfileAddress } = jobData;
|
||||||
const masterItems = await this.personalizationRepo.getAllMasterItems(logger);
|
const masterItems = await this.personalizationRepo.getAllMasterItems(logger);
|
||||||
logger.debug(`Retrieved ${masterItems.length} master items for AI matching.`);
|
logger.debug(`Retrieved ${masterItems.length} master items for AI matching.`);
|
||||||
|
|
||||||
const extractedData = await this.ai.extractCoreDataFromFlyerImage(
|
// BATCHING LOGIC: Process images in chunks to avoid hitting AI payload/token limits.
|
||||||
imagePaths,
|
const BATCH_SIZE = 4;
|
||||||
masterItems,
|
const batches = [];
|
||||||
submitterIp,
|
for (let i = 0; i < imagePaths.length; i += BATCH_SIZE) {
|
||||||
userProfileAddress,
|
batches.push(imagePaths.slice(i, i + BATCH_SIZE));
|
||||||
logger,
|
}
|
||||||
);
|
|
||||||
|
|
||||||
return this._validateAiData(extractedData, logger);
|
// Initialize container for merged data
|
||||||
|
const mergedData: ValidatedAiDataType = {
|
||||||
|
store_name: null,
|
||||||
|
valid_from: null,
|
||||||
|
valid_to: null,
|
||||||
|
store_address: null,
|
||||||
|
items: [],
|
||||||
|
};
|
||||||
|
|
||||||
|
logger.info(`Processing ${imagePaths.length} pages in ${batches.length} batches (Batch Size: ${BATCH_SIZE}).`);
|
||||||
|
|
||||||
|
for (const [index, batch] of batches.entries()) {
|
||||||
|
logger.info(`Processing batch ${index + 1}/${batches.length} (${batch.length} pages)...`);
|
||||||
|
|
||||||
|
// The AI service handles rate limiting internally (e.g., max 5 RPM).
|
||||||
|
// Processing these sequentially ensures we respect that limit.
|
||||||
|
const batchResult = await this.ai.extractCoreDataFromFlyerImage(
|
||||||
|
batch,
|
||||||
|
masterItems,
|
||||||
|
submitterIp,
|
||||||
|
userProfileAddress,
|
||||||
|
logger,
|
||||||
|
);
|
||||||
|
|
||||||
|
// MERGE LOGIC:
|
||||||
|
// 1. Metadata (Store Name, Dates): Prioritize the first batch (usually the cover page).
|
||||||
|
// If subsequent batches have data and the current is null, fill it in.
|
||||||
|
if (index === 0) {
|
||||||
|
mergedData.store_name = batchResult.store_name;
|
||||||
|
mergedData.valid_from = batchResult.valid_from;
|
||||||
|
mergedData.valid_to = batchResult.valid_to;
|
||||||
|
mergedData.store_address = batchResult.store_address;
|
||||||
|
} else {
|
||||||
|
if (!mergedData.store_name && batchResult.store_name) mergedData.store_name = batchResult.store_name;
|
||||||
|
if (!mergedData.valid_from && batchResult.valid_from) mergedData.valid_from = batchResult.valid_from;
|
||||||
|
if (!mergedData.valid_to && batchResult.valid_to) mergedData.valid_to = batchResult.valid_to;
|
||||||
|
if (!mergedData.store_address && batchResult.store_address) mergedData.store_address = batchResult.store_address;
|
||||||
|
}
|
||||||
|
|
||||||
|
// 2. Items: Append all found items to the master list.
|
||||||
|
mergedData.items.push(...batchResult.items);
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.info(`Batch processing complete. Total items extracted: ${mergedData.items.length}`);
|
||||||
|
|
||||||
|
// Validate the final merged dataset
|
||||||
|
return this._validateAiData(mergedData, logger);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -2,9 +2,11 @@
|
|||||||
import path from 'path';
|
import path from 'path';
|
||||||
import type { z } from 'zod';
|
import type { z } from 'zod';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import type { FlyerInsert, FlyerItemInsert, FlyerStatus } from '../types';
|
import type { FlyerInsert, FlyerItemInsert } from '../types';
|
||||||
import type { AiFlyerDataSchema, AiProcessorResult } from './flyerAiProcessor.server';
|
import type { AiProcessorResult } from './flyerAiProcessor.server'; // Keep this import for AiProcessorResult
|
||||||
|
import { AiFlyerDataSchema } from '../types/ai'; // Import consolidated schema
|
||||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||||
|
import { TransformationError } from './processingErrors';
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This class is responsible for transforming the validated data from the AI service
|
* This class is responsible for transforming the validated data from the AI service
|
||||||
@@ -56,41 +58,47 @@ export class FlyerDataTransformer {
|
|||||||
): Promise<{ flyerData: FlyerInsert; itemsForDb: FlyerItemInsert[] }> {
|
): Promise<{ flyerData: FlyerInsert; itemsForDb: FlyerItemInsert[] }> {
|
||||||
logger.info('Starting data transformation from AI output to database format.');
|
logger.info('Starting data transformation from AI output to database format.');
|
||||||
|
|
||||||
const { data: extractedData, needsReview } = aiResult;
|
try {
|
||||||
|
const { data: extractedData, needsReview } = aiResult;
|
||||||
|
|
||||||
const firstImage = imagePaths[0].path;
|
const firstImage = imagePaths[0].path;
|
||||||
const iconFileName = await generateFlyerIcon(
|
const iconFileName = await generateFlyerIcon(
|
||||||
firstImage,
|
firstImage,
|
||||||
path.join(path.dirname(firstImage), 'icons'),
|
path.join(path.dirname(firstImage), 'icons'),
|
||||||
logger,
|
logger,
|
||||||
);
|
);
|
||||||
|
|
||||||
const itemsForDb: FlyerItemInsert[] = extractedData.items.map((item) => this._normalizeItem(item));
|
const itemsForDb: FlyerItemInsert[] = extractedData.items.map((item) => this._normalizeItem(item));
|
||||||
|
|
||||||
const storeName = extractedData.store_name || 'Unknown Store (auto)';
|
const storeName = extractedData.store_name || 'Unknown Store (auto)';
|
||||||
if (!extractedData.store_name) {
|
if (!extractedData.store_name) {
|
||||||
logger.warn('AI did not return a store name. Using fallback "Unknown Store (auto)".');
|
logger.warn('AI did not return a store name. Using fallback "Unknown Store (auto)".');
|
||||||
|
}
|
||||||
|
|
||||||
|
const flyerData: FlyerInsert = {
|
||||||
|
file_name: originalFileName,
|
||||||
|
image_url: `/flyer-images/${path.basename(firstImage)}`,
|
||||||
|
icon_url: `/flyer-images/icons/${iconFileName}`,
|
||||||
|
checksum,
|
||||||
|
store_name: storeName,
|
||||||
|
valid_from: extractedData.valid_from,
|
||||||
|
valid_to: extractedData.valid_to,
|
||||||
|
store_address: extractedData.store_address, // The number of items is now calculated directly from the transformed data.
|
||||||
|
item_count: itemsForDb.length,
|
||||||
|
uploaded_by: userId,
|
||||||
|
status: needsReview ? 'needs_review' : 'processed',
|
||||||
|
};
|
||||||
|
|
||||||
|
logger.info(
|
||||||
|
{ itemCount: itemsForDb.length, storeName: flyerData.store_name },
|
||||||
|
'Data transformation complete.',
|
||||||
|
);
|
||||||
|
|
||||||
|
return { flyerData, itemsForDb };
|
||||||
|
} catch (err) {
|
||||||
|
logger.error({ err }, 'Transformation process failed');
|
||||||
|
// Wrap and rethrow with the new error class
|
||||||
|
throw new TransformationError('Flyer Data Transformation Failed');
|
||||||
}
|
}
|
||||||
|
|
||||||
const flyerData: FlyerInsert = {
|
|
||||||
file_name: originalFileName,
|
|
||||||
image_url: `/flyer-images/${path.basename(firstImage)}`,
|
|
||||||
icon_url: `/flyer-images/icons/${iconFileName}`,
|
|
||||||
checksum,
|
|
||||||
store_name: storeName,
|
|
||||||
valid_from: extractedData.valid_from,
|
|
||||||
valid_to: extractedData.valid_to,
|
|
||||||
store_address: extractedData.store_address, // The number of items is now calculated directly from the transformed data.
|
|
||||||
item_count: itemsForDb.length,
|
|
||||||
uploaded_by: userId,
|
|
||||||
status: needsReview ? 'needs_review' : 'processed',
|
|
||||||
};
|
|
||||||
|
|
||||||
logger.info(
|
|
||||||
{ itemCount: itemsForDb.length, storeName: flyerData.store_name },
|
|
||||||
'Data transformation complete.',
|
|
||||||
);
|
|
||||||
|
|
||||||
return { flyerData, itemsForDb };
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,12 +1,8 @@
|
|||||||
// src/services/flyerProcessingService.server.test.ts
|
// src/services/flyerProcessingService.server.test.ts
|
||||||
import { describe, it, expect, vi, beforeEach, type Mocked } from 'vitest';
|
import { describe, it, expect, vi, beforeEach, type Mocked } from 'vitest';
|
||||||
import sharp from 'sharp';
|
|
||||||
import { Job, UnrecoverableError } from 'bullmq';
|
import { Job, UnrecoverableError } from 'bullmq';
|
||||||
import type { Dirent } from 'node:fs';
|
import { AiFlyerDataSchema } from '../types/ai';
|
||||||
import type { Logger } from 'pino';
|
import type { FlyerInsert } from '../types';
|
||||||
import { z } from 'zod';
|
|
||||||
import { AiFlyerDataSchema } from './flyerAiProcessor.server';
|
|
||||||
import type { Flyer, FlyerInsert, FlyerItemInsert } from '../types';
|
|
||||||
import type { CleanupJobData, FlyerJobData } from '../types/job-data';
|
import type { CleanupJobData, FlyerJobData } from '../types/job-data';
|
||||||
|
|
||||||
// 1. Create hoisted mocks FIRST
|
// 1. Create hoisted mocks FIRST
|
||||||
|
|||||||
@@ -133,6 +133,12 @@ export class FlyerProcessingService {
|
|||||||
return { flyerId: flyer.flyer_id };
|
return { flyerId: flyer.flyer_id };
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
||||||
|
// Add detailed logging of the raw error object
|
||||||
|
if (error instanceof Error) {
|
||||||
|
logger.error({ err: error, stack: error.stack }, 'Raw error object in processJob catch block');
|
||||||
|
} else {
|
||||||
|
logger.error({ error }, 'Raw non-Error object in processJob catch block');
|
||||||
|
}
|
||||||
// This private method handles error reporting and re-throwing.
|
// This private method handles error reporting and re-throwing.
|
||||||
await this._reportErrorAndThrow(error, job, logger, stages);
|
await this._reportErrorAndThrow(error, job, logger, stages);
|
||||||
// This line is technically unreachable because the above method always throws,
|
// This line is technically unreachable because the above method always throws,
|
||||||
@@ -197,6 +203,14 @@ export class FlyerProcessingService {
|
|||||||
logger: Logger,
|
logger: Logger,
|
||||||
initialStages: ProcessingStage[],
|
initialStages: ProcessingStage[],
|
||||||
): Promise<never> {
|
): Promise<never> {
|
||||||
|
// Map specific error codes to their corresponding processing stage names.
|
||||||
|
// This is more maintainable than a long if/else if chain.
|
||||||
|
const errorCodeToStageMap = new Map<string, string>([
|
||||||
|
['PDF_CONVERSION_FAILED', 'Preparing Inputs'],
|
||||||
|
['UNSUPPORTED_FILE_TYPE', 'Preparing Inputs'],
|
||||||
|
['AI_VALIDATION_FAILED', 'Extracting Data with AI'],
|
||||||
|
['TRANSFORMATION_FAILED', 'Transforming AI Data'], // Add new mapping
|
||||||
|
]);
|
||||||
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
||||||
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
||||||
let stagesToReport: ProcessingStage[] = [...initialStages]; // Create a mutable copy
|
let stagesToReport: ProcessingStage[] = [...initialStages]; // Create a mutable copy
|
||||||
@@ -209,16 +223,15 @@ export class FlyerProcessingService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Determine which stage failed
|
// Determine which stage failed
|
||||||
let errorStageIndex = -1;
|
const failedStageName = errorCodeToStageMap.get(errorPayload.errorCode);
|
||||||
|
let errorStageIndex = failedStageName ? stagesToReport.findIndex(s => s.name === failedStageName) : -1;
|
||||||
|
|
||||||
// 1. Try to map specific error codes/messages to stages
|
// Fallback for generic errors not in the map. This is less robust and relies on string matching.
|
||||||
if (errorPayload.errorCode === 'PDF_CONVERSION_FAILED' || errorPayload.errorCode === 'UNSUPPORTED_FILE_TYPE') {
|
// A future improvement would be to wrap these in specific FlyerProcessingError subclasses.
|
||||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Preparing Inputs');
|
if (errorStageIndex === -1 && errorPayload.message.includes('Icon generation failed')) {
|
||||||
} else if (errorPayload.errorCode === 'AI_VALIDATION_FAILED') {
|
|
||||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Extracting Data with AI');
|
|
||||||
} else if (errorPayload.message.includes('Icon generation failed')) {
|
|
||||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Transforming AI Data');
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Transforming AI Data');
|
||||||
} else if (errorPayload.message.includes('Database transaction failed')) {
|
}
|
||||||
|
if (errorStageIndex === -1 && errorPayload.message.includes('Database transaction failed')) {
|
||||||
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Saving to Database');
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Saving to Database');
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -254,24 +267,16 @@ export class FlyerProcessingService {
|
|||||||
|
|
||||||
// Logging logic
|
// Logging logic
|
||||||
if (normalizedError instanceof FlyerProcessingError) {
|
if (normalizedError instanceof FlyerProcessingError) {
|
||||||
const logDetails: Record<string, any> = { err: normalizedError };
|
// Simplify log object creation
|
||||||
|
const logDetails: Record<string, any> = { ...errorPayload, err: normalizedError };
|
||||||
|
|
||||||
if (normalizedError instanceof AiDataValidationError) {
|
if (normalizedError instanceof AiDataValidationError) {
|
||||||
logDetails.validationErrors = normalizedError.validationErrors;
|
logDetails.validationErrors = normalizedError.validationErrors;
|
||||||
logDetails.rawData = normalizedError.rawData;
|
logDetails.rawData = normalizedError.rawData;
|
||||||
}
|
}
|
||||||
// Also include stderr for PdfConversionError in logs
|
|
||||||
if (normalizedError instanceof PdfConversionError) {
|
if (normalizedError instanceof PdfConversionError) {
|
||||||
logDetails.stderr = normalizedError.stderr;
|
logDetails.stderr = normalizedError.stderr;
|
||||||
}
|
}
|
||||||
// Include the errorPayload details in the log, but avoid duplicating err, validationErrors, rawData
|
|
||||||
Object.assign(logDetails, errorPayload);
|
|
||||||
// Remove the duplicated err property if it was assigned by Object.assign
|
|
||||||
if ('err' in logDetails && logDetails.err === normalizedError) {
|
|
||||||
// This check prevents accidental deletion if 'err' was a legitimate property of errorPayload
|
|
||||||
delete logDetails.err;
|
|
||||||
}
|
|
||||||
// Ensure the original error object is always passed as 'err' for consistency in logging
|
|
||||||
logDetails.err = normalizedError;
|
|
||||||
|
|
||||||
logger.error(logDetails, `A known processing error occurred: ${normalizedError.name}`);
|
logger.error(logDetails, `A known processing error occurred: ${normalizedError.name}`);
|
||||||
} else {
|
} else {
|
||||||
|
|||||||
@@ -62,6 +62,18 @@ export class AiDataValidationError extends FlyerProcessingError {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Error thrown when a transformation step fails.
|
||||||
|
*/
|
||||||
|
export class TransformationError extends FlyerProcessingError {
|
||||||
|
constructor(message: string) {
|
||||||
|
super(
|
||||||
|
message,
|
||||||
|
'TRANSFORMATION_FAILED',
|
||||||
|
'There was a problem transforming the flyer data. Please check the input.',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
/**
|
/**
|
||||||
* Error thrown when an image conversion fails (e.g., using sharp).
|
* Error thrown when an image conversion fails (e.g., using sharp).
|
||||||
*/
|
*/
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
// src/services/workers.server.ts
|
||||||
import { Worker, Job, UnrecoverableError } from 'bullmq';
|
import { Worker, Job, UnrecoverableError } from 'bullmq';
|
||||||
import fsPromises from 'node:fs/promises';
|
import fsPromises from 'node:fs/promises';
|
||||||
import { exec } from 'child_process';
|
import { exec } from 'child_process';
|
||||||
|
|||||||
215
src/tests/e2e/auth.e2e.test.ts
Normal file
215
src/tests/e2e/auth.e2e.test.ts
Normal file
@@ -0,0 +1,215 @@
|
|||||||
|
// src/tests/e2e/auth.e2e.test.ts
|
||||||
|
import { describe, it, expect, afterAll, beforeAll } from 'vitest';
|
||||||
|
import * as apiClient from '../../services/apiClient';
|
||||||
|
import { cleanupDb } from '../utils/cleanup';
|
||||||
|
import { createAndLoginUser, TEST_PASSWORD } from '../utils/testHelpers';
|
||||||
|
import type { UserProfile } from '../../types';
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @vitest-environment node
|
||||||
|
*/
|
||||||
|
|
||||||
|
describe('Authentication E2E Flow', () => {
|
||||||
|
let testUser: UserProfile;
|
||||||
|
const createdUserIds: string[] = [];
|
||||||
|
|
||||||
|
beforeAll(async () => {
|
||||||
|
// Create a user that can be used for login-related tests in this suite.
|
||||||
|
const { user } = await createAndLoginUser({
|
||||||
|
email: `e2e-login-user-${Date.now()}@example.com`,
|
||||||
|
fullName: 'E2E Login User',
|
||||||
|
// E2E tests use apiClient which doesn't need the `request` object.
|
||||||
|
});
|
||||||
|
testUser = user;
|
||||||
|
createdUserIds.push(user.user.user_id);
|
||||||
|
});
|
||||||
|
|
||||||
|
afterAll(async () => {
|
||||||
|
if (createdUserIds.length > 0) {
|
||||||
|
await cleanupDb({ userIds: createdUserIds });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('Registration Flow', () => {
|
||||||
|
it('should successfully register a new user', async () => {
|
||||||
|
const email = `e2e-register-success-${Date.now()}@example.com`;
|
||||||
|
const fullName = 'E2E Register User';
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const response = await apiClient.registerUser(email, TEST_PASSWORD, fullName);
|
||||||
|
const data = await response.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(201);
|
||||||
|
expect(data.message).toBe('User registered successfully!');
|
||||||
|
expect(data.userprofile).toBeDefined();
|
||||||
|
expect(data.userprofile.user.email).toBe(email);
|
||||||
|
expect(data.token).toBeTypeOf('string');
|
||||||
|
|
||||||
|
// Add to cleanup
|
||||||
|
createdUserIds.push(data.userprofile.user.user_id);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fail to register a user with a weak password', async () => {
|
||||||
|
const email = `e2e-register-weakpass-${Date.now()}@example.com`;
|
||||||
|
const weakPassword = '123';
|
||||||
|
|
||||||
|
// Act
|
||||||
|
const response = await apiClient.registerUser(email, weakPassword, 'Weak Pass User');
|
||||||
|
const errorData = await response.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(400);
|
||||||
|
expect(errorData.errors[0].message).toContain('Password must be at least 8 characters long.');
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fail to register a user with a duplicate email', async () => {
|
||||||
|
const email = `e2e-register-duplicate-${Date.now()}@example.com`;
|
||||||
|
|
||||||
|
// Act 1: Register the user successfully
|
||||||
|
const firstResponse = await apiClient.registerUser(email, TEST_PASSWORD, 'Duplicate User');
|
||||||
|
const firstData = await firstResponse.json();
|
||||||
|
expect(firstResponse.status).toBe(201);
|
||||||
|
createdUserIds.push(firstData.userprofile.user.user_id); // Add for cleanup
|
||||||
|
|
||||||
|
// Act 2: Attempt to register the same user again
|
||||||
|
const secondResponse = await apiClient.registerUser(email, TEST_PASSWORD, 'Duplicate User');
|
||||||
|
const errorData = await secondResponse.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(secondResponse.status).toBe(409); // Conflict
|
||||||
|
expect(errorData.message).toContain('A user with this email address already exists.');
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('Login Flow', () => {
|
||||||
|
it('should successfully log in a registered user', async () => {
|
||||||
|
// Act: Attempt to log in with the user created in beforeAll
|
||||||
|
const response = await apiClient.loginUser(testUser.user.email, TEST_PASSWORD, false);
|
||||||
|
const data = await response.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(data.userprofile).toBeDefined();
|
||||||
|
expect(data.userprofile.user.email).toBe(testUser.user.email);
|
||||||
|
expect(data.token).toBeTypeOf('string');
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fail to log in with an incorrect password', async () => {
|
||||||
|
// Act: Attempt to log in with the wrong password
|
||||||
|
const response = await apiClient.loginUser(testUser.user.email, 'wrong-password', false);
|
||||||
|
const errorData = await response.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(response.status).toBe(401);
|
||||||
|
expect(errorData.message).toBe('Incorrect email or password.');
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fail to log in with a non-existent email', async () => {
|
||||||
|
const response = await apiClient.loginUser('no-one-here@example.com', TEST_PASSWORD, false);
|
||||||
|
const errorData = await response.json();
|
||||||
|
|
||||||
|
expect(response.status).toBe(401);
|
||||||
|
expect(errorData.message).toBe('Incorrect email or password.');
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should be able to access a protected route after logging in', async () => {
|
||||||
|
// Arrange: Log in to get a token
|
||||||
|
const loginResponse = await apiClient.loginUser(testUser.user.email, TEST_PASSWORD, false);
|
||||||
|
const loginData = await loginResponse.json();
|
||||||
|
const token = loginData.token;
|
||||||
|
|
||||||
|
expect(loginResponse.status).toBe(200);
|
||||||
|
expect(token).toBeDefined();
|
||||||
|
|
||||||
|
// Act: Use the token to access a protected route
|
||||||
|
const profileResponse = await apiClient.getAuthenticatedUserProfile({ tokenOverride: token });
|
||||||
|
const profileData = await profileResponse.json();
|
||||||
|
|
||||||
|
// Assert
|
||||||
|
expect(profileResponse.status).toBe(200);
|
||||||
|
expect(profileData).toBeDefined();
|
||||||
|
expect(profileData.user.user_id).toBe(testUser.user.user_id);
|
||||||
|
expect(profileData.user.email).toBe(testUser.user.email);
|
||||||
|
expect(profileData.role).toBe('user');
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should allow an authenticated user to update their profile', async () => {
|
||||||
|
// Arrange: Log in to get a token
|
||||||
|
const loginResponse = await apiClient.loginUser(testUser.user.email, TEST_PASSWORD, false);
|
||||||
|
const loginData = await loginResponse.json();
|
||||||
|
const token = loginData.token;
|
||||||
|
expect(loginResponse.status).toBe(200);
|
||||||
|
|
||||||
|
const profileUpdates = {
|
||||||
|
full_name: 'E2E Updated Name',
|
||||||
|
avatar_url: 'https://www.projectium.com/updated-avatar.png',
|
||||||
|
};
|
||||||
|
|
||||||
|
// Act: Call the update endpoint
|
||||||
|
const updateResponse = await apiClient.updateUserProfile(profileUpdates, { tokenOverride: token });
|
||||||
|
const updatedProfileData = await updateResponse.json();
|
||||||
|
|
||||||
|
// Assert: Check the response from the update call
|
||||||
|
expect(updateResponse.status).toBe(200);
|
||||||
|
expect(updatedProfileData.full_name).toBe(profileUpdates.full_name);
|
||||||
|
expect(updatedProfileData.avatar_url).toBe(profileUpdates.avatar_url);
|
||||||
|
|
||||||
|
// Act 2: Fetch the profile again to verify persistence
|
||||||
|
const verifyResponse = await apiClient.getAuthenticatedUserProfile({ tokenOverride: token });
|
||||||
|
const verifiedProfileData = await verifyResponse.json();
|
||||||
|
|
||||||
|
// Assert 2: Check the fetched data
|
||||||
|
expect(verifiedProfileData.full_name).toBe(profileUpdates.full_name);
|
||||||
|
expect(verifiedProfileData.avatar_url).toBe(profileUpdates.avatar_url);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('Forgot/Reset Password Flow', () => {
|
||||||
|
it('should allow a user to reset their password and log in with the new one', async () => {
|
||||||
|
// Arrange: Create a user to reset the password for
|
||||||
|
const email = `e2e-reset-pass-${Date.now()}@example.com`;
|
||||||
|
const registerResponse = await apiClient.registerUser(email, TEST_PASSWORD, 'Reset Pass User');
|
||||||
|
const registerData = await registerResponse.json();
|
||||||
|
expect(registerResponse.status).toBe(201);
|
||||||
|
createdUserIds.push(registerData.userprofile.user.user_id);
|
||||||
|
|
||||||
|
// Act 1: Request a password reset.
|
||||||
|
// The test environment returns the token directly in the response for E2E testing.
|
||||||
|
const forgotResponse = await apiClient.requestPasswordReset(email);
|
||||||
|
const forgotData = await forgotResponse.json();
|
||||||
|
const resetToken = forgotData.token;
|
||||||
|
|
||||||
|
// Assert 1: Check that we received a token.
|
||||||
|
expect(forgotResponse.status).toBe(200);
|
||||||
|
expect(resetToken).toBeDefined();
|
||||||
|
expect(resetToken).toBeTypeOf('string');
|
||||||
|
|
||||||
|
// Act 2: Use the token to set a new password.
|
||||||
|
const newPassword = 'my-new-e2e-password-!@#$';
|
||||||
|
const resetResponse = await apiClient.resetPassword(resetToken, newPassword);
|
||||||
|
const resetData = await resetResponse.json();
|
||||||
|
|
||||||
|
// Assert 2: Check for a successful password reset message.
|
||||||
|
expect(resetResponse.status).toBe(200);
|
||||||
|
expect(resetData.message).toBe('Password has been reset successfully.');
|
||||||
|
|
||||||
|
// Act 3 & Assert 3 (Verification): Log in with the NEW password to confirm the change.
|
||||||
|
const loginResponse = await apiClient.loginUser(email, newPassword, false);
|
||||||
|
const loginData = await loginResponse.json();
|
||||||
|
|
||||||
|
expect(loginResponse.status).toBe(200);
|
||||||
|
expect(loginData.userprofile).toBeDefined();
|
||||||
|
expect(loginData.userprofile.user.email).toBe(email);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should return a generic success message for a non-existent email to prevent enumeration', async () => {
|
||||||
|
const nonExistentEmail = `non-existent-e2e-${Date.now()}@example.com`;
|
||||||
|
const response = await apiClient.requestPasswordReset(nonExistentEmail);
|
||||||
|
const data = await response.json();
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(data.message).toBe('If an account with that email exists, a password reset link has been sent.');
|
||||||
|
expect(data.token).toBeUndefined();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
@@ -164,8 +164,10 @@ describe('Admin API Routes Integration Tests', () => {
|
|||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
const flyerRes = await getPool().query(
|
const flyerRes = await getPool().query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||||
VALUES ($1, 'admin-test.jpg', 'http://test.com/img.jpg', 1, $2) RETURNING flyer_id`,
|
VALUES ($1, 'admin-test.jpg', 'https://example.com/flyer-images/asdmin-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||||
[testStoreId, `checksum-${Date.now()}-${Math.random()}`],
|
// The checksum must be a unique 64-character string to satisfy the DB constraint.
|
||||||
|
// We generate a dynamic string and pad it to 64 characters.
|
||||||
|
[testStoreId, `checksum-${Date.now()}-${Math.random()}`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
const flyerId = flyerRes.rows[0].flyer_id;
|
const flyerId = flyerRes.rows[0].flyer_id;
|
||||||
|
|
||||||
|
|||||||
@@ -21,16 +21,18 @@ const request = supertest(app);
|
|||||||
describe('Authentication API Integration', () => {
|
describe('Authentication API Integration', () => {
|
||||||
let testUserEmail: string;
|
let testUserEmail: string;
|
||||||
let testUser: UserProfile;
|
let testUser: UserProfile;
|
||||||
|
const createdUserIds: string[] = [];
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
// Use a unique email for this test suite to prevent collisions with other tests.
|
// Use a unique email for this test suite to prevent collisions with other tests.
|
||||||
const email = `auth-integration-test-${Date.now()}@example.com`;
|
const email = `auth-integration-test-${Date.now()}@example.com`;
|
||||||
({ user: testUser } = await createAndLoginUser({ email, fullName: 'Auth Test User', request }));
|
({ user: testUser } = await createAndLoginUser({ email, fullName: 'Auth Test User', request }));
|
||||||
testUserEmail = testUser.user.email;
|
testUserEmail = testUser.user.email;
|
||||||
|
createdUserIds.push(testUser.user.user_id);
|
||||||
});
|
});
|
||||||
|
|
||||||
afterAll(async () => {
|
afterAll(async () => {
|
||||||
await cleanupDb({ userIds: testUser ? [testUser.user.user_id] : [] });
|
await cleanupDb({ userIds: createdUserIds });
|
||||||
});
|
});
|
||||||
|
|
||||||
// This test migrates the logic from the old DevTestRunner.tsx component.
|
// This test migrates the logic from the old DevTestRunner.tsx component.
|
||||||
@@ -83,6 +85,38 @@ describe('Authentication API Integration', () => {
|
|||||||
expect(errorData.message).toBe('Incorrect email or password.');
|
expect(errorData.message).toBe('Incorrect email or password.');
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('should allow registration with an empty string for avatar_url and save it as null', async () => {
|
||||||
|
// Arrange: Define user data with an empty avatar_url.
|
||||||
|
const email = `empty-avatar-user-${Date.now()}@example.com`;
|
||||||
|
const userData = {
|
||||||
|
email,
|
||||||
|
password: TEST_PASSWORD,
|
||||||
|
full_name: 'Empty Avatar',
|
||||||
|
avatar_url: '',
|
||||||
|
};
|
||||||
|
|
||||||
|
// Act: Register the new user.
|
||||||
|
const registerResponse = await request.post('/api/auth/register').send(userData);
|
||||||
|
|
||||||
|
// Assert 1: Check that the registration was successful and the returned profile is correct.
|
||||||
|
expect(registerResponse.status).toBe(201);
|
||||||
|
const registeredProfile = registerResponse.body.userprofile;
|
||||||
|
const registeredToken = registerResponse.body.token;
|
||||||
|
expect(registeredProfile.user.email).toBe(email);
|
||||||
|
expect(registeredProfile.avatar_url).toBeNull(); // The API should return null for the avatar_url.
|
||||||
|
|
||||||
|
// Add the newly created user's ID to the array for cleanup in afterAll.
|
||||||
|
createdUserIds.push(registeredProfile.user.user_id);
|
||||||
|
|
||||||
|
// Assert 2 (Verification): Fetch the profile using the new token to confirm the value in the DB is null.
|
||||||
|
const profileResponse = await request
|
||||||
|
.get('/api/users/profile')
|
||||||
|
.set('Authorization', `Bearer ${registeredToken}`);
|
||||||
|
|
||||||
|
expect(profileResponse.status).toBe(200);
|
||||||
|
expect(profileResponse.body.avatar_url).toBeNull();
|
||||||
|
});
|
||||||
|
|
||||||
it('should successfully refresh an access token using a refresh token cookie', async () => {
|
it('should successfully refresh an access token using a refresh token cookie', async () => {
|
||||||
// Arrange: Log in to get a fresh, valid refresh token cookie for this specific test.
|
// Arrange: Log in to get a fresh, valid refresh token cookie for this specific test.
|
||||||
// This ensures the test is self-contained and not affected by other tests.
|
// This ensures the test is self-contained and not affected by other tests.
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
// src/tests/integration/flyer-processing.integration.test.ts
|
// src/tests/integration/flyer-processing.integration.test.ts
|
||||||
import { describe, it, expect, beforeAll, afterAll } from 'vitest';
|
import { describe, it, expect, beforeAll, afterAll, vi } from 'vitest';
|
||||||
import supertest from 'supertest';
|
import supertest from 'supertest';
|
||||||
import app from '../../../server';
|
import app from '../../../server';
|
||||||
import fs from 'node:fs/promises';
|
import fs from 'node:fs/promises';
|
||||||
@@ -8,7 +8,7 @@ import * as db from '../../services/db/index.db';
|
|||||||
import { getPool } from '../../services/db/connection.db';
|
import { getPool } from '../../services/db/connection.db';
|
||||||
import { generateFileChecksum } from '../../utils/checksum';
|
import { generateFileChecksum } from '../../utils/checksum';
|
||||||
import { logger } from '../../services/logger.server';
|
import { logger } from '../../services/logger.server';
|
||||||
import type { UserProfile } from '../../types';
|
import type { UserProfile, ExtractedFlyerItem } from '../../types';
|
||||||
import { createAndLoginUser } from '../utils/testHelpers';
|
import { createAndLoginUser } from '../utils/testHelpers';
|
||||||
import { cleanupDb } from '../utils/cleanup';
|
import { cleanupDb } from '../utils/cleanup';
|
||||||
import { cleanupFiles } from '../utils/cleanupFiles';
|
import { cleanupFiles } from '../utils/cleanupFiles';
|
||||||
@@ -16,12 +16,22 @@ import piexif from 'piexifjs';
|
|||||||
import exifParser from 'exif-parser';
|
import exifParser from 'exif-parser';
|
||||||
import sharp from 'sharp';
|
import sharp from 'sharp';
|
||||||
|
|
||||||
|
// Mock the AI service to prevent actual API calls during integration tests.
|
||||||
|
vi.mock('../../services/aiService.server', () => ({
|
||||||
|
aiService: {
|
||||||
|
extractCoreDataFromFlyerImage: vi.fn(),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
|
|
||||||
const request = supertest(app);
|
const request = supertest(app);
|
||||||
|
|
||||||
|
// Import the mocked service to control its behavior in tests.
|
||||||
|
import { aiService } from '../../services/aiService.server';
|
||||||
|
|
||||||
describe('Flyer Processing Background Job Integration Test', () => {
|
describe('Flyer Processing Background Job Integration Test', () => {
|
||||||
const createdUserIds: string[] = [];
|
const createdUserIds: string[] = [];
|
||||||
const createdFlyerIds: number[] = [];
|
const createdFlyerIds: number[] = [];
|
||||||
@@ -29,6 +39,23 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
// This setup is now simpler as the worker handles fetching master items.
|
// This setup is now simpler as the worker handles fetching master items.
|
||||||
|
// Setup default mock response for AI service
|
||||||
|
const mockItems: ExtractedFlyerItem[] = [
|
||||||
|
{
|
||||||
|
item: 'Mocked Integration Item',
|
||||||
|
price_display: '$1.99',
|
||||||
|
price_in_cents: 199,
|
||||||
|
quantity: 'each',
|
||||||
|
category_name: 'Mock Category',
|
||||||
|
},
|
||||||
|
];
|
||||||
|
vi.mocked(aiService.extractCoreDataFromFlyerImage).mockResolvedValue({
|
||||||
|
store_name: 'Mock Store',
|
||||||
|
valid_from: null,
|
||||||
|
valid_to: null,
|
||||||
|
store_address: null,
|
||||||
|
items: mockItems,
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
afterAll(async () => {
|
afterAll(async () => {
|
||||||
@@ -80,8 +107,9 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
|
|
||||||
// Act 2: Poll for the job status until it completes.
|
// Act 2: Poll for the job status until it completes.
|
||||||
let jobStatus;
|
let jobStatus;
|
||||||
const maxRetries = 30; // Poll for up to 90 seconds (30 * 3s)
|
const maxRetries = 60; // Poll for up to 180 seconds (60 * 3s)
|
||||||
for (let i = 0; i < maxRetries; i++) {
|
for (let i = 0; i < maxRetries; i++) {
|
||||||
|
console.log(`Polling attempt ${i + 1}...`);
|
||||||
await new Promise((resolve) => setTimeout(resolve, 3000)); // Wait 3 seconds between polls
|
await new Promise((resolve) => setTimeout(resolve, 3000)); // Wait 3 seconds between polls
|
||||||
const statusReq = request.get(`/api/ai/jobs/${jobId}/status`);
|
const statusReq = request.get(`/api/ai/jobs/${jobId}/status`);
|
||||||
if (token) {
|
if (token) {
|
||||||
@@ -89,6 +117,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
}
|
}
|
||||||
const statusResponse = await statusReq;
|
const statusResponse = await statusReq;
|
||||||
jobStatus = statusResponse.body;
|
jobStatus = statusResponse.body;
|
||||||
|
console.log(`Job status: ${JSON.stringify(jobStatus)}`);
|
||||||
if (jobStatus.state === 'completed' || jobStatus.state === 'failed') {
|
if (jobStatus.state === 'completed' || jobStatus.state === 'failed') {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@@ -147,12 +176,12 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
|
|
||||||
// Act & Assert
|
// Act & Assert
|
||||||
await runBackgroundProcessingTest(authUser, token);
|
await runBackgroundProcessingTest(authUser, token);
|
||||||
}, 120000); // Increase timeout to 120 seconds for this long-running test
|
}, 240000); // Increase timeout to 240 seconds for this long-running test
|
||||||
|
|
||||||
it('should successfully process a flyer for an ANONYMOUS user via the background queue', async () => {
|
it('should successfully process a flyer for an ANONYMOUS user via the background queue', async () => {
|
||||||
// Act & Assert: Call the test helper without a user or token.
|
// Act & Assert: Call the test helper without a user or token.
|
||||||
await runBackgroundProcessingTest();
|
await runBackgroundProcessingTest();
|
||||||
}, 120000); // Increase timeout to 120 seconds for this long-running test
|
}, 240000); // Increase timeout to 240 seconds for this long-running test
|
||||||
|
|
||||||
it(
|
it(
|
||||||
'should strip EXIF data from uploaded JPEG images during processing',
|
'should strip EXIF data from uploaded JPEG images during processing',
|
||||||
@@ -217,7 +246,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
console.error('[DEBUG] EXIF test job failed:', jobStatus.failedReason);
|
console.error('[DEBUG] EXIF test job failed:', jobStatus.failedReason);
|
||||||
}
|
}
|
||||||
expect(jobStatus?.state).toBe('completed');
|
expect(jobStatus?.state).toBe('completed');
|
||||||
const flyerId = jobStatus?.returnValue?.flyerId;
|
const flyerId = jobStatus?.data?.flyerId;
|
||||||
expect(flyerId).toBeTypeOf('number');
|
expect(flyerId).toBeTypeOf('number');
|
||||||
createdFlyerIds.push(flyerId);
|
createdFlyerIds.push(flyerId);
|
||||||
|
|
||||||
@@ -236,7 +265,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
expect(exifResult.tags).toEqual({});
|
expect(exifResult.tags).toEqual({});
|
||||||
expect(exifResult.tags.Software).toBeUndefined();
|
expect(exifResult.tags.Software).toBeUndefined();
|
||||||
},
|
},
|
||||||
120000,
|
240000,
|
||||||
);
|
);
|
||||||
|
|
||||||
it(
|
it(
|
||||||
@@ -303,7 +332,7 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
console.error('[DEBUG] PNG metadata test job failed:', jobStatus.failedReason);
|
console.error('[DEBUG] PNG metadata test job failed:', jobStatus.failedReason);
|
||||||
}
|
}
|
||||||
expect(jobStatus?.state).toBe('completed');
|
expect(jobStatus?.state).toBe('completed');
|
||||||
const flyerId = jobStatus?.returnValue?.flyerId;
|
const flyerId = jobStatus?.data?.flyerId;
|
||||||
expect(flyerId).toBeTypeOf('number');
|
expect(flyerId).toBeTypeOf('number');
|
||||||
createdFlyerIds.push(flyerId);
|
createdFlyerIds.push(flyerId);
|
||||||
|
|
||||||
@@ -320,6 +349,6 @@ describe('Flyer Processing Background Job Integration Test', () => {
|
|||||||
// The `exif` property should be undefined after the fix.
|
// The `exif` property should be undefined after the fix.
|
||||||
expect(savedImageMetadata.exif).toBeUndefined();
|
expect(savedImageMetadata.exif).toBeUndefined();
|
||||||
},
|
},
|
||||||
120000,
|
240000,
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -25,8 +25,8 @@ describe('Public Flyer API Routes Integration Tests', () => {
|
|||||||
|
|
||||||
const flyerRes = await getPool().query(
|
const flyerRes = await getPool().query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||||
VALUES ($1, 'integration-test.jpg', 'http://test.com/img.jpg', 1, $2) RETURNING flyer_id`,
|
VALUES ($1, 'integration-test.jpg', 'https://example.com/flyer-images/integration-test.jpg', 1, $2) RETURNING flyer_id`,
|
||||||
[storeId, `checksum-${Date.now()}`],
|
[storeId, `${Date.now().toString(16)}`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
createdFlyerId = flyerRes.rows[0].flyer_id;
|
createdFlyerId = flyerRes.rows[0].flyer_id;
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
// src/tests/integration/gamification.integration.test.ts
|
// src/tests/integration/gamification.integration.test.ts
|
||||||
import { describe, it, expect, beforeAll, afterAll } from 'vitest';
|
import { describe, it, expect, beforeAll, afterAll, vi } from 'vitest';
|
||||||
import supertest from 'supertest';
|
import supertest from 'supertest';
|
||||||
import app from '../../../server';
|
import app from '../../../server';
|
||||||
import path from 'path';
|
import path from 'path';
|
||||||
@@ -9,15 +9,31 @@ import { generateFileChecksum } from '../../utils/checksum';
|
|||||||
import * as db from '../../services/db/index.db';
|
import * as db from '../../services/db/index.db';
|
||||||
import { cleanupDb } from '../utils/cleanup';
|
import { cleanupDb } from '../utils/cleanup';
|
||||||
import { logger } from '../../services/logger.server';
|
import { logger } from '../../services/logger.server';
|
||||||
import type { UserProfile, UserAchievement, LeaderboardUser, Achievement } from '../../types';
|
import type {
|
||||||
|
UserProfile,
|
||||||
|
UserAchievement,
|
||||||
|
LeaderboardUser,
|
||||||
|
Achievement,
|
||||||
|
ExtractedFlyerItem,
|
||||||
|
} from '../../types';
|
||||||
import { cleanupFiles } from '../utils/cleanupFiles';
|
import { cleanupFiles } from '../utils/cleanupFiles';
|
||||||
|
|
||||||
|
// Mock the AI service to prevent actual API calls during integration tests.
|
||||||
|
vi.mock('../../services/aiService.server', () => ({
|
||||||
|
aiService: {
|
||||||
|
extractCoreDataFromFlyerImage: vi.fn(),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
|
|
||||||
const request = supertest(app);
|
const request = supertest(app);
|
||||||
|
|
||||||
|
// Import the mocked service to control its behavior in tests.
|
||||||
|
import { aiService } from '../../services/aiService.server';
|
||||||
|
|
||||||
describe('Gamification Flow Integration Test', () => {
|
describe('Gamification Flow Integration Test', () => {
|
||||||
let testUser: UserProfile;
|
let testUser: UserProfile;
|
||||||
let authToken: string;
|
let authToken: string;
|
||||||
@@ -44,6 +60,28 @@ describe('Gamification Flow Integration Test', () => {
|
|||||||
it(
|
it(
|
||||||
'should award the "First Upload" achievement after a user successfully uploads and processes their first flyer',
|
'should award the "First Upload" achievement after a user successfully uploads and processes their first flyer',
|
||||||
async () => {
|
async () => {
|
||||||
|
// --- Arrange: Mock AI Service Response ---
|
||||||
|
// This is crucial for making the integration test reliable. We don't want to
|
||||||
|
// depend on the external Gemini API, which has quotas and can be slow.
|
||||||
|
// By mocking this, we test our application's internal flow:
|
||||||
|
// API -> Queue -> Worker -> DB -> Gamification Logic
|
||||||
|
const mockExtractedItems: ExtractedFlyerItem[] = [
|
||||||
|
{
|
||||||
|
item: 'Integration Test Milk',
|
||||||
|
price_display: '$4.99',
|
||||||
|
price_in_cents: 499,
|
||||||
|
quantity: '2L',
|
||||||
|
category_name: 'Dairy',
|
||||||
|
},
|
||||||
|
];
|
||||||
|
vi.mocked(aiService.extractCoreDataFromFlyerImage).mockResolvedValue({
|
||||||
|
store_name: 'Gamification Test Store',
|
||||||
|
valid_from: null,
|
||||||
|
valid_to: null,
|
||||||
|
store_address: null,
|
||||||
|
items: mockExtractedItems,
|
||||||
|
});
|
||||||
|
|
||||||
// --- Arrange: Prepare a unique flyer file for upload ---
|
// --- Arrange: Prepare a unique flyer file for upload ---
|
||||||
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
const imagePath = path.resolve(__dirname, '../assets/test-flyer-image.jpg');
|
||||||
const imageBuffer = await fs.readFile(imagePath);
|
const imageBuffer = await fs.readFile(imagePath);
|
||||||
@@ -81,12 +119,17 @@ describe('Gamification Flow Integration Test', () => {
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if (!jobStatus) {
|
||||||
|
console.error('[DEBUG] Gamification test job timed out: No job status received.');
|
||||||
|
throw new Error('Gamification test job timed out: No job status received.');
|
||||||
|
}
|
||||||
|
|
||||||
// --- Assert 1: Verify the job completed successfully ---
|
// --- Assert 1: Verify the job completed successfully ---
|
||||||
if (jobStatus?.state === 'failed') {
|
if (jobStatus?.state === 'failed') {
|
||||||
console.error('[DEBUG] Gamification test job failed:', jobStatus.failedReason);
|
console.error('[DEBUG] Gamification test job failed:', jobStatus.failedReason);
|
||||||
}
|
}
|
||||||
expect(jobStatus?.state).toBe('completed');
|
expect(jobStatus?.state).toBe('completed');
|
||||||
|
|
||||||
const flyerId = jobStatus?.returnValue?.flyerId;
|
const flyerId = jobStatus?.returnValue?.flyerId;
|
||||||
expect(flyerId).toBeTypeOf('number');
|
expect(flyerId).toBeTypeOf('number');
|
||||||
createdFlyerIds.push(flyerId); // Track for cleanup
|
createdFlyerIds.push(flyerId); // Track for cleanup
|
||||||
|
|||||||
@@ -35,22 +35,22 @@ describe('Price History API Integration Test (/api/price-history)', () => {
|
|||||||
// 3. Create two flyers with different dates
|
// 3. Create two flyers with different dates
|
||||||
const flyerRes1 = await pool.query(
|
const flyerRes1 = await pool.query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||||
VALUES ($1, 'price-test-1.jpg', 'http://test.com/price-1.jpg', 1, $2, '2025-01-01') RETURNING flyer_id`,
|
VALUES ($1, 'price-test-1.jpg', 'https://example.com/flyer-images/price-test-1.jpg', 1, $2, '2025-01-01') RETURNING flyer_id`,
|
||||||
[storeId, `checksum-price-1-${Date.now()}`],
|
[storeId, `${Date.now().toString(16)}1`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
flyerId1 = flyerRes1.rows[0].flyer_id;
|
flyerId1 = flyerRes1.rows[0].flyer_id;
|
||||||
|
|
||||||
const flyerRes2 = await pool.query(
|
const flyerRes2 = await pool.query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||||
VALUES ($1, 'price-test-2.jpg', 'http://test.com/price-2.jpg', 1, $2, '2025-01-08') RETURNING flyer_id`,
|
VALUES ($1, 'price-test-2.jpg', 'https://example.com/flyer-images/price-test-2.jpg', 1, $2, '2025-01-08') RETURNING flyer_id`,
|
||||||
[storeId, `checksum-price-2-${Date.now()}`],
|
[storeId, `${Date.now().toString(16)}2`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
flyerId2 = flyerRes2.rows[0].flyer_id; // This was a duplicate, fixed.
|
flyerId2 = flyerRes2.rows[0].flyer_id; // This was a duplicate, fixed.
|
||||||
|
|
||||||
const flyerRes3 = await pool.query(
|
const flyerRes3 = await pool.query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum, valid_from)
|
||||||
VALUES ($1, 'price-test-3.jpg', 'http://test.com/price-3.jpg', 1, $2, '2025-01-15') RETURNING flyer_id`,
|
VALUES ($1, 'price-test-3.jpg', 'https://example.com/flyer-images/price-test-3.jpg', 1, $2, '2025-01-15') RETURNING flyer_id`,
|
||||||
[storeId, `checksum-price-3-${Date.now()}`],
|
[storeId, `${Date.now().toString(16)}3`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
flyerId3 = flyerRes3.rows[0].flyer_id;
|
flyerId3 = flyerRes3.rows[0].flyer_id;
|
||||||
|
|
||||||
|
|||||||
@@ -38,6 +38,7 @@ describe('Public API Routes Integration Tests', () => {
|
|||||||
email: userEmail,
|
email: userEmail,
|
||||||
password: 'a-Very-Strong-Password-123!',
|
password: 'a-Very-Strong-Password-123!',
|
||||||
fullName: 'Public Routes Test User',
|
fullName: 'Public Routes Test User',
|
||||||
|
request,
|
||||||
});
|
});
|
||||||
testUser = createdUser;
|
testUser = createdUser;
|
||||||
|
|
||||||
@@ -77,8 +78,8 @@ describe('Public API Routes Integration Tests', () => {
|
|||||||
testStoreId = storeRes.rows[0].store_id;
|
testStoreId = storeRes.rows[0].store_id;
|
||||||
const flyerRes = await pool.query(
|
const flyerRes = await pool.query(
|
||||||
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
`INSERT INTO public.flyers (store_id, file_name, image_url, item_count, checksum)
|
||||||
VALUES ($1, 'public-routes-test.jpg', 'http://test.com/public-routes.jpg', 1, $2) RETURNING *`,
|
VALUES ($1, 'public-routes-test.jpg', 'https://example.com/flyer-images/public-routes-test.jpg', 1, $2) RETURNING *`,
|
||||||
[testStoreId, `checksum-public-routes-${Date.now()}`],
|
[testStoreId, `${Date.now().toString(16)}`.padEnd(64, '0')],
|
||||||
);
|
);
|
||||||
testFlyer = flyerRes.rows[0];
|
testFlyer = flyerRes.rows[0];
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
// src/tests/integration/recipe.integration.test.ts
|
// src/tests/integration/recipe.integration.test.ts
|
||||||
import { describe, it, expect, beforeAll, afterAll } from 'vitest';
|
import { describe, it, expect, beforeAll, afterAll, vi } from 'vitest';
|
||||||
import supertest from 'supertest';
|
import supertest from 'supertest';
|
||||||
import app from '../../../server';
|
import app from '../../../server';
|
||||||
import { createAndLoginUser } from '../utils/testHelpers';
|
import { createAndLoginUser } from '../utils/testHelpers';
|
||||||
@@ -7,6 +7,14 @@ import { cleanupDb } from '../utils/cleanup';
|
|||||||
import type { UserProfile, Recipe, RecipeComment } from '../../types';
|
import type { UserProfile, Recipe, RecipeComment } from '../../types';
|
||||||
import { getPool } from '../../services/db/connection.db';
|
import { getPool } from '../../services/db/connection.db';
|
||||||
|
|
||||||
|
// Mock the AI service
|
||||||
|
vi.mock('../../services/aiService.server', () => ({
|
||||||
|
aiService: {
|
||||||
|
generateRecipeSuggestion: vi.fn(),
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
import { aiService } from '../../services/aiService.server';
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @vitest-environment node
|
* @vitest-environment node
|
||||||
*/
|
*/
|
||||||
@@ -124,4 +132,24 @@ describe('Recipe API Routes Integration Tests', () => {
|
|||||||
it.todo('should prevent a user from deleting another user\'s recipe');
|
it.todo('should prevent a user from deleting another user\'s recipe');
|
||||||
it.todo('should allow an authenticated user to post a comment on a recipe');
|
it.todo('should allow an authenticated user to post a comment on a recipe');
|
||||||
it.todo('should allow an authenticated user to fork a recipe');
|
it.todo('should allow an authenticated user to fork a recipe');
|
||||||
|
|
||||||
|
describe('POST /api/recipes/suggest', () => {
|
||||||
|
it('should return a recipe suggestion based on ingredients', async () => {
|
||||||
|
const ingredients = ['chicken', 'rice', 'broccoli'];
|
||||||
|
const mockSuggestion = 'Chicken and Broccoli Stir-fry with Rice';
|
||||||
|
vi.mocked(aiService.generateRecipeSuggestion).mockResolvedValue(mockSuggestion);
|
||||||
|
|
||||||
|
const response = await request
|
||||||
|
.post('/api/recipes/suggest')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send({ ingredients });
|
||||||
|
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(response.body).toEqual({ suggestion: mockSuggestion });
|
||||||
|
expect(aiService.generateRecipeSuggestion).toHaveBeenCalledWith(
|
||||||
|
ingredients,
|
||||||
|
expect.anything(),
|
||||||
|
);
|
||||||
|
});
|
||||||
|
});
|
||||||
});
|
});
|
||||||
@@ -75,6 +75,32 @@ describe('User API Routes Integration Tests', () => {
|
|||||||
expect(refetchedProfile.full_name).toBe('Updated Test User');
|
expect(refetchedProfile.full_name).toBe('Updated Test User');
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('should allow updating the profile with an empty string for avatar_url', async () => {
|
||||||
|
// Arrange: Define the profile updates.
|
||||||
|
const profileUpdates = {
|
||||||
|
full_name: 'Empty Avatar User',
|
||||||
|
avatar_url: '',
|
||||||
|
};
|
||||||
|
|
||||||
|
// Act: Call the update endpoint with the new data and the auth token.
|
||||||
|
const response = await request
|
||||||
|
.put('/api/users/profile')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send(profileUpdates);
|
||||||
|
const updatedProfile = response.body;
|
||||||
|
|
||||||
|
// Assert: Check that the returned profile reflects the changes.
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(updatedProfile.full_name).toBe('Empty Avatar User');
|
||||||
|
expect(updatedProfile.avatar_url).toBeNull();
|
||||||
|
|
||||||
|
// Also, fetch the profile again to ensure the change was persisted in the database as NULL.
|
||||||
|
const refetchResponse = await request
|
||||||
|
.get('/api/users/profile')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
expect(refetchResponse.body.avatar_url).toBeNull();
|
||||||
|
});
|
||||||
|
|
||||||
it('should update user preferences via PUT /api/users/profile/preferences', async () => {
|
it('should update user preferences via PUT /api/users/profile/preferences', async () => {
|
||||||
// Arrange: Define the preference updates.
|
// Arrange: Define the preference updates.
|
||||||
const preferenceUpdates = {
|
const preferenceUpdates = {
|
||||||
|
|||||||
@@ -12,11 +12,11 @@ import { cleanupDb } from '../utils/cleanup';
|
|||||||
|
|
||||||
const request = supertest(app);
|
const request = supertest(app);
|
||||||
|
|
||||||
let authToken = '';
|
|
||||||
let createdListId: number;
|
|
||||||
let testUser: UserProfile;
|
|
||||||
|
|
||||||
describe('User Routes Integration Tests (/api/users)', () => {
|
describe('User Routes Integration Tests (/api/users)', () => {
|
||||||
|
let authToken = '';
|
||||||
|
let testUser: UserProfile;
|
||||||
|
const createdUserIds: string[] = [];
|
||||||
|
|
||||||
// Authenticate once before all tests in this suite to get a JWT.
|
// Authenticate once before all tests in this suite to get a JWT.
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
// Use the helper to create and log in a user in one step.
|
// Use the helper to create and log in a user in one step.
|
||||||
@@ -26,10 +26,11 @@ describe('User Routes Integration Tests (/api/users)', () => {
|
|||||||
});
|
});
|
||||||
testUser = user;
|
testUser = user;
|
||||||
authToken = token;
|
authToken = token;
|
||||||
|
createdUserIds.push(user.user.user_id);
|
||||||
});
|
});
|
||||||
|
|
||||||
afterAll(async () => {
|
afterAll(async () => {
|
||||||
await cleanupDb({ userIds: testUser ? [testUser.user.user_id] : [] });
|
await cleanupDb({ userIds: createdUserIds });
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('GET /api/users/profile', () => {
|
describe('GET /api/users/profile', () => {
|
||||||
@@ -70,54 +71,6 @@ describe('User Routes Integration Tests (/api/users)', () => {
|
|||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('Shopping List CRUD', () => {
|
|
||||||
it('POST /api/users/shopping-lists should create a new shopping list', async () => {
|
|
||||||
const listName = `My Integration Test List ${Date.now()}`;
|
|
||||||
const response = await request
|
|
||||||
.post('/api/users/shopping-lists')
|
|
||||||
.set('Authorization', `Bearer ${authToken}`)
|
|
||||||
.send({ name: listName });
|
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
|
||||||
expect(response.body.name).toBe(listName);
|
|
||||||
expect(response.body.shopping_list_id).toBeDefined();
|
|
||||||
createdListId = response.body.shopping_list_id; // Save for the next test
|
|
||||||
});
|
|
||||||
|
|
||||||
it('GET /api/users/shopping-lists should retrieve the created shopping list', async () => {
|
|
||||||
const response = await request
|
|
||||||
.get('/api/users/shopping-lists')
|
|
||||||
.set('Authorization', `Bearer ${authToken}`);
|
|
||||||
|
|
||||||
expect(response.status).toBe(200);
|
|
||||||
expect(Array.isArray(response.body)).toBe(true);
|
|
||||||
const foundList = response.body.find(
|
|
||||||
(list: { shopping_list_id: number }) => list.shopping_list_id === createdListId,
|
|
||||||
);
|
|
||||||
expect(foundList).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('DELETE /api/users/shopping-lists/:listId should delete the shopping list', async () => {
|
|
||||||
expect(createdListId).toBeDefined(); // Ensure the previous test ran and set the ID
|
|
||||||
|
|
||||||
const response = await request
|
|
||||||
.delete(`/api/users/shopping-lists/${createdListId}`)
|
|
||||||
.set('Authorization', `Bearer ${authToken}`);
|
|
||||||
|
|
||||||
expect(response.status).toBe(204);
|
|
||||||
|
|
||||||
// Verify deletion
|
|
||||||
const verifyResponse = await request
|
|
||||||
.get('/api/users/shopping-lists')
|
|
||||||
.set('Authorization', `Bearer ${authToken}`);
|
|
||||||
|
|
||||||
const foundList = verifyResponse.body.find(
|
|
||||||
(list: { shopping_list_id: number }) => list.shopping_list_id === createdListId,
|
|
||||||
);
|
|
||||||
expect(foundList).toBeUndefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('PUT /api/users/profile/preferences', () => {
|
describe('PUT /api/users/profile/preferences', () => {
|
||||||
it('should update user preferences', async () => {
|
it('should update user preferences', async () => {
|
||||||
const preferences = { darkMode: true, unitSystem: 'metric' };
|
const preferences = { darkMode: true, unitSystem: 'metric' };
|
||||||
@@ -138,4 +91,164 @@ describe('User Routes Integration Tests (/api/users)', () => {
|
|||||||
expect(verifyResponse.body.preferences?.unitSystem).toBe('metric');
|
expect(verifyResponse.body.preferences?.unitSystem).toBe('metric');
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('Shopping Lists and Items', () => {
|
||||||
|
it('should create, retrieve, and delete a shopping list', async () => {
|
||||||
|
// 1. Create
|
||||||
|
const listName = `My Test List - ${Date.now()}`;
|
||||||
|
const createResponse = await request
|
||||||
|
.post('/api/users/shopping-lists')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send({ name: listName });
|
||||||
|
|
||||||
|
expect(createResponse.status).toBe(201);
|
||||||
|
expect(createResponse.body.name).toBe(listName);
|
||||||
|
const listId = createResponse.body.shopping_list_id;
|
||||||
|
expect(listId).toBeDefined();
|
||||||
|
|
||||||
|
// 2. Retrieve
|
||||||
|
const getResponse = await request
|
||||||
|
.get('/api/users/shopping-lists')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
|
||||||
|
expect(getResponse.status).toBe(200);
|
||||||
|
const foundList = getResponse.body.find(
|
||||||
|
(l: { shopping_list_id: number }) => l.shopping_list_id === listId,
|
||||||
|
);
|
||||||
|
expect(foundList).toBeDefined();
|
||||||
|
|
||||||
|
// 3. Delete
|
||||||
|
const deleteResponse = await request
|
||||||
|
.delete(`/api/users/shopping-lists/${listId}`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
expect(deleteResponse.status).toBe(204);
|
||||||
|
|
||||||
|
// 4. Verify Deletion
|
||||||
|
const verifyResponse = await request
|
||||||
|
.get('/api/users/shopping-lists')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
const notFoundList = verifyResponse.body.find(
|
||||||
|
(l: { shopping_list_id: number }) => l.shopping_list_id === listId,
|
||||||
|
);
|
||||||
|
expect(notFoundList).toBeUndefined();
|
||||||
|
});
|
||||||
|
|
||||||
|
it("should prevent a user from modifying another user's shopping list", async () => {
|
||||||
|
// Arrange: Create a shopping list owned by the primary testUser.
|
||||||
|
const listName = `Owner's List - ${Date.now()}`;
|
||||||
|
const createListResponse = await request
|
||||||
|
.post('/api/users/shopping-lists')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`) // Use owner's token
|
||||||
|
.send({ name: listName });
|
||||||
|
expect(createListResponse.status).toBe(201);
|
||||||
|
const listId = createListResponse.body.shopping_list_id;
|
||||||
|
|
||||||
|
// Arrange: Create a second, "malicious" user.
|
||||||
|
const maliciousEmail = `malicious-user-${Date.now()}@example.com`;
|
||||||
|
const { token: maliciousToken, user: maliciousUser } = await createAndLoginUser({
|
||||||
|
email: maliciousEmail,
|
||||||
|
fullName: 'Malicious User',
|
||||||
|
request,
|
||||||
|
});
|
||||||
|
createdUserIds.push(maliciousUser.user.user_id); // Ensure cleanup
|
||||||
|
|
||||||
|
// Act 1: Malicious user attempts to add an item to the owner's list.
|
||||||
|
const addItemResponse = await request
|
||||||
|
.post(`/api/users/shopping-lists/${listId}/items`)
|
||||||
|
.set('Authorization', `Bearer ${maliciousToken}`) // Use malicious user's token
|
||||||
|
.send({ customItemName: 'Malicious Item' });
|
||||||
|
|
||||||
|
// Assert 1: The request should fail. A 404 is expected because the list is not found for this user.
|
||||||
|
expect(addItemResponse.status).toBe(404);
|
||||||
|
expect(addItemResponse.body.message).toContain('Shopping list not found');
|
||||||
|
|
||||||
|
// Act 2: Malicious user attempts to delete the owner's list.
|
||||||
|
const deleteResponse = await request
|
||||||
|
.delete(`/api/users/shopping-lists/${listId}`)
|
||||||
|
.set('Authorization', `Bearer ${maliciousToken}`); // Use malicious user's token
|
||||||
|
|
||||||
|
// Assert 2: This should also fail with a 404.
|
||||||
|
expect(deleteResponse.status).toBe(404);
|
||||||
|
expect(deleteResponse.body.message).toContain('Shopping list not found');
|
||||||
|
|
||||||
|
// Act 3: Malicious user attempts to update an item on the owner's list.
|
||||||
|
// First, the owner adds an item.
|
||||||
|
const ownerAddItemResponse = await request
|
||||||
|
.post(`/api/users/shopping-lists/${listId}/items`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`) // Owner's token
|
||||||
|
.send({ customItemName: 'Legitimate Item' });
|
||||||
|
expect(ownerAddItemResponse.status).toBe(201);
|
||||||
|
const itemId = ownerAddItemResponse.body.shopping_list_item_id;
|
||||||
|
|
||||||
|
// Now, the malicious user tries to update it.
|
||||||
|
const updateItemResponse = await request
|
||||||
|
.put(`/api/users/shopping-lists/items/${itemId}`)
|
||||||
|
.set('Authorization', `Bearer ${maliciousToken}`) // Malicious token
|
||||||
|
.send({ is_purchased: true });
|
||||||
|
|
||||||
|
// Assert 3: This should also fail with a 404.
|
||||||
|
expect(updateItemResponse.status).toBe(404);
|
||||||
|
expect(updateItemResponse.body.message).toContain('Shopping list item not found');
|
||||||
|
|
||||||
|
// Cleanup the list created in this test
|
||||||
|
await request
|
||||||
|
.delete(`/api/users/shopping-lists/${listId}`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('Shopping List Item Management', () => {
|
||||||
|
let listId: number;
|
||||||
|
let itemId: number;
|
||||||
|
|
||||||
|
// Create a dedicated list for these item tests
|
||||||
|
beforeAll(async () => {
|
||||||
|
const response = await request
|
||||||
|
.post('/api/users/shopping-lists')
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send({ name: 'Item Test List' });
|
||||||
|
listId = response.body.shopping_list_id;
|
||||||
|
});
|
||||||
|
|
||||||
|
// Clean up the list after the item tests are done
|
||||||
|
afterAll(async () => {
|
||||||
|
if (listId) {
|
||||||
|
await request
|
||||||
|
.delete(`/api/users/shopping-lists/${listId}`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should add an item to a shopping list', async () => {
|
||||||
|
const response = await request
|
||||||
|
.post(`/api/users/shopping-lists/${listId}/items`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send({ customItemName: 'Test Item' });
|
||||||
|
|
||||||
|
expect(response.status).toBe(201);
|
||||||
|
expect(response.body.custom_item_name).toBe('Test Item');
|
||||||
|
expect(response.body.shopping_list_item_id).toBeDefined();
|
||||||
|
itemId = response.body.shopping_list_item_id; // Save for next tests
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should update an item in a shopping list', async () => {
|
||||||
|
const updates = { is_purchased: true, quantity: 5 };
|
||||||
|
const response = await request
|
||||||
|
.put(`/api/users/shopping-lists/items/${itemId}`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`)
|
||||||
|
.send(updates);
|
||||||
|
|
||||||
|
expect(response.status).toBe(200);
|
||||||
|
expect(response.body.is_purchased).toBe(true);
|
||||||
|
expect(response.body.quantity).toBe(5);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should delete an item from a shopping list', async () => {
|
||||||
|
const response = await request
|
||||||
|
.delete(`/api/users/shopping-lists/items/${itemId}`)
|
||||||
|
.set('Authorization', `Bearer ${authToken}`);
|
||||||
|
|
||||||
|
expect(response.status).toBe(204);
|
||||||
|
});
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -43,8 +43,6 @@ export const cleanupDb = async (ids: TestResourceIds) => {
|
|||||||
await pool.query('DELETE FROM public.user_watched_items WHERE user_id = ANY($1::uuid[])', [userIds]);
|
await pool.query('DELETE FROM public.user_watched_items WHERE user_id = ANY($1::uuid[])', [userIds]);
|
||||||
await pool.query('DELETE FROM public.user_achievements WHERE user_id = ANY($1::uuid[])', [userIds]);
|
await pool.query('DELETE FROM public.user_achievements WHERE user_id = ANY($1::uuid[])', [userIds]);
|
||||||
await pool.query('DELETE FROM public.activity_log WHERE user_id = ANY($1::uuid[])', [userIds]);
|
await pool.query('DELETE FROM public.activity_log WHERE user_id = ANY($1::uuid[])', [userIds]);
|
||||||
await pool.query('DELETE FROM public.user_refresh_tokens WHERE user_id = ANY($1::uuid[])', [userIds]);
|
|
||||||
await pool.query('DELETE FROM public.password_reset_tokens WHERE user_id = ANY($1::uuid[])', [userIds]);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// --- Stage 2: Delete parent records that other things depend on ---
|
// --- Stage 2: Delete parent records that other things depend on ---
|
||||||
|
|||||||
@@ -93,6 +93,8 @@ export const createMockUser = (overrides: Partial<User> = {}): User => {
|
|||||||
const defaultUser: User = {
|
const defaultUser: User = {
|
||||||
user_id: userId,
|
user_id: userId,
|
||||||
email: `${userId}@example.com`,
|
email: `${userId}@example.com`,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultUser, ...overrides };
|
return { ...defaultUser, ...overrides };
|
||||||
@@ -107,7 +109,7 @@ export const createMockUser = (overrides: Partial<User> = {}): User => {
|
|||||||
* @returns A complete and type-safe UserProfile object.
|
* @returns A complete and type-safe UserProfile object.
|
||||||
*/
|
*/
|
||||||
export const createMockUserProfile = (
|
export const createMockUserProfile = (
|
||||||
overrides: Partial<UserProfile & { user: Partial<User> }> = {},
|
overrides: Partial<Omit<UserProfile, 'user'>> & { user?: Partial<User> } = {},
|
||||||
): UserProfile => {
|
): UserProfile => {
|
||||||
// The user object is the source of truth for user_id and email.
|
// The user object is the source of truth for user_id and email.
|
||||||
const user = createMockUser(overrides.user);
|
const user = createMockUser(overrides.user);
|
||||||
@@ -119,10 +121,10 @@ export const createMockUserProfile = (
|
|||||||
avatar_url: null,
|
avatar_url: null,
|
||||||
preferences: {},
|
preferences: {},
|
||||||
address_id: null,
|
address_id: null,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
created_by: null,
|
created_by: null,
|
||||||
address: null,
|
address: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
user,
|
user,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -143,11 +145,11 @@ export const createMockStore = (overrides: Partial<Store> = {}): Store => {
|
|||||||
|
|
||||||
const defaultStore: Store = {
|
const defaultStore: Store = {
|
||||||
store_id: storeId,
|
store_id: storeId,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
name: 'Mock Store',
|
name: 'Mock Store',
|
||||||
logo_url: null,
|
logo_url: null,
|
||||||
created_by: null,
|
created_by: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultStore, ...overrides };
|
return { ...defaultStore, ...overrides };
|
||||||
@@ -167,12 +169,11 @@ export const createMockFlyer = (
|
|||||||
const flyerId = overrides.flyer_id ?? getNextId();
|
const flyerId = overrides.flyer_id ?? getNextId();
|
||||||
|
|
||||||
// Ensure the store_id is consistent between the flyer and the nested store object
|
// Ensure the store_id is consistent between the flyer and the nested store object
|
||||||
const storeOverrides = overrides.store || {};
|
const store = createMockStore({
|
||||||
if (overrides.store_id && !storeOverrides.store_id) {
|
...overrides.store,
|
||||||
storeOverrides.store_id = overrides.store_id;
|
// Prioritize the top-level store_id if provided
|
||||||
}
|
store_id: overrides.store_id ?? overrides.store?.store_id,
|
||||||
|
});
|
||||||
const store = createMockStore(storeOverrides);
|
|
||||||
|
|
||||||
// Determine the final file_name to generate dependent properties from.
|
// Determine the final file_name to generate dependent properties from.
|
||||||
const fileName = overrides.file_name ?? `flyer-${flyerId}.jpg`;
|
const fileName = overrides.file_name ?? `flyer-${flyerId}.jpg`;
|
||||||
@@ -190,8 +191,6 @@ export const createMockFlyer = (
|
|||||||
|
|
||||||
const defaultFlyer: Flyer = {
|
const defaultFlyer: Flyer = {
|
||||||
flyer_id: flyerId,
|
flyer_id: flyerId,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
file_name: fileName,
|
file_name: fileName,
|
||||||
image_url: `/flyer-images/${fileName}`,
|
image_url: `/flyer-images/${fileName}`,
|
||||||
icon_url: `/flyer-images/icons/icon-${fileName.replace(/\.[^/.]+$/, '.webp')}`,
|
icon_url: `/flyer-images/icons/icon-${fileName.replace(/\.[^/.]+$/, '.webp')}`,
|
||||||
@@ -203,6 +202,8 @@ export const createMockFlyer = (
|
|||||||
status: 'processed',
|
status: 'processed',
|
||||||
item_count: 50,
|
item_count: 50,
|
||||||
uploaded_by: null,
|
uploaded_by: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
store,
|
store,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -244,12 +245,12 @@ export const createMockBrand = (overrides: Partial<Brand> = {}): Brand => {
|
|||||||
|
|
||||||
const defaultBrand: Brand = {
|
const defaultBrand: Brand = {
|
||||||
brand_id: brandId,
|
brand_id: brandId,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
name: `Brand ${brandId}`,
|
name: `Brand ${brandId}`,
|
||||||
logo_url: null,
|
logo_url: null,
|
||||||
store_id: null,
|
store_id: null,
|
||||||
store_name: null,
|
store_name: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultBrand, ...overrides };
|
return { ...defaultBrand, ...overrides };
|
||||||
@@ -266,6 +267,8 @@ export const createMockCategory = (overrides: Partial<Category> = {}): Category
|
|||||||
const defaultCategory: Category = {
|
const defaultCategory: Category = {
|
||||||
category_id: categoryId,
|
category_id: categoryId,
|
||||||
name: `Category ${categoryId}`,
|
name: `Category ${categoryId}`,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultCategory, ...overrides };
|
return { ...defaultCategory, ...overrides };
|
||||||
@@ -319,7 +322,6 @@ export const createMockFlyerItem = (
|
|||||||
const defaultItem: FlyerItem = {
|
const defaultItem: FlyerItem = {
|
||||||
flyer_item_id: flyerItemId,
|
flyer_item_id: flyerItemId,
|
||||||
flyer_id: flyerId,
|
flyer_id: flyerId,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
item: 'Mock Item',
|
item: 'Mock Item',
|
||||||
price_display: '$1.99',
|
price_display: '$1.99',
|
||||||
price_in_cents: 199,
|
price_in_cents: 199,
|
||||||
@@ -327,6 +329,7 @@ export const createMockFlyerItem = (
|
|||||||
quantity: 'each',
|
quantity: 'each',
|
||||||
view_count: 0,
|
view_count: 0,
|
||||||
click_count: 0,
|
click_count: 0,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
updated_at: new Date().toISOString(),
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -358,11 +361,11 @@ export const createMockRecipe = (
|
|||||||
rating_count: 50,
|
rating_count: 50,
|
||||||
fork_count: 10,
|
fork_count: 10,
|
||||||
status: 'public',
|
status: 'public',
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
prep_time_minutes: 15,
|
prep_time_minutes: 15,
|
||||||
cook_time_minutes: 30,
|
cook_time_minutes: 30,
|
||||||
servings: 4,
|
servings: 4,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const {
|
const {
|
||||||
@@ -412,6 +415,8 @@ export const createMockRecipeIngredient = (
|
|||||||
master_item_id: masterItemId,
|
master_item_id: masterItemId,
|
||||||
quantity: 1,
|
quantity: 1,
|
||||||
unit: 'cup',
|
unit: 'cup',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { master_item: _, ...itemOverrides } = overrides;
|
const { master_item: _, ...itemOverrides } = overrides;
|
||||||
@@ -432,6 +437,7 @@ export const createMockRecipeComment = (overrides: Partial<RecipeComment> = {}):
|
|||||||
content: 'This is a mock comment.',
|
content: 'This is a mock comment.',
|
||||||
status: 'visible',
|
status: 'visible',
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
user_full_name: 'Mock User', // This was correct
|
user_full_name: 'Mock User', // This was correct
|
||||||
user_avatar_url: undefined,
|
user_avatar_url: undefined,
|
||||||
};
|
};
|
||||||
@@ -452,6 +458,8 @@ export const createMockPlannedMeal = (overrides: Partial<PlannedMeal> = {}): Pla
|
|||||||
plan_date: new Date().toISOString().split('T')[0],
|
plan_date: new Date().toISOString().split('T')[0],
|
||||||
meal_type: 'dinner',
|
meal_type: 'dinner',
|
||||||
servings_to_cook: 4,
|
servings_to_cook: 4,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultMeal, ...overrides };
|
return { ...defaultMeal, ...overrides };
|
||||||
@@ -476,6 +484,7 @@ export const createMockMenuPlan = (
|
|||||||
start_date: new Date().toISOString().split('T')[0],
|
start_date: new Date().toISOString().split('T')[0],
|
||||||
end_date: new Date(Date.now() + 7 * 24 * 60 * 60 * 1000).toISOString().split('T')[0],
|
end_date: new Date(Date.now() + 7 * 24 * 60 * 60 * 1000).toISOString().split('T')[0],
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { planned_meals: mealsOverrides, ...planOverrides } = overrides;
|
const { planned_meals: mealsOverrides, ...planOverrides } = overrides;
|
||||||
@@ -661,23 +670,22 @@ export const createMockMasterGroceryItem = (
|
|||||||
overrides: Partial<MasterGroceryItem> & { category?: Partial<Category> } = {},
|
overrides: Partial<MasterGroceryItem> & { category?: Partial<Category> } = {},
|
||||||
): MasterGroceryItem => {
|
): MasterGroceryItem => {
|
||||||
// Ensure category_id is consistent between the item and the nested category object
|
// Ensure category_id is consistent between the item and the nested category object
|
||||||
const categoryOverrides = overrides.category || {};
|
const category = createMockCategory({
|
||||||
if (overrides.category_id && !categoryOverrides.category_id) {
|
...overrides.category,
|
||||||
categoryOverrides.category_id = overrides.category_id;
|
// Prioritize the top-level category_id if provided
|
||||||
}
|
category_id: overrides.category_id ?? overrides.category?.category_id,
|
||||||
|
});
|
||||||
const category = createMockCategory(categoryOverrides);
|
|
||||||
|
|
||||||
const defaultItem: MasterGroceryItem = {
|
const defaultItem: MasterGroceryItem = {
|
||||||
master_grocery_item_id: getNextId(),
|
master_grocery_item_id: getNextId(),
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
name: 'Mock Master Item',
|
name: 'Mock Master Item',
|
||||||
category_id: category.category_id,
|
category_id: category.category_id,
|
||||||
category_name: category.name,
|
category_name: category.name,
|
||||||
is_allergen: false,
|
is_allergen: false,
|
||||||
allergy_info: null,
|
allergy_info: null,
|
||||||
created_by: null,
|
created_by: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { category: _, ...itemOverrides } = overrides;
|
const { category: _, ...itemOverrides } = overrides;
|
||||||
@@ -729,9 +737,9 @@ export const createMockShoppingList = (
|
|||||||
shopping_list_id: shoppingListId,
|
shopping_list_id: shoppingListId,
|
||||||
user_id: `user-${getNextId()}`,
|
user_id: `user-${getNextId()}`,
|
||||||
name: 'My Mock List',
|
name: 'My Mock List',
|
||||||
|
items: [],
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
updated_at: new Date().toISOString(),
|
updated_at: new Date().toISOString(),
|
||||||
items: [],
|
|
||||||
};
|
};
|
||||||
|
|
||||||
if (overrides.items) {
|
if (overrides.items) {
|
||||||
@@ -767,15 +775,12 @@ export const createMockShoppingListItem = (
|
|||||||
shopping_list_id: shoppingListId,
|
shopping_list_id: shoppingListId,
|
||||||
custom_item_name: 'Mock Shopping List Item',
|
custom_item_name: 'Mock Shopping List Item',
|
||||||
quantity: 1,
|
quantity: 1,
|
||||||
is_purchased: false,
|
is_purchased: false, // This was correct
|
||||||
added_at: new Date().toISOString(),
|
added_at: new Date().toISOString(),
|
||||||
updated_at: new Date().toISOString(),
|
updated_at: new Date().toISOString(),
|
||||||
|
master_item_id: masterItemId,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (masterItemId) {
|
|
||||||
defaultItem.master_item_id = masterItemId;
|
|
||||||
}
|
|
||||||
|
|
||||||
const { master_item: masterItemOverride, ...itemOverrides } = overrides;
|
const { master_item: masterItemOverride, ...itemOverrides } = overrides;
|
||||||
const result = { ...defaultItem, ...itemOverrides };
|
const result = { ...defaultItem, ...itemOverrides };
|
||||||
|
|
||||||
@@ -805,6 +810,8 @@ export const createMockShoppingTripItem = (
|
|||||||
master_item_name: masterItemId ? (overrides.master_item?.name ?? 'Mock Master Item') : null,
|
master_item_name: masterItemId ? (overrides.master_item?.name ?? 'Mock Master Item') : null,
|
||||||
quantity: 1,
|
quantity: 1,
|
||||||
price_paid_cents: 199,
|
price_paid_cents: 199,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { master_item: _, ...itemOverrides } = overrides;
|
const { master_item: _, ...itemOverrides } = overrides;
|
||||||
@@ -829,6 +836,7 @@ export const createMockShoppingTrip = (
|
|||||||
completed_at: new Date().toISOString(),
|
completed_at: new Date().toISOString(),
|
||||||
total_spent_cents: 0,
|
total_spent_cents: 0,
|
||||||
items: [],
|
items: [],
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { items: itemsOverrides, ...tripOverrides } = overrides;
|
const { items: itemsOverrides, ...tripOverrides } = overrides;
|
||||||
@@ -864,6 +872,8 @@ export const createMockReceiptItem = (overrides: Partial<ReceiptItem> = {}): Rec
|
|||||||
master_item_id: null,
|
master_item_id: null,
|
||||||
product_id: null,
|
product_id: null,
|
||||||
status: 'unmatched',
|
status: 'unmatched',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultItem, ...overrides };
|
return { ...defaultItem, ...overrides };
|
||||||
@@ -888,8 +898,9 @@ export const createMockReceipt = (
|
|||||||
total_amount_cents: null,
|
total_amount_cents: null,
|
||||||
status: 'pending',
|
status: 'pending',
|
||||||
raw_text: null,
|
raw_text: null,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
processed_at: null,
|
processed_at: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
const { items: itemsOverrides, ...receiptOverrides } = overrides;
|
const { items: itemsOverrides, ...receiptOverrides } = overrides;
|
||||||
@@ -916,6 +927,8 @@ export const createMockDietaryRestriction = (
|
|||||||
dietary_restriction_id: 1,
|
dietary_restriction_id: 1,
|
||||||
name: 'Vegetarian',
|
name: 'Vegetarian',
|
||||||
type: 'diet',
|
type: 'diet',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
...overrides,
|
...overrides,
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
@@ -955,6 +968,8 @@ export const createMockItemPriceHistory = (
|
|||||||
max_price_in_cents: 399,
|
max_price_in_cents: 399,
|
||||||
avg_price_in_cents: 299,
|
avg_price_in_cents: 299,
|
||||||
data_points_count: 10,
|
data_points_count: 10,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultHistory, ...overrides };
|
return { ...defaultHistory, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1008,6 +1023,7 @@ export const createMockRecipeCollection = (
|
|||||||
name: 'My Favorite Recipes',
|
name: 'My Favorite Recipes',
|
||||||
description: 'A collection of mock recipes.',
|
description: 'A collection of mock recipes.',
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultCollection, ...overrides };
|
return { ...defaultCollection, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1027,6 +1043,7 @@ export const createMockSharedShoppingList = (
|
|||||||
shared_with_user_id: `user-${getNextId()}`,
|
shared_with_user_id: `user-${getNextId()}`,
|
||||||
permission_level: 'view',
|
permission_level: 'view',
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultSharedList, ...overrides };
|
return { ...defaultSharedList, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1118,6 +1135,7 @@ export const createMockUserAlert = (overrides: Partial<UserAlert> = {}): UserAle
|
|||||||
threshold_value: 499,
|
threshold_value: 499,
|
||||||
is_active: true,
|
is_active: true,
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultAlert, ...overrides };
|
return { ...defaultAlert, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1140,6 +1158,7 @@ export const createMockUserSubmittedPrice = (
|
|||||||
upvotes: 0,
|
upvotes: 0,
|
||||||
downvotes: 0,
|
downvotes: 0,
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultPrice, ...overrides };
|
return { ...defaultPrice, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1157,6 +1176,7 @@ export const createMockRecipeRating = (overrides: Partial<RecipeRating> = {}): R
|
|||||||
rating: 5,
|
rating: 5,
|
||||||
comment: 'Great recipe!',
|
comment: 'Great recipe!',
|
||||||
created_at: new Date().toISOString(),
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultRating, ...overrides };
|
return { ...defaultRating, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1171,6 +1191,8 @@ export const createMockTag = (overrides: Partial<Tag> = {}): Tag => {
|
|||||||
const defaultTag: Tag = {
|
const defaultTag: Tag = {
|
||||||
tag_id: tagId,
|
tag_id: tagId,
|
||||||
name: `Tag ${tagId}`,
|
name: `Tag ${tagId}`,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultTag, ...overrides };
|
return { ...defaultTag, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1188,6 +1210,8 @@ export const createMockPantryLocation = (
|
|||||||
pantry_location_id: locationId,
|
pantry_location_id: locationId,
|
||||||
user_id: `user-${getNextId()}`,
|
user_id: `user-${getNextId()}`,
|
||||||
name: `Location ${locationId}`,
|
name: `Location ${locationId}`,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
return { ...defaultLocation, ...overrides };
|
return { ...defaultLocation, ...overrides };
|
||||||
};
|
};
|
||||||
@@ -1228,6 +1252,7 @@ export const createMockUserDietaryRestriction = (
|
|||||||
const defaultUserRestriction: UserDietaryRestriction = {
|
const defaultUserRestriction: UserDietaryRestriction = {
|
||||||
user_id: userId,
|
user_id: userId,
|
||||||
restriction_id: restrictionId,
|
restriction_id: restrictionId,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultUserRestriction, ...overrides };
|
return { ...defaultUserRestriction, ...overrides };
|
||||||
@@ -1245,12 +1270,14 @@ export const createMockUserAppliance = (
|
|||||||
const userId = overrides.user_id ?? overrides.user?.user_id ?? `user-${getNextId()}`;
|
const userId = overrides.user_id ?? overrides.user?.user_id ?? `user-${getNextId()}`;
|
||||||
const applianceId = overrides.appliance_id ?? overrides.appliance?.appliance_id ?? getNextId();
|
const applianceId = overrides.appliance_id ?? overrides.appliance?.appliance_id ?? getNextId();
|
||||||
|
|
||||||
const defaultUserAppliance: UserAppliance = {
|
const defaultUserAppliance = {
|
||||||
user_id: userId,
|
user_id: userId,
|
||||||
appliance_id: applianceId,
|
appliance_id: applianceId,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultUserAppliance, ...overrides };
|
// The 'as UserAppliance' cast is necessary because TypeScript can't guarantee that the spread of a Partial<T> results in a complete T.
|
||||||
|
return { ...defaultUserAppliance, ...overrides } as UserAppliance;
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -1266,13 +1293,13 @@ export const createMockAddress = (overrides: Partial<Address> = {}): Address =>
|
|||||||
province_state: 'BC',
|
province_state: 'BC',
|
||||||
postal_code: 'V8T 1A1',
|
postal_code: 'V8T 1A1',
|
||||||
country: 'CA',
|
country: 'CA',
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
// Optional fields
|
// Optional fields
|
||||||
address_line_2: null,
|
address_line_2: null,
|
||||||
latitude: null,
|
latitude: null,
|
||||||
longitude: null,
|
longitude: null,
|
||||||
location: null,
|
location: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultAddress, ...overrides };
|
return { ...defaultAddress, ...overrides };
|
||||||
@@ -1309,8 +1336,6 @@ export const createMockUserWithPasswordHash = (
|
|||||||
*/
|
*/
|
||||||
export const createMockProfile = (overrides: Partial<Profile> = {}): Profile => {
|
export const createMockProfile = (overrides: Partial<Profile> = {}): Profile => {
|
||||||
const defaultProfile: Profile = {
|
const defaultProfile: Profile = {
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
full_name: 'Mock Profile User',
|
full_name: 'Mock Profile User',
|
||||||
avatar_url: null,
|
avatar_url: null,
|
||||||
address_id: null,
|
address_id: null,
|
||||||
@@ -1319,6 +1344,8 @@ export const createMockProfile = (overrides: Partial<Profile> = {}): Profile =>
|
|||||||
preferences: {},
|
preferences: {},
|
||||||
created_by: null,
|
created_by: null,
|
||||||
updated_by: null,
|
updated_by: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultProfile, ...overrides };
|
return { ...defaultProfile, ...overrides };
|
||||||
@@ -1376,14 +1403,14 @@ export const createMockUnmatchedFlyerItem = (
|
|||||||
const defaultItem: UnmatchedFlyerItem = {
|
const defaultItem: UnmatchedFlyerItem = {
|
||||||
unmatched_flyer_item_id: getNextId(),
|
unmatched_flyer_item_id: getNextId(),
|
||||||
status: 'pending',
|
status: 'pending',
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
reviewed_at: null,
|
reviewed_at: null,
|
||||||
flyer_item_id: getNextId(),
|
flyer_item_id: getNextId(),
|
||||||
flyer_item_name: 'Mystery Product',
|
flyer_item_name: 'Mystery Product',
|
||||||
price_display: '$?.??',
|
price_display: '$?.??',
|
||||||
flyer_id: getNextId(),
|
flyer_id: getNextId(),
|
||||||
store_name: 'Random Store',
|
store_name: 'Random Store',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultItem, ...overrides };
|
return { ...defaultItem, ...overrides };
|
||||||
@@ -1400,10 +1427,10 @@ export const createMockAdminUserView = (overrides: Partial<AdminUserView> = {}):
|
|||||||
const defaultUserView: AdminUserView = {
|
const defaultUserView: AdminUserView = {
|
||||||
user_id: userId,
|
user_id: userId,
|
||||||
email: `${userId}@example.com`,
|
email: `${userId}@example.com`,
|
||||||
created_at: new Date().toISOString(),
|
|
||||||
role: 'user',
|
role: 'user',
|
||||||
full_name: 'Mock User',
|
full_name: 'Mock User',
|
||||||
avatar_url: null,
|
avatar_url: null,
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
};
|
};
|
||||||
|
|
||||||
return { ...defaultUserView, ...overrides };
|
return { ...defaultUserView, ...overrides };
|
||||||
@@ -1450,6 +1477,8 @@ export const createMockAppliance = (overrides: Partial<Appliance> = {}): Applian
|
|||||||
return {
|
return {
|
||||||
appliance_id: 1,
|
appliance_id: 1,
|
||||||
name: 'Oven',
|
name: 'Oven',
|
||||||
|
created_at: new Date().toISOString(),
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
...overrides,
|
...overrides,
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
@@ -1482,7 +1511,7 @@ export const createMockAddressPayload = (overrides: Partial<Address> = {}): Part
|
|||||||
...overrides,
|
...overrides,
|
||||||
});
|
});
|
||||||
|
|
||||||
export const createMockSearchQueryPayload = (overrides: Partial<Omit<SearchQuery, 'search_query_id' | 'id' | 'created_at' | 'user_id'>> = {}): Omit<SearchQuery, 'search_query_id' | 'id' | 'created_at' | 'user_id'> => ({
|
export const createMockSearchQueryPayload = (overrides: Partial<Omit<SearchQuery, 'search_query_id' | 'created_at' | 'updated_at' | 'user_id'>> = {}): Omit<SearchQuery, 'search_query_id' | 'created_at' | 'updated_at' | 'user_id'> => ({
|
||||||
query_text: 'mock search',
|
query_text: 'mock search',
|
||||||
result_count: 5,
|
result_count: 5,
|
||||||
was_successful: true,
|
was_successful: true,
|
||||||
|
|||||||
486
src/types.ts
486
src/types.ts
@@ -1,31 +1,31 @@
|
|||||||
// src/types.ts
|
// src/types.ts
|
||||||
export interface Store {
|
export interface Store {
|
||||||
store_id: number;
|
readonly store_id: number;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
name: string;
|
name: string;
|
||||||
logo_url?: string | null;
|
logo_url?: string | null;
|
||||||
created_by?: string | null;
|
readonly created_by?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export type FlyerStatus = 'processed' | 'needs_review' | 'archived';
|
export type FlyerStatus = 'processed' | 'needs_review' | 'archived';
|
||||||
|
|
||||||
export interface Flyer {
|
export interface Flyer {
|
||||||
flyer_id: number;
|
readonly flyer_id: number;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
file_name: string;
|
file_name: string;
|
||||||
image_url: string;
|
image_url: string;
|
||||||
icon_url?: string | null; // URL for the 64x64 icon version of the flyer
|
icon_url?: string | null; // URL for the 64x64 icon version of the flyer
|
||||||
checksum?: string;
|
readonly checksum?: string;
|
||||||
store_id?: number;
|
readonly store_id?: number;
|
||||||
valid_from?: string | null;
|
valid_from?: string | null;
|
||||||
valid_to?: string | null;
|
valid_to?: string | null;
|
||||||
store_address?: string | null;
|
store_address?: string | null;
|
||||||
status: FlyerStatus;
|
status: FlyerStatus;
|
||||||
item_count: number;
|
item_count: number;
|
||||||
uploaded_by?: string | null; // UUID of the user who uploaded it, can be null for anonymous uploads
|
readonly uploaded_by?: string | null; // UUID of the user who uploaded it, can be null for anonymous uploads
|
||||||
store?: Store;
|
store?: Store;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -67,62 +67,64 @@ export interface UnitPrice {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface FlyerItem {
|
export interface FlyerItem {
|
||||||
flyer_item_id: number;
|
readonly flyer_item_id: number;
|
||||||
flyer_id: number;
|
readonly flyer_id: number;
|
||||||
created_at: string;
|
|
||||||
item: string;
|
item: string;
|
||||||
price_display: string;
|
price_display: string;
|
||||||
price_in_cents?: number | null;
|
price_in_cents?: number | null;
|
||||||
quantity?: string;
|
quantity?: string;
|
||||||
quantity_num?: number | null;
|
quantity_num?: number | null;
|
||||||
master_item_id?: number;
|
master_item_id?: number; // Can be updated by admin correction
|
||||||
master_item_name?: string | null;
|
master_item_name?: string | null;
|
||||||
category_id?: number | null;
|
category_id?: number | null; // Can be updated by admin correction
|
||||||
category_name?: string | null;
|
category_name?: string | null;
|
||||||
unit_price?: UnitPrice | null;
|
unit_price?: UnitPrice | null;
|
||||||
product_id?: number | null;
|
product_id?: number | null; // Can be updated by admin correction
|
||||||
view_count: number;
|
readonly view_count: number;
|
||||||
click_count: number;
|
readonly click_count: number;
|
||||||
updated_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface MasterGroceryItem {
|
export interface MasterGroceryItem {
|
||||||
master_grocery_item_id: number;
|
readonly master_grocery_item_id: number;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
name: string;
|
name: string;
|
||||||
category_id?: number | null;
|
category_id?: number | null; // Can be updated by admin
|
||||||
category_name?: string | null;
|
category_name?: string | null;
|
||||||
is_allergen?: boolean;
|
is_allergen?: boolean;
|
||||||
allergy_info?: unknown | null; // JSONB
|
allergy_info?: unknown | null; // JSONB
|
||||||
created_by?: string | null;
|
readonly created_by?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Category {
|
export interface Category {
|
||||||
category_id: number;
|
readonly category_id: number;
|
||||||
name: string;
|
name: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Brand {
|
export interface Brand {
|
||||||
brand_id: number;
|
readonly brand_id: number;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
name: string;
|
name: string;
|
||||||
logo_url?: string | null;
|
logo_url?: string | null;
|
||||||
store_id?: number | null;
|
readonly store_id?: number | null;
|
||||||
store_name?: string | null;
|
store_name?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Product {
|
export interface Product {
|
||||||
product_id: number;
|
readonly product_id: number;
|
||||||
created_at: string;
|
readonly master_item_id: number;
|
||||||
updated_at: string;
|
readonly brand_id?: number | null;
|
||||||
master_item_id: number;
|
|
||||||
brand_id?: number | null;
|
|
||||||
name: string;
|
name: string;
|
||||||
description?: string | null;
|
description?: string | null;
|
||||||
size?: string | null;
|
size?: string | null;
|
||||||
upc_code?: string | null;
|
upc_code?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface DealItem {
|
export interface DealItem {
|
||||||
@@ -137,8 +139,10 @@ export interface DealItem {
|
|||||||
|
|
||||||
// User-specific types
|
// User-specific types
|
||||||
export interface User {
|
export interface User {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
email: string;
|
email: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -147,27 +151,25 @@ export interface User {
|
|||||||
*/
|
*/
|
||||||
export interface UserWithPasswordHash extends User {
|
export interface UserWithPasswordHash extends User {
|
||||||
password_hash: string | null;
|
password_hash: string | null;
|
||||||
failed_login_attempts: number;
|
readonly failed_login_attempts: number;
|
||||||
last_failed_login: string | null; // TIMESTAMPTZ
|
readonly last_failed_login: string | null; // TIMESTAMPTZ
|
||||||
last_login_at?: string | null; // TIMESTAMPTZ
|
readonly last_login_at?: string | null; // TIMESTAMPTZ
|
||||||
last_login_ip?: string | null;
|
readonly last_login_ip?: string | null;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
}
|
}
|
||||||
export interface Profile {
|
export interface Profile {
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
full_name?: string | null;
|
full_name?: string | null;
|
||||||
avatar_url?: string | null;
|
avatar_url?: string | null;
|
||||||
address_id?: number | null;
|
address_id?: number | null; // Can be updated
|
||||||
points: number;
|
readonly points: number;
|
||||||
role: 'admin' | 'user';
|
readonly role: 'admin' | 'user';
|
||||||
preferences?: {
|
preferences?: {
|
||||||
darkMode?: boolean;
|
darkMode?: boolean;
|
||||||
unitSystem?: 'metric' | 'imperial';
|
unitSystem?: 'metric' | 'imperial';
|
||||||
} | null;
|
} | null;
|
||||||
created_by?: string | null;
|
readonly created_by?: string | null;
|
||||||
updated_by?: string | null;
|
readonly updated_by?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -181,16 +183,16 @@ export type UserProfile = Profile & {
|
|||||||
};
|
};
|
||||||
|
|
||||||
export interface SuggestedCorrection {
|
export interface SuggestedCorrection {
|
||||||
suggested_correction_id: number;
|
readonly suggested_correction_id: number;
|
||||||
flyer_item_id: number;
|
readonly flyer_item_id: number;
|
||||||
user_id: string;
|
readonly user_id: string;
|
||||||
correction_type: string;
|
correction_type: string;
|
||||||
suggested_value: string;
|
suggested_value: string;
|
||||||
status: 'pending' | 'approved' | 'rejected';
|
status: 'pending' | 'approved' | 'rejected';
|
||||||
created_at: string;
|
readonly reviewed_at?: string | null;
|
||||||
updated_at: string;
|
|
||||||
reviewed_at?: string | null;
|
|
||||||
reviewed_notes?: string | null;
|
reviewed_notes?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
// Joined data
|
// Joined data
|
||||||
user_email?: string;
|
user_email?: string;
|
||||||
flyer_item_name?: string;
|
flyer_item_name?: string;
|
||||||
@@ -210,43 +212,44 @@ export interface UserDataExport {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface UserAlert {
|
export interface UserAlert {
|
||||||
user_alert_id: number;
|
readonly user_alert_id: number;
|
||||||
user_watched_item_id: number;
|
readonly user_watched_item_id: number;
|
||||||
alert_type: 'PRICE_BELOW' | 'PERCENT_OFF_AVERAGE';
|
alert_type: 'PRICE_BELOW' | 'PERCENT_OFF_AVERAGE';
|
||||||
threshold_value: number;
|
threshold_value: number;
|
||||||
is_active: boolean;
|
is_active: boolean;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Notification {
|
export interface Notification {
|
||||||
notification_id: number;
|
readonly notification_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
content: string;
|
content: string;
|
||||||
link_url?: string | null;
|
link_url?: string | null;
|
||||||
is_read: boolean;
|
is_read: boolean;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
updated_at: string;
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ShoppingList {
|
export interface ShoppingList {
|
||||||
shopping_list_id: number;
|
readonly shopping_list_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
name: string;
|
name: string;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
items: ShoppingListItem[]; // Nested items
|
items: ShoppingListItem[]; // Nested items
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ShoppingListItem {
|
export interface ShoppingListItem {
|
||||||
shopping_list_item_id: number;
|
readonly shopping_list_item_id: number;
|
||||||
shopping_list_id: number;
|
readonly shopping_list_id: number;
|
||||||
master_item_id?: number | null;
|
readonly master_item_id?: number | null;
|
||||||
custom_item_name?: string | null;
|
custom_item_name?: string | null;
|
||||||
quantity: number;
|
quantity: number;
|
||||||
is_purchased: boolean;
|
is_purchased: boolean;
|
||||||
notes?: string | null;
|
notes?: string | null;
|
||||||
added_at: string;
|
readonly added_at: string;
|
||||||
updated_at: string;
|
readonly updated_at: string;
|
||||||
// Joined data for display
|
// Joined data for display
|
||||||
master_item?: {
|
master_item?: {
|
||||||
name: string;
|
name: string;
|
||||||
@@ -254,25 +257,29 @@ export interface ShoppingListItem {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface UserSubmittedPrice {
|
export interface UserSubmittedPrice {
|
||||||
user_submitted_price_id: number;
|
readonly user_submitted_price_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
store_id: number;
|
readonly store_id: number;
|
||||||
price_in_cents: number;
|
price_in_cents: number;
|
||||||
photo_url?: string | null;
|
photo_url?: string | null;
|
||||||
upvotes: number;
|
readonly upvotes: number;
|
||||||
downvotes: number;
|
readonly downvotes: number;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ItemPriceHistory {
|
export interface ItemPriceHistory {
|
||||||
item_price_history_id: number;
|
readonly item_price_history_id: number;
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
summary_date: string; // DATE
|
summary_date: string; // DATE
|
||||||
|
readonly store_location_id?: number | null;
|
||||||
min_price_in_cents?: number | null;
|
min_price_in_cents?: number | null;
|
||||||
max_price_in_cents?: number | null;
|
max_price_in_cents?: number | null;
|
||||||
avg_price_in_cents?: number | null;
|
avg_price_in_cents?: number | null;
|
||||||
data_points_count: number;
|
data_points_count: number;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -286,15 +293,17 @@ export interface HistoricalPriceDataPoint {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface MasterItemAlias {
|
export interface MasterItemAlias {
|
||||||
master_item_alias_id: number;
|
readonly master_item_alias_id: number;
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
alias: string;
|
alias: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Recipe {
|
export interface Recipe {
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
user_id?: string | null; // UUID
|
readonly user_id?: string | null; // UUID
|
||||||
original_recipe_id?: number | null;
|
readonly original_recipe_id?: number | null;
|
||||||
name: string;
|
name: string;
|
||||||
description?: string | null;
|
description?: string | null;
|
||||||
instructions?: string | null;
|
instructions?: string | null;
|
||||||
@@ -306,34 +315,40 @@ export interface Recipe {
|
|||||||
protein_grams?: number | null;
|
protein_grams?: number | null;
|
||||||
fat_grams?: number | null;
|
fat_grams?: number | null;
|
||||||
carb_grams?: number | null;
|
carb_grams?: number | null;
|
||||||
avg_rating: number;
|
readonly avg_rating: number;
|
||||||
status: 'private' | 'pending_review' | 'public' | 'rejected';
|
status: 'private' | 'pending_review' | 'public' | 'rejected';
|
||||||
rating_count: number;
|
readonly rating_count: number;
|
||||||
fork_count: number;
|
readonly fork_count: number;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
comments?: RecipeComment[];
|
comments?: RecipeComment[];
|
||||||
ingredients?: RecipeIngredient[];
|
ingredients?: RecipeIngredient[];
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeIngredient {
|
export interface RecipeIngredient {
|
||||||
recipe_ingredient_id: number;
|
readonly recipe_ingredient_id: number;
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
quantity: number;
|
quantity: number;
|
||||||
unit: string;
|
unit: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeIngredientSubstitution {
|
export interface RecipeIngredientSubstitution {
|
||||||
recipe_ingredient_substitution_id: number;
|
readonly recipe_ingredient_substitution_id: number;
|
||||||
recipe_ingredient_id: number;
|
readonly recipe_ingredient_id: number;
|
||||||
substitute_master_item_id: number;
|
readonly substitute_master_item_id: number;
|
||||||
notes?: string | null;
|
notes?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Tag {
|
export interface Tag {
|
||||||
tag_id: number;
|
readonly tag_id: number;
|
||||||
name: string;
|
name: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeTag {
|
export interface RecipeTag {
|
||||||
@@ -342,147 +357,165 @@ export interface RecipeTag {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeRating {
|
export interface RecipeRating {
|
||||||
recipe_rating_id: number;
|
readonly recipe_rating_id: number;
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
rating: number;
|
rating: number;
|
||||||
comment?: string | null;
|
comment?: string | null;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeComment {
|
export interface RecipeComment {
|
||||||
recipe_comment_id: number;
|
readonly recipe_comment_id: number;
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
parent_comment_id?: number | null;
|
readonly parent_comment_id?: number | null;
|
||||||
content: string;
|
content: string;
|
||||||
status: 'visible' | 'hidden' | 'reported';
|
status: 'visible' | 'hidden' | 'reported';
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
updated_at?: string | null;
|
readonly updated_at: string;
|
||||||
user_full_name?: string; // Joined data
|
user_full_name?: string; // Joined data
|
||||||
user_avatar_url?: string; // Joined data
|
user_avatar_url?: string; // Joined data
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface MenuPlan {
|
export interface MenuPlan {
|
||||||
menu_plan_id: number;
|
readonly menu_plan_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
name: string;
|
name: string;
|
||||||
start_date: string; // DATE
|
start_date: string; // DATE
|
||||||
end_date: string; // DATE
|
end_date: string; // DATE
|
||||||
created_at: string;
|
|
||||||
planned_meals?: PlannedMeal[];
|
planned_meals?: PlannedMeal[];
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface SharedMenuPlan {
|
export interface SharedMenuPlan {
|
||||||
shared_menu_plan_id: number;
|
readonly shared_menu_plan_id: number;
|
||||||
menu_plan_id: number;
|
readonly menu_plan_id: number;
|
||||||
shared_by_user_id: string; // UUID
|
readonly shared_by_user_id: string; // UUID
|
||||||
shared_with_user_id: string; // UUID
|
readonly shared_with_user_id: string; // UUID
|
||||||
permission_level: 'view' | 'edit';
|
permission_level: 'view' | 'edit';
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface PlannedMeal {
|
export interface PlannedMeal {
|
||||||
planned_meal_id: number;
|
readonly planned_meal_id: number;
|
||||||
menu_plan_id: number;
|
readonly menu_plan_id: number;
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
plan_date: string; // DATE
|
plan_date: string; // DATE
|
||||||
meal_type: string;
|
meal_type: string;
|
||||||
servings_to_cook?: number | null;
|
servings_to_cook?: number | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface PantryItem {
|
export interface PantryItem {
|
||||||
pantry_item_id: number;
|
readonly pantry_item_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
quantity: number;
|
quantity: number;
|
||||||
unit?: string | null;
|
unit?: string | null;
|
||||||
best_before_date?: string | null; // DATE
|
best_before_date?: string | null; // DATE
|
||||||
pantry_location_id?: number | null;
|
pantry_location_id?: number | null;
|
||||||
notification_sent_at?: string | null; // TIMESTAMPTZ
|
readonly notification_sent_at?: string | null; // TIMESTAMPTZ
|
||||||
updated_at: string;
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface UserItemAlias {
|
export interface UserItemAlias {
|
||||||
user_item_alias_id: number;
|
readonly user_item_alias_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
alias: string;
|
alias: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface FavoriteRecipe {
|
export interface FavoriteRecipe {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface FavoriteStore {
|
export interface FavoriteStore {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
store_id: number;
|
readonly store_id: number;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeCollection {
|
export interface RecipeCollection {
|
||||||
recipe_collection_id: number;
|
readonly recipe_collection_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
name: string;
|
name: string;
|
||||||
description?: string | null;
|
description?: string | null;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeCollectionItem {
|
export interface RecipeCollectionItem {
|
||||||
collection_id: number;
|
readonly collection_id: number;
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
added_at: string;
|
readonly added_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface SharedShoppingList {
|
export interface SharedShoppingList {
|
||||||
shared_shopping_list_id: number;
|
readonly shared_shopping_list_id: number;
|
||||||
shopping_list_id: number;
|
readonly shopping_list_id: number;
|
||||||
shared_by_user_id: string; // UUID
|
readonly shared_by_user_id: string; // UUID
|
||||||
shared_with_user_id: string; // UUID
|
readonly shared_with_user_id: string; // UUID
|
||||||
permission_level: 'view' | 'edit';
|
permission_level: 'view' | 'edit';
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface SharedRecipeCollection {
|
export interface SharedRecipeCollection {
|
||||||
shared_collection_id: number;
|
readonly shared_collection_id: number;
|
||||||
recipe_collection_id: number;
|
readonly recipe_collection_id: number;
|
||||||
shared_by_user_id: string; // UUID
|
readonly shared_by_user_id: string; // UUID
|
||||||
shared_with_user_id: string; // UUID
|
readonly shared_with_user_id: string; // UUID
|
||||||
permission_level: 'view' | 'edit';
|
permission_level: 'view' | 'edit';
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface DietaryRestriction {
|
export interface DietaryRestriction {
|
||||||
dietary_restriction_id: number;
|
readonly dietary_restriction_id: number;
|
||||||
name: string;
|
name: string;
|
||||||
type: 'diet' | 'allergy';
|
type: 'diet' | 'allergy';
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface UserDietaryRestriction {
|
export interface UserDietaryRestriction {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
restriction_id: number;
|
readonly restriction_id: number;
|
||||||
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Appliance {
|
export interface Appliance {
|
||||||
appliance_id: number;
|
readonly appliance_id: number;
|
||||||
name: string;
|
name: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface UserAppliance {
|
export interface UserAppliance {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
appliance_id: number;
|
readonly appliance_id: number;
|
||||||
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface RecipeAppliance {
|
export interface RecipeAppliance {
|
||||||
recipe_id: number;
|
readonly recipe_id: number;
|
||||||
appliance_id: number;
|
readonly appliance_id: number;
|
||||||
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface UserFollow {
|
export interface UserFollow {
|
||||||
follower_id: string; // UUID
|
readonly follower_id: string; // UUID
|
||||||
following_id: string; // UUID
|
readonly following_id: string; // UUID
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* The list of possible actions for an activity log.
|
* The list of possible actions for an activity log.
|
||||||
@@ -501,13 +534,13 @@ export type ActivityLogAction =
|
|||||||
* Base interface for all log items, containing common properties.
|
* Base interface for all log items, containing common properties.
|
||||||
*/
|
*/
|
||||||
interface ActivityLogItemBase {
|
interface ActivityLogItemBase {
|
||||||
activity_log_id: number;
|
readonly activity_log_id: number;
|
||||||
user_id: string | null;
|
readonly user_id: string | null;
|
||||||
action: string;
|
action: string;
|
||||||
display_text: string;
|
display_text: string;
|
||||||
created_at: string;
|
|
||||||
updated_at: string;
|
|
||||||
icon?: string | null;
|
icon?: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
// Joined data for display in feeds
|
// Joined data for display in feeds
|
||||||
user_full_name?: string;
|
user_full_name?: string;
|
||||||
user_avatar_url?: string;
|
user_avatar_url?: string;
|
||||||
@@ -566,68 +599,77 @@ export type ActivityLogItem =
|
|||||||
| ListSharedLog;
|
| ListSharedLog;
|
||||||
|
|
||||||
export interface PantryLocation {
|
export interface PantryLocation {
|
||||||
pantry_location_id: number;
|
readonly pantry_location_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
name: string;
|
name: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface SearchQuery {
|
export interface SearchQuery {
|
||||||
search_query_id: number;
|
readonly search_query_id: number;
|
||||||
user_id?: string | null; // UUID
|
readonly user_id?: string | null; // UUID
|
||||||
query_text: string;
|
query_text: string;
|
||||||
result_count?: number | null;
|
result_count?: number | null;
|
||||||
was_successful?: boolean | null;
|
was_successful?: boolean | null;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ShoppingTripItem {
|
export interface ShoppingTripItem {
|
||||||
shopping_trip_item_id: number;
|
readonly shopping_trip_item_id: number;
|
||||||
shopping_trip_id: number;
|
readonly shopping_trip_id: number;
|
||||||
master_item_id?: number | null;
|
readonly master_item_id?: number | null;
|
||||||
custom_item_name?: string | null;
|
custom_item_name?: string | null;
|
||||||
quantity: number;
|
quantity: number;
|
||||||
price_paid_cents?: number | null;
|
price_paid_cents?: number | null;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
// Joined data for display
|
// Joined data for display
|
||||||
master_item_name?: string | null;
|
master_item_name?: string | null;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ShoppingTrip {
|
export interface ShoppingTrip {
|
||||||
shopping_trip_id: number;
|
readonly shopping_trip_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
shopping_list_id?: number | null;
|
readonly shopping_list_id?: number | null;
|
||||||
completed_at: string;
|
readonly completed_at: string;
|
||||||
total_spent_cents?: number | null;
|
total_spent_cents?: number | null;
|
||||||
items: ShoppingTripItem[]; // Nested items
|
items: ShoppingTripItem[]; // Nested items
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Receipt {
|
export interface Receipt {
|
||||||
receipt_id: number;
|
readonly receipt_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
store_id?: number | null;
|
store_id?: number | null;
|
||||||
receipt_image_url: string;
|
receipt_image_url: string;
|
||||||
transaction_date?: string | null;
|
transaction_date?: string | null;
|
||||||
total_amount_cents?: number | null;
|
total_amount_cents?: number | null;
|
||||||
status: 'pending' | 'processing' | 'completed' | 'failed';
|
status: 'pending' | 'processing' | 'completed' | 'failed';
|
||||||
raw_text?: string | null;
|
raw_text?: string | null;
|
||||||
created_at: string;
|
readonly processed_at?: string | null;
|
||||||
processed_at?: string | null;
|
|
||||||
items?: ReceiptItem[];
|
items?: ReceiptItem[];
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ReceiptItem {
|
export interface ReceiptItem {
|
||||||
receipt_item_id: number;
|
readonly receipt_item_id: number;
|
||||||
receipt_id: number;
|
readonly receipt_id: number;
|
||||||
raw_item_description: string;
|
raw_item_description: string;
|
||||||
quantity: number;
|
quantity: number;
|
||||||
price_paid_cents: number;
|
price_paid_cents: number;
|
||||||
master_item_id?: number | null;
|
master_item_id?: number | null; // Can be updated by admin correction
|
||||||
product_id?: number | null;
|
product_id?: number | null; // Can be updated by admin correction
|
||||||
status: 'unmatched' | 'matched' | 'needs_review' | 'ignored';
|
status: 'unmatched' | 'matched' | 'needs_review' | 'ignored';
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface ReceiptDeal {
|
export interface ReceiptDeal {
|
||||||
receipt_item_id: number;
|
readonly receipt_item_id: number;
|
||||||
master_item_id: number;
|
readonly master_item_id: number;
|
||||||
item_name: string;
|
item_name: string;
|
||||||
price_paid_cents: number;
|
price_paid_cents: number;
|
||||||
current_best_price_in_cents: number;
|
current_best_price_in_cents: number;
|
||||||
@@ -646,13 +688,15 @@ export interface GeoJSONPoint {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export interface StoreLocation {
|
export interface StoreLocation {
|
||||||
store_location_id: number;
|
readonly store_location_id: number;
|
||||||
store_id?: number | null;
|
readonly store_id?: number | null;
|
||||||
address_id: number;
|
readonly address_id: number;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface Address {
|
export interface Address {
|
||||||
address_id: number;
|
readonly address_id: number;
|
||||||
address_line_1: string;
|
address_line_1: string;
|
||||||
address_line_2?: string | null;
|
address_line_2?: string | null;
|
||||||
city: string;
|
city: string;
|
||||||
@@ -661,14 +705,16 @@ export interface Address {
|
|||||||
country: string;
|
country: string;
|
||||||
latitude?: number | null;
|
latitude?: number | null;
|
||||||
longitude?: number | null;
|
longitude?: number | null;
|
||||||
location?: GeoJSONPoint | null;
|
readonly location?: GeoJSONPoint | null;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
updated_at: string;
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface FlyerLocation {
|
export interface FlyerLocation {
|
||||||
flyer_id: number;
|
readonly flyer_id: number;
|
||||||
store_location_id: number;
|
readonly store_location_id: number;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export enum AnalysisType {
|
export enum AnalysisType {
|
||||||
@@ -878,30 +924,30 @@ export interface MenuPlanShoppingListItem {
|
|||||||
* Returned by `getUnmatchedFlyerItems`.
|
* Returned by `getUnmatchedFlyerItems`.
|
||||||
*/
|
*/
|
||||||
export interface UnmatchedFlyerItem {
|
export interface UnmatchedFlyerItem {
|
||||||
unmatched_flyer_item_id: number;
|
readonly unmatched_flyer_item_id: number;
|
||||||
status: 'pending' | 'resolved' | 'ignored'; // 'resolved' is used instead of 'reviewed' from the DB for clarity
|
status: 'pending' | 'resolved' | 'ignored'; // 'resolved' is used instead of 'reviewed' from the DB for clarity
|
||||||
created_at: string; // Date string
|
readonly reviewed_at?: string | null;
|
||||||
updated_at: string;
|
readonly flyer_item_id: number;
|
||||||
reviewed_at?: string | null;
|
|
||||||
flyer_item_id: number;
|
|
||||||
flyer_item_name: string;
|
flyer_item_name: string;
|
||||||
price_display: string;
|
price_display: string;
|
||||||
flyer_id: number;
|
flyer_id: number;
|
||||||
store_name: string;
|
store_name: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Represents a user-defined budget for tracking grocery spending.
|
* Represents a user-defined budget for tracking grocery spending.
|
||||||
*/
|
*/
|
||||||
export interface Budget {
|
export interface Budget {
|
||||||
budget_id: number;
|
readonly budget_id: number;
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
name: string;
|
name: string;
|
||||||
amount_cents: number;
|
amount_cents: number;
|
||||||
period: 'weekly' | 'monthly';
|
period: 'weekly' | 'monthly';
|
||||||
start_date: string; // DATE
|
start_date: string; // DATE
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
updated_at: string;
|
readonly updated_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -918,21 +964,21 @@ export interface SpendingByCategory {
|
|||||||
* Represents a single defined achievement in the system.
|
* Represents a single defined achievement in the system.
|
||||||
*/
|
*/
|
||||||
export interface Achievement {
|
export interface Achievement {
|
||||||
achievement_id: number;
|
readonly achievement_id: number;
|
||||||
name: string;
|
name: string;
|
||||||
description: string;
|
description: string;
|
||||||
icon?: string | null;
|
icon?: string | null;
|
||||||
points_value: number;
|
points_value: number;
|
||||||
created_at: string;
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Represents an achievement that has been awarded to a user.
|
* Represents an achievement that has been awarded to a user.
|
||||||
*/
|
*/
|
||||||
export interface UserAchievement {
|
export interface UserAchievement {
|
||||||
user_id: string; // UUID
|
readonly user_id: string; // UUID
|
||||||
achievement_id: number;
|
readonly achievement_id: number;
|
||||||
achieved_at: string; // TIMESTAMPTZ
|
readonly achieved_at: string; // TIMESTAMPTZ
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -940,11 +986,11 @@ export interface UserAchievement {
|
|||||||
* Returned by the `getLeaderboard` database function.
|
* Returned by the `getLeaderboard` database function.
|
||||||
*/
|
*/
|
||||||
export interface LeaderboardUser {
|
export interface LeaderboardUser {
|
||||||
user_id: string;
|
readonly user_id: string;
|
||||||
full_name: string | null;
|
full_name: string | null;
|
||||||
avatar_url: string | null;
|
avatar_url: string | null;
|
||||||
points: number;
|
points: number;
|
||||||
rank: string; // RANK() returns a bigint, which the pg driver returns as a string.
|
readonly rank: string; // RANK() returns a bigint, which the pg driver returns as a string.
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -952,12 +998,12 @@ export interface LeaderboardUser {
|
|||||||
* This is a public-facing type and does not include sensitive fields.
|
* This is a public-facing type and does not include sensitive fields.
|
||||||
*/
|
*/
|
||||||
export interface AdminUserView {
|
export interface AdminUserView {
|
||||||
user_id: string;
|
readonly user_id: string;
|
||||||
email: string;
|
email: string;
|
||||||
created_at: string;
|
|
||||||
role: 'admin' | 'user';
|
role: 'admin' | 'user';
|
||||||
full_name: string | null;
|
full_name: string | null;
|
||||||
avatar_url: string | null;
|
avatar_url: string | null;
|
||||||
|
readonly created_at: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface PriceHistoryData {
|
export interface PriceHistoryData {
|
||||||
@@ -965,3 +1011,23 @@ export interface PriceHistoryData {
|
|||||||
price_in_cents: number;
|
price_in_cents: number;
|
||||||
date: string; // ISO date string
|
date: string; // ISO date string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface UserReaction {
|
||||||
|
readonly reaction_id: number;
|
||||||
|
readonly user_id: string; // UUID
|
||||||
|
readonly entity_type: string;
|
||||||
|
readonly entity_id: string;
|
||||||
|
reaction_type: string;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface UnitConversion {
|
||||||
|
readonly unit_conversion_id: number;
|
||||||
|
readonly master_item_id: number;
|
||||||
|
from_unit: string;
|
||||||
|
to_unit: string;
|
||||||
|
factor: number;
|
||||||
|
readonly created_at: string;
|
||||||
|
readonly updated_at: string;
|
||||||
|
}
|
||||||
|
|||||||
29
src/types/ai.ts
Normal file
29
src/types/ai.ts
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
// src/types/ai.ts
|
||||||
|
|
||||||
|
import { z } from 'zod';
|
||||||
|
|
||||||
|
// Helper for consistent required string validation (handles missing/null/empty)
|
||||||
|
// This is moved here as it's directly related to the schemas.
|
||||||
|
export const requiredString = (message: string) =>
|
||||||
|
z.preprocess((val) => val ?? '', z.string().min(1, message));
|
||||||
|
|
||||||
|
// --- Zod Schemas for AI Response Validation ---
|
||||||
|
// These schemas define the expected structure of data returned by the AI.
|
||||||
|
// They are used for validation and type inference across multiple services.
|
||||||
|
|
||||||
|
export const ExtractedFlyerItemSchema = z.object({
|
||||||
|
item: z.string().nullable(),
|
||||||
|
price_display: z.string().nullable(),
|
||||||
|
price_in_cents: z.number().nullable(),
|
||||||
|
quantity: z.string().nullable(),
|
||||||
|
category_name: z.string().nullable(),
|
||||||
|
master_item_id: z.number().nullish(), // .nullish() allows null or undefined
|
||||||
|
});
|
||||||
|
|
||||||
|
export const AiFlyerDataSchema = z.object({
|
||||||
|
store_name: z.string().nullable(),
|
||||||
|
valid_from: z.string().nullable(),
|
||||||
|
valid_to: z.string().nullable(),
|
||||||
|
store_address: z.string().nullable(),
|
||||||
|
items: z.array(ExtractedFlyerItemSchema),
|
||||||
|
});
|
||||||
Reference in New Issue
Block a user