Compare commits
6 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4b0a172c35 | ||
| e8c894d5cf | |||
| 6c8fd4b126 | |||
|
|
a1f52544d0 | ||
| 2334359756 | |||
| 406954ca06 |
@@ -6,4 +6,4 @@
|
||||
"tabWidth": 2,
|
||||
"useTabs": false,
|
||||
"endOfLine": "auto"
|
||||
}
|
||||
}
|
||||
|
||||
130
README.md
130
README.md
@@ -2,7 +2,7 @@
|
||||
|
||||
Flyer Crawler is a web application that uses the Google Gemini AI to extract, analyze, and manage data from grocery store flyers. Users can upload flyer images or PDFs, and the application will automatically identify items, prices, and sale dates, storing the structured data in a PostgreSQL database for historical analysis, price tracking, and personalized deal alerts.
|
||||
|
||||
We are working on an app to help people save money, by finding good deals that are only advertized in store flyers/ads. So, the primary purpose of the site is to make uploading flyers as easy as possible and as accurate as possible, and to store peoples needs, so sales can be matched to needs.
|
||||
We are working on an app to help people save money, by finding good deals that are only advertized in store flyers/ads. So, the primary purpose of the site is to make uploading flyers as easy as possible and as accurate as possible, and to store peoples needs, so sales can be matched to needs.
|
||||
|
||||
## Features
|
||||
|
||||
@@ -45,9 +45,9 @@ This project is configured to run in a CI/CD environment and does not use `.env`
|
||||
|
||||
1. **Set up a PostgreSQL database instance.**
|
||||
2. **Run the Database Schema**:
|
||||
- Connect to your database using a tool like `psql` or DBeaver.
|
||||
- Open `sql/schema.sql.txt`, copy its entire contents, and execute it against your database.
|
||||
- This will create all necessary tables, functions, and relationships.
|
||||
- Connect to your database using a tool like `psql` or DBeaver.
|
||||
- Open `sql/schema.sql.txt`, copy its entire contents, and execute it against your database.
|
||||
- This will create all necessary tables, functions, and relationships.
|
||||
|
||||
### Step 2: Install Dependencies and Run the Application
|
||||
|
||||
@@ -79,11 +79,11 @@ sudo nano /etc/nginx/mime.types
|
||||
|
||||
change
|
||||
|
||||
application/javascript js;
|
||||
application/javascript js;
|
||||
|
||||
TO
|
||||
|
||||
application/javascript js mjs;
|
||||
application/javascript js mjs;
|
||||
|
||||
RESTART NGINX
|
||||
|
||||
@@ -95,7 +95,7 @@ actually the proper change was to do this in the /etc/nginx/sites-available/flye
|
||||
## for OAuth
|
||||
|
||||
1. Get Google OAuth Credentials
|
||||
This is a crucial step that you must do outside the codebase:
|
||||
This is a crucial step that you must do outside the codebase:
|
||||
|
||||
Go to the Google Cloud Console.
|
||||
|
||||
@@ -112,7 +112,7 @@ Under Authorized redirect URIs, click ADD URI and enter the URL where Google wil
|
||||
Click Create. You will be given a Client ID and a Client Secret.
|
||||
|
||||
2. Get GitHub OAuth Credentials
|
||||
You'll need to obtain a Client ID and Client Secret from GitHub:
|
||||
You'll need to obtain a Client ID and Client Secret from GitHub:
|
||||
|
||||
Go to your GitHub profile settings.
|
||||
|
||||
@@ -133,21 +133,23 @@ You will be given a Client ID and a Client Secret.
|
||||
|
||||
psql -h localhost -U flyer_crawler_user -d "flyer-crawler-prod" -W
|
||||
|
||||
|
||||
## postgis
|
||||
|
||||
flyer-crawler-prod=> SELECT version();
|
||||
version
|
||||
------------------------------------------------------------------------------------------------------------------------------------------
|
||||
PostgreSQL 14.19 (Ubuntu 14.19-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04.2) 11.4.0, 64-bit
|
||||
version
|
||||
|
||||
---
|
||||
|
||||
PostgreSQL 14.19 (Ubuntu 14.19-0ubuntu0.22.04.1) on x86_64-pc-linux-gnu, compiled by gcc (Ubuntu 11.4.0-1ubuntu1~22.04.2) 11.4.0, 64-bit
|
||||
(1 row)
|
||||
|
||||
flyer-crawler-prod=> SELECT PostGIS_Full_Version();
|
||||
postgis_full_version
|
||||
--------------------------------------------------------------------------------------------------------------------------------------------------------------------
|
||||
POSTGIS="3.2.0 c3e3cc0" [EXTENSION] PGSQL="140" GEOS="3.10.2-CAPI-1.16.0" PROJ="8.2.1" LIBXML="2.9.12" LIBJSON="0.15" LIBPROTOBUF="1.3.3" WAGYU="0.5.0 (Internal)"
|
||||
(1 row)
|
||||
postgis_full_version
|
||||
|
||||
---
|
||||
|
||||
POSTGIS="3.2.0 c3e3cc0" [EXTENSION] PGSQL="140" GEOS="3.10.2-CAPI-1.16.0" PROJ="8.2.1" LIBXML="2.9.12" LIBJSON="0.15" LIBPROTOBUF="1.3.3" WAGYU="0.5.0 (Internal)"
|
||||
(1 row)
|
||||
|
||||
## production postgres setup
|
||||
|
||||
@@ -201,9 +203,13 @@ Step 4: Seed the Admin Account (If Needed)
|
||||
Your application has a separate script to create the initial admin user. To run it, you must first set the required environment variables in your shell session.
|
||||
|
||||
bash
|
||||
|
||||
# Set variables for the current session
|
||||
|
||||
export DB_USER=flyer_crawler_user DB_PASSWORD=your_password DB_NAME="flyer-crawler-prod" ...
|
||||
|
||||
# Run the seeding script
|
||||
|
||||
npx tsx src/db/seed_admin_account.ts
|
||||
Your production database is now ready!
|
||||
|
||||
@@ -284,8 +290,6 @@ Test Execution: Your tests run against this clean, isolated schema.
|
||||
|
||||
This approach is faster, more reliable, and removes the need for sudo access within the CI pipeline.
|
||||
|
||||
|
||||
|
||||
gitea-runner@projectium:~$ pm2 install pm2-logrotate
|
||||
[PM2][Module] Installing NPM pm2-logrotate module
|
||||
[PM2][Module] Calling [NPM] to install pm2-logrotate ...
|
||||
@@ -293,7 +297,7 @@ gitea-runner@projectium:~$ pm2 install pm2-logrotate
|
||||
added 161 packages in 5s
|
||||
|
||||
21 packages are looking for funding
|
||||
run `npm fund` for details
|
||||
run `npm fund` for details
|
||||
npm notice
|
||||
npm notice New patch version of npm available! 11.6.3 -> 11.6.4
|
||||
npm notice Changelog: https://github.com/npm/cli/releases/tag/v11.6.4
|
||||
@@ -308,23 +312,23 @@ $ pm2 set pm2-logrotate:retain 30
|
||||
$ pm2 set pm2-logrotate:compress false
|
||||
$ pm2 set pm2-logrotate:dateFormat YYYY-MM-DD_HH-mm-ss
|
||||
$ pm2 set pm2-logrotate:workerInterval 30
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 * * *
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 \* \* _
|
||||
$ pm2 set pm2-logrotate:rotateModule true
|
||||
Modules configuration. Copy/Paste line to edit values.
|
||||
[PM2][Module] Module successfully installed and launched
|
||||
[PM2][Module] Checkout module options: `$ pm2 conf`
|
||||
┌────┬───────────────────────────────────┬─────────────┬─────────┬─────────┬──────────┬────────┬──────┬───────────┬──────────┬──────────┬──────────┬──────────┐
|
||||
│ id │ name │ namespace │ version │ mode │ pid │ uptime │ ↺ │ status │ cpu │ mem │ user │ watching │
|
||||
│ id │ name │ namespace │ version │ mode │ pid │ uptime │ ↺ │ status │ cpu │ mem │ user │ watching │
|
||||
├────┼───────────────────────────────────┼─────────────┼─────────┼─────────┼──────────┼────────┼──────┼───────────┼──────────┼──────────┼──────────┼──────────┤
|
||||
│ 2 │ flyer-crawler-analytics-worker │ default │ 0.0.0 │ fork │ 3846981 │ 7m │ 5 │ online │ 0% │ 55.8mb │ git… │ disabled │
|
||||
│ 11 │ flyer-crawler-api │ default │ 0.0.0 │ fork │ 3846987 │ 7m │ 0 │ online │ 0% │ 59.0mb │ git… │ disabled │
|
||||
│ 12 │ flyer-crawler-worker │ default │ 0.0.0 │ fork │ 3846988 │ 7m │ 0 │ online │ 0% │ 54.2mb │ git… │ disabled │
|
||||
│ 2 │ flyer-crawler-analytics-worker │ default │ 0.0.0 │ fork │ 3846981 │ 7m │ 5 │ online │ 0% │ 55.8mb │ git… │ disabled │
|
||||
│ 11 │ flyer-crawler-api │ default │ 0.0.0 │ fork │ 3846987 │ 7m │ 0 │ online │ 0% │ 59.0mb │ git… │ disabled │
|
||||
│ 12 │ flyer-crawler-worker │ default │ 0.0.0 │ fork │ 3846988 │ 7m │ 0 │ online │ 0% │ 54.2mb │ git… │ disabled │
|
||||
└────┴───────────────────────────────────┴─────────────┴─────────┴─────────┴──────────┴────────┴──────┴───────────┴──────────┴──────────┴──────────┴──────────┘
|
||||
Module
|
||||
┌────┬──────────────────────────────┬───────────────┬──────────┬──────────┬──────┬──────────┬──────────┬──────────┐
|
||||
│ id │ module │ version │ pid │ status │ ↺ │ cpu │ mem │ user │
|
||||
│ id │ module │ version │ pid │ status │ ↺ │ cpu │ mem │ user │
|
||||
├────┼──────────────────────────────┼───────────────┼──────────┼──────────┼──────┼──────────┼──────────┼──────────┤
|
||||
│ 13 │ pm2-logrotate │ 3.0.0 │ 3848878 │ online │ 0 │ 0% │ 20.1mb │ git… │
|
||||
│ 13 │ pm2-logrotate │ 3.0.0 │ 3848878 │ online │ 0 │ 0% │ 20.1mb │ git… │
|
||||
└────┴──────────────────────────────┴───────────────┴──────────┴──────────┴──────┴──────────┴──────────┴──────────┘
|
||||
gitea-runner@projectium:~$ pm2 set pm2-logrotate:max_size 10M
|
||||
[PM2] Module pm2-logrotate restarted
|
||||
@@ -335,7 +339,7 @@ $ pm2 set pm2-logrotate:retain 30
|
||||
$ pm2 set pm2-logrotate:compress false
|
||||
$ pm2 set pm2-logrotate:dateFormat YYYY-MM-DD_HH-mm-ss
|
||||
$ pm2 set pm2-logrotate:workerInterval 30
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 * * *
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 _ \* _
|
||||
$ pm2 set pm2-logrotate:rotateModule true
|
||||
gitea-runner@projectium:~$ pm2 set pm2-logrotate:retain 14
|
||||
[PM2] Module pm2-logrotate restarted
|
||||
@@ -346,33 +350,31 @@ $ pm2 set pm2-logrotate:retain 14
|
||||
$ pm2 set pm2-logrotate:compress false
|
||||
$ pm2 set pm2-logrotate:dateFormat YYYY-MM-DD_HH-mm-ss
|
||||
$ pm2 set pm2-logrotate:workerInterval 30
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 * * *
|
||||
$ pm2 set pm2-logrotate:rotateInterval 0 0 _ \* \*
|
||||
$ pm2 set pm2-logrotate:rotateModule true
|
||||
gitea-runner@projectium:~$
|
||||
|
||||
|
||||
|
||||
|
||||
## dev server setup:
|
||||
|
||||
Here are the steps to set up the development environment on Windows using Podman with an Ubuntu container:
|
||||
|
||||
1. Install Prerequisites on Windows
|
||||
Install WSL 2: Podman on Windows relies on the Windows Subsystem for Linux. Install it by running wsl --install in an administrator PowerShell.
|
||||
Install Podman Desktop: Download and install Podman Desktop for Windows.
|
||||
Install WSL 2: Podman on Windows relies on the Windows Subsystem for Linux. Install it by running wsl --install in an administrator PowerShell.
|
||||
Install Podman Desktop: Download and install Podman Desktop for Windows.
|
||||
|
||||
2. Set Up Podman
|
||||
Initialize Podman: Launch Podman Desktop. It will automatically set up its WSL 2 machine.
|
||||
Start Podman: Ensure the Podman machine is running from the Podman Desktop interface.
|
||||
Initialize Podman: Launch Podman Desktop. It will automatically set up its WSL 2 machine.
|
||||
Start Podman: Ensure the Podman machine is running from the Podman Desktop interface.
|
||||
|
||||
3. Set Up the Ubuntu Container
|
||||
- Pull Ubuntu Image: Open a PowerShell or command prompt and pull the latest Ubuntu image:
|
||||
podman pull ubuntu:latest
|
||||
- Create a Podman Volume: Create a volume to persist node_modules and avoid installing them every time the container starts.
|
||||
podman volume create node_modules_cache
|
||||
- Run the Ubuntu Container: Start a new container with the project directory mounted and the necessary ports forwarded.
|
||||
- Open a terminal in your project's root directory on Windows.
|
||||
- Run the following command, replacing D:\gitea\flyer-crawler.projectium.com\flyer-crawler.projectium.com with the full path to your project:
|
||||
|
||||
- Pull Ubuntu Image: Open a PowerShell or command prompt and pull the latest Ubuntu image:
|
||||
podman pull ubuntu:latest
|
||||
- Create a Podman Volume: Create a volume to persist node_modules and avoid installing them every time the container starts.
|
||||
podman volume create node_modules_cache
|
||||
- Run the Ubuntu Container: Start a new container with the project directory mounted and the necessary ports forwarded.
|
||||
- Open a terminal in your project's root directory on Windows.
|
||||
- Run the following command, replacing D:\gitea\flyer-crawler.projectium.com\flyer-crawler.projectium.com with the full path to your project:
|
||||
|
||||
podman run -it -p 3001:3001 -p 5173:5173 --name flyer-dev -v "D:\gitea\flyer-crawler.projectium.com\flyer-crawler.projectium.com:/app" -v "node_modules_cache:/app/node_modules" ubuntu:latest
|
||||
|
||||
@@ -383,46 +385,40 @@ podman run -it -p 3001:3001 -p 5173:5173 --name flyer-dev -v "D:\gitea\flyer-cra
|
||||
-v "node_modules_cache:/app/node_modules": Mounts the named volume for node_modules.
|
||||
|
||||
4. Configure the Ubuntu Environment
|
||||
You are now inside the Ubuntu container's shell.
|
||||
You are now inside the Ubuntu container's shell.
|
||||
|
||||
- Update Package Lists:
|
||||
apt-get update
|
||||
- Install Dependencies: Install curl, git, and nodejs (which includes npm).
|
||||
apt-get install -y curl git
|
||||
curl -sL https://deb.nodesource.com/setup_20.x | bash -
|
||||
apt-get install -y nodejs
|
||||
- Navigate to Project Directory:
|
||||
cd /app
|
||||
- Update Package Lists:
|
||||
apt-get update
|
||||
- Install Dependencies: Install curl, git, and nodejs (which includes npm).
|
||||
apt-get install -y curl git
|
||||
curl -sL https://deb.nodesource.com/setup_20.x | bash -
|
||||
apt-get install -y nodejs
|
||||
- Navigate to Project Directory:
|
||||
cd /app
|
||||
|
||||
- Install Project Dependencies:
|
||||
npm install
|
||||
- Install Project Dependencies:
|
||||
npm install
|
||||
|
||||
5. Run the Development Server
|
||||
- Start the Application:
|
||||
npm run dev
|
||||
npm run dev
|
||||
|
||||
6. Accessing the Application
|
||||
- Frontend: Open your browser and go to http://localhost:5173.
|
||||
- Backend: The frontend will make API calls to http://localhost:3001.
|
||||
|
||||
- Frontend: Open your browser and go to http://localhost:5173.
|
||||
- Backend: The frontend will make API calls to http://localhost:3001.
|
||||
|
||||
Managing the Environment
|
||||
- Stopping the Container: Press Ctrl+C in the container terminal, then type exit.
|
||||
- Restarting the Container:
|
||||
podman start -a -i flyer-dev
|
||||
|
||||
|
||||
- Stopping the Container: Press Ctrl+C in the container terminal, then type exit.
|
||||
- Restarting the Container:
|
||||
podman start -a -i flyer-dev
|
||||
|
||||
## for me:
|
||||
|
||||
cd /mnt/d/gitea/flyer-crawler.projectium.com/flyer-crawler.projectium.com
|
||||
podman run -it -p 3001:3001 -p 5173:5173 --name flyer-dev -v "$(pwd):/app" -v "node_modules_cache:/app/node_modules" ubuntu:latest
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
rate limiting
|
||||
|
||||
respect the AI service's rate limits, making it more stable and robust. You can adjust the GEMINI_RPM environment variable in your production environment as needed without changing the code.
|
||||
respect the AI service's rate limits, making it more stable and robust. You can adjust the GEMINI_RPM environment variable in your production environment as needed without changing the code.
|
||||
|
||||
@@ -15,7 +15,7 @@ module.exports = {
|
||||
args: 'server.ts', // tsx will execute this file
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production', // Set the Node.js environment to production
|
||||
NODE_ENV: 'production', // Set the Node.js environment to production
|
||||
name: 'flyer-crawler-api',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
},
|
||||
@@ -63,4 +63,4 @@ module.exports = {
|
||||
},
|
||||
},
|
||||
],
|
||||
};
|
||||
};
|
||||
|
||||
@@ -1,21 +1,21 @@
|
||||
import globals from "globals";
|
||||
import tseslint from "typescript-eslint";
|
||||
import pluginReact from "eslint-plugin-react";
|
||||
import pluginReactHooks from "eslint-plugin-react-hooks";
|
||||
import pluginReactRefresh from "eslint-plugin-react-refresh";
|
||||
import globals from 'globals';
|
||||
import tseslint from 'typescript-eslint';
|
||||
import pluginReact from 'eslint-plugin-react';
|
||||
import pluginReactHooks from 'eslint-plugin-react-hooks';
|
||||
import pluginReactRefresh from 'eslint-plugin-react-refresh';
|
||||
|
||||
export default tseslint.config(
|
||||
{
|
||||
// Global ignores
|
||||
ignores: ["dist", ".gitea", "node_modules", "*.cjs"],
|
||||
ignores: ['dist', '.gitea', 'node_modules', '*.cjs'],
|
||||
},
|
||||
{
|
||||
// All files
|
||||
files: ["**/*.{js,mjs,cjs,ts,jsx,tsx}"],
|
||||
files: ['**/*.{js,mjs,cjs,ts,jsx,tsx}'],
|
||||
plugins: {
|
||||
react: pluginReact,
|
||||
"react-hooks": pluginReactHooks,
|
||||
"react-refresh": pluginReactRefresh,
|
||||
'react-hooks': pluginReactHooks,
|
||||
'react-refresh': pluginReactRefresh,
|
||||
},
|
||||
languageOptions: {
|
||||
globals: {
|
||||
@@ -24,12 +24,9 @@ export default tseslint.config(
|
||||
},
|
||||
},
|
||||
rules: {
|
||||
"react-refresh/only-export-components": [
|
||||
"warn",
|
||||
{ allowConstantExport: true },
|
||||
],
|
||||
'react-refresh/only-export-components': ['warn', { allowConstantExport: true }],
|
||||
},
|
||||
},
|
||||
// TypeScript files
|
||||
...tseslint.configs.recommended,
|
||||
);
|
||||
);
|
||||
|
||||
4
express.d.ts
vendored
4
express.d.ts
vendored
@@ -1,4 +1,4 @@
|
||||
// src/types/express.d.ts
|
||||
// express.d.ts
|
||||
import { Logger } from 'pino';
|
||||
|
||||
/**
|
||||
@@ -12,4 +12,4 @@ declare global {
|
||||
log: Logger;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
38
index.html
38
index.html
@@ -1,20 +1,20 @@
|
||||
<!DOCTYPE html>
|
||||
<!doctype html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>Grocery Flyer AI Analyzer</title>
|
||||
<style>
|
||||
@import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;500;600;700&display=swap');
|
||||
body {
|
||||
font-family: 'Inter', sans-serif;
|
||||
}
|
||||
</style>
|
||||
<!-- The stylesheet will be injected here by Vite during the build process -->
|
||||
</head>
|
||||
<body>
|
||||
<div id="root"></div>
|
||||
<!-- Vite will inject the correct <script> tag here during the build process -->
|
||||
<script type="module" src="/src/index.tsx"></script>
|
||||
</body>
|
||||
</html>
|
||||
<head>
|
||||
<meta charset="UTF-8" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>Grocery Flyer AI Analyzer</title>
|
||||
<style>
|
||||
@import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;500;600;700&display=swap');
|
||||
body {
|
||||
font-family: 'Inter', sans-serif;
|
||||
}
|
||||
</style>
|
||||
<!-- The stylesheet will be injected here by Vite during the build process -->
|
||||
</head>
|
||||
<body>
|
||||
<div id="root"></div>
|
||||
<!-- Vite will inject the correct <script> tag here during the build process -->
|
||||
<script type="module" src="/src/index.tsx"></script>
|
||||
</body>
|
||||
</html>
|
||||
|
||||
@@ -1,8 +1,5 @@
|
||||
{
|
||||
"name": "Flyer Crawler",
|
||||
"description": "Upload a grocery store flyer image to extract item details, prices, and quantities using AI. Get insights, meal plans, and compare prices to save money on your shopping.",
|
||||
"requestFramePermissions": [
|
||||
"geolocation",
|
||||
"microphone"
|
||||
]
|
||||
}
|
||||
"requestFramePermissions": ["geolocation", "microphone"]
|
||||
}
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.0.19",
|
||||
"version": "0.0.21",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.0.19",
|
||||
"version": "0.0.21",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.0.19",
|
||||
"version": "0.0.21",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -10,10 +10,13 @@ const tailwindConfigPath = path.resolve(process.cwd(), 'tailwind.config.js');
|
||||
console.log(`[POSTCSS] Attempting to use Tailwind config at: ${tailwindConfigPath}`);
|
||||
|
||||
// Log to prove the imported config object is what we expect
|
||||
console.log('[POSTCSS] Imported tailwind.config.js object:', JSON.stringify(tailwindConfig, null, 2));
|
||||
console.log(
|
||||
'[POSTCSS] Imported tailwind.config.js object:',
|
||||
JSON.stringify(tailwindConfig, null, 2),
|
||||
);
|
||||
|
||||
export default {
|
||||
plugins: {
|
||||
'@tailwindcss/postcss': {}, // The empty object is correct.
|
||||
},
|
||||
};
|
||||
};
|
||||
|
||||
@@ -1030,11 +1030,61 @@ DROP FUNCTION IF EXISTS public.fork_recipe(UUID, BIGINT);
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.fork_recipe(p_user_id UUID, p_original_recipe_id BIGINT)
|
||||
RETURNS SETOF public.recipes
|
||||
LANGUAGE sql
|
||||
LANGUAGE plpgsql
|
||||
SECURITY INVOKER
|
||||
AS $$
|
||||
-- The entire forking logic is now encapsulated in a single, atomic database function.
|
||||
SELECT * FROM public.fork_recipe(p_user_id, p_original_recipe_id);
|
||||
DECLARE
|
||||
new_recipe_id BIGINT;
|
||||
BEGIN
|
||||
-- 1. Create a copy of the recipe, linking it to the new user and the original recipe.
|
||||
INSERT INTO public.recipes (
|
||||
user_id,
|
||||
original_recipe_id,
|
||||
name,
|
||||
description,
|
||||
instructions,
|
||||
prep_time_minutes,
|
||||
cook_time_minutes,
|
||||
servings,
|
||||
photo_url,
|
||||
calories_per_serving,
|
||||
protein_grams,
|
||||
fat_grams,
|
||||
carb_grams,
|
||||
status -- Forked recipes should be private by default
|
||||
)
|
||||
SELECT
|
||||
p_user_id,
|
||||
p_original_recipe_id,
|
||||
original.name || ' (Fork)', -- Append '(Fork)' to distinguish it
|
||||
original.description,
|
||||
original.instructions,
|
||||
original.prep_time_minutes,
|
||||
original.cook_time_minutes,
|
||||
original.servings,
|
||||
original.photo_url,
|
||||
original.calories_per_serving,
|
||||
original.protein_grams,
|
||||
original.fat_grams,
|
||||
original.carb_grams,
|
||||
'private'
|
||||
FROM public.recipes AS original
|
||||
WHERE original.recipe_id = p_original_recipe_id
|
||||
RETURNING recipe_id INTO new_recipe_id;
|
||||
|
||||
-- If the original recipe didn't exist, new_recipe_id will be null.
|
||||
IF new_recipe_id IS NULL THEN
|
||||
RETURN;
|
||||
END IF;
|
||||
|
||||
-- 2. Copy all ingredients, tags, and appliances from the original recipe to the new one.
|
||||
INSERT INTO public.recipe_ingredients (recipe_id, master_item_id, quantity, unit) SELECT new_recipe_id, master_item_id, quantity, unit FROM public.recipe_ingredients WHERE recipe_id = p_original_recipe_id;
|
||||
INSERT INTO public.recipe_tags (recipe_id, tag_id) SELECT new_recipe_id, tag_id FROM public.recipe_tags WHERE recipe_id = p_original_recipe_id;
|
||||
INSERT INTO public.recipe_appliances (recipe_id, appliance_id) SELECT new_recipe_id, appliance_id FROM public.recipe_appliances WHERE recipe_id = p_original_recipe_id;
|
||||
|
||||
-- 3. Return the newly created recipe record.
|
||||
RETURN QUERY SELECT * FROM public.recipes WHERE recipe_id = new_recipe_id;
|
||||
END;
|
||||
$$;
|
||||
|
||||
|
||||
@@ -1566,4 +1616,3 @@ BEGIN
|
||||
bp.price_rank = 1;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
|
||||
@@ -8,7 +8,23 @@
|
||||
-- It is idempotent, meaning it can be run multiple times without causing errors.
|
||||
|
||||
-- 1. Pre-populate the master grocery items dictionary.
|
||||
-- This block links generic items to their respective categories.
|
||||
-- This MUST run after populating categories.
|
||||
-- Renumbered to 2.
|
||||
|
||||
-- 2. Pre-populate the categories table from a predefined list.
|
||||
-- Renumbered to 1. This MUST run before populating master_grocery_items.
|
||||
DO $$
|
||||
BEGIN
|
||||
INSERT INTO public.categories (name) VALUES
|
||||
('Fruits & Vegetables'), ('Meat & Seafood'), ('Dairy & Eggs'), ('Bakery & Bread'),
|
||||
('Pantry & Dry Goods'), ('Beverages'), ('Frozen Foods'), ('Snacks'), ('Household & Cleaning'),
|
||||
('Personal Care & Health'), ('Baby & Child'), ('Pet Supplies'), ('Deli & Prepared Foods'),
|
||||
('Canned Goods'), ('Condiments & Spices'), ('Breakfast & Cereal'), ('Organic'),
|
||||
('International Foods'), ('Other/Miscellaneous')
|
||||
ON CONFLICT (name) DO NOTHING;
|
||||
END $$;
|
||||
|
||||
-- 2. Pre-populate the master grocery items dictionary.
|
||||
DO $$
|
||||
DECLARE
|
||||
fv_cat_id BIGINT; ms_cat_id BIGINT; de_cat_id BIGINT; bb_cat_id BIGINT; pdg_cat_id BIGINT;
|
||||
@@ -53,18 +69,6 @@ BEGIN
|
||||
ON CONFLICT (name) DO NOTHING;
|
||||
END $$;
|
||||
|
||||
-- 2. Pre-populate the categories table from a predefined list.
|
||||
DO $$
|
||||
BEGIN
|
||||
INSERT INTO public.categories (name) VALUES
|
||||
('Fruits & Vegetables'), ('Meat & Seafood'), ('Dairy & Eggs'), ('Bakery & Bread'),
|
||||
('Pantry & Dry Goods'), ('Beverages'), ('Frozen Foods'), ('Snacks'), ('Household & Cleaning'),
|
||||
('Personal Care & Health'), ('Baby & Child'), ('Pet Supplies'), ('Deli & Prepared Foods'),
|
||||
('Canned Goods'), ('Condiments & Spices'), ('Breakfast & Cereal'), ('Organic'),
|
||||
('International Foods'), ('Other/Miscellaneous')
|
||||
ON CONFLICT (name) DO NOTHING;
|
||||
END $$;
|
||||
|
||||
-- 3. Pre-populate the brands and products tables.
|
||||
-- This block adds common brands and links them to specific products.
|
||||
DO $$
|
||||
|
||||
@@ -162,6 +162,38 @@ CREATE TABLE IF NOT EXISTS public.user_watched_items (
|
||||
COMMENT ON TABLE public.user_watched_items IS 'A linking table that represents a user''s personal watchlist of grocery items.';
|
||||
CREATE INDEX IF NOT EXISTS idx_user_watched_items_master_item_id ON public.user_watched_items(master_item_id);
|
||||
|
||||
-- 23. Store brand information. (Moved up due to dependency in flyer_items)
|
||||
CREATE TABLE IF NOT EXISTS public.brands (
|
||||
brand_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
name TEXT NOT NULL UNIQUE,
|
||||
logo_url TEXT,
|
||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||
|
||||
-- 24. For specific products, linking a master item with a brand and size. (Moved up due to dependency in flyer_items)
|
||||
CREATE TABLE IF NOT EXISTS public.products (
|
||||
product_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||
brand_id BIGINT REFERENCES public.brands(brand_id) ON DELETE SET NULL,
|
||||
name TEXT NOT NULL,
|
||||
description TEXT,
|
||||
size TEXT,
|
||||
upc_code TEXT UNIQUE,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||
COMMENT ON COLUMN public.products.brand_id IS 'Can be null for generic/store-brand items.';
|
||||
COMMENT ON COLUMN public.products.name IS 'Prime Raised without Antibiotics Chicken Breast.';
|
||||
COMMENT ON COLUMN public.products.size IS 'e.g., "4L", "500g".';
|
||||
CREATE INDEX IF NOT EXISTS idx_products_master_item_id ON public.products(master_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_products_brand_id ON public.products(brand_id);
|
||||
|
||||
-- 9. The 'flyer_items' table. This stores individual items from flyers.
|
||||
CREATE TABLE IF NOT EXISTS public.flyer_items (
|
||||
flyer_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
@@ -409,38 +441,6 @@ COMMENT ON TABLE public.unmatched_flyer_items IS 'A queue for reviewing flyer it
|
||||
CREATE INDEX IF NOT EXISTS idx_unmatched_flyer_items_flyer_item_id ON public.unmatched_flyer_items(flyer_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_unmatched_flyer_items_pending ON public.unmatched_flyer_items (created_at) WHERE status = 'pending';
|
||||
|
||||
-- 23. Store brand information.
|
||||
CREATE TABLE IF NOT EXISTS public.brands (
|
||||
brand_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
name TEXT NOT NULL UNIQUE,
|
||||
logo_url TEXT,
|
||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||
|
||||
-- 24. For specific products, linking a master item with a brand and size.
|
||||
CREATE TABLE IF NOT EXISTS public.products (
|
||||
product_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||
brand_id BIGINT REFERENCES public.brands(brand_id) ON DELETE SET NULL,
|
||||
name TEXT NOT NULL,
|
||||
description TEXT,
|
||||
size TEXT,
|
||||
upc_code TEXT UNIQUE,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||
COMMENT ON COLUMN public.products.brand_id IS 'Can be null for generic/store-brand items.';
|
||||
COMMENT ON COLUMN public.products.name IS 'Prime Raised without Antibiotics Chicken Breast.';
|
||||
COMMENT ON COLUMN public.products.size IS 'e.g., "4L", "500g".';
|
||||
CREATE INDEX IF NOT EXISTS idx_products_master_item_id ON public.products(master_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_products_brand_id ON public.products(brand_id);
|
||||
|
||||
-- 25. Linking table for when one flyer is valid for multiple locations.
|
||||
CREATE TABLE IF NOT EXISTS public.flyer_locations (
|
||||
flyer_id BIGINT NOT NULL REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
||||
|
||||
@@ -178,6 +178,38 @@ CREATE TABLE IF NOT EXISTS public.user_watched_items (
|
||||
COMMENT ON TABLE public.user_watched_items IS 'A linking table that represents a user''s personal watchlist of grocery items.';
|
||||
CREATE INDEX IF NOT EXISTS idx_user_watched_items_master_item_id ON public.user_watched_items(master_item_id);
|
||||
|
||||
-- 23. Store brand information. (Moved up due to dependency in flyer_items)
|
||||
CREATE TABLE IF NOT EXISTS public.brands (
|
||||
brand_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
name TEXT NOT NULL UNIQUE,
|
||||
logo_url TEXT,
|
||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||
|
||||
-- 24. For specific products, linking a master item with a brand and size. (Moved up due to dependency in flyer_items)
|
||||
CREATE TABLE IF NOT EXISTS public.products (
|
||||
product_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||
brand_id BIGINT REFERENCES public.brands(brand_id) ON DELETE SET NULL,
|
||||
name TEXT NOT NULL,
|
||||
description TEXT,
|
||||
size TEXT,
|
||||
upc_code TEXT UNIQUE,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||
COMMENT ON COLUMN public.products.brand_id IS 'Can be null for generic/store-brand items.';
|
||||
COMMENT ON COLUMN public.products.name IS 'Prime Raised without Antibiotics Chicken Breast.';
|
||||
COMMENT ON COLUMN public.products.size IS 'e.g., "4L", "500g".';
|
||||
CREATE INDEX IF NOT EXISTS idx_products_master_item_id ON public.products(master_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_products_brand_id ON public.products(brand_id);
|
||||
|
||||
-- 9. The 'flyer_items' table. This stores individual items from flyers.
|
||||
CREATE TABLE IF NOT EXISTS public.flyer_items (
|
||||
flyer_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
@@ -424,38 +456,6 @@ COMMENT ON TABLE public.unmatched_flyer_items IS 'A queue for reviewing flyer it
|
||||
CREATE INDEX IF NOT EXISTS idx_unmatched_flyer_items_flyer_item_id ON public.unmatched_flyer_items(flyer_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_unmatched_flyer_items_pending ON public.unmatched_flyer_items (created_at) WHERE status = 'pending';
|
||||
|
||||
-- 23. Store brand information.
|
||||
CREATE TABLE IF NOT EXISTS public.brands (
|
||||
brand_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
name TEXT NOT NULL UNIQUE,
|
||||
logo_url TEXT,
|
||||
store_id BIGINT REFERENCES public.stores(store_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.brands IS 'Stores brand names like "Coca-Cola", "Maple Leaf", or "Kraft".';
|
||||
COMMENT ON COLUMN public.brands.store_id IS 'If this is a store-specific brand (e.g., President''s Choice), this links to the parent store.';
|
||||
|
||||
-- 24. For specific products, linking a master item with a brand and size.
|
||||
CREATE TABLE IF NOT EXISTS public.products (
|
||||
product_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
master_item_id BIGINT NOT NULL REFERENCES public.master_grocery_items(master_grocery_item_id) ON DELETE CASCADE,
|
||||
brand_id BIGINT REFERENCES public.brands(brand_id) ON DELETE SET NULL,
|
||||
name TEXT NOT NULL,
|
||||
description TEXT,
|
||||
size TEXT,
|
||||
upc_code TEXT UNIQUE,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
updated_at TIMESTAMPTZ DEFAULT now() NOT NULL
|
||||
);
|
||||
COMMENT ON TABLE public.products IS 'Represents a specific, sellable product, combining a generic item with a brand and size.';
|
||||
COMMENT ON COLUMN public.products.upc_code IS 'Universal Product Code, if available, for exact product matching.';
|
||||
COMMENT ON COLUMN public.products.brand_id IS 'Can be null for generic/store-brand items.';
|
||||
COMMENT ON COLUMN public.products.name IS 'Prime Raised without Antibiotics Chicken Breast.';
|
||||
COMMENT ON COLUMN public.products.size IS 'e.g., "4L", "500g".';
|
||||
CREATE INDEX IF NOT EXISTS idx_products_master_item_id ON public.products(master_item_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_products_brand_id ON public.products(brand_id);
|
||||
|
||||
-- 25. Linking table for when one flyer is valid for multiple locations.
|
||||
CREATE TABLE IF NOT EXISTS public.flyer_locations (
|
||||
flyer_id BIGINT NOT NULL REFERENCES public.flyers(flyer_id) ON DELETE CASCADE,
|
||||
@@ -2128,11 +2128,61 @@ DROP FUNCTION IF EXISTS public.fork_recipe(UUID, BIGINT);
|
||||
|
||||
CREATE OR REPLACE FUNCTION public.fork_recipe(p_user_id UUID, p_original_recipe_id BIGINT)
|
||||
RETURNS SETOF public.recipes
|
||||
LANGUAGE sql
|
||||
LANGUAGE plpgsql
|
||||
SECURITY INVOKER
|
||||
AS $$
|
||||
-- The entire forking logic is now encapsulated in a single, atomic database function.
|
||||
SELECT * FROM public.fork_recipe(p_user_id, p_original_recipe_id);
|
||||
DECLARE
|
||||
new_recipe_id BIGINT;
|
||||
BEGIN
|
||||
-- 1. Create a copy of the recipe, linking it to the new user and the original recipe.
|
||||
INSERT INTO public.recipes (
|
||||
user_id,
|
||||
original_recipe_id,
|
||||
name,
|
||||
description,
|
||||
instructions,
|
||||
prep_time_minutes,
|
||||
cook_time_minutes,
|
||||
servings,
|
||||
photo_url,
|
||||
calories_per_serving,
|
||||
protein_grams,
|
||||
fat_grams,
|
||||
carb_grams,
|
||||
status -- Forked recipes should be private by default
|
||||
)
|
||||
SELECT
|
||||
p_user_id,
|
||||
p_original_recipe_id,
|
||||
original.name || ' (Fork)', -- Append '(Fork)' to distinguish it
|
||||
original.description,
|
||||
original.instructions,
|
||||
original.prep_time_minutes,
|
||||
original.cook_time_minutes,
|
||||
original.servings,
|
||||
original.photo_url,
|
||||
original.calories_per_serving,
|
||||
original.protein_grams,
|
||||
original.fat_grams,
|
||||
original.carb_grams,
|
||||
'private'
|
||||
FROM public.recipes AS original
|
||||
WHERE original.recipe_id = p_original_recipe_id
|
||||
RETURNING recipe_id INTO new_recipe_id;
|
||||
|
||||
-- If the original recipe didn't exist, new_recipe_id will be null.
|
||||
IF new_recipe_id IS NULL THEN
|
||||
RETURN;
|
||||
END IF;
|
||||
|
||||
-- 2. Copy all ingredients, tags, and appliances from the original recipe to the new one.
|
||||
INSERT INTO public.recipe_ingredients (recipe_id, master_item_id, quantity, unit) SELECT new_recipe_id, master_item_id, quantity, unit FROM public.recipe_ingredients WHERE recipe_id = p_original_recipe_id;
|
||||
INSERT INTO public.recipe_tags (recipe_id, tag_id) SELECT new_recipe_id, tag_id FROM public.recipe_tags WHERE recipe_id = p_original_recipe_id;
|
||||
INSERT INTO public.recipe_appliances (recipe_id, appliance_id) SELECT new_recipe_id, appliance_id FROM public.recipe_appliances WHERE recipe_id = p_original_recipe_id;
|
||||
|
||||
-- 3. Return the newly created recipe record.
|
||||
RETURN QUERY SELECT * FROM public.recipes WHERE recipe_id = new_recipe_id;
|
||||
END;
|
||||
$$;
|
||||
|
||||
|
||||
|
||||
@@ -16,4 +16,4 @@ const config = {
|
||||
},
|
||||
};
|
||||
|
||||
export default config;
|
||||
export default config;
|
||||
|
||||
202
src/db/seed.ts
202
src/db/seed.ts
@@ -6,10 +6,11 @@
|
||||
* DO NOT run this on a production database.
|
||||
*/
|
||||
|
||||
import { Pool } from 'pg';
|
||||
import { Pool, PoolClient } from 'pg';
|
||||
import fs from 'node:fs/promises';
|
||||
import path from 'node:path';
|
||||
import bcrypt from 'bcrypt';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { CATEGORIES } from '../types';
|
||||
|
||||
const pool = new Pool({
|
||||
user: process.env.DB_USER,
|
||||
@@ -20,81 +21,55 @@ const pool = new Pool({
|
||||
});
|
||||
|
||||
async function main() {
|
||||
// Declare client outside the try block so it's accessible in the finally block.
|
||||
let client;
|
||||
let client: PoolClient | undefined;
|
||||
|
||||
try {
|
||||
client = await pool.connect();
|
||||
logger.info('Connected to the database for seeding.');
|
||||
await client.query('BEGIN');
|
||||
|
||||
// 1. Clean the database
|
||||
logger.info('--- Wiping existing data... ---');
|
||||
// Using TRUNCATE ... RESTART IDENTITY CASCADE is a powerful way to clean all tables
|
||||
// and reset auto-incrementing keys, while respecting foreign key relationships.
|
||||
const tablesRes = await client.query(`
|
||||
SELECT tablename
|
||||
FROM pg_tables
|
||||
WHERE schemaname = 'public'
|
||||
-- Exclude PostGIS system tables from truncation to avoid permission errors.
|
||||
AND tablename NOT IN ('spatial_ref_sys', 'geometry_columns')
|
||||
`);
|
||||
const tables = tablesRes.rows.map((row) => `"${row.tablename}"`).join(', ');
|
||||
if (tables) {
|
||||
await client.query(`TRUNCATE ${tables} RESTART IDENTITY CASCADE`);
|
||||
logger.info('All tables in public schema have been truncated.');
|
||||
}
|
||||
// 1. Clean the database by dropping and recreating the schema
|
||||
logger.info('--- Wiping and rebuilding schema... ---');
|
||||
const dropScriptPath = path.resolve(process.cwd(), 'sql/drop_tables.sql');
|
||||
const dropSql = await fs.readFile(dropScriptPath, 'utf-8');
|
||||
await client.query(dropSql);
|
||||
logger.info('All tables dropped successfully.');
|
||||
|
||||
// 2. Seed Categories
|
||||
logger.info('--- Seeding Categories... ---');
|
||||
const categoryQuery = `INSERT INTO public.categories (name) VALUES ${CATEGORIES.map((_, i) => `($${i + 1})`).join(', ')} RETURNING category_id, name`;
|
||||
const seededCategories = (
|
||||
await client.query<{ category_id: number; name: string }>(categoryQuery, CATEGORIES)
|
||||
).rows;
|
||||
const categoryMap = new Map(seededCategories.map((c) => [c.name, c.category_id]));
|
||||
logger.info(`Seeded ${seededCategories.length} categories.`);
|
||||
const schemaScriptPath = path.resolve(process.cwd(), 'sql/master_schema_rollup.sql');
|
||||
const schemaSql = await fs.readFile(schemaScriptPath, 'utf-8');
|
||||
await client.query(schemaSql);
|
||||
logger.info(
|
||||
'Schema rebuilt and static data seeded successfully from master_schema_rollup.sql.',
|
||||
);
|
||||
|
||||
// 3. Seed Stores
|
||||
// 2. Seed Additional Stores (if any beyond what's in the rollup)
|
||||
logger.info('--- Seeding Stores... ---');
|
||||
const stores = ['Safeway', 'No Frills', 'Costco', 'Superstore'];
|
||||
const storeQuery = `INSERT INTO public.stores (name) VALUES ${stores.map((_, i) => `($${i + 1})`).join(', ')} RETURNING store_id, name`;
|
||||
const seededStores = (
|
||||
await client.query<{ store_id: number; name: string }>(storeQuery, stores)
|
||||
const storeQuery = `INSERT INTO public.stores (name) VALUES ${stores.map((_, i) => `($${i + 1})`).join(', ')} ON CONFLICT (name) DO NOTHING RETURNING store_id, name`;
|
||||
await client.query<{ store_id: number; name: string }>(storeQuery, stores);
|
||||
const allStores = (
|
||||
await client.query<{ store_id: number; name: string }>(
|
||||
'SELECT store_id, name FROM public.stores',
|
||||
)
|
||||
).rows;
|
||||
const storeMap = new Map(seededStores.map((s) => [s.name, s.store_id]));
|
||||
logger.info(`Seeded ${seededStores.length} stores.`);
|
||||
|
||||
// 4. Seed Master Grocery Items
|
||||
logger.info('--- Seeding Master Grocery Items... ---');
|
||||
const masterItems = [
|
||||
{ name: 'Chicken Breast, Boneless Skinless', category: 'Meat & Seafood' },
|
||||
{ name: 'Ground Beef, Lean', category: 'Meat & Seafood' },
|
||||
{ name: 'Avocado', category: 'Fruits & Vegetables' },
|
||||
{ name: 'Bananas', category: 'Fruits & Vegetables' },
|
||||
{ name: 'Broccoli', category: 'Fruits & Vegetables' },
|
||||
{ name: 'Cheddar Cheese, Block', category: 'Dairy & Eggs' },
|
||||
{ name: 'Milk, 2%', category: 'Dairy & Eggs' },
|
||||
{ name: 'Eggs, Large', category: 'Dairy & Eggs' },
|
||||
{ name: 'Whole Wheat Bread', category: 'Bakery & Bread' },
|
||||
{ name: 'Pasta, Spaghetti', category: 'Pantry & Dry Goods' },
|
||||
{ name: 'Canned Tomatoes, Diced', category: 'Canned Goods' },
|
||||
{ name: 'Coca-Cola, 12-pack', category: 'Beverages' },
|
||||
{ name: 'Frozen Pizza', category: 'Frozen Foods' },
|
||||
{ name: 'Paper Towels', category: 'Household & Cleaning' },
|
||||
];
|
||||
const masterItemValues = masterItems
|
||||
.map((item) => `('${item.name.replace(/'/g, "''")}', ${categoryMap.get(item.category)})`)
|
||||
.join(', ');
|
||||
const masterItemQuery = `INSERT INTO public.master_grocery_items (name, category_id) VALUES ${masterItemValues} RETURNING master_grocery_item_id, name`;
|
||||
const seededMasterItems = (
|
||||
await client.query<{ master_grocery_item_id: number; name: string }>(masterItemQuery)
|
||||
).rows;
|
||||
const masterItemMap = new Map(
|
||||
seededMasterItems.map((item) => [item.name, item.master_grocery_item_id]),
|
||||
const storeMap = new Map(
|
||||
allStores.map((s: { name: string; store_id: number }) => [s.name, s.store_id]),
|
||||
);
|
||||
logger.info(`Seeded ${seededMasterItems.length} master grocery items.`);
|
||||
logger.info(`Seeded/verified ${allStores.length} total stores.`);
|
||||
|
||||
// 5. Seed Users & Profiles
|
||||
// Fetch maps for items seeded by the master rollup script
|
||||
const masterItemMap = new Map(
|
||||
(
|
||||
await client.query<{ master_grocery_item_id: number; name: string }>(
|
||||
'SELECT master_grocery_item_id, name FROM public.master_grocery_items',
|
||||
)
|
||||
).rows.map((item: { name: string; master_grocery_item_id: number }) => [
|
||||
item.name,
|
||||
item.master_grocery_item_id,
|
||||
]),
|
||||
);
|
||||
|
||||
// 3. Seed Users & Profiles
|
||||
logger.info('--- Seeding Users & Profiles... ---');
|
||||
const saltRounds = 10;
|
||||
const adminPassHash = await bcrypt.hash('adminpass', saltRounds);
|
||||
@@ -126,7 +101,7 @@ async function main() {
|
||||
const userId = userRes.rows[0].user_id;
|
||||
logger.info('Seeded regular user (user@example.com / userpass)');
|
||||
|
||||
// 6. Seed a Flyer
|
||||
// 4. Seed a Flyer
|
||||
logger.info('--- Seeding a Sample Flyer... ---');
|
||||
const today = new Date();
|
||||
const validFrom = new Date(today);
|
||||
@@ -146,29 +121,29 @@ async function main() {
|
||||
const flyerId = flyerRes.rows[0].flyer_id;
|
||||
logger.info(`Seeded flyer for Safeway (ID: ${flyerId}).`);
|
||||
|
||||
// 7. Seed Flyer Items
|
||||
// 5. Seed Flyer Items
|
||||
logger.info('--- Seeding Flyer Items... ---');
|
||||
const flyerItems = [
|
||||
{
|
||||
name: 'Chicken Breast, Boneless Skinless',
|
||||
name: 'chicken breast',
|
||||
price_display: '$3.99 /lb',
|
||||
price_in_cents: 399,
|
||||
quantity: 'per lb',
|
||||
master_item_id: masterItemMap.get('Chicken Breast, Boneless Skinless'),
|
||||
master_item_id: masterItemMap.get('chicken breast'),
|
||||
},
|
||||
{
|
||||
name: 'Avocado',
|
||||
name: 'avocados',
|
||||
price_display: '2 for $5.00',
|
||||
price_in_cents: 250,
|
||||
quantity: 'each',
|
||||
master_item_id: masterItemMap.get('Avocado'),
|
||||
master_item_id: masterItemMap.get('avocados'),
|
||||
},
|
||||
{
|
||||
name: 'Coca-Cola 12-pack',
|
||||
name: 'soda',
|
||||
price_display: '$6.99',
|
||||
price_in_cents: 699,
|
||||
quantity: '12x355ml',
|
||||
master_item_id: masterItemMap.get('Coca-Cola, 12-pack'),
|
||||
master_item_id: masterItemMap.get('soda'),
|
||||
},
|
||||
{
|
||||
name: 'Unmatched Sample Item',
|
||||
@@ -194,12 +169,12 @@ async function main() {
|
||||
}
|
||||
logger.info(`Seeded ${flyerItems.length} items for the Safeway flyer.`);
|
||||
|
||||
// 8. Seed Watched Items for the user
|
||||
// 6. Seed Watched Items for the user
|
||||
logger.info('--- Seeding Watched Items... ---');
|
||||
const watchedItemIds = [
|
||||
masterItemMap.get('Chicken Breast, Boneless Skinless'),
|
||||
masterItemMap.get('Avocado'),
|
||||
masterItemMap.get('Ground Beef, Lean'),
|
||||
masterItemMap.get('chicken breast'),
|
||||
masterItemMap.get('avocados'),
|
||||
masterItemMap.get('ground beef'),
|
||||
];
|
||||
for (const itemId of watchedItemIds) {
|
||||
if (itemId) {
|
||||
@@ -211,7 +186,7 @@ async function main() {
|
||||
}
|
||||
logger.info(`Seeded ${watchedItemIds.length} watched items for Test User.`);
|
||||
|
||||
// 9. Seed a Shopping List
|
||||
// 7. Seed a Shopping List
|
||||
logger.info('--- Seeding a Shopping List... ---');
|
||||
const listRes = await client.query<{ shopping_list_id: number }>(
|
||||
'INSERT INTO public.shopping_lists (user_id, name) VALUES ($1, $2) RETURNING shopping_list_id',
|
||||
@@ -220,8 +195,8 @@ async function main() {
|
||||
const listId = listRes.rows[0].shopping_list_id;
|
||||
|
||||
const shoppingListItems = [
|
||||
{ master_item_id: masterItemMap.get('Milk, 2%'), quantity: 1 },
|
||||
{ master_item_id: masterItemMap.get('Eggs, Large'), quantity: 1 },
|
||||
{ master_item_id: masterItemMap.get('milk'), quantity: 1 },
|
||||
{ master_item_id: masterItemMap.get('eggs'), quantity: 1 },
|
||||
{ custom_item_name: 'Specialty Hot Sauce', quantity: 1 },
|
||||
];
|
||||
|
||||
@@ -235,75 +210,6 @@ async function main() {
|
||||
`Seeded shopping list "Weekly Groceries" with ${shoppingListItems.length} items for Test User.`,
|
||||
);
|
||||
|
||||
// 10. Seed Brands
|
||||
logger.info('--- Seeding Brands... ---');
|
||||
const brands = [
|
||||
'Coca-Cola',
|
||||
'Kraft',
|
||||
'Maple Leaf',
|
||||
"Dempster's",
|
||||
'No Name',
|
||||
"President's Choice",
|
||||
];
|
||||
const brandQuery = `INSERT INTO public.brands (name) VALUES ${brands.map((_, i) => `($${i + 1})`).join(', ')} ON CONFLICT (name) DO NOTHING`;
|
||||
await client.query(brandQuery, brands);
|
||||
logger.info(`Seeded ${brands.length} brands.`);
|
||||
|
||||
// Link store-specific brands
|
||||
const loblawsId = storeMap.get('Loblaws');
|
||||
if (loblawsId) {
|
||||
await client.query('UPDATE public.brands SET store_id = $1 WHERE name = $2 OR name = $3', [
|
||||
loblawsId,
|
||||
'No Name',
|
||||
"President's Choice",
|
||||
]);
|
||||
logger.info('Linked store brands to Loblaws.');
|
||||
}
|
||||
|
||||
// 11. Seed Recipes
|
||||
logger.info('--- Seeding Recipes... ---');
|
||||
const recipes = [
|
||||
{
|
||||
name: 'Simple Chicken and Rice',
|
||||
description: 'A quick and healthy weeknight meal.',
|
||||
instructions: '1. Cook rice. 2. Cook chicken. 3. Combine.',
|
||||
prep: 10,
|
||||
cook: 20,
|
||||
servings: 4,
|
||||
},
|
||||
{
|
||||
name: 'Classic Spaghetti Bolognese',
|
||||
description: 'A rich and hearty meat sauce.',
|
||||
instructions: '1. Brown beef. 2. Add sauce. 3. Simmer.',
|
||||
prep: 15,
|
||||
cook: 45,
|
||||
servings: 6,
|
||||
},
|
||||
{
|
||||
name: 'Vegetable Stir-fry',
|
||||
description: 'A fast and flavorful vegetarian meal.',
|
||||
instructions: '1. Chop veggies. 2. Stir-fry. 3. Add sauce.',
|
||||
prep: 10,
|
||||
cook: 10,
|
||||
servings: 3,
|
||||
},
|
||||
];
|
||||
for (const recipe of recipes) {
|
||||
await client.query(
|
||||
`INSERT INTO public.recipes (name, description, instructions, prep_time_minutes, cook_time_minutes, servings, status)
|
||||
VALUES ($1, $2, $3, $4, $5, $6, 'public') ON CONFLICT (name) WHERE user_id IS NULL DO NOTHING`,
|
||||
[
|
||||
recipe.name,
|
||||
recipe.description,
|
||||
recipe.instructions,
|
||||
recipe.prep,
|
||||
recipe.cook,
|
||||
recipe.servings,
|
||||
],
|
||||
);
|
||||
}
|
||||
logger.info(`Seeded ${recipes.length} recipes.`);
|
||||
|
||||
// --- SEED SCRIPT DEBUG LOGGING ---
|
||||
// Corrected the query to be unambiguous by specifying the table alias for each column.
|
||||
// `id` and `email` come from the `users` table (u), and `role` comes from the `profiles` table (p).
|
||||
|
||||
@@ -4,7 +4,7 @@
|
||||
This single directive replaces @tailwind base, components, and utilities.
|
||||
It is the new entry point for all of Tailwind's generated CSS.
|
||||
*/
|
||||
@import "tailwindcss";
|
||||
@import 'tailwindcss';
|
||||
|
||||
/*
|
||||
This is the new v4 directive that tells the @tailwindcss/postcss plugin
|
||||
@@ -12,4 +12,3 @@
|
||||
Since tailwind.config.js is in the root and this is in src/, the path is '../tailwind.config.js'.
|
||||
*/
|
||||
@config '../tailwind.config.js';
|
||||
|
||||
|
||||
@@ -8,17 +8,16 @@ import './index.css';
|
||||
|
||||
const rootElement = document.getElementById('root');
|
||||
if (!rootElement) {
|
||||
throw new Error("Could not find root element to mount to");
|
||||
throw new Error('Could not find root element to mount to');
|
||||
}
|
||||
|
||||
const root = ReactDOM.createRoot(rootElement);
|
||||
root.render(
|
||||
<React.StrictMode>
|
||||
<BrowserRouter>
|
||||
<BrowserRouter>
|
||||
<AppProviders>
|
||||
<App />
|
||||
</AppProviders>
|
||||
</BrowserRouter>
|
||||
</React.StrictMode>
|
||||
</BrowserRouter>
|
||||
</React.StrictMode>,
|
||||
);
|
||||
|
||||
@@ -5,4 +5,4 @@ import toast from 'react-hot-toast';
|
||||
// This intermediate file allows us to mock 'src/lib/toast' reliably in tests
|
||||
// without wrestling with the internal structure of the 'react-hot-toast' package.
|
||||
export * from 'react-hot-toast';
|
||||
export default toast;
|
||||
export default toast;
|
||||
|
||||
76
src/types.ts
76
src/types.ts
@@ -52,7 +52,10 @@ export type FlyerDbInsert = Omit<FlyerInsert, 'store_name'> & { store_id: number
|
||||
* Represents the data required to insert a new flyer item into the database.
|
||||
* It's a subset of the full FlyerItem type.
|
||||
*/
|
||||
export type FlyerItemInsert = Omit<FlyerItem, 'flyer_item_id' | 'flyer_id' | 'created_at' | 'updated_at'>;
|
||||
export type FlyerItemInsert = Omit<
|
||||
FlyerItem,
|
||||
'flyer_item_id' | 'flyer_id' | 'created_at' | 'updated_at'
|
||||
>;
|
||||
|
||||
export interface UnitPrice {
|
||||
value: number;
|
||||
@@ -163,13 +166,12 @@ export interface Profile {
|
||||
updated_by?: string | null;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Represents the combined user and profile data object returned by the backend's /users/profile endpoint.
|
||||
* It embeds the User object within the Profile object.
|
||||
* It also includes the full Address object if one is associated with the profile.
|
||||
*/
|
||||
export type UserProfile = Profile & {
|
||||
export type UserProfile = Profile & {
|
||||
user: User;
|
||||
address?: Address | null;
|
||||
};
|
||||
@@ -325,7 +327,6 @@ export interface RecipeIngredientSubstitution {
|
||||
notes?: string | null;
|
||||
}
|
||||
|
||||
|
||||
export interface Tag {
|
||||
tag_id: number;
|
||||
name: string;
|
||||
@@ -718,7 +719,10 @@ export type AiAnalysisAction =
|
||||
// Dispatched when an analysis that returns a simple string succeeds.
|
||||
| { type: 'FETCH_SUCCESS_TEXT'; payload: { analysisType: AnalysisType; data: string } }
|
||||
// Dispatched when an analysis that returns text and sources succeeds.
|
||||
| { type: 'FETCH_SUCCESS_GROUNDED'; payload: { analysisType: AnalysisType; data: GroundedResponse } }
|
||||
| {
|
||||
type: 'FETCH_SUCCESS_GROUNDED';
|
||||
payload: { analysisType: AnalysisType; data: GroundedResponse };
|
||||
}
|
||||
// Dispatched when the image generation succeeds.
|
||||
| { type: 'FETCH_SUCCESS_IMAGE'; payload: { data: string } }
|
||||
// Dispatched when any analysis fails.
|
||||
@@ -738,11 +742,25 @@ export interface ProcessingStage {
|
||||
}
|
||||
|
||||
export const CATEGORIES = [
|
||||
'Fruits & Vegetables', 'Meat & Seafood', 'Dairy & Eggs', 'Bakery & Bread',
|
||||
'Pantry & Dry Goods', 'Beverages', 'Frozen Foods', 'Snacks', 'Household & Cleaning',
|
||||
'Personal Care & Health', 'Baby & Child', 'Pet Supplies', 'Deli & Prepared Foods',
|
||||
'Canned Goods', 'Condiments & Spices', 'Breakfast & Cereal', 'Organic',
|
||||
'International Foods', 'Other/Miscellaneous'
|
||||
'Fruits & Vegetables',
|
||||
'Meat & Seafood',
|
||||
'Dairy & Eggs',
|
||||
'Bakery & Bread',
|
||||
'Pantry & Dry Goods',
|
||||
'Beverages',
|
||||
'Frozen Foods',
|
||||
'Snacks',
|
||||
'Household & Cleaning',
|
||||
'Personal Care & Health',
|
||||
'Baby & Child',
|
||||
'Pet Supplies',
|
||||
'Deli & Prepared Foods',
|
||||
'Canned Goods',
|
||||
'Condiments & Spices',
|
||||
'Breakfast & Cereal',
|
||||
'Organic',
|
||||
'International Foods',
|
||||
'Other/Miscellaneous',
|
||||
];
|
||||
|
||||
/**
|
||||
@@ -750,11 +768,11 @@ export const CATEGORIES = [
|
||||
* This is the structure returned from the backend to the frontend.
|
||||
*/
|
||||
export interface ExtractedCoreData {
|
||||
store_name: string;
|
||||
valid_from: string | null;
|
||||
valid_to: string | null;
|
||||
store_address: string | null;
|
||||
items: ExtractedFlyerItem[];
|
||||
store_name: string;
|
||||
valid_from: string | null;
|
||||
valid_to: string | null;
|
||||
store_address: string | null;
|
||||
items: ExtractedFlyerItem[];
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -776,19 +794,19 @@ export interface ExtractedFlyerItem {
|
||||
* Represents the logo data extracted from a flyer by the AI service.
|
||||
*/
|
||||
export interface ExtractedLogoData {
|
||||
store_logo_base_64: string | null;
|
||||
store_logo_base_64: string | null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Represents the data extracted from a receipt image by the AI service.
|
||||
*/
|
||||
export interface ExtractedReceiptData {
|
||||
raw_text: string;
|
||||
items: {
|
||||
raw_item_description: string;
|
||||
quantity: number;
|
||||
price_paid_cents: number;
|
||||
}[];
|
||||
raw_text: string;
|
||||
items: {
|
||||
raw_item_description: string;
|
||||
quantity: number;
|
||||
price_paid_cents: number;
|
||||
}[];
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -930,10 +948,10 @@ export interface LeaderboardUser {
|
||||
* This is a public-facing type and does not include sensitive fields.
|
||||
*/
|
||||
export interface AdminUserView {
|
||||
user_id: string;
|
||||
email: string;
|
||||
created_at: string;
|
||||
role: 'admin' | 'user';
|
||||
full_name: string | null;
|
||||
avatar_url: string | null;
|
||||
}
|
||||
user_id: string;
|
||||
email: string;
|
||||
created_at: string;
|
||||
role: 'admin' | 'user';
|
||||
full_name: string | null;
|
||||
avatar_url: string | null;
|
||||
}
|
||||
|
||||
2
src/vite-env.d.ts
vendored
2
src/vite-env.d.ts
vendored
@@ -10,4 +10,4 @@ interface ImportMetaEnv {
|
||||
|
||||
interface ImportMeta {
|
||||
readonly env: ImportMetaEnv;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -2,4 +2,4 @@
|
||||
|
||||
// This file can be used for global setup logic that applies to ALL test projects
|
||||
// defined in the workspace. Since our unit and integration tests have distinct
|
||||
// setup requirements, this file is currently empty.
|
||||
// setup requirements, this file is currently empty.
|
||||
|
||||
@@ -4,8 +4,5 @@ console.log('--- [EXECUTION PROOF] tailwind.config.js is being loaded. ---');
|
||||
|
||||
/** @type {import('tailwindcss').Config} */
|
||||
export default {
|
||||
content: [
|
||||
"./index.html",
|
||||
"./src/**/*.{js,ts,jsx,tsx}",
|
||||
],
|
||||
};
|
||||
content: ['./index.html', './src/**/*.{js,ts,jsx,tsx}'],
|
||||
};
|
||||
|
||||
@@ -17,12 +17,10 @@
|
||||
"jsx": "react-jsx",
|
||||
// This line makes Vitest's global APIs (describe, it, expect) available everywhere
|
||||
// without needing to import them.
|
||||
"types": [
|
||||
"vitest/globals"
|
||||
]
|
||||
"types": ["vitest/globals"]
|
||||
},
|
||||
// This is the most important part: It tells TypeScript to include ALL files
|
||||
// within the 'src' directory, including your new 'vite-env.d.ts' file.
|
||||
"include": ["src"],
|
||||
"references": [{ "path": "./tsconfig.node.json" }]
|
||||
}
|
||||
}
|
||||
|
||||
@@ -9,5 +9,10 @@
|
||||
"strict": true, // It's good practice to keep tooling config strict
|
||||
"types": ["node"]
|
||||
},
|
||||
"include": ["vite.config.ts", "vitest.config.ts", "vitest.config.integration.ts", "vitest.workspace.ts"]
|
||||
}
|
||||
"include": [
|
||||
"vite.config.ts",
|
||||
"vitest.config.ts",
|
||||
"vitest.config.integration.ts",
|
||||
"vitest.workspace.ts"
|
||||
]
|
||||
}
|
||||
|
||||
@@ -22,7 +22,7 @@ export default defineConfig({
|
||||
resolve: {
|
||||
alias: {
|
||||
// Use __dirname for a more robust path resolution
|
||||
'@': path.resolve(__dirname, './src'),
|
||||
'@': path.resolve(__dirname, './src'),
|
||||
// This alias ensures that any import of 'services/logger' is resolved
|
||||
// to the browser-safe client version during the Vite build process.
|
||||
// Server-side code should explicitly import 'services/logger.server'.
|
||||
@@ -51,7 +51,7 @@ export default defineConfig({
|
||||
'**/node_modules/**',
|
||||
'**/dist/**',
|
||||
'src/tests/integration/**', // Exclude the entire integration test directory
|
||||
'**/*.e2e.test.ts'
|
||||
'**/*.e2e.test.ts',
|
||||
],
|
||||
// Disable file parallelism to run tests sequentially (replaces --no-threads)
|
||||
fileParallelism: false,
|
||||
@@ -60,10 +60,12 @@ export default defineConfig({
|
||||
reporter: [
|
||||
// Add maxCols to suggest a wider output for the text summary.
|
||||
['text', { maxCols: 200 }],
|
||||
'html', 'json'],
|
||||
'html',
|
||||
'json',
|
||||
],
|
||||
// hanging-process reporter helps identify tests that do not exit properly - comes at a high cost tho
|
||||
//reporter: ['verbose', 'html', 'json', 'hanging-process'],
|
||||
reportsDirectory: './.coverage/unit',
|
||||
reportsDirectory: './.coverage/unit',
|
||||
clean: true,
|
||||
reportOnFailure: true, // This ensures the report generates even if tests fail
|
||||
include: ['src/**/*.{ts,tsx}'],
|
||||
@@ -71,19 +73,19 @@ export default defineConfig({
|
||||
// By excluding scripts, setup files, and type definitions, we get a more accurate
|
||||
// picture of the test coverage for the actual application logic.
|
||||
exclude: [
|
||||
'src/index.tsx', // Application entry point
|
||||
'src/main.tsx', // A common alternative entry point name
|
||||
'src/index.tsx', // Application entry point
|
||||
'src/main.tsx', // A common alternative entry point name
|
||||
'src/types.ts',
|
||||
'src/tests/**', // Exclude all test setup and helper files
|
||||
'src/vitest.setup.ts', // Global test setup config
|
||||
'src/vitest.setup.ts', // Global test setup config
|
||||
'src/**/*.test.{ts,tsx}', // Exclude test files themselves
|
||||
'src/**/*.stories.{ts,tsx}', // Exclude Storybook stories
|
||||
'src/**/*.d.ts', // Exclude type definition files
|
||||
'src/components/icons/**', // Exclude icon components if they are simple wrappers
|
||||
'src/db/seed.ts', // Database seeding script
|
||||
'src/db/seed_admin_account.ts', // Database seeding script
|
||||
'src/db/backup_user.ts', // Database backup script
|
||||
'src/db/seed.ts', // Database seeding script
|
||||
'src/db/seed_admin_account.ts', // Database seeding script
|
||||
'src/db/backup_user.ts', // Database backup script
|
||||
],
|
||||
},
|
||||
},
|
||||
});
|
||||
});
|
||||
|
||||
@@ -6,7 +6,7 @@ import viteConfig from './vite.config';
|
||||
// Ensure NODE_ENV is set to 'test' for all Vitest runs.
|
||||
process.env.NODE_ENV = 'test';
|
||||
|
||||
// 1. Separate the 'test' config (which has Unit Test settings)
|
||||
// 1. Separate the 'test' config (which has Unit Test settings)
|
||||
// from the rest of the general Vite config (plugins, aliases, etc.)
|
||||
// DEBUG: Use console.error to ensure logs appear in CI/CD output
|
||||
console.error('[DEBUG] Loading vitest.config.integration.ts...');
|
||||
@@ -30,38 +30,41 @@ console.error('[DEBUG] Base vite config keys:', Object.keys(baseViteConfig));
|
||||
* It MERGES with the main vite.config.ts to inherit plugins and aliases,
|
||||
* then overrides the test-specific settings for a Node.js environment.
|
||||
*/
|
||||
const finalConfig = mergeConfig(baseViteConfig, defineConfig({
|
||||
test: {
|
||||
// Override settings from the main config for this specific test project.
|
||||
name: 'integration',
|
||||
environment: 'node',
|
||||
// Point specifically to the new integration tests directory.
|
||||
// This pattern will match any test file inside `src/tests/integration/`.
|
||||
include: ['src/tests/integration/**/*.test.{ts,tsx}'],
|
||||
// CRITICAL: We must override the `exclude` property from the base vite.config.ts.
|
||||
// Otherwise, the inherited `exclude` rule will prevent any integration tests from running.
|
||||
// Setting it to an empty array removes all exclusion rules for this project.
|
||||
exclude: [],
|
||||
// This setup script starts the backend server before tests run.
|
||||
globalSetup: './src/tests/setup/integration-global-setup.ts',
|
||||
// The default timeout is 5000ms (5 seconds)
|
||||
testTimeout: 60000, // Increased timeout for server startup and API calls, especially AI services.
|
||||
// "singleThread: true" is removed in modern Vitest.
|
||||
// Use fileParallelism: false to ensure test files run one by one to prevent port conflicts.
|
||||
fileParallelism: false,
|
||||
coverage: {
|
||||
provider: 'v8',
|
||||
// We remove 'text' here. The final text report will be generated by `nyc` after merging.
|
||||
reporter: ['html', 'json-summary', 'json'],
|
||||
reportsDirectory: '.coverage/integration',
|
||||
reportOnFailure: true, // This ensures the report generates even if tests fail
|
||||
clean: true,
|
||||
const finalConfig = mergeConfig(
|
||||
baseViteConfig,
|
||||
defineConfig({
|
||||
test: {
|
||||
// Override settings from the main config for this specific test project.
|
||||
name: 'integration',
|
||||
environment: 'node',
|
||||
// Point specifically to the new integration tests directory.
|
||||
// This pattern will match any test file inside `src/tests/integration/`.
|
||||
include: ['src/tests/integration/**/*.test.{ts,tsx}'],
|
||||
// CRITICAL: We must override the `exclude` property from the base vite.config.ts.
|
||||
// Otherwise, the inherited `exclude` rule will prevent any integration tests from running.
|
||||
// Setting it to an empty array removes all exclusion rules for this project.
|
||||
exclude: [],
|
||||
// This setup script starts the backend server before tests run.
|
||||
globalSetup: './src/tests/setup/integration-global-setup.ts',
|
||||
// The default timeout is 5000ms (5 seconds)
|
||||
testTimeout: 60000, // Increased timeout for server startup and API calls, especially AI services.
|
||||
// "singleThread: true" is removed in modern Vitest.
|
||||
// Use fileParallelism: false to ensure test files run one by one to prevent port conflicts.
|
||||
fileParallelism: false,
|
||||
coverage: {
|
||||
provider: 'v8',
|
||||
// We remove 'text' here. The final text report will be generated by `nyc` after merging.
|
||||
reporter: ['html', 'json-summary', 'json'],
|
||||
reportsDirectory: '.coverage/integration',
|
||||
reportOnFailure: true, // This ensures the report generates even if tests fail
|
||||
clean: true,
|
||||
},
|
||||
},
|
||||
}
|
||||
}));
|
||||
}),
|
||||
);
|
||||
|
||||
console.error('[DEBUG] Integration Final Config - INCLUDE:', finalConfig.test?.include);
|
||||
console.error('[DEBUG] Integration Final Config - EXCLUDE:', finalConfig.test?.exclude);
|
||||
console.error('[DEBUG] ----------------------------------\n');
|
||||
|
||||
export default finalConfig;
|
||||
export default finalConfig;
|
||||
|
||||
@@ -7,8 +7,11 @@
|
||||
* - Integration tests are defined in `vitest.config.integration.ts` and run in a 'node' environment.
|
||||
*/
|
||||
export default [
|
||||
// DEBUGGING LOG
|
||||
((): string => { console.error('\n[DEBUG] Loading vitest.workspace.ts'); return ''; })(),
|
||||
'vite.config.ts', // Defines the 'unit' test project
|
||||
// DEBUGGING LOG
|
||||
((): string => {
|
||||
console.error('\n[DEBUG] Loading vitest.workspace.ts');
|
||||
return '';
|
||||
})(),
|
||||
'vite.config.ts', // Defines the 'unit' test project
|
||||
'vitest.config.integration.ts', // Defines the 'integration' test project
|
||||
];
|
||||
];
|
||||
|
||||
Reference in New Issue
Block a user