Compare commits
11 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
57215e2778 | ||
| 2c1de24e9a | |||
| c8baff7aac | |||
| de3f21a7ec | |||
|
|
c6adbf79e7 | ||
| 7399a27600 | |||
|
|
68aadcaa4e | ||
| 971d2c3fa7 | |||
|
|
daaacfde5e | ||
| 7ac8fe1d29 | |||
| a2462dfb6b |
@@ -93,8 +93,9 @@ jobs:
|
||||
exit 1
|
||||
fi
|
||||
GITEA_SERVER_URL="https://gitea.projectium.com"
|
||||
COMMIT_MESSAGE=$(git log -1 --pretty=%s)
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD)" \
|
||||
COMMIT_MESSAGE=$(git log -1 --grep="\[skip ci\]" --invert-grep --pretty=%s)
|
||||
PACKAGE_VERSION=$(node -p "require('./package.json').version")
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD):$PACKAGE_VERSION" \
|
||||
VITE_APP_COMMIT_URL="$GITEA_SERVER_URL/${{ gitea.repository }}/commit/${{ gitea.sha }}" \
|
||||
VITE_APP_COMMIT_MESSAGE="$COMMIT_MESSAGE" \
|
||||
VITE_API_BASE_URL=/api VITE_API_KEY=${{ secrets.VITE_GOOGLE_GENAI_API_KEY }} npm run build
|
||||
@@ -162,7 +163,12 @@ jobs:
|
||||
echo "Updating schema hash in production database..."
|
||||
CURRENT_HASH=$(cat sql/master_schema_rollup.sql | dos2unix | sha256sum | awk '{ print $1 }')
|
||||
PGPASSWORD="$DB_PASSWORD" psql -v ON_ERROR_STOP=1 -h "$DB_HOST" -p 5432 -U "$DB_USER" -d "$DB_NAME" -c \
|
||||
"INSERT INTO public.schema_info (environment, schema_hash, deployed_at) VALUES ('production', '$CURRENT_HASH', NOW())
|
||||
"CREATE TABLE IF NOT EXISTS public.schema_info (
|
||||
environment VARCHAR(50) PRIMARY KEY,
|
||||
schema_hash VARCHAR(64) NOT NULL,
|
||||
deployed_at TIMESTAMP DEFAULT NOW()
|
||||
);
|
||||
INSERT INTO public.schema_info (environment, schema_hash, deployed_at) VALUES ('production', '$CURRENT_HASH', NOW())
|
||||
ON CONFLICT (environment) DO UPDATE SET schema_hash = EXCLUDED.schema_hash, deployed_at = NOW();"
|
||||
|
||||
UPDATED_HASH=$(PGPASSWORD="$DB_PASSWORD" psql -v ON_ERROR_STOP=1 -h "$DB_HOST" -p 5432 -U "$DB_USER" -d "$DB_NAME" -c "SELECT schema_hash FROM public.schema_info WHERE environment = 'production';" -t -A)
|
||||
|
||||
@@ -282,6 +282,9 @@ jobs:
|
||||
if [ -z "$DEPLOYED_HASH" ]; then
|
||||
echo "WARNING: No schema hash found in the test database."
|
||||
echo "This is expected for a first-time deployment. The hash will be set after a successful deployment."
|
||||
echo "--- Debug: Dumping schema_info table ---"
|
||||
PGPASSWORD="$DB_PASSWORD" psql -v ON_ERROR_STOP=0 -h "$DB_HOST" -p 5432 -U "$DB_USER" -d "$DB_NAME" -c "SELECT * FROM public.schema_info;" || true
|
||||
echo "----------------------------------------"
|
||||
# We allow the deployment to continue, but a manual schema update is required.
|
||||
# You could choose to fail here by adding `exit 1`.
|
||||
elif [ "$CURRENT_HASH" != "$DEPLOYED_HASH" ]; then
|
||||
@@ -305,8 +308,9 @@ jobs:
|
||||
fi
|
||||
|
||||
GITEA_SERVER_URL="https://gitea.projectium.com" # Your Gitea instance URL
|
||||
COMMIT_MESSAGE=$(git log -1 --pretty=%s)
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD)" \
|
||||
COMMIT_MESSAGE=$(git log -1 --grep="\[skip ci\]" --invert-grep --pretty=%s)
|
||||
PACKAGE_VERSION=$(node -p "require('./package.json').version")
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD):$PACKAGE_VERSION" \
|
||||
VITE_APP_COMMIT_URL="$GITEA_SERVER_URL/${{ gitea.repository }}/commit/${{ gitea.sha }}" \
|
||||
VITE_APP_COMMIT_MESSAGE="$COMMIT_MESSAGE" \
|
||||
VITE_API_BASE_URL="https://flyer-crawler-test.projectium.com/api" VITE_API_KEY=${{ secrets.VITE_GOOGLE_GENAI_API_KEY_TEST }} npm run build
|
||||
@@ -380,7 +384,12 @@ jobs:
|
||||
echo "Updating schema hash in test database..."
|
||||
CURRENT_HASH=$(cat sql/master_schema_rollup.sql | dos2unix | sha256sum | awk '{ print $1 }')
|
||||
PGPASSWORD="$DB_PASSWORD" psql -v ON_ERROR_STOP=1 -h "$DB_HOST" -p 5432 -U "$DB_USER" -d "$DB_NAME" -c \
|
||||
"INSERT INTO public.schema_info (environment, schema_hash, deployed_at) VALUES ('test', '$CURRENT_HASH', NOW())
|
||||
"CREATE TABLE IF NOT EXISTS public.schema_info (
|
||||
environment VARCHAR(50) PRIMARY KEY,
|
||||
schema_hash VARCHAR(64) NOT NULL,
|
||||
deployed_at TIMESTAMP DEFAULT NOW()
|
||||
);
|
||||
INSERT INTO public.schema_info (environment, schema_hash, deployed_at) VALUES ('test', '$CURRENT_HASH', NOW())
|
||||
ON CONFLICT (environment) DO UPDATE SET schema_hash = EXCLUDED.schema_hash, deployed_at = NOW();"
|
||||
|
||||
# Verify the hash was updated
|
||||
|
||||
@@ -92,8 +92,9 @@ jobs:
|
||||
exit 1
|
||||
fi
|
||||
GITEA_SERVER_URL="https://gitea.projectium.com"
|
||||
COMMIT_MESSAGE=$(git log -1 --pretty=%s)
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD)" \
|
||||
COMMIT_MESSAGE=$(git log -1 --grep="\[skip ci\]" --invert-grep --pretty=%s)
|
||||
PACKAGE_VERSION=$(node -p "require('./package.json').version")
|
||||
VITE_APP_VERSION="$(date +'%Y%m%d-%H%M'):$(git rev-parse --short HEAD):$PACKAGE_VERSION" \
|
||||
VITE_APP_COMMIT_URL="$GITEA_SERVER_URL/${{ gitea.repository }}/commit/${{ gitea.sha }}" \
|
||||
VITE_APP_COMMIT_MESSAGE="$COMMIT_MESSAGE" \
|
||||
VITE_API_BASE_URL=/api VITE_API_KEY=${{ secrets.VITE_GOOGLE_GENAI_API_KEY }} npm run build
|
||||
|
||||
@@ -18,12 +18,46 @@ module.exports = {
|
||||
NODE_ENV: 'production', // Set the Node.js environment to production
|
||||
name: 'flyer-crawler-api',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
NODE_ENV: 'test', // Set to 'test' to match the environment purpose and disable pino-pretty
|
||||
name: 'flyer-crawler-api-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Development Environment Settings
|
||||
env_development: {
|
||||
@@ -31,6 +65,23 @@ module.exports = {
|
||||
name: 'flyer-crawler-api-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
},
|
||||
{
|
||||
@@ -43,12 +94,46 @@ module.exports = {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Development Environment Settings
|
||||
env_development: {
|
||||
@@ -56,6 +141,23 @@ module.exports = {
|
||||
name: 'flyer-crawler-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
},
|
||||
{
|
||||
@@ -68,12 +170,46 @@ module.exports = {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-analytics-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
// Development Environment Settings
|
||||
env_development: {
|
||||
@@ -81,6 +217,23 @@ module.exports = {
|
||||
name: 'flyer-crawler-analytics-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
DB_NAME: process.env.DB_NAME,
|
||||
REDIS_URL: process.env.REDIS_URL,
|
||||
REDIS_PASSWORD: process.env.REDIS_PASSWORD,
|
||||
FRONTEND_URL: process.env.FRONTEND_URL,
|
||||
JWT_SECRET: process.env.JWT_SECRET,
|
||||
GEMINI_API_KEY: process.env.GEMINI_API_KEY,
|
||||
GOOGLE_MAPS_API_KEY: process.env.GOOGLE_MAPS_API_KEY,
|
||||
SMTP_HOST: process.env.SMTP_HOST,
|
||||
SMTP_PORT: process.env.SMTP_PORT,
|
||||
SMTP_SECURE: process.env.SMTP_SECURE,
|
||||
SMTP_USER: process.env.SMTP_USER,
|
||||
SMTP_PASS: process.env.SMTP_PASS,
|
||||
SMTP_FROM_EMAIL: process.env.SMTP_FROM_EMAIL,
|
||||
},
|
||||
},
|
||||
],
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.0",
|
||||
"version": "0.1.4",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.0",
|
||||
"version": "0.1.4",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.1.0",
|
||||
"version": "0.1.4",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -36,7 +36,7 @@ vi.mock('pdfjs-dist', () => ({
|
||||
// Mock the new config module
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: { version: '1.0.0', commitMessage: 'Initial commit', commitUrl: '#' },
|
||||
app: { version: '20250101-1200:abc1234:1.0.0', commitMessage: 'Initial commit', commitUrl: '#' },
|
||||
google: { mapsEmbedApiKey: 'mock-key' },
|
||||
},
|
||||
}));
|
||||
@@ -588,11 +588,11 @@ describe('App Component', () => {
|
||||
// Mock the config module for this specific test
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: { version: '1.0.1', commitMessage: 'New feature!', commitUrl: '#' },
|
||||
app: { version: '20250101-1200:abc1234:1.0.1', commitMessage: 'New feature!', commitUrl: '#' },
|
||||
google: { mapsEmbedApiKey: 'mock-key' },
|
||||
},
|
||||
}));
|
||||
localStorageMock.setItem('lastSeenVersion', '1.0.0');
|
||||
localStorageMock.setItem('lastSeenVersion', '20250101-1200:abc1234:1.0.0');
|
||||
renderApp();
|
||||
await expect(screen.findByTestId('whats-new-modal-mock')).resolves.toBeInTheDocument();
|
||||
});
|
||||
@@ -741,7 +741,7 @@ describe('App Component', () => {
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: {
|
||||
version: '2.0.0',
|
||||
version: '20250101-1200:abc1234:2.0.0',
|
||||
commitMessage: 'A new version!',
|
||||
commitUrl: 'http://example.com/commit/2.0.0',
|
||||
},
|
||||
@@ -752,14 +752,14 @@ describe('App Component', () => {
|
||||
|
||||
it('should display the version number and commit link', () => {
|
||||
renderApp();
|
||||
const versionLink = screen.getByText(`Version: 2.0.0`);
|
||||
const versionLink = screen.getByText(`Version: 20250101-1200:abc1234:2.0.0`);
|
||||
expect(versionLink).toBeInTheDocument();
|
||||
expect(versionLink).toHaveAttribute('href', 'http://example.com/commit/2.0.0');
|
||||
});
|
||||
|
||||
it('should open the "What\'s New" modal when the question mark icon is clicked', async () => {
|
||||
// Pre-set the localStorage to prevent the modal from opening automatically
|
||||
localStorageMock.setItem('lastSeenVersion', '2.0.0');
|
||||
localStorageMock.setItem('lastSeenVersion', '20250101-1200:abc1234:2.0.0');
|
||||
|
||||
renderApp();
|
||||
expect(screen.queryByTestId('whats-new-modal-mock')).not.toBeInTheDocument();
|
||||
|
||||
@@ -44,7 +44,7 @@ export const FlyerCorrectionTool: React.FC<FlyerCorrectionToolProps> = ({
|
||||
})
|
||||
.catch((err) => {
|
||||
console.error('[DEBUG] FlyerCorrectionTool: Failed to fetch image.', { err });
|
||||
logger.error('Failed to fetch image for correction tool', { error: err });
|
||||
logger.error({ error: err }, 'Failed to fetch image for correction tool');
|
||||
notifyError('Could not load the image for correction.');
|
||||
});
|
||||
}
|
||||
@@ -164,7 +164,7 @@ export const FlyerCorrectionTool: React.FC<FlyerCorrectionToolProps> = ({
|
||||
const msg = err instanceof Error ? err.message : 'An unknown error occurred.';
|
||||
console.error('[DEBUG] handleRescan: Caught an error.', { error: err });
|
||||
notifyError(msg);
|
||||
logger.error('Error during rescan:', { error: err });
|
||||
logger.error({ error: err }, 'Error during rescan:');
|
||||
} finally {
|
||||
console.debug('[DEBUG] handleRescan: Finished. Setting isProcessing=false.');
|
||||
setIsProcessing(false);
|
||||
|
||||
@@ -73,12 +73,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle file upload and start polling', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for upload and polling.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-123' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue(
|
||||
new Response(JSON.stringify({ state: 'active', progress: { message: 'Checking...' } })),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-123' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'active',
|
||||
progress: { message: 'Checking...' },
|
||||
});
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering component and preparing file.');
|
||||
renderComponent();
|
||||
@@ -131,12 +130,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle file upload via drag and drop', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for drag and drop.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-dnd' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue(
|
||||
new Response(JSON.stringify({ state: 'active', progress: { message: 'Dropped...' } })),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-dnd' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'active',
|
||||
progress: { message: 'Dropped...' },
|
||||
});
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering component and preparing file for drop.');
|
||||
renderComponent();
|
||||
@@ -159,16 +157,10 @@ describe('FlyerUploader', () => {
|
||||
it('should poll for status, complete successfully, and redirect', async () => {
|
||||
const onProcessingComplete = vi.fn();
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock sequence for polling.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-123' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-123' });
|
||||
mockedAiApiClient.getJobStatus
|
||||
.mockResolvedValueOnce(
|
||||
new Response(JSON.stringify({ state: 'active', progress: { message: 'Analyzing...' } })),
|
||||
)
|
||||
.mockResolvedValueOnce(
|
||||
new Response(JSON.stringify({ state: 'completed', returnValue: { flyerId: 42 } })),
|
||||
);
|
||||
.mockResolvedValueOnce({ state: 'active', progress: { message: 'Analyzing...' } })
|
||||
.mockResolvedValueOnce({ state: 'completed', returnValue: { flyerId: 42 } });
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering component and uploading file.');
|
||||
renderComponent(onProcessingComplete);
|
||||
@@ -229,12 +221,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle a failed job', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for a failed job.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-fail' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue(
|
||||
new Response(JSON.stringify({ state: 'failed', failedReason: 'AI model exploded' })),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-fail' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'failed',
|
||||
failedReason: 'AI model exploded',
|
||||
});
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering and uploading.');
|
||||
renderComponent();
|
||||
@@ -260,11 +251,82 @@ describe('FlyerUploader', () => {
|
||||
console.log('--- [TEST LOG] ---: 6. "Upload Another" button confirmed.');
|
||||
});
|
||||
|
||||
it('should clear the polling timeout when a job fails', async () => {
|
||||
const clearTimeoutSpy = vi.spyOn(global, 'clearTimeout');
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for failed job timeout clearance.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-fail-timeout' });
|
||||
|
||||
// We need at least one 'active' response to establish a timeout loop so we have something to clear
|
||||
mockedAiApiClient.getJobStatus
|
||||
.mockResolvedValueOnce({ state: 'active', progress: { message: 'Working...' } })
|
||||
.mockResolvedValueOnce({ state: 'failed', failedReason: 'Fatal Error' });
|
||||
|
||||
renderComponent();
|
||||
const file = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const input = screen.getByLabelText(/click to select a file/i);
|
||||
|
||||
fireEvent.change(input, { target: { files: [file] } });
|
||||
|
||||
// Wait for the first poll to complete and UI to update to "Working..."
|
||||
await screen.findByText('Working...');
|
||||
|
||||
// Advance time to trigger the second poll
|
||||
await act(async () => {
|
||||
vi.advanceTimersByTime(3000);
|
||||
});
|
||||
|
||||
// Wait for the failure UI
|
||||
await screen.findByText(/Processing failed: Fatal Error/i);
|
||||
|
||||
// Verify clearTimeout was called
|
||||
expect(clearTimeoutSpy).toHaveBeenCalled();
|
||||
|
||||
// Verify no further polling occurs
|
||||
const callsBefore = mockedAiApiClient.getJobStatus.mock.calls.length;
|
||||
await act(async () => {
|
||||
vi.advanceTimersByTime(10000);
|
||||
});
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(callsBefore);
|
||||
|
||||
clearTimeoutSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should clear the polling timeout when the component unmounts', async () => {
|
||||
const clearTimeoutSpy = vi.spyOn(global, 'clearTimeout');
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for unmount timeout clearance.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-unmount' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'active',
|
||||
progress: { message: 'Polling...' },
|
||||
});
|
||||
|
||||
const { unmount } = renderComponent();
|
||||
const file = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const input = screen.getByLabelText(/click to select a file/i);
|
||||
|
||||
fireEvent.change(input, { target: { files: [file] } });
|
||||
|
||||
// Wait for the first poll to complete and the UI to show the polling state
|
||||
await screen.findByText('Polling...');
|
||||
|
||||
// Now that we are in a polling state (and a timeout is set), unmount the component
|
||||
console.log('--- [TEST LOG] ---: 2. Unmounting component to trigger cleanup effect.');
|
||||
unmount();
|
||||
|
||||
// Verify that the cleanup function in the useEffect hook was called
|
||||
expect(clearTimeoutSpy).toHaveBeenCalled();
|
||||
console.log('--- [TEST LOG] ---: 3. clearTimeout confirmed.');
|
||||
|
||||
clearTimeoutSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle a duplicate flyer error (409)', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for 409 duplicate error.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ flyerId: 99, message: 'Duplicate' }), { status: 409 }),
|
||||
);
|
||||
// The API client now throws a structured error for non-2xx responses.
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockRejectedValue({
|
||||
status: 409,
|
||||
body: { flyerId: 99, message: 'Duplicate' },
|
||||
});
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering and uploading.');
|
||||
renderComponent();
|
||||
@@ -295,12 +357,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should allow the user to stop watching progress', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for infinite polling.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-stop' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue(
|
||||
new Response(JSON.stringify({ state: 'active', progress: { message: 'Analyzing...' } })),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-stop' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'active',
|
||||
progress: { message: 'Analyzing...' },
|
||||
} as any);
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering and uploading.');
|
||||
renderComponent();
|
||||
@@ -362,9 +423,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle a generic network error during upload', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for generic upload error.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockRejectedValue(
|
||||
new Error('Network Error During Upload'),
|
||||
);
|
||||
// Simulate a structured error from the API client
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockRejectedValue({
|
||||
status: 500,
|
||||
body: { message: 'Network Error During Upload' },
|
||||
});
|
||||
renderComponent();
|
||||
const file = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const input = screen.getByLabelText(/click to select a file/i);
|
||||
@@ -379,9 +442,7 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle a generic network error during polling', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for polling error.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-poll-fail' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-poll-fail' });
|
||||
mockedAiApiClient.getJobStatus.mockRejectedValue(new Error('Polling Network Error'));
|
||||
|
||||
renderComponent();
|
||||
@@ -398,11 +459,9 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle a completed job with a missing flyerId', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for malformed completion payload.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue(
|
||||
new Response(JSON.stringify({ jobId: 'job-no-flyerid' }), { status: 200 }),
|
||||
);
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-no-flyerid' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue(
|
||||
new Response(JSON.stringify({ state: 'completed', returnValue: {} })), // No flyerId
|
||||
{ state: 'completed', returnValue: {} }, // No flyerId
|
||||
);
|
||||
|
||||
renderComponent();
|
||||
@@ -419,6 +478,27 @@ describe('FlyerUploader', () => {
|
||||
console.log('--- [TEST LOG] ---: 4. Assertions passed.');
|
||||
});
|
||||
|
||||
it('should handle a non-JSON response during polling', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for non-JSON response.');
|
||||
// The actual function would throw, so we mock the rejection.
|
||||
// The new getJobStatus would throw an error like "Failed to parse JSON..."
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-bad-json' });
|
||||
mockedAiApiClient.getJobStatus.mockRejectedValue(
|
||||
new Error('Failed to parse JSON response from server. Body: <html>502 Bad Gateway</html>'),
|
||||
);
|
||||
|
||||
renderComponent();
|
||||
const file = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const input = screen.getByLabelText(/click to select a file/i);
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Firing file change event.');
|
||||
fireEvent.change(input, { target: { files: [file] } });
|
||||
|
||||
console.log('--- [TEST LOG] ---: 3. Awaiting error message.');
|
||||
expect(await screen.findByText(/Failed to parse JSON response from server/i)).toBeInTheDocument();
|
||||
console.log('--- [TEST LOG] ---: 4. Assertions passed.');
|
||||
});
|
||||
|
||||
it('should do nothing if the file input is cancelled', () => {
|
||||
renderComponent();
|
||||
const input = screen.getByLabelText(/click to select a file/i);
|
||||
|
||||
@@ -60,14 +60,8 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
const pollStatus = async () => {
|
||||
console.debug(`[DEBUG] pollStatus(): Polling for jobId: ${jobId}`);
|
||||
try {
|
||||
const statusResponse = await getJobStatus(jobId);
|
||||
console.debug(`[DEBUG] pollStatus(): API response status: ${statusResponse.status}`);
|
||||
if (!statusResponse.ok) {
|
||||
throw new Error(`Failed to get job status (HTTP ${statusResponse.status})`);
|
||||
}
|
||||
|
||||
const job = await statusResponse.json();
|
||||
console.debug('[DEBUG] pollStatus(): Job status received:', job);
|
||||
const job = await getJobStatus(jobId); // Now returns parsed JSON directly
|
||||
console.debug('[DEBUG] pollStatus(): Job status received:', job); // The rest of the logic remains the same
|
||||
|
||||
if (job.progress) {
|
||||
setProcessingStages(job.progress.stages || []);
|
||||
@@ -97,7 +91,13 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
console.debug(
|
||||
`[DEBUG] pollStatus(): Job state is "failed". Reason: ${job.failedReason}`,
|
||||
);
|
||||
// Explicitly clear any pending timeout to stop the polling loop immediately.
|
||||
if (pollingTimeoutRef.current) {
|
||||
clearTimeout(pollingTimeoutRef.current);
|
||||
}
|
||||
setErrorMessage(`Processing failed: ${job.failedReason || 'Unknown error'}`);
|
||||
// Clear any stale "in-progress" messages to avoid user confusion.
|
||||
setStatusMessage(null);
|
||||
setProcessingState('error');
|
||||
break;
|
||||
|
||||
@@ -112,7 +112,7 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
break;
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error('Error during polling:', { error });
|
||||
logger.error({ error }, 'Error during polling:');
|
||||
setErrorMessage(
|
||||
error instanceof Error ? error.message : 'An unexpected error occurred during polling.',
|
||||
);
|
||||
@@ -150,29 +150,24 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
`[DEBUG] processFile(): Checksum generated: ${checksum}. Calling uploadAndProcessFlyer.`,
|
||||
);
|
||||
|
||||
const startResponse = await uploadAndProcessFlyer(file, checksum);
|
||||
console.debug(`[DEBUG] processFile(): Upload response status: ${startResponse.status}`);
|
||||
|
||||
if (!startResponse.ok) {
|
||||
const errorData = await startResponse.json();
|
||||
console.debug('[DEBUG] processFile(): Upload failed. Error data:', errorData);
|
||||
if (startResponse.status === 409 && errorData.flyerId) {
|
||||
setErrorMessage(`This flyer has already been processed. You can view it here:`);
|
||||
setDuplicateFlyerId(errorData.flyerId);
|
||||
} else {
|
||||
setErrorMessage(errorData.message || `Upload failed with status ${startResponse.status}`);
|
||||
}
|
||||
setProcessingState('error');
|
||||
return;
|
||||
}
|
||||
|
||||
const { jobId: newJobId } = await startResponse.json();
|
||||
// The API client now returns parsed JSON on success or throws a structured error on failure.
|
||||
const { jobId: newJobId } = await uploadAndProcessFlyer(file, checksum);
|
||||
console.debug(`[DEBUG] processFile(): Upload successful. Received jobId: ${newJobId}`);
|
||||
setJobId(newJobId);
|
||||
setProcessingState('polling');
|
||||
} catch (error) {
|
||||
logger.error('An unexpected error occurred during file upload:', { error });
|
||||
setErrorMessage(error instanceof Error ? error.message : 'An unexpected error occurred.');
|
||||
} catch (error: any) {
|
||||
// Handle the structured error thrown by the API client.
|
||||
logger.error({ error }, 'An error occurred during file upload:');
|
||||
// Handle 409 Conflict for duplicate flyers
|
||||
if (error?.status === 409 && error.body?.flyerId) {
|
||||
setErrorMessage(`This flyer has already been processed. You can view it here:`);
|
||||
setDuplicateFlyerId(error.body.flyerId);
|
||||
} else {
|
||||
// Handle other errors (e.g., validation, server errors)
|
||||
const message =
|
||||
error?.body?.message || error?.message || 'An unexpected error occurred during upload.';
|
||||
setErrorMessage(message);
|
||||
}
|
||||
setProcessingState('error');
|
||||
}
|
||||
}, []);
|
||||
|
||||
@@ -1,94 +1,53 @@
|
||||
// src/middleware/errorHandler.ts
|
||||
import { Request, Response, NextFunction } from 'express';
|
||||
import {
|
||||
DatabaseError,
|
||||
UniqueConstraintError,
|
||||
ForeignKeyConstraintError,
|
||||
NotFoundError,
|
||||
ValidationError,
|
||||
ValidationIssue,
|
||||
} from '../services/db/errors.db';
|
||||
import crypto from 'crypto';
|
||||
import { ZodError } from 'zod';
|
||||
import { NotFoundError, UniqueConstraintError, ValidationError } from '../services/db/errors.db';
|
||||
import { logger } from '../services/logger.server';
|
||||
|
||||
interface HttpError extends Error {
|
||||
status?: number;
|
||||
}
|
||||
|
||||
export const errorHandler = (err: HttpError, req: Request, res: Response, next: NextFunction) => {
|
||||
// If the response headers have already been sent, we must delegate to the default Express error handler.
|
||||
/**
|
||||
* A centralized error handling middleware for the Express application.
|
||||
* This middleware should be the LAST `app.use()` call to catch all errors from previous routes and middleware.
|
||||
*
|
||||
* It standardizes error responses and ensures consistent logging.
|
||||
*/
|
||||
export const errorHandler = (err: Error, req: Request, res: Response, next: NextFunction) => {
|
||||
// If headers have already been sent, delegate to the default Express error handler.
|
||||
if (res.headersSent) {
|
||||
return next(err);
|
||||
}
|
||||
|
||||
// The pino-http middleware guarantees that `req.log` will be available.
|
||||
const log = req.log;
|
||||
// Use the request-scoped logger if available, otherwise fall back to the global logger.
|
||||
const log = req.log || logger;
|
||||
|
||||
// --- 1. Determine Final Status Code and Message ---
|
||||
let statusCode = err.status ?? 500;
|
||||
const message = err.message;
|
||||
let validationIssues: ValidationIssue[] | undefined;
|
||||
let errorId: string | undefined;
|
||||
|
||||
// Refine the status code for known error types. Check for most specific types first.
|
||||
if (err instanceof UniqueConstraintError) {
|
||||
statusCode = 409; // Conflict
|
||||
} else if (err instanceof NotFoundError) {
|
||||
statusCode = 404;
|
||||
} else if (err instanceof ForeignKeyConstraintError) {
|
||||
statusCode = 400;
|
||||
} else if (err instanceof ValidationError) {
|
||||
statusCode = 400;
|
||||
validationIssues = err.validationErrors;
|
||||
} else if (err instanceof DatabaseError) {
|
||||
// This is a generic fallback for other database errors that are not the specific subclasses above.
|
||||
statusCode = err.status;
|
||||
} else if (err.name === 'UnauthorizedError') {
|
||||
statusCode = err.status || 401;
|
||||
// --- Handle Zod Validation Errors ---
|
||||
if (err instanceof ZodError) {
|
||||
log.warn({ err: err.flatten() }, 'Request validation failed');
|
||||
return res.status(400).json({
|
||||
message: 'The request data is invalid.',
|
||||
errors: err.errors.map((e) => ({ path: e.path, message: e.message })),
|
||||
});
|
||||
}
|
||||
|
||||
// --- 2. Log Based on Final Status Code ---
|
||||
// Log the full error details for debugging, especially for server errors.
|
||||
if (statusCode >= 500) {
|
||||
errorId = crypto.randomBytes(4).toString('hex');
|
||||
// The request-scoped logger already contains user, IP, and request_id.
|
||||
// We add the full error and the request object itself.
|
||||
// Pino's `redact` config will automatically sanitize sensitive fields in `req`.
|
||||
log.error(
|
||||
{
|
||||
err,
|
||||
errorId,
|
||||
req: { method: req.method, url: req.originalUrl, headers: req.headers, body: req.body },
|
||||
},
|
||||
`Unhandled API Error (ID: ${errorId})`,
|
||||
);
|
||||
} else {
|
||||
// For 4xx errors, log at a lower level (e.g., 'warn') to avoid flooding error trackers.
|
||||
// We include the validation errors in the log context if they exist.
|
||||
log.warn(
|
||||
{
|
||||
err,
|
||||
validationErrors: validationIssues, // Add validation issues to the log object
|
||||
statusCode,
|
||||
},
|
||||
`Client Error on ${req.method} ${req.path}: ${message}`,
|
||||
);
|
||||
// --- Handle Custom Operational Errors ---
|
||||
if (err instanceof NotFoundError) {
|
||||
log.info({ err }, 'Resource not found');
|
||||
return res.status(404).json({ message: err.message });
|
||||
}
|
||||
|
||||
// --- TEST ENVIRONMENT DEBUGGING ---
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
console.error('--- [TEST] UNHANDLED ERROR ---', err);
|
||||
if (err instanceof UniqueConstraintError || err instanceof ValidationError) {
|
||||
log.warn({ err }, 'Constraint or validation error occurred');
|
||||
return res.status(400).json({ message: err.message });
|
||||
}
|
||||
|
||||
// --- 3. Send Response ---
|
||||
// In production, send a generic message for 5xx errors.
|
||||
// In dev/test, send the actual error message for easier debugging.
|
||||
const responseMessage =
|
||||
statusCode >= 500 && process.env.NODE_ENV === 'production'
|
||||
? `An unexpected server error occurred. Please reference error ID: ${errorId}`
|
||||
: message;
|
||||
// --- Handle Generic Errors ---
|
||||
// Log the full error object for debugging. The pino logger will handle redaction.
|
||||
log.error({ err }, 'An unhandled error occurred in an Express route');
|
||||
|
||||
res.status(statusCode).json({
|
||||
message: responseMessage,
|
||||
...(validationIssues && { errors: validationIssues }), // Conditionally add the 'errors' array if it exists
|
||||
});
|
||||
};
|
||||
// In production, send a generic message to avoid leaking implementation details.
|
||||
if (process.env.NODE_ENV === 'production') {
|
||||
return res.status(500).json({ message: 'An internal server error occurred.' });
|
||||
}
|
||||
|
||||
// In development, send more details for easier debugging.
|
||||
return res.status(500).json({ message: err.message, stack: err.stack });
|
||||
};
|
||||
@@ -135,6 +135,7 @@ router.get('/corrections', async (req, res, next: NextFunction) => {
|
||||
const corrections = await db.adminRepo.getSuggestedCorrections(req.log);
|
||||
res.json(corrections);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching suggested corrections');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -144,6 +145,7 @@ router.get('/brands', async (req, res, next: NextFunction) => {
|
||||
const brands = await db.flyerRepo.getAllBrands(req.log);
|
||||
res.json(brands);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching brands');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -153,6 +155,7 @@ router.get('/stats', async (req, res, next: NextFunction) => {
|
||||
const stats = await db.adminRepo.getApplicationStats(req.log);
|
||||
res.json(stats);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching application stats');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -162,6 +165,7 @@ router.get('/stats/daily', async (req, res, next: NextFunction) => {
|
||||
const dailyStats = await db.adminRepo.getDailyStatsForLast30Days(req.log);
|
||||
res.json(dailyStats);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching daily stats');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -176,6 +180,7 @@ router.post(
|
||||
await db.adminRepo.approveCorrection(params.id, req.log); // params.id is now safely typed as number
|
||||
res.status(200).json({ message: 'Correction approved successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error approving correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -191,6 +196,7 @@ router.post(
|
||||
await db.adminRepo.rejectCorrection(params.id, req.log); // params.id is now safely typed as number
|
||||
res.status(200).json({ message: 'Correction rejected successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error rejecting correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -210,6 +216,7 @@ router.put(
|
||||
);
|
||||
res.status(200).json(updatedCorrection);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating suggested correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -225,6 +232,7 @@ router.put(
|
||||
const updatedRecipe = await db.adminRepo.updateRecipeStatus(params.id, body.status, req.log); // This is still a standalone function in admin.db.ts
|
||||
res.status(200).json(updatedRecipe);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating recipe status');
|
||||
next(error); // Pass all errors to the central error handler
|
||||
}
|
||||
},
|
||||
@@ -250,6 +258,7 @@ router.post(
|
||||
logger.info({ brandId: params.id, logoUrl }, `Brand logo updated for brand ID: ${params.id}`);
|
||||
res.status(200).json({ message: 'Brand logo updated successfully.', logoUrl });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating brand logo');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -260,6 +269,7 @@ router.get('/unmatched-items', async (req, res, next: NextFunction) => {
|
||||
const items = await db.adminRepo.getUnmatchedFlyerItems(req.log);
|
||||
res.json(items);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching unmatched items');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -279,6 +289,7 @@ router.delete(
|
||||
await db.recipeRepo.deleteRecipe(params.recipeId, userProfile.user.user_id, true, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error deleting recipe');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -297,6 +308,7 @@ router.delete(
|
||||
await db.flyerRepo.deleteFlyer(params.flyerId, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error deleting flyer');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -316,6 +328,7 @@ router.put(
|
||||
); // This is still a standalone function in admin.db.ts
|
||||
res.status(200).json(updatedComment);
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error updating comment status');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -326,6 +339,7 @@ router.get('/users', async (req, res, next: NextFunction) => {
|
||||
const users = await db.adminRepo.getAllUsers(req.log);
|
||||
res.json(users);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching users');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -345,6 +359,7 @@ router.get(
|
||||
const logs = await db.adminRepo.getActivityLog(limit, offset, req.log);
|
||||
res.json(logs);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching activity log');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -360,6 +375,7 @@ router.get(
|
||||
const user = await db.userRepo.findUserProfileById(params.id, req.log);
|
||||
res.json(user);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching user profile');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -395,6 +411,7 @@ router.delete(
|
||||
await db.userRepo.deleteUserById(params.id, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error deleting user');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -478,6 +495,7 @@ router.post(
|
||||
.status(202)
|
||||
.json({ message: `File cleanup job for flyer ID ${params.flyerId} has been enqueued.` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing cleanup job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -500,6 +518,7 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
||||
.status(202)
|
||||
.json({ message: `Failing test job has been enqueued successfully. Job ID: ${job.id}` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing failing job');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -572,6 +591,7 @@ router.get('/queues/status', async (req: Request, res: Response, next: NextFunct
|
||||
);
|
||||
res.json(queueStatuses);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching queue statuses');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -620,6 +640,7 @@ router.post(
|
||||
);
|
||||
res.status(200).json({ message: `Job ${jobId} has been successfully marked for retry.` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error retrying job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -651,6 +672,7 @@ router.post(
|
||||
.status(202)
|
||||
.json({ message: 'Successfully enqueued weekly analytics job.', jobId: job.id });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing weekly analytics job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -535,6 +535,27 @@ describe('AI Routes (/api/ai)', () => {
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toBe('Root Store');
|
||||
});
|
||||
|
||||
it('should default item quantity to 1 if missing', async () => {
|
||||
const payloadMissingQuantity = {
|
||||
checksum: 'qty-checksum',
|
||||
originalFileName: 'flyer-qty.jpg',
|
||||
extractedData: {
|
||||
store_name: 'Qty Store',
|
||||
items: [{ name: 'Item without qty', price: 100 }],
|
||||
},
|
||||
};
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/flyers/process')
|
||||
.field('data', JSON.stringify(payloadMissingQuantity))
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
const itemsArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][1];
|
||||
expect(itemsArg[0].quantity).toBe(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('POST /check-flyer', () => {
|
||||
|
||||
@@ -424,6 +424,7 @@ router.post(
|
||||
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
||||
...item,
|
||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
||||
quantity: item.quantity ?? 1, // Default to 1 to satisfy DB constraint
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
updated_at: new Date().toISOString(),
|
||||
|
||||
@@ -134,8 +134,8 @@ router.post(
|
||||
// If the email is a duplicate, return a 409 Conflict status.
|
||||
return res.status(409).json({ message: error.message });
|
||||
}
|
||||
// The createUser method now handles its own transaction logging, so we just log the route failure.
|
||||
logger.error({ error }, `User registration route failed for email: ${email}.`);
|
||||
// Pass the error to the centralized handler
|
||||
return next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -108,6 +108,7 @@ router.post(
|
||||
const items = await db.flyerRepo.getFlyerItemsForFlyers(body.flyerIds, req.log);
|
||||
res.json(items);
|
||||
} catch (error) {
|
||||
req.log.error({ error }, 'Error fetching batch flyer items');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -127,6 +128,7 @@ router.post(
|
||||
const count = await db.flyerRepo.countFlyerItemsForFlyers(body.flyerIds ?? [], req.log);
|
||||
res.json({ count });
|
||||
} catch (error) {
|
||||
req.log.error({ error }, 'Error counting batch flyer items');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -39,10 +39,7 @@ router.get('/db-schema', validateRequest(emptySchema), async (req, res, next: Ne
|
||||
}
|
||||
return res.status(200).json({ success: true, message: 'All required database tables exist.' });
|
||||
} catch (error: unknown) {
|
||||
logger.error(
|
||||
{ error: error instanceof Error ? error.message : error },
|
||||
'Error during DB schema check:',
|
||||
);
|
||||
logger.error({ error }, 'Error during DB schema check:');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -133,6 +130,7 @@ router.get(
|
||||
}
|
||||
throw new Error(`Unexpected Redis ping response: ${reply}`); // This will be caught below
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error checking Redis health');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -46,7 +46,6 @@ router.get(
|
||||
}
|
||||
|
||||
// Check if there was output to stderr, even if the exit code was 0 (success).
|
||||
// This handles warnings or non-fatal errors that should arguably be treated as failures in this context.
|
||||
if (stderr && stderr.trim().length > 0) {
|
||||
logger.error({ stderr }, '[API /pm2-status] PM2 executed but produced stderr:');
|
||||
return next(new Error(`PM2 command produced an error: ${stderr}`));
|
||||
@@ -86,6 +85,7 @@ router.post(
|
||||
|
||||
res.json(coordinates);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error geocoding address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -77,7 +77,7 @@ router.use(passport.authenticate('jwt', { session: false }));
|
||||
// Ensure the directory for avatar uploads exists.
|
||||
const avatarUploadDir = path.join(process.cwd(), 'public', 'uploads', 'avatars');
|
||||
fs.mkdir(avatarUploadDir, { recursive: true }).catch((err) => {
|
||||
logger.error('Failed to create avatar upload directory:', err);
|
||||
logger.error({ err }, 'Failed to create avatar upload directory');
|
||||
});
|
||||
|
||||
// Define multer storage configuration. The `req.user` object will be available
|
||||
@@ -122,6 +122,7 @@ router.post(
|
||||
);
|
||||
res.json(updatedProfile);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error uploading avatar');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -151,6 +152,7 @@ router.get(
|
||||
);
|
||||
res.json(notifications);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching notifications');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -168,6 +170,7 @@ router.post(
|
||||
await db.notificationRepo.markAllNotificationsAsRead(userProfile.user.user_id, req.log);
|
||||
res.status(204).send(); // No Content
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error marking all notifications as read');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -193,6 +196,7 @@ router.post(
|
||||
);
|
||||
res.status(204).send(); // Success, no content to return
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error marking notification as read');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -345,11 +349,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to add watched item');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -453,11 +453,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to create shopping list');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -516,12 +512,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
params: req.params,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, params: req.params, body: req.body }, 'Failed to add shopping list item');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -661,11 +652,7 @@ router.put(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to set user dietary restrictions');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -709,11 +696,7 @@ router.put(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to set user appliances');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -743,6 +726,7 @@ router.get(
|
||||
const address = await db.addressRepo.getAddressById(addressId, req.log); // This will throw NotFoundError if not found
|
||||
res.json(address);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching user address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -781,6 +765,7 @@ router.put(
|
||||
const addressId = await userService.upsertUserAddress(userProfile, addressData, req.log); // This was a duplicate, fixed.
|
||||
res.status(200).json({ message: 'Address updated successfully', address_id: addressId });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating user address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -51,9 +51,7 @@ export class AiAnalysisService {
|
||||
// Normalize sources to a consistent format.
|
||||
const mappedSources = (response.sources || []).map(
|
||||
(s: RawSource) =>
|
||||
(s.web
|
||||
? { uri: s.web.uri || '', title: s.web.title || 'Untitled' }
|
||||
: { uri: '', title: 'Untitled' }) as Source,
|
||||
(s.web ? { uri: s.web.uri || '', title: s.web.title || 'Untitled' } : { uri: '', title: 'Untitled' }) as Source,
|
||||
);
|
||||
return { ...response, sources: mappedSources };
|
||||
}
|
||||
@@ -84,9 +82,7 @@ export class AiAnalysisService {
|
||||
// Normalize sources to a consistent format.
|
||||
const mappedSources = (response.sources || []).map(
|
||||
(s: RawSource) =>
|
||||
(s.web
|
||||
? { uri: s.web.uri || '', title: s.web.title || 'Untitled' }
|
||||
: { uri: '', title: 'Untitled' }) as Source,
|
||||
(s.web ? { uri: s.web.uri || '', title: s.web.title || 'Untitled' } : { uri: '', title: 'Untitled' }) as Source,
|
||||
);
|
||||
return { ...response, sources: mappedSources };
|
||||
}
|
||||
|
||||
@@ -4,7 +4,13 @@
|
||||
* It communicates with the application's own backend endpoints, which then securely
|
||||
* call the Google AI services. This ensures no API keys are exposed on the client.
|
||||
*/
|
||||
import type { FlyerItem, Store, MasterGroceryItem } from '../types';
|
||||
import type {
|
||||
FlyerItem,
|
||||
Store,
|
||||
MasterGroceryItem,
|
||||
ProcessingStage,
|
||||
GroundedResponse,
|
||||
} from '../types';
|
||||
import { logger } from './logger.client';
|
||||
import { apiFetch } from './apiClient';
|
||||
|
||||
@@ -20,14 +26,14 @@ export const uploadAndProcessFlyer = async (
|
||||
file: File,
|
||||
checksum: string,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
): Promise<{ jobId: string }> => {
|
||||
const formData = new FormData();
|
||||
formData.append('flyerFile', file);
|
||||
formData.append('checksum', checksum);
|
||||
|
||||
logger.info(`[aiApiClient] Starting background processing for file: ${file.name}`);
|
||||
|
||||
return apiFetch(
|
||||
const response = await apiFetch(
|
||||
'/ai/upload-and-process',
|
||||
{
|
||||
method: 'POST',
|
||||
@@ -35,20 +41,73 @@ export const uploadAndProcessFlyer = async (
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
|
||||
if (!response.ok) {
|
||||
let errorBody;
|
||||
try {
|
||||
errorBody = await response.json();
|
||||
} catch (e) {
|
||||
errorBody = { message: await response.text() };
|
||||
}
|
||||
// Throw a structured error so the component can inspect the status and body
|
||||
throw { status: response.status, body: errorBody };
|
||||
}
|
||||
|
||||
return response.json();
|
||||
};
|
||||
|
||||
// Define the expected shape of the job status response
|
||||
export interface JobStatus {
|
||||
id: string;
|
||||
state: 'completed' | 'failed' | 'active' | 'waiting' | 'delayed' | 'paused';
|
||||
progress: {
|
||||
stages?: ProcessingStage[];
|
||||
estimatedTimeRemaining?: number;
|
||||
message?: string;
|
||||
} | null;
|
||||
returnValue: {
|
||||
flyerId?: number;
|
||||
} | null;
|
||||
failedReason: string | null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Fetches the status of a background processing job.
|
||||
* This is the second step in the new background processing flow.
|
||||
* @param jobId The ID of the job to check.
|
||||
* @param tokenOverride Optional token for testing.
|
||||
* @returns A promise that resolves to the API response with the job's status.
|
||||
* @returns A promise that resolves to the parsed job status object.
|
||||
* @throws An error if the network request fails or if the response is not valid JSON.
|
||||
*/
|
||||
export const getJobStatus = async (jobId: string, tokenOverride?: string): Promise<Response> => {
|
||||
return apiFetch(`/ai/jobs/${jobId}/status`, {}, { tokenOverride });
|
||||
export const getJobStatus = async (
|
||||
jobId: string,
|
||||
tokenOverride?: string,
|
||||
): Promise<JobStatus> => {
|
||||
const response = await apiFetch(`/ai/jobs/${jobId}/status`, {}, { tokenOverride });
|
||||
|
||||
if (!response.ok) {
|
||||
let errorText = `API Error: ${response.status} ${response.statusText}`;
|
||||
try {
|
||||
const errorBody = await response.text();
|
||||
if (errorBody) errorText = `API Error ${response.status}: ${errorBody}`;
|
||||
} catch (e) {
|
||||
// ignore if reading body fails
|
||||
}
|
||||
throw new Error(errorText);
|
||||
}
|
||||
|
||||
try {
|
||||
return await response.json();
|
||||
} catch (error) {
|
||||
const rawText = await response.text();
|
||||
throw new Error(`Failed to parse JSON response from server. Body: ${rawText}`);
|
||||
}
|
||||
};
|
||||
|
||||
export const isImageAFlyer = async (imageFile: File, tokenOverride?: string): Promise<Response> => {
|
||||
export const isImageAFlyer = (
|
||||
imageFile: File,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
const formData = new FormData();
|
||||
formData.append('image', imageFile);
|
||||
|
||||
@@ -64,7 +123,7 @@ export const isImageAFlyer = async (imageFile: File, tokenOverride?: string): Pr
|
||||
);
|
||||
};
|
||||
|
||||
export const extractAddressFromImage = async (
|
||||
export const extractAddressFromImage = (
|
||||
imageFile: File,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
@@ -81,7 +140,7 @@ export const extractAddressFromImage = async (
|
||||
);
|
||||
};
|
||||
|
||||
export const extractLogoFromImage = async (
|
||||
export const extractLogoFromImage = (
|
||||
imageFiles: File[],
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
@@ -100,7 +159,7 @@ export const extractLogoFromImage = async (
|
||||
);
|
||||
};
|
||||
|
||||
export const getQuickInsights = async (
|
||||
export const getQuickInsights = (
|
||||
items: Partial<FlyerItem>[],
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
@@ -117,7 +176,7 @@ export const getQuickInsights = async (
|
||||
);
|
||||
};
|
||||
|
||||
export const getDeepDiveAnalysis = async (
|
||||
export const getDeepDiveAnalysis = (
|
||||
items: Partial<FlyerItem>[],
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
@@ -134,7 +193,7 @@ export const getDeepDiveAnalysis = async (
|
||||
);
|
||||
};
|
||||
|
||||
export const searchWeb = async (
|
||||
export const searchWeb = (
|
||||
query: string,
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
@@ -179,7 +238,7 @@ export const planTripWithMaps = async (
|
||||
* @param prompt A description of the image to generate (e.g., a meal plan).
|
||||
* @returns A base64-encoded string of the generated PNG image.
|
||||
*/
|
||||
export const generateImageFromText = async (
|
||||
export const generateImageFromText = (
|
||||
prompt: string,
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
@@ -202,7 +261,7 @@ export const generateImageFromText = async (
|
||||
* @param text The text to be spoken.
|
||||
* @returns A base64-encoded string of the raw audio data.
|
||||
*/
|
||||
export const generateSpeechFromText = async (
|
||||
export const generateSpeechFromText = (
|
||||
text: string,
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
@@ -259,7 +318,7 @@ export const startVoiceSession = (callbacks: {
|
||||
* @param tokenOverride Optional token for testing.
|
||||
* @returns A promise that resolves to the API response containing the extracted text.
|
||||
*/
|
||||
export const rescanImageArea = async (
|
||||
export const rescanImageArea = (
|
||||
imageFile: File,
|
||||
cropArea: { x: number; y: number; width: number; height: number },
|
||||
extractionType: 'store_name' | 'dates' | 'item_details',
|
||||
@@ -270,7 +329,11 @@ export const rescanImageArea = async (
|
||||
formData.append('cropArea', JSON.stringify(cropArea));
|
||||
formData.append('extractionType', extractionType);
|
||||
|
||||
return apiFetch('/ai/rescan-area', { method: 'POST', body: formData }, { tokenOverride });
|
||||
return apiFetch(
|
||||
'/ai/rescan-area',
|
||||
{ method: 'POST', body: formData },
|
||||
{ tokenOverride },
|
||||
);
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -278,7 +341,7 @@ export const rescanImageArea = async (
|
||||
* @param watchedItems An array of the user's watched master grocery items.
|
||||
* @returns A promise that resolves to the raw `Response` object from the API.
|
||||
*/
|
||||
export const compareWatchedItemPrices = async (
|
||||
export const compareWatchedItemPrices = (
|
||||
watchedItems: MasterGroceryItem[],
|
||||
signal?: AbortSignal,
|
||||
): Promise<Response> => {
|
||||
@@ -292,5 +355,4 @@ export const compareWatchedItemPrices = async (
|
||||
body: JSON.stringify({ items: watchedItems }),
|
||||
},
|
||||
{ signal },
|
||||
);
|
||||
};
|
||||
)};
|
||||
|
||||
@@ -166,6 +166,127 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Model Fallback Logic', () => {
|
||||
const originalEnv = process.env;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.unstubAllEnvs();
|
||||
process.env = { ...originalEnv, GEMINI_API_KEY: 'test-key' };
|
||||
vi.resetModules(); // Re-import to use the new env var and re-instantiate the service
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env = originalEnv;
|
||||
vi.unstubAllEnvs();
|
||||
});
|
||||
|
||||
it('should try the next model if the first one fails with a quota error', async () => {
|
||||
// Arrange
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
|
||||
const quotaError = new Error('User rate limit exceeded due to quota');
|
||||
const successResponse = { text: 'Success from fallback model', candidates: [] };
|
||||
|
||||
// Mock the generateContent function to fail on the first call and succeed on the second
|
||||
mockGenerateContent.mockRejectedValueOnce(quotaError).mockResolvedValueOnce(successResponse);
|
||||
|
||||
const request = { contents: [{ parts: [{ text: 'test prompt' }] }] };
|
||||
|
||||
// Act
|
||||
const result = await (serviceWithFallback as any).aiClient.generateContent(request);
|
||||
|
||||
// Assert
|
||||
expect(result).toEqual(successResponse);
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(2);
|
||||
|
||||
// Check first call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, {
|
||||
model: 'gemini-2.5-flash',
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check second call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, {
|
||||
model: 'gemini-3-flash',
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check that a warning was logged
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.stringContaining(
|
||||
"Model 'gemini-2.5-flash' failed due to quota/rate limit. Trying next model.",
|
||||
),
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw immediately for non-retriable errors', async () => {
|
||||
// Arrange
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
|
||||
const nonRetriableError = new Error('Invalid API Key');
|
||||
mockGenerateContent.mockRejectedValueOnce(nonRetriableError);
|
||||
|
||||
const request = { contents: [{ parts: [{ text: 'test prompt' }] }] };
|
||||
|
||||
// Act & Assert
|
||||
await expect((serviceWithFallback as any).aiClient.generateContent(request)).rejects.toThrow(
|
||||
'Invalid API Key',
|
||||
);
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(1);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: nonRetriableError },
|
||||
`[AIService Adapter] Model 'gemini-2.5-flash' failed with a non-retriable error.`,
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw the last error if all models fail', async () => {
|
||||
// Arrange
|
||||
const { AIService } = await import('./aiService.server');
|
||||
const { logger } = await import('./logger.server');
|
||||
const serviceWithFallback = new AIService(logger);
|
||||
|
||||
const quotaError1 = new Error('Quota exhausted for model 1');
|
||||
const quotaError2 = new Error('429 Too Many Requests for model 2');
|
||||
const quotaError3 = new Error('RESOURCE_EXHAUSTED for model 3');
|
||||
|
||||
mockGenerateContent
|
||||
.mockRejectedValueOnce(quotaError1)
|
||||
.mockRejectedValueOnce(quotaError2)
|
||||
.mockRejectedValueOnce(quotaError3);
|
||||
|
||||
const request = { contents: [{ parts: [{ text: 'test prompt' }] }] };
|
||||
|
||||
// Act & Assert
|
||||
await expect((serviceWithFallback as any).aiClient.generateContent(request)).rejects.toThrow(
|
||||
quotaError3,
|
||||
);
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(3);
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, {
|
||||
model: 'gemini-2.5-flash',
|
||||
...request,
|
||||
});
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, {
|
||||
model: 'gemini-3-flash',
|
||||
...request,
|
||||
});
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(3, {
|
||||
model: 'gemini-2.5-flash-lite',
|
||||
...request,
|
||||
});
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ lastError: quotaError3 },
|
||||
'[AIService Adapter] All AI models failed. Throwing last known error.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('extractItemsFromReceiptImage', () => {
|
||||
it('should extract items from a valid AI response', async () => {
|
||||
const mockAiResponseText = `[
|
||||
|
||||
@@ -72,6 +72,7 @@ export class AIService {
|
||||
private fs: IFileSystem;
|
||||
private rateLimiter: <T>(fn: () => Promise<T>) => Promise<T>;
|
||||
private logger: Logger;
|
||||
private readonly models = ['gemini-2.5-flash', 'gemini-3-flash', 'gemini-2.5-flash-lite'];
|
||||
|
||||
constructor(logger: Logger, aiClient?: IAiClient, fs?: IFileSystem) {
|
||||
this.logger = logger;
|
||||
@@ -121,17 +122,11 @@ export class AIService {
|
||||
);
|
||||
}
|
||||
|
||||
// do not change "gemini-2.5-flash" - this is correct
|
||||
const modelName = 'gemini-2.5-flash';
|
||||
|
||||
// We create a shim/adapter that matches the old structure but uses the new SDK call pattern.
|
||||
// This preserves the dependency injection pattern used throughout the class.
|
||||
this.aiClient = genAI
|
||||
? {
|
||||
generateContent: async (request) => {
|
||||
// The model name is now injected here, into every call, as the new SDK requires.
|
||||
// Architectural guard clause: All requests from this service must have content.
|
||||
// This prevents sending invalid requests to the API and satisfies TypeScript's strictness.
|
||||
if (!request.contents || request.contents.length === 0) {
|
||||
this.logger.error(
|
||||
{ request },
|
||||
@@ -140,14 +135,7 @@ export class AIService {
|
||||
throw new Error('AIService.generateContent requires at least one content element.');
|
||||
}
|
||||
|
||||
// Architectural Fix: After the guard clause, assign the guaranteed-to-exist element
|
||||
// to a new constant. This provides a definitive type-safe variable for the compiler.
|
||||
const firstContent = request.contents[0];
|
||||
this.logger.debug(
|
||||
{ modelName, requestParts: firstContent.parts?.length ?? 0 },
|
||||
'[AIService] Calling actual generateContent via adapter.',
|
||||
);
|
||||
return genAI.models.generateContent({ model: modelName, ...request });
|
||||
return this._generateWithFallback(genAI, request);
|
||||
},
|
||||
}
|
||||
: {
|
||||
@@ -182,6 +170,54 @@ export class AIService {
|
||||
this.logger.info('---------------- [AIService] Constructor End ----------------');
|
||||
}
|
||||
|
||||
private async _generateWithFallback(
|
||||
genAI: GoogleGenAI,
|
||||
request: { contents: Content[]; tools?: Tool[] },
|
||||
): Promise<GenerateContentResponse> {
|
||||
let lastError: Error | null = null;
|
||||
|
||||
for (const modelName of this.models) {
|
||||
try {
|
||||
this.logger.info(
|
||||
`[AIService Adapter] Attempting to generate content with model: ${modelName}`,
|
||||
);
|
||||
const result = await genAI.models.generateContent({ model: modelName, ...request });
|
||||
// If the call succeeds, return the result immediately.
|
||||
return result;
|
||||
} catch (error: unknown) {
|
||||
lastError = error instanceof Error ? error : new Error(String(error));
|
||||
const errorMessage = (lastError.message || '').toLowerCase(); // Make case-insensitive
|
||||
|
||||
// Check for specific error messages indicating quota issues or model unavailability.
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') || // HTTP 429 Too Many Requests
|
||||
errorMessage.includes('resource_exhausted') || // Make case-insensitive
|
||||
errorMessage.includes('model is overloaded')
|
||||
) {
|
||||
this.logger.warn(
|
||||
`[AIService Adapter] Model '${modelName}' failed due to quota/rate limit. Trying next model. Error: ${errorMessage}`,
|
||||
);
|
||||
continue; // Try the next model in the list.
|
||||
} else {
|
||||
// For other errors (e.g., invalid input, safety settings), fail immediately.
|
||||
this.logger.error(
|
||||
{ error: lastError },
|
||||
`[AIService Adapter] Model '${modelName}' failed with a non-retriable error.`,
|
||||
);
|
||||
throw lastError;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// If all models in the list have failed, throw the last error encountered.
|
||||
this.logger.error(
|
||||
{ lastError },
|
||||
'[AIService Adapter] All AI models failed. Throwing last known error.',
|
||||
);
|
||||
throw lastError || new Error('All AI models failed to generate content.');
|
||||
}
|
||||
|
||||
private async serverFileToGenerativePart(path: string, mimeType: string) {
|
||||
const fileData = await this.fs.readFile(path);
|
||||
return {
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
// src/services/apiClient.ts
|
||||
import { Profile, ShoppingListItem, SearchQuery, Budget, Address } from '../types';
|
||||
import { logger } from './logger.client';
|
||||
import { eventBus } from './eventBus';
|
||||
|
||||
// This constant should point to your backend API.
|
||||
// It's often a good practice to store this in an environment variable.
|
||||
@@ -62,12 +63,12 @@ const refreshToken = async (): Promise<string> => {
|
||||
logger.info('Successfully refreshed access token.');
|
||||
return data.token;
|
||||
} catch (error) {
|
||||
logger.error('Failed to refresh token. User will be logged out.', { error });
|
||||
logger.error({ error }, 'Failed to refresh token. User session has expired.');
|
||||
// Only perform browser-specific actions if in the browser environment.
|
||||
if (typeof window !== 'undefined') {
|
||||
localStorage.removeItem('authToken');
|
||||
// A hard redirect is a simple way to reset the app state to logged-out.
|
||||
// window.location.href = '/'; // Removed to allow the caller to handle session expiry.
|
||||
// Dispatch a global event that the UI layer can listen for to handle session expiry.
|
||||
eventBus.dispatch('sessionExpired');
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
@@ -144,9 +145,8 @@ export const apiFetch = async (
|
||||
// --- DEBUG LOGGING for failed requests ---
|
||||
if (!response.ok) {
|
||||
const responseText = await response.clone().text();
|
||||
logger.error(
|
||||
`apiFetch: Request to ${fullUrl} failed with status ${response.status}. Response body:`,
|
||||
responseText,
|
||||
logger.error({ url: fullUrl, status: response.status, body: responseText },
|
||||
'apiFetch: Request failed',
|
||||
);
|
||||
}
|
||||
// --- END DEBUG LOGGING ---
|
||||
|
||||
31
src/services/eventBus.ts
Normal file
31
src/services/eventBus.ts
Normal file
@@ -0,0 +1,31 @@
|
||||
// src/services/eventBus.ts
|
||||
|
||||
/**
|
||||
* A simple, generic event bus for cross-component communication without direct coupling.
|
||||
* This is particularly useful for broadcasting application-wide events, such as session expiry.
|
||||
*/
|
||||
|
||||
type EventCallback = (data?: any) => void;
|
||||
|
||||
class EventBus {
|
||||
private listeners: { [key: string]: EventCallback[] } = {};
|
||||
|
||||
on(event: string, callback: EventCallback): void {
|
||||
if (!this.listeners[event]) {
|
||||
this.listeners[event] = [];
|
||||
}
|
||||
this.listeners[event].push(callback);
|
||||
}
|
||||
|
||||
off(event: string, callback: EventCallback): void {
|
||||
if (!this.listeners[event]) return;
|
||||
this.listeners[event] = this.listeners[event].filter((l) => l !== callback);
|
||||
}
|
||||
|
||||
dispatch(event: string, data?: any): void {
|
||||
if (!this.listeners[event]) return;
|
||||
this.listeners[event].forEach((callback) => callback(data));
|
||||
}
|
||||
}
|
||||
|
||||
export const eventBus = new EventBus();
|
||||
@@ -25,10 +25,7 @@ export class GeocodingService {
|
||||
return JSON.parse(cached);
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error, cacheKey },
|
||||
'Redis GET or JSON.parse command failed. Proceeding without cache.',
|
||||
);
|
||||
logger.error({ err: error, cacheKey }, 'Redis GET or JSON.parse command failed. Proceeding without cache.');
|
||||
}
|
||||
|
||||
if (process.env.GOOGLE_MAPS_API_KEY) {
|
||||
@@ -44,8 +41,8 @@ export class GeocodingService {
|
||||
);
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error },
|
||||
'An error occurred while calling the Google Maps Geocoding API. Falling back to Nominatim.',
|
||||
{ err: error },
|
||||
'An error occurred while calling the Google Maps Geocoding API. Falling back to Nominatim.'
|
||||
);
|
||||
}
|
||||
} else {
|
||||
@@ -72,10 +69,7 @@ export class GeocodingService {
|
||||
try {
|
||||
await redis.set(cacheKey, JSON.stringify(result), 'EX', 60 * 60 * 24 * 30); // Cache for 30 days
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error, cacheKey },
|
||||
'Redis SET command failed. Result will not be cached.',
|
||||
);
|
||||
logger.error({ err: error, cacheKey }, 'Redis SET command failed. Result will not be cached.');
|
||||
}
|
||||
}
|
||||
|
||||
@@ -98,10 +92,7 @@ export class GeocodingService {
|
||||
logger.info(`Successfully deleted ${totalDeleted} geocode cache entries.`);
|
||||
return totalDeleted;
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error },
|
||||
'Failed to clear geocode cache from Redis.',
|
||||
);
|
||||
logger.error({ err: error }, 'Failed to clear geocode cache from Redis.');
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
// src/services/queueService.server.ts
|
||||
import { Queue, Worker, Job } from 'bullmq';
|
||||
import { Queue, Worker, Job, UnrecoverableError } from 'bullmq';
|
||||
import IORedis from 'ioredis'; // Correctly imported
|
||||
import fsPromises from 'node:fs/promises';
|
||||
import { exec } from 'child_process';
|
||||
@@ -185,9 +185,26 @@ const attachWorkerEventListeners = (worker: Worker) => {
|
||||
|
||||
export const flyerWorker = new Worker<FlyerJobData>(
|
||||
'flyer-processing', // Must match the queue name
|
||||
(job) => {
|
||||
// The processJob method creates its own job-specific logger internally.
|
||||
return flyerProcessingService.processJob(job);
|
||||
async (job) => {
|
||||
try {
|
||||
// The processJob method creates its own job-specific logger internally.
|
||||
return await flyerProcessingService.processJob(job);
|
||||
} catch (error: any) {
|
||||
// Check for quota errors or other unrecoverable errors from the AI service
|
||||
const errorMessage = error?.message || '';
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') ||
|
||||
errorMessage.includes('RESOURCE_EXHAUSTED')
|
||||
) {
|
||||
logger.error(
|
||||
{ err: error, jobId: job.id },
|
||||
'[FlyerWorker] Unrecoverable quota error detected. Failing job immediately.',
|
||||
);
|
||||
throw new UnrecoverableError(errorMessage);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
@@ -207,13 +224,9 @@ export const emailWorker = new Worker<EmailJobData>(
|
||||
try {
|
||||
await emailService.sendEmail(job.data, jobLogger);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging to capture the full error object, including the stack trace.
|
||||
// This provides more context for debugging than just logging the message.
|
||||
logger.error(
|
||||
{
|
||||
// Log the full error object for better diagnostics. // The patch requested this specific error handling.
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
// Also include the job data for context.
|
||||
err: error,
|
||||
jobData: job.data,
|
||||
},
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
@@ -249,11 +262,7 @@ export const analyticsWorker = new Worker<AnalyticsJobData>(
|
||||
logger.info(`[AnalyticsWorker] Successfully generated report for ${reportDate}.`);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
jobData: job.data,
|
||||
},
|
||||
logger.error({ err: error, jobData: job.data },
|
||||
`[AnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
@@ -315,9 +324,7 @@ export const cleanupWorker = new Worker<CleanupJobData>(
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
},
|
||||
{ err: error },
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
@@ -350,10 +357,7 @@ export const weeklyAnalyticsWorker = new Worker<WeeklyAnalyticsJobData>(
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
jobData: job.data,
|
||||
},
|
||||
{ err: error, jobData: job.data },
|
||||
`[WeeklyAnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
|
||||
@@ -56,15 +56,15 @@ describe('Price History API Integration Test (/api/price-history)', () => {
|
||||
|
||||
// 4. Create flyer items linking the master item to the flyers with prices
|
||||
await pool.query(
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display) VALUES ($1, $2, 'Apples', 199, '$1.99')`,
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display, quantity) VALUES ($1, $2, 'Apples', 199, '$1.99', '1')`,
|
||||
[flyerId1, masterItemId],
|
||||
);
|
||||
await pool.query(
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display) VALUES ($1, $2, 'Apples', 249, '$2.49')`,
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display, quantity) VALUES ($1, $2, 'Apples', 249, '$2.49', '1')`,
|
||||
[flyerId2, masterItemId],
|
||||
);
|
||||
await pool.query(
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display) VALUES ($1, $2, 'Apples', 299, '$2.99')`,
|
||||
`INSERT INTO public.flyer_items (flyer_id, master_item_id, item, price_in_cents, price_display, quantity) VALUES ($1, $2, 'Apples', 299, '$2.99', '1')`,
|
||||
[flyerId3, masterItemId],
|
||||
);
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user