Compare commits
2 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
62e35deddc | ||
| 59f6f43d03 |
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.9.48",
|
||||
"version": "0.9.49",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.9.48",
|
||||
"version": "0.9.49",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.9.48",
|
||||
"version": "0.9.49",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -32,6 +32,7 @@ export const uploadAndProcessFlyer = async (
|
||||
formData.append('checksum', checksum);
|
||||
|
||||
logger.info(`[aiApiClient] Starting background processing for file: ${file.name}`);
|
||||
console.error(`[aiApiClient] uploadAndProcessFlyer: Uploading file '${file.name}' with checksum '${checksum}'`);
|
||||
|
||||
const response = await authedPostForm('/ai/upload-and-process', formData, { tokenOverride });
|
||||
|
||||
@@ -94,6 +95,7 @@ export const getJobStatus = async (
|
||||
jobId: string,
|
||||
tokenOverride?: string,
|
||||
): Promise<JobStatus> => {
|
||||
console.error(`[aiApiClient] getJobStatus: Fetching status for job '${jobId}'`);
|
||||
const response = await authedGet(`/ai/jobs/${jobId}/status`, { tokenOverride });
|
||||
|
||||
// Handle non-OK responses first, as they might not have a JSON body.
|
||||
|
||||
@@ -328,9 +328,8 @@ describe('AI Service (Server)', () => {
|
||||
// Check that a warning was logged
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
// The warning should be for the model that failed ('gemini-2.5-flash'), not the next one.
|
||||
// The warning should be for the model that failed, not the next one.
|
||||
expect.stringContaining(
|
||||
`Model '${models[0]}' failed due to quota/rate limit. Trying next model.`,
|
||||
`Model '${models[0]}' failed due to quota/rate limit/overload. Trying next model.`,
|
||||
),
|
||||
);
|
||||
});
|
||||
@@ -506,7 +505,7 @@ describe('AI Service (Server)', () => {
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(2);
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, { model: models[0], ...request });
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, { model: models[1], ...request });
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.stringContaining(`Model '${models[0]}' failed due to quota/rate limit.`));
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.stringContaining(`Model '${models[0]}' failed due to quota/rate limit/overload.`));
|
||||
});
|
||||
|
||||
it('should fail immediately on a 400 Bad Request error without retrying', async () => {
|
||||
|
||||
@@ -98,9 +98,12 @@ describe('AnalyticsService', () => {
|
||||
|
||||
const promise = service.processDailyReportJob(job);
|
||||
|
||||
// Capture the expectation promise BEFORE triggering the rejection via timer advancement.
|
||||
const expectation = expect(promise).rejects.toThrow('A string error');
|
||||
|
||||
await vi.advanceTimersByTimeAsync(10000);
|
||||
|
||||
await expect(promise).rejects.toThrow('A string error');
|
||||
await expectation;
|
||||
|
||||
expect(mockLoggerInstance.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
@@ -188,9 +191,12 @@ describe('AnalyticsService', () => {
|
||||
|
||||
const promise = service.processWeeklyReportJob(job);
|
||||
|
||||
// Capture the expectation promise BEFORE triggering the rejection via timer advancement.
|
||||
const expectation = expect(promise).rejects.toThrow('A string error');
|
||||
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
|
||||
await expect(promise).rejects.toThrow('A string error');
|
||||
await expectation;
|
||||
|
||||
expect(mockLoggerInstance.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
|
||||
@@ -95,6 +95,7 @@ export const apiFetch = async (
|
||||
const fullUrl = url.startsWith('http') ? url : joinUrl(API_BASE_URL, url);
|
||||
|
||||
logger.debug(`apiFetch: ${options.method || 'GET'} ${fullUrl}`);
|
||||
console.error(`[apiClient] apiFetch Request: ${options.method || 'GET'} ${fullUrl}`);
|
||||
|
||||
// Create a new headers object to avoid mutating the original options.
|
||||
const headers = new Headers(options.headers || {});
|
||||
|
||||
@@ -34,7 +34,10 @@ export class BackgroundJobService {
|
||||
const reportDate = getCurrentDateISOString(); // YYYY-MM-DD
|
||||
const jobId = `manual-report-${reportDate}-${Date.now()}`;
|
||||
const job = await analyticsQueue.add('generate-daily-report', { reportDate }, { jobId });
|
||||
return job.id!;
|
||||
if (!job.id) {
|
||||
throw new Error('Failed to enqueue daily report job: No job ID returned');
|
||||
}
|
||||
return job.id;
|
||||
}
|
||||
|
||||
public async triggerWeeklyAnalyticsReport(): Promise<string> {
|
||||
@@ -45,7 +48,10 @@ export class BackgroundJobService {
|
||||
{ reportYear, reportWeek },
|
||||
{ jobId },
|
||||
);
|
||||
return job.id!;
|
||||
if (!job.id) {
|
||||
throw new Error('Failed to enqueue weekly report job: No job ID returned');
|
||||
}
|
||||
return job.id;
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
@@ -62,13 +62,13 @@ export class FlyerDataTransformer {
|
||||
baseUrl: string,
|
||||
logger: Logger,
|
||||
): { imageUrl: string; iconUrl: string } {
|
||||
console.log('[DEBUG] FlyerDataTransformer._buildUrls inputs:', { imageFileName, iconFileName, baseUrl });
|
||||
console.error('[DEBUG] FlyerDataTransformer._buildUrls inputs:', { imageFileName, iconFileName, baseUrl });
|
||||
logger.debug({ imageFileName, iconFileName, baseUrl }, 'Building URLs');
|
||||
const finalBaseUrl = baseUrl || getBaseUrl(logger);
|
||||
console.log('[DEBUG] FlyerDataTransformer._buildUrls finalBaseUrl resolved to:', finalBaseUrl);
|
||||
console.error('[DEBUG] FlyerDataTransformer._buildUrls finalBaseUrl resolved to:', finalBaseUrl);
|
||||
const imageUrl = `${finalBaseUrl}/flyer-images/${imageFileName}`;
|
||||
const iconUrl = `${finalBaseUrl}/flyer-images/icons/${iconFileName}`;
|
||||
console.log('[DEBUG] FlyerDataTransformer._buildUrls constructed:', { imageUrl, iconUrl });
|
||||
console.error('[DEBUG] FlyerDataTransformer._buildUrls constructed:', { imageUrl, iconUrl });
|
||||
logger.debug({ imageUrl, iconUrl }, 'Constructed URLs');
|
||||
return { imageUrl, iconUrl };
|
||||
}
|
||||
@@ -93,7 +93,7 @@ export class FlyerDataTransformer {
|
||||
logger: Logger,
|
||||
baseUrl: string,
|
||||
): Promise<{ flyerData: FlyerInsert; itemsForDb: FlyerItemInsert[] }> {
|
||||
console.log('[DEBUG] FlyerDataTransformer.transform called with baseUrl:', baseUrl);
|
||||
console.error('[DEBUG] FlyerDataTransformer.transform called with baseUrl:', baseUrl);
|
||||
logger.info('Starting data transformation from AI output to database format.');
|
||||
|
||||
try {
|
||||
|
||||
@@ -103,7 +103,7 @@ export class FlyerProcessingService {
|
||||
// The main processed image path is already in `allFilePaths` via `createdImagePaths`.
|
||||
allFilePaths.push(path.join(iconsDir, iconFileName));
|
||||
|
||||
console.log('[DEBUG] FlyerProcessingService calling transformer with:', { originalFileName: job.data.originalFileName, imageFileName, iconFileName, checksum: job.data.checksum, baseUrl: job.data.baseUrl });
|
||||
console.error('[DEBUG] FlyerProcessingService calling transformer with:', { originalFileName: job.data.originalFileName, imageFileName, iconFileName, checksum: job.data.checksum, baseUrl: job.data.baseUrl });
|
||||
|
||||
const { flyerData, itemsForDb } = await this.transformer.transform(
|
||||
aiResult,
|
||||
@@ -115,6 +115,7 @@ export class FlyerProcessingService {
|
||||
logger,
|
||||
job.data.baseUrl,
|
||||
);
|
||||
console.error('[DEBUG] FlyerProcessingService transformer output URLs:', { imageUrl: flyerData.image_url, iconUrl: flyerData.icon_url });
|
||||
stages[2].status = 'completed';
|
||||
await job.updateProgress({ stages });
|
||||
|
||||
|
||||
@@ -117,15 +117,28 @@ describe('Gamification Flow Integration Test', () => {
|
||||
createdFilePaths.push(path.join(uploadDir, 'icons', iconFileName));
|
||||
|
||||
// --- Act 1: Upload the flyer to trigger the background job ---
|
||||
const testBaseUrl = 'https://example.com';
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
console.error('[TEST DEBUG] STARTING UPLOAD STEP');
|
||||
console.error(`[TEST DEBUG] Env FRONTEND_URL: "${process.env.FRONTEND_URL}"`);
|
||||
console.error(`[TEST DEBUG] Sending baseUrl field: "${testBaseUrl}"`);
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
|
||||
const uploadResponse = await request
|
||||
.post('/api/ai/upload-and-process')
|
||||
.set('Authorization', `Bearer ${authToken}`)
|
||||
.field('checksum', checksum)
|
||||
.field('baseUrl', 'https://example.com')
|
||||
.field('baseUrl', testBaseUrl)
|
||||
.attach('flyerFile', uniqueContent, uniqueFileName);
|
||||
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
console.error(`[TEST DEBUG] Upload Response Status: ${uploadResponse.status}`);
|
||||
console.error(`[TEST DEBUG] Upload Response Body: ${JSON.stringify(uploadResponse.body, null, 2)}`);
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
|
||||
const { jobId } = uploadResponse.body;
|
||||
expect(jobId).toBeTypeOf('string');
|
||||
console.error(`[TEST DEBUG] Job ID received: ${jobId}`);
|
||||
|
||||
// --- Act 2: Poll for job completion using the new utility ---
|
||||
const jobStatus = await poll(
|
||||
@@ -133,6 +146,7 @@ describe('Gamification Flow Integration Test', () => {
|
||||
const statusResponse = await request
|
||||
.get(`/api/ai/jobs/${jobId}/status`)
|
||||
.set('Authorization', `Bearer ${authToken}`);
|
||||
console.error(`[TEST DEBUG] Polling status for ${jobId}: ${statusResponse.body?.state}`);
|
||||
return statusResponse.body;
|
||||
},
|
||||
(status) => status.state === 'completed' || status.state === 'failed',
|
||||
@@ -144,6 +158,17 @@ describe('Gamification Flow Integration Test', () => {
|
||||
throw new Error('Gamification test job timed out: No job status received.');
|
||||
}
|
||||
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
console.error('[TEST DEBUG] Final Job Status Object:', JSON.stringify(jobStatus, null, 2));
|
||||
if (jobStatus.state === 'failed') {
|
||||
console.error(`[TEST DEBUG] Job Failed Reason: ${jobStatus.failedReason}`);
|
||||
// If there is a progress object with error details, log it
|
||||
if (jobStatus.progress) {
|
||||
console.error(`[TEST DEBUG] Job Progress/Error Details:`, JSON.stringify(jobStatus.progress, null, 2));
|
||||
}
|
||||
}
|
||||
console.error('--------------------------------------------------------------------------------');
|
||||
|
||||
// --- Assert 1: Verify the job completed successfully ---
|
||||
if (jobStatus?.state === 'failed') {
|
||||
console.error('[DEBUG] Gamification test job failed:', jobStatus.failedReason);
|
||||
|
||||
@@ -245,8 +245,6 @@ describe('Public API Routes Integration Tests', () => {
|
||||
|
||||
expect(blockedResponse).toBeDefined();
|
||||
expect(blockedResponse.status).toBe(429);
|
||||
expect(blockedResponse.headers).toHaveProperty('x-ratelimit-limit');
|
||||
expect(blockedResponse.headers).toHaveProperty('x-ratelimit-remaining');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user