在前后端开发联调前 的提交20260223

This commit is contained in:
Neo
2026-02-23 23:02:20 +08:00
parent 254ccb1e77
commit fafc95e64c
1142 changed files with 10366960 additions and 36957 deletions

View File

@@ -0,0 +1,114 @@
- [P20260221-120503] 2026-02-21 12:05:03 +0800
- summary: CONTEXT TRANSFER: We are continuing a conversation that had gotten too long. Here is a summary: --- ## TASK 1: Modify "D…
- prompt:
```text
CONTEXT TRANSFER: We are continuing a conversation that had gotten too long. Here is a summary:
---
## TASK 1: Modify "Data Flow Structure Analysis" Hook — Whitelist Refactoring
**STATUS**: done
**USER QUERIES**: 1 (original request), continued across context transfers
**DETAILS**:
- Added global whitelist constants and helper functions at top of `scripts/ops/gen_dataflow_report.py`: `WHITELIST_ETL_META_COLS`, `WHITELIST_DWD_SCD2_COLS`, `WHITELIST_API_NESTED_PREFIXES`, plus `is_whitelist_etl_meta()`, `is_whitelist_scd2()`, `is_whitelist_api_nested()`, `whitelist_reason()`
- Refactored `_write_field_diff_report()` (Section 1.1) to separate whitelist items from non-whitelist items in all 4 categories, with `_write_whitelist_summary()` for collapsed display
- Refactored Section 3 per-table detail functions:
- `_write_api_section()`: Filters out whitelist API nested fields from table, shows folded summary with reason
- `_write_ods_section()`: Filters out whitelist ETL meta columns from table, shows folded summary
- `_write_dwd_section()`: Filters out whitelist SCD2 columns from table, shows folded summary
- Removed all uses of "跳过" (skip) — only "白名单" (whitelist) used
- Updated hook file to v4.0.0 with new prompt describing whitelist rules
- Both files pass `py_compile` and `getDiagnostics` checks
- Report has NOT been regenerated end-to-end yet (only the consistency check script was run)
**FILEPATHS**:
- `scripts/ops/gen_dataflow_report.py` (modified)
- `.kiro/hooks/dataflow-analyze.kiro.hook` (modified, v4.0.0)
---
## TASK 2: Create New Hook "ETL Data Consistency Check"
**STATUS**: in-progress
**USER QUERIES**: 1 (original request), continued across context transfers
**DETAILS**:
- Created `.kiro/hooks/etl-data-consistency.kiro.hook` (v1.0.0, userTriggered)
- Created `scripts/ops/etl_consistency_check.py` with full implementation
- Fixed bugs found during testing:
- Fixed `COUNT(DISTINCT id)` SQL alias bug — added `AS cnt` to both occurrences in `check_api_vs_ods()` and `check_ods_vs_dwd()`
- Fixed `load_api_json_records()` — rewrote to handle ETL framework's nested JSON format: `{pages: [{response: {data: {<list_field>: [...]}}}]}` where the data list field name varies by endpoint (e.g., `tenantMemberInfos` not `list`)
- Added `get_field_stats()` function — batch per-field statistics (NULL rate, min/max/avg for numerics, date ranges, text lengths, bool distribution, distinct counts for small tables) similar to `field_level_report.py`
- Added `_fmt_field_stat()` and `_write_field_stats_table()` helpers in `generate_report()` for collapsible field stats tables
- Integrated field stats into report: ODS stats in Section 2.2, DWD stats in Section 3.2
- Successfully ran end-to-end: `uv run python scripts/ops/etl_consistency_check.py` — produced `consistency_check_20260221_120249.md`
- Report shows real data: 22 ODS tasks checked, API JSON parsed correctly (19/22 with data), field-level stats included
- 3 tasks show "⚠️ 无 API JSON" (ODS_ASSISTANT_ABOLISH, ODS_RECHARGE_SETTLE, ODS_STORE_GOODS_SALES) — these may have empty JSON directories or different structures
**What still needs attention**:
- The report is functionally complete and tested end-to-end
- ODS_ASSISTANT_ABOLISH has 0 records in both API and ODS — this is expected (no cancellation data)
- ODS_RECHARGE_SETTLE and ODS_STORE_GOODS_SALES show "无 API JSON" — may need investigation of their JSON structure if they should have data
- `settlement_records` shows 0/2 field match with 2 API-only and 66 ODS-only — the API response structure for this endpoint may need special handling
- Temporary diagnostic scripts were created and should be cleaned up: `scripts/ops/_check_json_dirs.py`, `scripts/ops/_check_json_content.py`, `scripts/ops/_check_json_pages.py`, `scripts/ops/_check_json_response.py`
**NEXT STEPS**:
- Clean up temporary diagnostic scripts (`_check_json_*.py`)
- Optionally investigate the 3 "无 API JSON" cases and `settlement_records` field mismatch
- Run `gen_dataflow_report.py` to verify Task 1's whitelist changes produce correct output
- Consider running `/audit` since `scripts/ops/` files were modified (governance rule)
**FILEPATHS**:
- `scripts/ops/etl_consistency_check.py` (new, tested)
- `.kiro/hooks/etl-data-consistency.kiro.hook` (new)
- `export/ETL-Connectors/feiqiu/REPORTS/consistency_check_20260221_120249.md` (generated report)
- `scripts/ops/_check_json_dirs.py` (temp, delete)
- `scripts/ops/_check_json_content.py` (temp, delete)
- `scripts/ops/_check_json_pages.py` (temp, delete)
- `scripts/ops/_check_json_response.py` (temp, delete)
---
## USER CORRECTIONS AND INSTRUCTIONS:
- Language: All explanatory text must be in simplified Chinese (简体中文)
- Never use "跳过" (skip) — only use "白名单" (whitelist)
- Whitelist fields must still be checked/counted — they are only folded in report display
- Output paths must come from `.env` environment variables — never hardcode
- Scripts in `scripts/ops/` use `_env_paths.py` for path resolution or direct `os.environ.get()` with error on missing
- Database connections use `PG_DSN` from `.env` (readonly mode for reports)
- The project uses `uv run python` to execute scripts
- `scripts/ops/` scripts must `load_dotenv` from root `.env` and raise errors on missing env vars
- JSON files use ETL framework wrapper format: `{task_code, run_id, pages: [{page, request, response: {data: {<varying_field>: [...]}}}]}`
- Existing mapping tables (`ODS_TASK_TO_TABLE`, `DWD_TO_ODS`) are defined in `scripts/ops/blackbox_test_report.py` and were replicated in the new script
- Background context document: `export/ETL-Connectors/feiqiu/REPORTS/context_handoff_task2.md`
## Files to read
- `scripts/ops/etl_consistency_check.py` (main deliverable for Task 2, needs cleanup verification)
- `scripts/ops/gen_dataflow_report.py` (Task 1 deliverable, needs end-to-end test)
- `.kiro/hooks/dataflow-analyze.kiro.hook` (Task 1 hook)
- `.kiro/hooks/etl-data-consistency.kiro.hook` (Task 2 hook)
- `export/ETL-Connectors/feiqiu/REPORTS/consistency_check_20260221_120249.md` (latest generated report to verify quality)
- `export/ETL-Connectors/feiqiu/REPORTS/context_handoff_task2.md` (background context)
- `scripts/ops/_env_paths.py` (shared path utility)
USER QUERIES(most recent first):
1. Output:
Command executed successfully with no output.
Exit Code: 0
2. Output:
Command executed successfully with no output.
Exit Code: 0
---
METADATA:
The previous conversation had 2 messages.
INSTRUCTIONS:
Continue working until the user query has been fully addressed. Do not ask for clarification - proceed with the work based on the context provided.
IMPORTANT: you need to read from the files to Read section
```