All checks were successful
Build and Deploy Log Center / build-and-deploy (push) Successful in 1m30s
202 lines
7.3 KiB
Python
202 lines
7.3 KiB
Python
from fastapi import FastAPI, Depends, HTTPException, Query
|
|
from fastapi.middleware.cors import CORSMiddleware
|
|
from sqlmodel.ext.asyncio.session import AsyncSession
|
|
from sqlmodel import select, func
|
|
from .database import init_db, get_session
|
|
from .models import ErrorLog, ErrorLogCreate, LogStatus
|
|
from datetime import datetime, timedelta
|
|
from typing import Optional, List
|
|
import hashlib
|
|
import json
|
|
|
|
app = FastAPI(title="Log Center & AIOps Control Plane")
|
|
|
|
# CORS for frontend
|
|
app.add_middleware(
|
|
CORSMiddleware,
|
|
allow_origins=["*"], # In production, restrict to your domain
|
|
allow_credentials=True,
|
|
allow_methods=["*"],
|
|
allow_headers=["*"],
|
|
)
|
|
|
|
@app.on_event("startup")
|
|
async def on_startup():
|
|
await init_db()
|
|
|
|
def generate_fingerprint(log: ErrorLogCreate) -> str:
|
|
# Minimal fingerprinting: project + error_type + file + line
|
|
raw = f"{log.project_id}|{log.error.get('type')}|{log.error.get('file_path')}|{log.error.get('line_number')}"
|
|
return hashlib.md5(raw.encode()).hexdigest()
|
|
|
|
# ==================== Log Reporting ====================
|
|
@app.post("/api/v1/logs/report", tags=["Logs"])
|
|
async def report_log(log_data: ErrorLogCreate, session: AsyncSession = Depends(get_session)):
|
|
fingerprint = generate_fingerprint(log_data)
|
|
|
|
# Check deduplication
|
|
statement = select(ErrorLog).where(ErrorLog.fingerprint == fingerprint)
|
|
results = await session.exec(statement)
|
|
existing_log = results.first()
|
|
|
|
if existing_log:
|
|
# If exists and not resolved, just ignore or update count (implied)
|
|
if existing_log.status not in [LogStatus.DEPLOYED, LogStatus.FIXED, LogStatus.VERIFIED]:
|
|
return {"message": "Log deduplicated", "id": existing_log.id, "status": existing_log.status}
|
|
# If it was resolved but happened again -> Regression! Reset to NEW?
|
|
existing_log.status = LogStatus.NEW
|
|
existing_log.timestamp = log_data.timestamp or datetime.utcnow()
|
|
existing_log.retry_count = 0 # Reset retries for new occurrence
|
|
session.add(existing_log)
|
|
await session.commit()
|
|
await session.refresh(existing_log)
|
|
return {"message": "Regression detected, reopened", "id": existing_log.id}
|
|
|
|
# Create new
|
|
new_log = ErrorLog(
|
|
project_id=log_data.project_id,
|
|
environment=log_data.environment,
|
|
level=log_data.level,
|
|
error_type=log_data.error.get("type"),
|
|
error_message=log_data.error.get("message"),
|
|
file_path=log_data.error.get("file_path"),
|
|
line_number=log_data.error.get("line_number"),
|
|
stack_trace=log_data.error.get("stack_trace"),
|
|
context=log_data.context,
|
|
version=log_data.version,
|
|
commit_hash=log_data.commit_hash,
|
|
fingerprint=fingerprint,
|
|
timestamp=log_data.timestamp or datetime.utcnow()
|
|
)
|
|
|
|
session.add(new_log)
|
|
await session.commit()
|
|
await session.refresh(new_log)
|
|
|
|
return {"message": "Log reported", "id": new_log.id}
|
|
|
|
# ==================== Agent Tasks ====================
|
|
@app.get("/api/v1/tasks/pending", tags=["Tasks"])
|
|
async def get_pending_tasks(project_id: str = None, session: AsyncSession = Depends(get_session)):
|
|
query = select(ErrorLog).where(ErrorLog.status == LogStatus.NEW)
|
|
if project_id:
|
|
query = query.where(ErrorLog.project_id == project_id)
|
|
|
|
results = await session.exec(query)
|
|
return results.all()
|
|
|
|
@app.patch("/api/v1/tasks/{task_id}/status", tags=["Tasks"])
|
|
async def update_task_status(task_id: int, status: LogStatus, session: AsyncSession = Depends(get_session)):
|
|
statement = select(ErrorLog).where(ErrorLog.id == task_id)
|
|
results = await session.exec(statement)
|
|
task = results.first()
|
|
|
|
if not task:
|
|
raise HTTPException(status_code=404, detail="Task not found")
|
|
|
|
task.status = status
|
|
session.add(task)
|
|
await session.commit()
|
|
await session.refresh(task)
|
|
|
|
return {"message": "Status updated", "id": task.id, "status": task.status}
|
|
|
|
# ==================== Dashboard APIs ====================
|
|
@app.get("/api/v1/dashboard/stats", tags=["Dashboard"])
|
|
async def get_dashboard_stats(session: AsyncSession = Depends(get_session)):
|
|
"""Get overall statistics for dashboard"""
|
|
today = datetime.utcnow().replace(hour=0, minute=0, second=0, microsecond=0)
|
|
|
|
# Total bugs
|
|
total_query = select(func.count(ErrorLog.id))
|
|
total_result = await session.exec(total_query)
|
|
total_bugs = total_result.one()
|
|
|
|
# Today's new bugs
|
|
today_query = select(func.count(ErrorLog.id)).where(ErrorLog.timestamp >= today)
|
|
today_result = await session.exec(today_query)
|
|
today_bugs = today_result.one()
|
|
|
|
# Count by status
|
|
status_counts = {}
|
|
for status in LogStatus:
|
|
count_query = select(func.count(ErrorLog.id)).where(ErrorLog.status == status)
|
|
count_result = await session.exec(count_query)
|
|
status_counts[status.value] = count_result.one()
|
|
|
|
# Fixed rate = (FIXED + VERIFIED + DEPLOYED) / Total
|
|
fixed_count = status_counts.get("FIXED", 0) + status_counts.get("VERIFIED", 0) + status_counts.get("DEPLOYED", 0)
|
|
fix_rate = round((fixed_count / total_bugs * 100), 2) if total_bugs > 0 else 0
|
|
|
|
return {
|
|
"total_bugs": total_bugs,
|
|
"today_bugs": today_bugs,
|
|
"fix_rate": fix_rate,
|
|
"status_distribution": status_counts
|
|
}
|
|
|
|
@app.get("/api/v1/bugs", tags=["Dashboard"])
|
|
async def get_bugs_list(
|
|
page: int = Query(1, ge=1),
|
|
page_size: int = Query(20, ge=1, le=100),
|
|
status: Optional[LogStatus] = None,
|
|
project_id: Optional[str] = None,
|
|
session: AsyncSession = Depends(get_session)
|
|
):
|
|
"""Get paginated list of bugs with optional filters"""
|
|
query = select(ErrorLog).order_by(ErrorLog.timestamp.desc())
|
|
|
|
if status:
|
|
query = query.where(ErrorLog.status == status)
|
|
if project_id:
|
|
query = query.where(ErrorLog.project_id == project_id)
|
|
|
|
# Pagination
|
|
offset = (page - 1) * page_size
|
|
query = query.offset(offset).limit(page_size)
|
|
|
|
results = await session.exec(query)
|
|
bugs = results.all()
|
|
|
|
# Get total count for pagination info
|
|
count_query = select(func.count(ErrorLog.id))
|
|
if status:
|
|
count_query = count_query.where(ErrorLog.status == status)
|
|
if project_id:
|
|
count_query = count_query.where(ErrorLog.project_id == project_id)
|
|
count_result = await session.exec(count_query)
|
|
total = count_result.one()
|
|
|
|
return {
|
|
"items": bugs,
|
|
"total": total,
|
|
"page": page,
|
|
"page_size": page_size,
|
|
"total_pages": (total + page_size - 1) // page_size
|
|
}
|
|
|
|
@app.get("/api/v1/bugs/{bug_id}", tags=["Dashboard"])
|
|
async def get_bug_detail(bug_id: int, session: AsyncSession = Depends(get_session)):
|
|
"""Get detailed information about a specific bug"""
|
|
statement = select(ErrorLog).where(ErrorLog.id == bug_id)
|
|
results = await session.exec(statement)
|
|
bug = results.first()
|
|
|
|
if not bug:
|
|
raise HTTPException(status_code=404, detail="Bug not found")
|
|
|
|
return bug
|
|
|
|
@app.get("/api/v1/projects", tags=["Dashboard"])
|
|
async def get_projects(session: AsyncSession = Depends(get_session)):
|
|
"""Get list of all unique project IDs"""
|
|
query = select(ErrorLog.project_id).distinct()
|
|
results = await session.exec(query)
|
|
projects = results.all()
|
|
return {"projects": projects}
|
|
|
|
@app.get("/", tags=["Health"])
|
|
async def health_check():
|
|
return {"status": "ok"}
|
|
|