Backend Go: - Remplacement complet des anciennes migrations par la base V1 alignée sur ORIGIN. - Durcissement global du parsing JSON (BindAndValidateJSON + RespondWithAppError). - Sécurisation de config.go, CORS, statuts de santé et monitoring. - Implémentation des transactions P0 (RBAC, duplication de playlists, social toggles). - Ajout d’un job worker structuré (emails, analytics, thumbnails) + tests associés. - Nouvelle doc backend : AUDIT_CONFIG, BACKEND_CONFIG, AUTH_PASSWORD_RESET, JOB_WORKER_*. Chat server (Rust): - Refonte du pipeline JWT + sécurité, audit et rate limiting avancé. - Implémentation complète du cycle de message (read receipts, delivered, edit/delete, typing). - Nettoyage des panics, gestion d’erreurs robuste, logs structurés. - Migrations chat alignées sur le schéma UUID et nouvelles features. Stream server (Rust): - Refonte du moteur de streaming (encoding pipeline + HLS) et des modules core. - Transactions P0 pour les jobs et segments, garanties d’atomicité. - Documentation détaillée de la pipeline (AUDIT_STREAM_*, DESIGN_STREAM_PIPELINE, TRANSACTIONS_P0_IMPLEMENTATION). Documentation & audits: - TRIAGE.md et AUDIT_STABILITY.md à jour avec l’état réel des 3 services. - Cartographie complète des migrations et des transactions (DB_MIGRATIONS_*, DB_TRANSACTION_PLAN, AUDIT_DB_TRANSACTIONS, TRANSACTION_TESTS_PHASE3). - Scripts de reset et de cleanup pour la lab DB et la V1. Ce commit fige l’ensemble du travail de stabilisation P0 (UUID, backend, chat et stream) avant les phases suivantes (Coherence Guardian, WS hardening, etc.).
349 lines
7.8 KiB
Go
349 lines
7.8 KiB
Go
package handlers
|
|
|
|
import (
|
|
"context"
|
|
"net/http"
|
|
"runtime"
|
|
"time"
|
|
|
|
"github.com/gin-gonic/gin"
|
|
"github.com/redis/go-redis/v9"
|
|
"go.uber.org/zap"
|
|
"gorm.io/gorm"
|
|
|
|
"veza-backend-api/internal/database"
|
|
"veza-backend-api/internal/monitoring"
|
|
)
|
|
|
|
var (
|
|
// startTime tracks when the server started
|
|
startTime = time.Now()
|
|
)
|
|
|
|
// StatusResponse représente la réponse complète du status endpoint
|
|
type StatusResponse struct {
|
|
Status string `json:"status"`
|
|
UptimeSec int64 `json:"uptime_seconds"`
|
|
Services map[string]ServiceInfo `json:"services"`
|
|
Version string `json:"version"`
|
|
GitCommit string `json:"git_commit"`
|
|
BuildTime string `json:"build_time"`
|
|
Environment string `json:"environment,omitempty"`
|
|
}
|
|
|
|
// ServiceInfo représente l'état d'un service
|
|
type ServiceInfo struct {
|
|
Status string `json:"status"`
|
|
Latency float64 `json:"latency_ms,omitempty"`
|
|
Message string `json:"message,omitempty"`
|
|
}
|
|
|
|
// StatusHandler gère les endpoints de status
|
|
type StatusHandler struct {
|
|
db *gorm.DB
|
|
logger *zap.Logger
|
|
redis *redis.Client
|
|
chatServerURL string
|
|
streamServerURL string
|
|
version string
|
|
gitCommit string
|
|
buildTime string
|
|
environment string
|
|
}
|
|
|
|
// NewStatusHandler crée un nouveau handler de status
|
|
func NewStatusHandler(
|
|
db *gorm.DB,
|
|
logger *zap.Logger,
|
|
redisClient interface{},
|
|
chatServerURL string,
|
|
streamServerURL string,
|
|
version string,
|
|
gitCommit string,
|
|
buildTime string,
|
|
environment string,
|
|
) *StatusHandler {
|
|
h := &StatusHandler{
|
|
db: db,
|
|
logger: logger,
|
|
chatServerURL: chatServerURL,
|
|
streamServerURL: streamServerURL,
|
|
version: version,
|
|
gitCommit: gitCommit,
|
|
buildTime: buildTime,
|
|
environment: environment,
|
|
}
|
|
|
|
// Type assertion for Redis
|
|
if r, ok := redisClient.(*redis.Client); ok {
|
|
h.redis = r
|
|
}
|
|
|
|
return h
|
|
}
|
|
|
|
// GetStatus retourne le status complet de l'application
|
|
func (h *StatusHandler) GetStatus(c *gin.Context) {
|
|
response := StatusResponse{
|
|
Status: "ok",
|
|
UptimeSec: int64(time.Since(startTime).Seconds()),
|
|
Services: make(map[string]ServiceInfo),
|
|
Version: h.version,
|
|
GitCommit: h.gitCommit,
|
|
BuildTime: h.buildTime,
|
|
}
|
|
|
|
if h.environment != "" {
|
|
response.Environment = h.environment
|
|
}
|
|
|
|
// Check database
|
|
dbInfo := h.checkDatabase()
|
|
response.Services["database"] = dbInfo
|
|
|
|
// Check Redis
|
|
redisInfo := h.checkRedis()
|
|
response.Services["redis"] = redisInfo
|
|
|
|
// Check chat server (if configured)
|
|
if h.chatServerURL != "" {
|
|
chatInfo := h.checkChatServer(c.Request.Context())
|
|
response.Services["chat_server"] = chatInfo
|
|
}
|
|
|
|
// Check stream server (if configured)
|
|
if h.streamServerURL != "" {
|
|
streamInfo := h.checkStreamServer(c.Request.Context())
|
|
response.Services["stream_server"] = streamInfo
|
|
}
|
|
|
|
// Déterminer le statut global
|
|
globalStatus := "ok"
|
|
for _, service := range response.Services {
|
|
if service.Status == "error" {
|
|
globalStatus = "degraded"
|
|
break
|
|
}
|
|
if service.Status == "slow" {
|
|
if globalStatus != "degraded" {
|
|
globalStatus = "degraded"
|
|
}
|
|
}
|
|
}
|
|
response.Status = globalStatus
|
|
|
|
statusCode := http.StatusOK
|
|
if globalStatus == "degraded" {
|
|
statusCode = http.StatusServiceUnavailable
|
|
}
|
|
|
|
c.JSON(statusCode, response)
|
|
}
|
|
|
|
// checkDatabase vérifie la connexion à la base de données
|
|
func (h *StatusHandler) checkDatabase() ServiceInfo {
|
|
start := time.Now()
|
|
|
|
err := database.IsConnectionHealthy(h.db, 5*time.Second)
|
|
duration := time.Since(start)
|
|
latencyMs := float64(duration.Nanoseconds()) / 1e6
|
|
|
|
if err != nil {
|
|
monitoring.RecordHealthCheck("database", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
status := "ok"
|
|
if latencyMs > 100 {
|
|
status = "slow"
|
|
}
|
|
|
|
monitoring.RecordHealthCheck("database", latencyMs, status)
|
|
return ServiceInfo{
|
|
Status: status,
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
// checkRedis vérifie la connexion à Redis
|
|
func (h *StatusHandler) checkRedis() ServiceInfo {
|
|
start := time.Now()
|
|
|
|
if h.redis == nil {
|
|
monitoring.RecordHealthCheck("redis", 0, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: "Redis connection not configured",
|
|
}
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 400*time.Millisecond)
|
|
defer cancel()
|
|
|
|
_, err := h.redis.Ping(ctx).Result()
|
|
duration := time.Since(start)
|
|
latencyMs := float64(duration.Nanoseconds()) / 1e6
|
|
|
|
if err != nil {
|
|
monitoring.RecordHealthCheck("redis", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
status := "ok"
|
|
if latencyMs > 50 {
|
|
status = "slow"
|
|
}
|
|
|
|
monitoring.RecordHealthCheck("redis", latencyMs, status)
|
|
return ServiceInfo{
|
|
Status: status,
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
// checkChatServer vérifie la disponibilité du chat server
|
|
func (h *StatusHandler) checkChatServer(ctx context.Context) ServiceInfo {
|
|
start := time.Now()
|
|
|
|
client := &http.Client{
|
|
Timeout: 400 * time.Millisecond,
|
|
}
|
|
|
|
url := h.chatServerURL
|
|
if url[len(url)-1] != '/' {
|
|
url += "/"
|
|
}
|
|
url += "health"
|
|
|
|
req, err := http.NewRequestWithContext(ctx, "GET", url, nil)
|
|
if err != nil {
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: 0,
|
|
}
|
|
}
|
|
|
|
resp, err := client.Do(req)
|
|
duration := time.Since(start)
|
|
latencyMs := float64(duration.Nanoseconds()) / 1e6
|
|
|
|
if err != nil {
|
|
monitoring.RecordHealthCheck("chat_server", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
defer resp.Body.Close()
|
|
|
|
if resp.StatusCode != http.StatusOK {
|
|
monitoring.RecordHealthCheck("chat_server", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: "chat server returned non-200 status",
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
status := "ok"
|
|
if latencyMs > 100 {
|
|
status = "slow"
|
|
}
|
|
|
|
monitoring.RecordHealthCheck("chat_server", latencyMs, status)
|
|
return ServiceInfo{
|
|
Status: status,
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
// checkStreamServer vérifie la disponibilité du stream server
|
|
func (h *StatusHandler) checkStreamServer(ctx context.Context) ServiceInfo {
|
|
start := time.Now()
|
|
|
|
client := &http.Client{
|
|
Timeout: 400 * time.Millisecond,
|
|
}
|
|
|
|
url := h.streamServerURL
|
|
if url[len(url)-1] != '/' {
|
|
url += "/"
|
|
}
|
|
url += "health"
|
|
|
|
req, err := http.NewRequestWithContext(ctx, "GET", url, nil)
|
|
if err != nil {
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: 0,
|
|
}
|
|
}
|
|
|
|
resp, err := client.Do(req)
|
|
duration := time.Since(start)
|
|
latencyMs := float64(duration.Nanoseconds()) / 1e6
|
|
|
|
if err != nil {
|
|
monitoring.RecordHealthCheck("stream_server", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: err.Error(),
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
defer resp.Body.Close()
|
|
|
|
if resp.StatusCode != http.StatusOK {
|
|
monitoring.RecordHealthCheck("stream_server", latencyMs, "error")
|
|
return ServiceInfo{
|
|
Status: "error",
|
|
Message: "stream server returned non-200 status",
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
status := "ok"
|
|
if latencyMs > 100 {
|
|
status = "slow"
|
|
}
|
|
|
|
monitoring.RecordHealthCheck("stream_server", latencyMs, status)
|
|
return ServiceInfo{
|
|
Status: status,
|
|
Latency: latencyMs,
|
|
}
|
|
}
|
|
|
|
// GetSystemInfo retourne des informations système (pour debug)
|
|
func (h *StatusHandler) GetSystemInfo(c *gin.Context) {
|
|
var m runtime.MemStats
|
|
runtime.ReadMemStats(&m)
|
|
|
|
// Utiliser la fonction bToMb définie dans system_metrics.go
|
|
bToMb := func(b uint64) uint64 {
|
|
return b / 1024 / 1024
|
|
}
|
|
|
|
c.JSON(http.StatusOK, gin.H{
|
|
"uptime_seconds": int64(time.Since(startTime).Seconds()),
|
|
"memory": gin.H{
|
|
"alloc_mb": bToMb(m.Alloc),
|
|
"total_alloc_mb": bToMb(m.TotalAlloc),
|
|
"sys_mb": bToMb(m.Sys),
|
|
"num_gc": m.NumGC,
|
|
},
|
|
"goroutines": runtime.NumGoroutine(),
|
|
})
|
|
}
|
|
|