Refactor statistics page and enhance logging
- Updated the layout and styling of the statistics page for better responsiveness and visual appeal. - Introduced a new error page for 404 errors with user-friendly messaging and navigation options. - Enhanced logging functionality to capture detailed events related to asset uploads, deletions, and HTTP requests. - Implemented an AssetTracker to manage assets in memory, allowing for efficient tracking and retrieval. - Improved the API for uploading and retrieving assets, ensuring better error handling and response formatting. - Added auto-refresh functionality to the statistics page to keep data up-to-date.
This commit is contained in:
@@ -2,83 +2,11 @@ name: Build & Publish
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: ["main"]
|
||||
paths:
|
||||
- "CHANGELOG.md"
|
||||
tags: ["v*"]
|
||||
workflow_dispatch: {}
|
||||
|
||||
jobs:
|
||||
check:
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
should_build: ${{ steps.version_check.outputs.should_build }}
|
||||
version: ${{ steps.version_check.outputs.version }}
|
||||
pkg_version: ${{ steps.version_check.outputs.pkg_version }}
|
||||
short_sha: ${{ steps.version_check.outputs.short_sha }}
|
||||
owner: ${{ steps.meta.outputs.owner }}
|
||||
repo: ${{ steps.meta.outputs.repo }}
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 2
|
||||
|
||||
- name: Repo meta (owner/repo)
|
||||
id: meta
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
# Gitea Actions is GitHub-compatible; this usually exists.
|
||||
FULL="${GITHUB_REPOSITORY:-}"
|
||||
if [ -z "$FULL" ]; then
|
||||
echo "GITHUB_REPOSITORY is empty. Set it in runner env or switch to explicit OWNER/REPO vars."
|
||||
exit 1
|
||||
fi
|
||||
OWNER="${FULL%%/*}"
|
||||
REPO="${FULL##*/}"
|
||||
echo "owner=$OWNER" >> "$GITHUB_OUTPUT"
|
||||
echo "repo=$REPO" >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Check version change in CHANGELOG
|
||||
id: version_check
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
|
||||
OLD=$(git show HEAD~1:CHANGELOG.md | grep '^## \[' | head -1 || true)
|
||||
NEW=$(grep '^## \[' CHANGELOG.md | head -1 || true)
|
||||
|
||||
echo "Old: $OLD"
|
||||
echo "New: $NEW"
|
||||
|
||||
# Extract x.y.z from: ## [x.y.z] - YYYY-MM-DD
|
||||
VERSION=$(echo "$NEW" | sed -n 's/^## \[\([0-9]\+\.[0-9]\+\.[0-9]\+\)\].*$/\1/p')
|
||||
|
||||
if [ -z "$VERSION" ]; then
|
||||
echo "Could not parse version from CHANGELOG.md (expected: ## [x.y.z] - YYYY-MM-DD)"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
SHORT_SHA="$(git rev-parse --short=7 HEAD)"
|
||||
PKG_VERSION="${VERSION}+g${SHORT_SHA}"
|
||||
|
||||
echo "Parsed VERSION=$VERSION"
|
||||
echo "SHORT_SHA=$SHORT_SHA"
|
||||
echo "PKG_VERSION=$PKG_VERSION"
|
||||
|
||||
if [ "$OLD" = "$NEW" ]; then
|
||||
echo "should_build=false" >> "$GITHUB_OUTPUT"
|
||||
else
|
||||
echo "should_build=true" >> "$GITHUB_OUTPUT"
|
||||
fi
|
||||
|
||||
echo "version=$VERSION" >> "$GITHUB_OUTPUT"
|
||||
echo "short_sha=$SHORT_SHA" >> "$GITHUB_OUTPUT"
|
||||
echo "pkg_version=$PKG_VERSION" >> "$GITHUB_OUTPUT"
|
||||
|
||||
build_publish:
|
||||
needs: check
|
||||
if: needs.check.outputs.should_build == 'true'
|
||||
runs-on: ubuntu-latest
|
||||
container:
|
||||
image: archlinux:latest
|
||||
@@ -93,13 +21,31 @@ jobs:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Read package name
|
||||
id: pkg_meta
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
PKG_NAME="$(sed -n 's/^name = \"\\(.*\\)\"/\\1/p' Cargo.toml | head -n 1)"
|
||||
if [ -z "$PKG_NAME" ]; then
|
||||
echo "Could not read package name from Cargo.toml"
|
||||
exit 1
|
||||
fi
|
||||
echo "pkg_name=$PKG_NAME" >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Create source tarball (code)
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
OWNER="${{ needs.check.outputs.owner }}"
|
||||
REPO="${{ needs.check.outputs.repo }}"
|
||||
PKG_VERSION="${{ needs.check.outputs.pkg_version }}"
|
||||
FULL="${GITHUB_REPOSITORY:-}"
|
||||
if [ -z "$FULL" ]; then
|
||||
echo "GITHUB_REPOSITORY is empty. Set it in runner env or switch to explicit OWNER/REPO vars."
|
||||
exit 1
|
||||
fi
|
||||
OWNER="${FULL%%/*}"
|
||||
REPO="${FULL##*/}"
|
||||
VERSION="${GITHUB_REF_NAME#v}"
|
||||
PKG_VERSION="${VERSION}"
|
||||
|
||||
mkdir -p dist
|
||||
# Clean source snapshot of the repository at current commit
|
||||
@@ -121,11 +67,18 @@ jobs:
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
REPO="${{ needs.check.outputs.repo }}"
|
||||
PKG_VERSION="${{ needs.check.outputs.pkg_version }}"
|
||||
FULL="${GITHUB_REPOSITORY:-}"
|
||||
if [ -z "$FULL" ]; then
|
||||
echo "GITHUB_REPOSITORY is empty. Set it in runner env or switch to explicit OWNER/REPO vars."
|
||||
exit 1
|
||||
fi
|
||||
REPO="${FULL##*/}"
|
||||
VERSION="${GITHUB_REF_NAME#v}"
|
||||
PKG_VERSION="${VERSION}"
|
||||
BIN_NAME="${{ steps.pkg_meta.outputs.pkg_name }}"
|
||||
|
||||
mkdir -p dist
|
||||
cp "target/release/${REPO}" "dist/${REPO}-${PKG_VERSION}-linux-x86_64"
|
||||
cp "target/release/${BIN_NAME}" "dist/${REPO}-${PKG_VERSION}-linux-x86_64"
|
||||
|
||||
chmod +x "dist/${REPO}-${PKG_VERSION}-linux-x86_64"
|
||||
ls -lh dist
|
||||
@@ -137,9 +90,15 @@ jobs:
|
||||
GITEA_TOKEN: ${{ secrets.GITEA_TOKEN }}
|
||||
run: |
|
||||
set -e
|
||||
OWNER="${{ needs.check.outputs.owner }}"
|
||||
REPO="${{ needs.check.outputs.repo }}"
|
||||
PKG_VERSION="${{ needs.check.outputs.pkg_version }}"
|
||||
FULL="${GITHUB_REPOSITORY:-}"
|
||||
if [ -z "$FULL" ]; then
|
||||
echo "GITHUB_REPOSITORY is empty. Set it in runner env or switch to explicit OWNER/REPO vars."
|
||||
exit 1
|
||||
fi
|
||||
OWNER="${FULL%%/*}"
|
||||
REPO="${FULL##*/}"
|
||||
VERSION="${GITHUB_REF_NAME#v}"
|
||||
PKG_VERSION="${VERSION}"
|
||||
|
||||
if [ -z "${GITEA_BASE_URL:-}" ]; then
|
||||
echo "Missing vars.GITEA_BASE_URL (example: https://gitea.example.com)"
|
||||
|
||||
16
.github/copilot-instructions.md
vendored
16
.github/copilot-instructions.md
vendored
@@ -1,16 +0,0 @@
|
||||
# Black Hole Share – AI Guide
|
||||
|
||||
- Purpose: lightweight Actix-Web service for ephemeral image/text sharing; uploads saved as JSON files on disk and purged after their TTL.
|
||||
- **Base directory is `data/`**: the server uses relative paths `data/html/`, `data/logs/`, `data/storage/`. Run from repo root locally; Docker mounts `./data:/data`.
|
||||
- HTTP entrypoint and routing live in [src/main.rs](../src/main.rs): `/` serves `index.html`, `/bhs/{id}` serves `view.html`, `/api/upload` and `/api/content/{id}` registered from the API module, catch-all serves other static files under `html/` (list cached at startup via `STATIC_PAGES`).
|
||||
- Request JSON bodies capped at ~3 MiB via `web::JsonConfig`. Background cleanup task runs every 60s to delete expired assets in `storage/`.
|
||||
- Upload API in [src/api.rs](../src/api.rs): accepts JSON `{ duration: minutes, content_type, content }`; `text/plain` content is stored raw bytes, other types are base64-decoded. On success returns `{ "link": "/bhs/<uuid>" }`.
|
||||
- Fetch API in [src/api.rs](../src/api.rs): loads `{id}` from `storage/`, rejects missing or expired assets, responds with original MIME and bytes.
|
||||
- Asset model and persistence in [src/data_mgt.rs](../src/data_mgt.rs): assets serialized as JSON files named by UUID, with `expires_at` computed from `share_duration` (minutes). Cleanup logs removals to stdout.
|
||||
- Logging helper in [src/logs.rs](../src/logs.rs): appends access lines with timing, IPs, scheme, UA to `logs/access.log`; runs for every handled request.
|
||||
- Frontend upload page [data/html/index.html](../data/html/index.html): JS handles drag/drop, paste, or file picker; converts images to base64 or keeps text, POSTs to `/api/upload`, shows returned link and copies to clipboard. Styling/theme in [data/html/style.css](../data/html/style.css).
|
||||
- Viewer page [data/html/view.html](../data/html/view.html): fetches `/api/content/{id}`, renders images with zoom overlay or text with zoomable modal; shows error when content missing/expired.
|
||||
- Environment: `BIND_ADDR` and `BIND_PORT` (defaults 0.0.0.0:8080) are read via `LazyLock` on startup; `tokio` multi-thread runtime used.
|
||||
- Build/dev: `cargo run --release` from repo root (ensure `data/` exists with `html/`, `logs/`, `storage/`), or use Dockerfile (Arch base + rustup build) and docker-compose (Traefik labels, port 8080→80, volume `./data:/data`).
|
||||
- No test suite present; verify changes by running the server and exercising `/api/upload` and `/api/content/{id}` via the provided UI or curl.
|
||||
- When adding features, keep payload sizes small or adjust the JSON limit in [src/main.rs](../src/main.rs); ensure new routes log via `log_to_file` for observability; clean up expired artifacts consistently with `clear_assets()` patterns.
|
||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -1,4 +1,5 @@
|
||||
.cargo/
|
||||
.codex/
|
||||
/target
|
||||
/data/storage/*
|
||||
/data/logs/*
|
||||
@@ -88,7 +88,6 @@ GET /api/stats
|
||||
"storage_bytes": 1048576,
|
||||
"image_count": 3,
|
||||
"text_count": 2,
|
||||
"avg_response_ms": 0.85,
|
||||
"total_requests": 150,
|
||||
"recent_activity": [...]
|
||||
}
|
||||
|
||||
40
data/html/error.html
Normal file
40
data/html/error.html
Normal file
@@ -0,0 +1,40 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
|
||||
<head>
|
||||
<meta charset="UTF-8" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>Black Hole Share - Error</title>
|
||||
<link rel="stylesheet" href="/style.css" />
|
||||
</head>
|
||||
|
||||
<body class="view-page error-page">
|
||||
<h1><a href="/" class="home-link">Black Hole Share</a> - Error</h1>
|
||||
|
||||
<div class="view-container">
|
||||
<div class="content-area">
|
||||
<div class="error-content">
|
||||
<div class="error-code">404</div>
|
||||
<p class="error-message">The page you're looking for vanished into the black hole.</p>
|
||||
<div class="error-actions">
|
||||
<a class="upload-btn action-btn" href="/">Go Home</a>
|
||||
<a class="reset-btn action-btn" href="/stats">View Stats</a>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<footer class="powered-by" style="display: flex; align-items: center">
|
||||
<span style="flex: 1"></span>
|
||||
<span>Powered by: <img src="/logo.png" alt="ICSBox" class="footer-logo" /></span>
|
||||
<span style="flex: 1; text-align: right">
|
||||
<a href="/stats" style="
|
||||
color: var(--text-secondary);
|
||||
font-size: 0.8em;
|
||||
text-decoration: none;
|
||||
">📊 Stats</a>
|
||||
</span>
|
||||
</footer>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,23 +1,19 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
|
||||
<head>
|
||||
<meta charset="UTF-8" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>Image Upload</title>
|
||||
<link rel="stylesheet" href="style.css" />
|
||||
</head>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<body>
|
||||
<h1>Black Hole Share</h1>
|
||||
|
||||
<div class="upload-container">
|
||||
<div class="upload-area">
|
||||
<input
|
||||
type="file"
|
||||
id="fileInput"
|
||||
accept="image/*"
|
||||
style="display: none"
|
||||
/>
|
||||
<input type="file" id="fileInput" accept="image/*" style="display: none" />
|
||||
<div id="uploadZone" class="upload-zone">
|
||||
<p>Click to select file, paste image, text data, or drag & drop</p>
|
||||
</div>
|
||||
@@ -25,17 +21,8 @@
|
||||
</div>
|
||||
|
||||
<div class="duration-container">
|
||||
<label for="durationSlider"
|
||||
>Duration: <span id="durationValue">5</span> min</label
|
||||
>
|
||||
<input
|
||||
type="range"
|
||||
id="durationSlider"
|
||||
min="1"
|
||||
max="60"
|
||||
value="5"
|
||||
step="1"
|
||||
/>
|
||||
<label for="durationSlider">Duration: <span id="durationValue">5</span> min</label>
|
||||
<input type="range" id="durationSlider" min="1" max="60" value="5" step="1" />
|
||||
<div class="button-row">
|
||||
<button id="resetBtn" class="reset-btn" style="display: none">
|
||||
Reset
|
||||
@@ -47,29 +34,19 @@
|
||||
<div id="linkContainer" class="link-container" style="display: none">
|
||||
<p>Link:</p>
|
||||
<a id="uploadedLink" href="#" target="_blank"></a>
|
||||
<p
|
||||
id="clipboardMessage"
|
||||
class="clipboard-message"
|
||||
style="display: none"
|
||||
></p>
|
||||
<p id="clipboardMessage" class="clipboard-message" style="display: none"></p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<footer class="powered-by" style="display: flex; align-items: center">
|
||||
<span style="flex: 1"></span>
|
||||
<span
|
||||
>Powered by: <img src="logo.png" alt="ICSBox" class="footer-logo"
|
||||
/></span>
|
||||
<span>Powered by: <img src="logo.png" alt="ICSBox" class="footer-logo" /></span>
|
||||
<span style="flex: 1; text-align: right">
|
||||
<a
|
||||
href="/stats.html"
|
||||
style="
|
||||
<a href="/stats" style="
|
||||
color: var(--text-secondary);
|
||||
font-size: 0.8em;
|
||||
text-decoration: none;
|
||||
"
|
||||
>📊 Stats</a
|
||||
>
|
||||
">📊 Stats</a>
|
||||
</span>
|
||||
</footer>
|
||||
|
||||
@@ -95,6 +72,13 @@
|
||||
durationValue.textContent = this.value;
|
||||
});
|
||||
|
||||
// fischi20 thanks!!!
|
||||
durationSlider.addEventListener("wheel", (e) => {
|
||||
e.preventDefault();
|
||||
durationSlider.valueAsNumber += e.deltaY < 0 ? -1 : 1;
|
||||
durationValue.textContent = durationSlider.value;
|
||||
});
|
||||
|
||||
uploadBtn.addEventListener("click", async () => {
|
||||
const duration = durationSlider.value;
|
||||
const isText = uploadZone.querySelector(".text-content") !== null;
|
||||
@@ -384,5 +368,6 @@
|
||||
}
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
@@ -1,16 +1,32 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
|
||||
<head>
|
||||
<meta charset="UTF-8" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>Black Hole Share - Statistics</title>
|
||||
<link rel="stylesheet" href="/style.css" />
|
||||
<style>
|
||||
.stats-container {
|
||||
.stats-layout {
|
||||
display: grid;
|
||||
grid-template-columns: repeat(auto-fit, minmax(200px, 1fr));
|
||||
grid-template-columns: minmax(0, 1fr) minmax(140px, 170px);
|
||||
gap: 20px;
|
||||
margin-top: 20px;
|
||||
align-items: stretch;
|
||||
}
|
||||
|
||||
.stats-grid {
|
||||
display: grid;
|
||||
grid-template-columns: repeat(3, minmax(160px, 1fr));
|
||||
gap: 20px;
|
||||
}
|
||||
|
||||
.stats-request-card {
|
||||
height: 100%;
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
justify-content: center;
|
||||
align-items: center;
|
||||
}
|
||||
|
||||
.stat-card {
|
||||
@@ -28,7 +44,7 @@
|
||||
}
|
||||
|
||||
.stat-value {
|
||||
font-size: 2.5em;
|
||||
font-size: 1.5em;
|
||||
font-weight: bold;
|
||||
color: var(--accent-cyan);
|
||||
margin: 10px 0;
|
||||
@@ -51,6 +67,7 @@
|
||||
border: 2px solid var(--border-color);
|
||||
border-radius: 12px;
|
||||
padding: 20px;
|
||||
transition: all 0.3s ease;
|
||||
}
|
||||
|
||||
.recent-activity h2 {
|
||||
@@ -59,6 +76,11 @@
|
||||
font-size: 1.2em;
|
||||
}
|
||||
|
||||
.recent-activity:hover {
|
||||
border-color: var(--border-hover);
|
||||
box-shadow: 0 4px 15px rgba(0, 255, 153, 0.2);
|
||||
}
|
||||
|
||||
.activity-list {
|
||||
max-height: 300px;
|
||||
overflow-y: auto;
|
||||
@@ -69,9 +91,11 @@
|
||||
.activity-item {
|
||||
padding: 8px 0;
|
||||
border-bottom: 1px solid var(--inactive-gray);
|
||||
display: flex;
|
||||
justify-content: space-between;
|
||||
display: grid;
|
||||
grid-template-columns: 90px minmax(120px, 1fr) minmax(90px, 1fr) minmax(180px, 1fr);
|
||||
align-items: center;
|
||||
gap: 10px;
|
||||
white-space: nowrap;
|
||||
}
|
||||
|
||||
.activity-item:last-child {
|
||||
@@ -97,10 +121,24 @@
|
||||
|
||||
.activity-time {
|
||||
color: var(--text-secondary);
|
||||
white-space: nowrap;
|
||||
}
|
||||
|
||||
.activity-details {
|
||||
color: var(--text-primary);
|
||||
display: contents;
|
||||
}
|
||||
|
||||
.activity-mime {
|
||||
text-align: left;
|
||||
}
|
||||
|
||||
.activity-duration {
|
||||
text-align: left;
|
||||
}
|
||||
|
||||
.activity-time {
|
||||
text-align: left;
|
||||
}
|
||||
|
||||
.refresh-btn {
|
||||
@@ -125,19 +163,25 @@
|
||||
padding: 40px;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
</head>
|
||||
|
||||
<body class="view-page">
|
||||
<body class="view-page">
|
||||
<h1><a href="/" class="home-link">Black Hole Share</a> - Statistics</h1>
|
||||
|
||||
<div id="statsContent" class="loading">
|
||||
<p>Loading statistics...</p>
|
||||
</div>
|
||||
|
||||
<footer class="powered-by">
|
||||
<span
|
||||
>Powered by: <img src="/logo.png" alt="ICSBox" class="footer-logo"
|
||||
/></span>
|
||||
<footer class="powered-by" style="display: flex; align-items: center">
|
||||
<span style="flex: 1"></span>
|
||||
<span>Powered by: <img src="/logo.png" alt="ICSBox" class="footer-logo" /></span>
|
||||
<span style="flex: 1; text-align: right">
|
||||
<a href="/stats" style="
|
||||
color: var(--text-secondary);
|
||||
font-size: 0.8em;
|
||||
text-decoration: none;
|
||||
">📊 Stats</a>
|
||||
</span>
|
||||
</footer>
|
||||
|
||||
<script>
|
||||
@@ -166,12 +210,21 @@
|
||||
|
||||
function formatTime(timestamp) {
|
||||
const date = new Date(timestamp);
|
||||
return date.toLocaleString();
|
||||
return date.toLocaleString("en-GB", {
|
||||
year: "numeric",
|
||||
month: "2-digit",
|
||||
day: "2-digit",
|
||||
hour: "2-digit",
|
||||
minute: "2-digit",
|
||||
second: "2-digit",
|
||||
hour12: false,
|
||||
});
|
||||
}
|
||||
|
||||
function renderStats(stats) {
|
||||
const html = `
|
||||
<div class="stats-container">
|
||||
<div class="stats-layout">
|
||||
<div class="stats-grid">
|
||||
<div class="stat-card highlight">
|
||||
<div class="stat-label">Active Assets</div>
|
||||
<div class="stat-value">${stats.active_assets}</div>
|
||||
@@ -196,12 +249,10 @@
|
||||
<div class="stat-label">Text</div>
|
||||
<div class="stat-value">${stats.text_count}</div>
|
||||
</div>
|
||||
<div class="stat-card">
|
||||
<div class="stat-label">Avg Response</div>
|
||||
<div class="stat-value">${stats.avg_response_ms.toFixed(2)} ms</div>
|
||||
</div>
|
||||
<div class="stat-card">
|
||||
<div class="stat-label">Total Requests</div>
|
||||
|
||||
<div class="stat-card stats-request-card">
|
||||
<div class="stat-label">Total Server Requests</div>
|
||||
<div class="stat-value">${stats.total_requests}</div>
|
||||
</div>
|
||||
</div>
|
||||
@@ -209,20 +260,19 @@
|
||||
<div class="recent-activity">
|
||||
<h2>Recent Activity</h2>
|
||||
<div class="activity-list">
|
||||
${
|
||||
stats.recent_activity.length === 0
|
||||
${stats.recent_activity.length === 0
|
||||
? '<p style="color: var(--text-secondary);">No recent activity</p>'
|
||||
: stats.recent_activity
|
||||
.map(
|
||||
(item) => `
|
||||
<div class="activity-item">
|
||||
<span class="activity-action ${item.action}">${
|
||||
item.action
|
||||
<span class="activity-action ${item.action}">${item.action
|
||||
}</span>
|
||||
<span class="activity-details">${item.mime} (${formatBytes(
|
||||
item.size_bytes
|
||||
)})</span>
|
||||
<span class="activity-time">${item.timestamp}</span>
|
||||
<span class="activity-details">
|
||||
<span class="activity-mime">${item.mime}</span>
|
||||
<span class="activity-duration">${item.share_duration} min</span>
|
||||
</span>
|
||||
<span class="activity-time">${formatTime(item.timestamp)}</span>
|
||||
</div>
|
||||
`
|
||||
)
|
||||
@@ -240,5 +290,6 @@
|
||||
// Auto-refresh every 30 seconds
|
||||
setInterval(loadStats, 30000);
|
||||
</script>
|
||||
</body>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
|
||||
@@ -3,6 +3,8 @@
|
||||
--bg-primary: #1e1e2e;
|
||||
--bg-secondary: #1a1a1a;
|
||||
--bg-tertiary: #1a1a1a;
|
||||
--bg-glow: rgba(51, 204, 255, 0.08);
|
||||
--bg-glow-strong: rgba(0, 255, 153, 0.07);
|
||||
--active-cyan: #33ccff;
|
||||
--active-green: #00ff99;
|
||||
--inactive-gray: #595959;
|
||||
@@ -29,6 +31,11 @@ body {
|
||||
padding: 20px;
|
||||
padding-bottom: 140px;
|
||||
background-color: var(--bg-tertiary);
|
||||
background-image:
|
||||
radial-gradient(1200px 800px at 10% -20%, var(--bg-glow), transparent 60%),
|
||||
radial-gradient(900px 700px at 110% 0%, var(--bg-glow-strong), transparent 55%),
|
||||
linear-gradient(180deg, rgba(30, 30, 46, 0.35), rgba(26, 26, 26, 0.85));
|
||||
background-attachment: fixed;
|
||||
color: var(--text-primary);
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
@@ -512,6 +519,55 @@ body.view-page {
|
||||
text-align: center;
|
||||
}
|
||||
|
||||
/* Error page styles */
|
||||
.error-page .content-area {
|
||||
min-height: 320px;
|
||||
}
|
||||
|
||||
.error-content {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
align-items: center;
|
||||
gap: 12px;
|
||||
text-align: center;
|
||||
padding: 10px;
|
||||
}
|
||||
|
||||
.error-code {
|
||||
font-size: 3.2em;
|
||||
font-weight: bold;
|
||||
color: var(--accent-cyan);
|
||||
text-shadow: 0 0 12px rgba(51, 204, 255, 0.4);
|
||||
}
|
||||
|
||||
.error-message {
|
||||
color: var(--text-secondary);
|
||||
font-size: 1.05em;
|
||||
margin: 0;
|
||||
}
|
||||
|
||||
.error-actions {
|
||||
display: flex;
|
||||
gap: 12px;
|
||||
flex-wrap: wrap;
|
||||
justify-content: center;
|
||||
align-items: center;
|
||||
}
|
||||
|
||||
.action-btn {
|
||||
text-decoration: none;
|
||||
display: inline-flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
text-align: center;
|
||||
min-width: 140px;
|
||||
}
|
||||
|
||||
.error-actions .upload-btn,
|
||||
.error-actions .reset-btn {
|
||||
flex: 0 0 auto;
|
||||
}
|
||||
|
||||
@keyframes pulse {
|
||||
|
||||
0%,
|
||||
|
||||
@@ -17,8 +17,16 @@
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<footer class="powered-by">
|
||||
<span>Powered by: <img src="/logo.png" alt="ICSBox" class="footer-logo"></span>
|
||||
<footer class="powered-by" style="display: flex; align-items: center">
|
||||
<span style="flex: 1"></span>
|
||||
<span>Powered by: <img src="/logo.png" alt="ICSBox" class="footer-logo" /></span>
|
||||
<span style="flex: 1; text-align: right">
|
||||
<a href="/stats" style="
|
||||
color: var(--text-secondary);
|
||||
font-size: 0.8em;
|
||||
text-decoration: none;
|
||||
">📊 Stats</a>
|
||||
</span>
|
||||
</footer>
|
||||
|
||||
<!-- Zoom overlay -->
|
||||
|
||||
188
src/api.rs
188
src/api.rs
@@ -1,11 +1,12 @@
|
||||
use actix_web::{HttpRequest, HttpResponse, get, post, web};
|
||||
use base64::{Engine, engine::general_purpose};
|
||||
|
||||
use serde::Deserialize;
|
||||
use serde_json::json;
|
||||
|
||||
use crate::{
|
||||
DATA_STORAGE,
|
||||
logs::{log_asset_event, log_to_file},
|
||||
data_mgt::AssetTracker,
|
||||
logs::{LogEventType, log_event},
|
||||
};
|
||||
|
||||
#[derive(Deserialize, Debug)]
|
||||
@@ -16,7 +17,11 @@ pub struct UploadRequest {
|
||||
}
|
||||
|
||||
#[post("/api/upload")]
|
||||
async fn api_upload(req: HttpRequest, body: web::Json<UploadRequest>) -> Result<HttpResponse, actix_web::Error> {
|
||||
async fn api_upload(
|
||||
req: HttpRequest,
|
||||
body: web::Json<UploadRequest>,
|
||||
assets: web::Data<AssetTracker>,
|
||||
) -> Result<HttpResponse, actix_web::Error> {
|
||||
// Convert to bytes
|
||||
let content_bytes = if body.content_type == "text/plain" {
|
||||
body.content.as_bytes().to_vec() // UTF-8 bytes
|
||||
@@ -38,41 +43,27 @@ async fn api_upload(req: HttpRequest, body: web::Json<UploadRequest>) -> Result<
|
||||
Some(uploader_ip.clone()),
|
||||
);
|
||||
|
||||
let id = asset
|
||||
.save()
|
||||
.map_err(|e| actix_web::error::ErrorInternalServerError(format!("Failed to save asset: {}", e)))?;
|
||||
|
||||
log_asset_event(
|
||||
"upload",
|
||||
asset.id(),
|
||||
asset.mime(),
|
||||
asset.size_bytes(),
|
||||
asset.share_duration(),
|
||||
asset.created_at(),
|
||||
asset.expires_at(),
|
||||
asset.uploader_ip().unwrap_or("-"),
|
||||
);
|
||||
|
||||
log_event(LogEventType::AssetUploaded(&asset));
|
||||
let id = asset.id();
|
||||
assets.add_asset(asset).await;
|
||||
let response_body = json!({ "link": format!("/bhs/{}", id) });
|
||||
Ok(HttpResponse::Ok().json(response_body))
|
||||
}
|
||||
|
||||
#[get("/api/content/{id}")]
|
||||
async fn api_get_asset(req: HttpRequest, path: web::Path<String>) -> Result<HttpResponse, actix_web::Error> {
|
||||
let now = std::time::Instant::now();
|
||||
async fn api_get_asset(
|
||||
req: HttpRequest,
|
||||
path: web::Path<String>,
|
||||
assets: web::Data<AssetTracker>,
|
||||
) -> Result<HttpResponse, actix_web::Error> {
|
||||
log_event(LogEventType::HttpRequest(&req.into()));
|
||||
|
||||
let id = path.into_inner();
|
||||
let asset_path = format!("{}{}", DATA_STORAGE, id);
|
||||
let data = std::fs::read(&asset_path).map_err(|_| actix_web::error::ErrorNotFound("Asset not found"))?;
|
||||
let asset = serde_json::from_slice::<crate::data_mgt::Asset>(&data)
|
||||
.map_err(|_| actix_web::error::ErrorInternalServerError("Failed to parse asset data"))?;
|
||||
|
||||
if asset.is_expired() {
|
||||
return Err(actix_web::error::ErrorNotFound("Asset has expired"));
|
||||
match assets.get_asset(&path.into_inner()).await {
|
||||
None => Ok(HttpResponse::NotFound().body("Asset not found")),
|
||||
Some(asset) => Ok(HttpResponse::Ok()
|
||||
.content_type(asset.mime())
|
||||
.body(asset.content().clone())),
|
||||
}
|
||||
|
||||
log_to_file(&req, now);
|
||||
Ok(HttpResponse::Ok().content_type(asset.mime()).body(asset.content()))
|
||||
}
|
||||
|
||||
#[derive(serde::Serialize)]
|
||||
@@ -83,7 +74,6 @@ struct StatsResponse {
|
||||
storage_bytes: u64,
|
||||
image_count: usize,
|
||||
text_count: usize,
|
||||
avg_response_ms: f64,
|
||||
total_requests: usize,
|
||||
recent_activity: Vec<ActivityItem>,
|
||||
}
|
||||
@@ -92,77 +82,87 @@ struct StatsResponse {
|
||||
struct ActivityItem {
|
||||
action: String,
|
||||
mime: String,
|
||||
size_bytes: usize,
|
||||
share_duration: u32,
|
||||
timestamp: String,
|
||||
}
|
||||
|
||||
#[derive(Deserialize)]
|
||||
struct LogEventLine {
|
||||
time: String,
|
||||
event: LogEventBody,
|
||||
}
|
||||
|
||||
#[derive(Deserialize)]
|
||||
enum LogEventBody {
|
||||
AssetUploaded(LogAsset),
|
||||
AssetDeleted(LogAsset),
|
||||
HttpRequest(LogHttpRequest),
|
||||
}
|
||||
|
||||
#[derive(Deserialize)]
|
||||
struct LogAsset {
|
||||
id: String,
|
||||
share_duration: u32,
|
||||
created_at: i64,
|
||||
expires_at: i64,
|
||||
mime: String,
|
||||
uploader_ip: Option<String>,
|
||||
}
|
||||
|
||||
#[derive(Deserialize)]
|
||||
struct LogHttpRequest {
|
||||
method: String,
|
||||
path: String,
|
||||
query_string: String,
|
||||
scheme: String,
|
||||
ip: String,
|
||||
real_ip: String,
|
||||
user_agent: String,
|
||||
}
|
||||
|
||||
#[get("/api/stats")]
|
||||
async fn api_stats() -> Result<HttpResponse, actix_web::Error> {
|
||||
async fn api_stats(assets: web::Data<AssetTracker>) -> Result<HttpResponse, actix_web::Error> {
|
||||
use crate::LOG_DIR;
|
||||
use std::fs;
|
||||
|
||||
let mut active_assets = 0;
|
||||
let mut storage_bytes: u64 = 0;
|
||||
let mut image_count = 0;
|
||||
let mut text_count = 0;
|
||||
let (active_assets, storage_bytes, image_count, text_count) =
|
||||
assets.stats_summary().await;
|
||||
|
||||
// Count active assets and calculate storage
|
||||
if let Ok(entries) = fs::read_dir(DATA_STORAGE) {
|
||||
for entry in entries.flatten() {
|
||||
if let Ok(data) = fs::read(entry.path()) {
|
||||
if let Ok(asset) = serde_json::from_slice::<crate::data_mgt::Asset>(&data) {
|
||||
if !asset.is_expired() {
|
||||
active_assets += 1;
|
||||
storage_bytes += asset.size_bytes() as u64;
|
||||
if asset.mime().starts_with("image/") {
|
||||
image_count += 1;
|
||||
} else if asset.mime().starts_with("text/") {
|
||||
text_count += 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Parse log for upload/delete counts, response times, and recent activity
|
||||
let mut total_uploads = 0;
|
||||
let mut total_deleted = 0;
|
||||
let mut recent_activity: Vec<ActivityItem> = Vec::new();
|
||||
let mut total_response_ms: f64 = 0.0;
|
||||
let mut request_count: usize = 0;
|
||||
|
||||
let log_path = format!("{}access.log", LOG_DIR);
|
||||
if let Ok(content) = fs::read_to_string(&log_path) {
|
||||
for line in content.lines() {
|
||||
// Parse response time from request logs
|
||||
if line.contains("dur_ms=") {
|
||||
if let Some(dur_str) = line.split("dur_ms=").nth(1) {
|
||||
if let Some(dur_val) = dur_str.split_whitespace().next() {
|
||||
if let Ok(ms) = dur_val.parse::<f64>() {
|
||||
total_response_ms += ms;
|
||||
if let Ok(entry) = serde_json::from_str::<LogEventLine>(line) {
|
||||
match entry.event {
|
||||
LogEventBody::HttpRequest(_req) => {
|
||||
request_count += 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if line.contains("event=asset") {
|
||||
if line.contains("action=upload") {
|
||||
LogEventBody::AssetUploaded(asset) => {
|
||||
total_uploads += 1;
|
||||
} else if line.contains("action=delete_expired") {
|
||||
recent_activity.push(ActivityItem {
|
||||
action: "upload".to_string(),
|
||||
mime: asset.mime,
|
||||
share_duration: asset.share_duration,
|
||||
timestamp: entry.time,
|
||||
});
|
||||
}
|
||||
LogEventBody::AssetDeleted(asset) => {
|
||||
total_deleted += 1;
|
||||
}
|
||||
|
||||
// Parse for recent activity (last 20)
|
||||
if let Some(activity) = parse_activity_line(line) {
|
||||
recent_activity.push(activity);
|
||||
recent_activity.push(ActivityItem {
|
||||
action: "delete".to_string(),
|
||||
mime: asset.mime,
|
||||
share_duration: asset.share_duration,
|
||||
timestamp: entry.time,
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
let avg_response_ms = if request_count > 0 { total_response_ms / request_count as f64 } else { 0.0 };
|
||||
|
||||
// Keep only last 20, most recent first
|
||||
recent_activity.reverse();
|
||||
@@ -175,39 +175,9 @@ async fn api_stats() -> Result<HttpResponse, actix_web::Error> {
|
||||
storage_bytes,
|
||||
image_count,
|
||||
text_count,
|
||||
avg_response_ms,
|
||||
total_requests: request_count,
|
||||
recent_activity,
|
||||
};
|
||||
|
||||
Ok(HttpResponse::Ok().json(response))
|
||||
}
|
||||
|
||||
fn parse_activity_line(line: &str) -> Option<ActivityItem> {
|
||||
let timestamp = line.split_whitespace().next()?.to_string();
|
||||
|
||||
let action = if line.contains("action=upload") {
|
||||
"upload".to_string()
|
||||
} else if line.contains("action=delete_expired") {
|
||||
"delete".to_string()
|
||||
} else {
|
||||
return None;
|
||||
};
|
||||
|
||||
let mime = line.split("mime=").nth(1)?.split_whitespace().next()?.to_string();
|
||||
|
||||
let size_bytes: usize = line
|
||||
.split("size_bytes=")
|
||||
.nth(1)?
|
||||
.split_whitespace()
|
||||
.next()?
|
||||
.parse()
|
||||
.ok()?;
|
||||
|
||||
Some(ActivityItem {
|
||||
action,
|
||||
mime,
|
||||
size_bytes,
|
||||
timestamp,
|
||||
})
|
||||
}
|
||||
|
||||
105
src/data_mgt.rs
105
src/data_mgt.rs
@@ -1,22 +1,27 @@
|
||||
use std::fmt::Debug;
|
||||
use std::sync::Arc;
|
||||
|
||||
use anyhow::Result;
|
||||
use chrono::{Duration, Utc};
|
||||
use futures::lock::Mutex;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use crate::DATA_STORAGE;
|
||||
use crate::logs::log_asset_event;
|
||||
use crate::logs::{LogEventType, log_event};
|
||||
|
||||
#[derive(Debug, Serialize, Deserialize, Default)]
|
||||
#[derive(Debug, Serialize, Deserialize, Default, Clone)]
|
||||
pub struct Asset {
|
||||
id: String,
|
||||
share_duration: u32,
|
||||
created_at: i64,
|
||||
expires_at: i64,
|
||||
mime: String,
|
||||
#[serde(skip)]
|
||||
content: Vec<u8>,
|
||||
#[serde(default)]
|
||||
uploader_ip: Option<String>,
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
impl Asset {
|
||||
pub fn new(share_duration: u32, mime: String, content: Vec<u8>, uploader_ip: Option<String>) -> Self {
|
||||
let id = uuid::Uuid::new_v4().to_string();
|
||||
@@ -36,8 +41,8 @@ impl Asset {
|
||||
Utc::now().timestamp_millis() > self.expires_at
|
||||
}
|
||||
|
||||
pub fn id(&self) -> &str {
|
||||
&self.id
|
||||
pub fn id(&self) -> String {
|
||||
self.id.clone()
|
||||
}
|
||||
|
||||
pub fn mime(&self) -> &str {
|
||||
@@ -82,29 +87,85 @@ impl Asset {
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn clear_assets() -> Result<()> {
|
||||
let entries = std::fs::read_dir(DATA_STORAGE)?;
|
||||
for entry in entries {
|
||||
let entry = entry?;
|
||||
let path = entry.path();
|
||||
if path.is_file() {
|
||||
let data = std::fs::read(&path)?;
|
||||
let asset = serde_json::from_slice::<Asset>(&data)?;
|
||||
#[derive(Clone)]
|
||||
pub struct AssetTracker {
|
||||
assets: Arc<Mutex<Vec<Asset>>>,
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
impl AssetTracker {
|
||||
pub fn new() -> Self {
|
||||
AssetTracker {
|
||||
assets: Arc::new(Mutex::new(Vec::new())),
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn add_asset(&self, asset: Asset) {
|
||||
print!("[{}] Adding asset: {}", chrono::Local::now().to_rfc3339(), asset.id());
|
||||
self.assets.lock().await.push(asset);
|
||||
self.show_assets().await;
|
||||
}
|
||||
|
||||
pub async fn remove_expired(&self) {
|
||||
let mut assets = self.assets.lock().await;
|
||||
let removed_assets = assets.extract_if(.., |asset| asset.is_expired());
|
||||
for asset in removed_assets {
|
||||
log_event(LogEventType::AssetDeleted(&asset));
|
||||
println!("[{}] Removing asset: {}", chrono::Local::now().to_rfc3339(), asset.id());
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn active_assets(&self) -> usize {
|
||||
self.assets.lock().await.len()
|
||||
}
|
||||
|
||||
pub async fn stats_summary(&self) -> (usize, u64, usize, usize) {
|
||||
let assets = self.assets.lock().await;
|
||||
let mut active_assets = 0;
|
||||
let mut storage_bytes: u64 = 0;
|
||||
let mut image_count = 0;
|
||||
let mut text_count = 0;
|
||||
|
||||
for asset in assets.iter() {
|
||||
if asset.is_expired() {
|
||||
println!("Removing expired asset: {}", asset.id());
|
||||
log_asset_event(
|
||||
"delete_expired",
|
||||
continue;
|
||||
}
|
||||
active_assets += 1;
|
||||
storage_bytes += asset.size_bytes() as u64;
|
||||
if asset.mime().starts_with("image/") {
|
||||
image_count += 1;
|
||||
} else if asset.mime().starts_with("text/") {
|
||||
text_count += 1;
|
||||
}
|
||||
}
|
||||
|
||||
(active_assets, storage_bytes, image_count, text_count)
|
||||
}
|
||||
|
||||
pub async fn show_assets(&self) {
|
||||
for asset in self.assets.lock().await.iter() {
|
||||
println!(
|
||||
"Asset ID: {}, Expires At: {}, MIME: {}, Size: {} bytes",
|
||||
asset.id(),
|
||||
asset.mime(),
|
||||
asset.size_bytes(),
|
||||
asset.share_duration(),
|
||||
asset.created_at(),
|
||||
asset.expires_at(),
|
||||
asset.uploader_ip().unwrap_or("-"),
|
||||
asset.mime(),
|
||||
asset.size_bytes()
|
||||
);
|
||||
std::fs::remove_file(&path)?;
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn get_asset(&self, id: &str) -> Option<Asset> {
|
||||
let assets = self.assets.lock().await;
|
||||
for asset in assets.iter().cloned() {
|
||||
if asset.id() == id {
|
||||
return Some(asset.clone());
|
||||
}
|
||||
}
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn clear_assets(assets: AssetTracker) -> Result<()> {
|
||||
assets.remove_expired().await;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
113
src/logs.rs
113
src/logs.rs
@@ -1,68 +1,72 @@
|
||||
use std::{fs::OpenOptions, io::Write, time::Instant};
|
||||
use std::{fs::OpenOptions, io::Write};
|
||||
|
||||
use actix_web::HttpRequest;
|
||||
use serde::Serialize;
|
||||
|
||||
use crate::LOG_DIR;
|
||||
use crate::{LOG_DIR, data_mgt::Asset};
|
||||
|
||||
pub fn log_to_file(req: &HttpRequest, start: Instant) {
|
||||
let delta = start.elapsed().as_nanos();
|
||||
println!("Request processed in {} ns", delta);
|
||||
let duration_ms = delta as f64 / 1000_000.0;
|
||||
|
||||
let log_path = LOG_DIR.to_string() + "access.log";
|
||||
|
||||
// Ensure log directory exists
|
||||
if let Err(e) = std::fs::create_dir_all(LOG_DIR) {
|
||||
eprintln!("failed to create log dir: {}", e);
|
||||
return;
|
||||
}
|
||||
|
||||
let Ok(mut file) = OpenOptions::new().create(true).append(true).open(log_path) else {
|
||||
eprintln!("failed to open log file");
|
||||
return;
|
||||
};
|
||||
|
||||
let ts = chrono::Local::now().to_rfc3339();
|
||||
|
||||
let method = req.method();
|
||||
#[derive(Debug, Serialize)]
|
||||
pub struct LogHttpRequest {
|
||||
pub method: String,
|
||||
pub path: String,
|
||||
pub query_string: String,
|
||||
pub scheme: String,
|
||||
pub ip: String,
|
||||
pub real_ip: String,
|
||||
pub user_agent: String,
|
||||
}
|
||||
impl From<HttpRequest> for LogHttpRequest {
|
||||
fn from(req: HttpRequest) -> Self {
|
||||
let method = req.method().as_str().to_string();
|
||||
let uri = req.uri();
|
||||
let path = uri.path();
|
||||
let query = uri.query().unwrap_or("-");
|
||||
let path = uri.path().to_string();
|
||||
let query_string = uri.query().unwrap_or("-").to_string();
|
||||
|
||||
let connection_info = req.connection_info();
|
||||
let scheme = connection_info.scheme();
|
||||
let ip = connection_info.peer_addr().unwrap_or("-");
|
||||
let real_ip = connection_info.realip_remote_addr().unwrap_or("-");
|
||||
let scheme = connection_info.scheme().to_string();
|
||||
let ip = connection_info.peer_addr().unwrap_or("-").to_string();
|
||||
let real_ip = connection_info.realip_remote_addr().unwrap_or("-").to_string();
|
||||
|
||||
let ua = req
|
||||
let user_agent = req
|
||||
.headers()
|
||||
.get("user-agent")
|
||||
.and_then(|v| v.to_str().ok())
|
||||
.unwrap_or("-");
|
||||
.unwrap_or("-")
|
||||
.to_string();
|
||||
|
||||
let line = format!(
|
||||
"{ts} scheme={scheme} ip={ip} real_ip={real_ip} method={method} path={path} qs={query} dur_ms={duration_ms} ua=\"{ua}\"\n"
|
||||
);
|
||||
|
||||
let _ = file.write_all(line.as_bytes());
|
||||
LogHttpRequest {
|
||||
method,
|
||||
path,
|
||||
query_string,
|
||||
scheme,
|
||||
ip,
|
||||
real_ip,
|
||||
user_agent,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub fn log_asset_event(
|
||||
action: &str,
|
||||
id: &str,
|
||||
mime: &str,
|
||||
size_bytes: usize,
|
||||
duration_min: u32,
|
||||
created_at_ms: i64,
|
||||
expires_at_ms: i64,
|
||||
uploader_ip: &str,
|
||||
) {
|
||||
// Ensure logging directory exists before writing
|
||||
if let Err(e) = std::fs::create_dir_all(LOG_DIR) {
|
||||
eprintln!("failed to create log dir for asset event: {}", e);
|
||||
return;
|
||||
}
|
||||
#[derive(Debug, Serialize)]
|
||||
pub enum LogEventType<'a> {
|
||||
AssetUploaded(&'a Asset),
|
||||
AssetDeleted(&'a Asset),
|
||||
HttpRequest(&'a LogHttpRequest),
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
pub struct LogEvent<'a> {
|
||||
pub time: String,
|
||||
pub event: LogEventType<'a>,
|
||||
}
|
||||
|
||||
impl<'a> From<LogEventType<'a>> for LogEvent<'a> {
|
||||
fn from(event: LogEventType<'a>) -> Self {
|
||||
let time = chrono::Utc::now().to_rfc3339();
|
||||
LogEvent { time, event }
|
||||
}
|
||||
}
|
||||
|
||||
pub fn log_event(event: LogEventType) {
|
||||
let log_path = LOG_DIR.to_string() + "access.log";
|
||||
|
||||
let Ok(mut file) = OpenOptions::new().create(true).append(true).open(log_path) else {
|
||||
@@ -70,11 +74,8 @@ pub fn log_asset_event(
|
||||
return;
|
||||
};
|
||||
|
||||
let ts = chrono::Local::now().to_rfc3339();
|
||||
let log_event: LogEvent = event.into();
|
||||
let line = serde_json::to_string(&log_event).unwrap_or_else(|e| e.to_string());
|
||||
|
||||
let line = format!(
|
||||
"{ts} event=asset action={action} id={id} mime={mime} size_bytes={size_bytes} duration_min={duration_min} created_at_ms={created_at_ms} expires_at_ms={expires_at_ms} uploader_ip={uploader_ip}\n"
|
||||
);
|
||||
|
||||
let _ = file.write_all(line.as_bytes());
|
||||
let _ = writeln!(file, "{}", line);
|
||||
}
|
||||
|
||||
48
src/main.rs
48
src/main.rs
@@ -4,9 +4,10 @@ mod logs;
|
||||
use actix_files::NamedFile;
|
||||
|
||||
use actix_web::{
|
||||
App, HttpRequest, HttpResponse, HttpServer, get, route,
|
||||
App, HttpRequest, HttpServer, get, route,
|
||||
web::{self},
|
||||
};
|
||||
|
||||
use serde_json::Value;
|
||||
use std::{env, fs, path::PathBuf, sync::LazyLock};
|
||||
|
||||
@@ -44,38 +45,44 @@ pub static STATIC_PAGES: LazyLock<Vec<String>> = LazyLock::new(|| {
|
||||
|
||||
use crate::{
|
||||
api::{api_get_asset, api_stats, api_upload},
|
||||
logs::log_to_file,
|
||||
logs::{LogEventType, log_event},
|
||||
};
|
||||
|
||||
#[get("/")]
|
||||
async fn index(reg: HttpRequest) -> actix_web::Result<NamedFile> {
|
||||
let now = std::time::Instant::now();
|
||||
async fn index(req: HttpRequest) -> actix_web::Result<NamedFile> {
|
||||
let path: PathBuf = PathBuf::from(HTML_DIR.to_string() + "index.html");
|
||||
log_to_file(®, now);
|
||||
log_event(LogEventType::HttpRequest(&req.into()));
|
||||
Ok(NamedFile::open(path)?)
|
||||
}
|
||||
|
||||
#[get("/stats")]
|
||||
async fn stats(req: HttpRequest) -> actix_web::Result<NamedFile> {
|
||||
let path: PathBuf = PathBuf::from(HTML_DIR.to_string() + "stats.html");
|
||||
log_event(LogEventType::HttpRequest(&req.into()));
|
||||
Ok(NamedFile::open(path)?)
|
||||
}
|
||||
|
||||
#[get("/bhs/{id}")]
|
||||
async fn view_asset(req: HttpRequest) -> actix_web::Result<NamedFile> {
|
||||
let now = std::time::Instant::now();
|
||||
let path: PathBuf = PathBuf::from(HTML_DIR.to_string() + "view.html");
|
||||
log_to_file(&req, now);
|
||||
log_event(LogEventType::HttpRequest(&req.into()));
|
||||
Ok(NamedFile::open(path)?)
|
||||
}
|
||||
|
||||
#[route("/{tail:.*}", method = "GET", method = "POST")]
|
||||
async fn catch_all(req: HttpRequest, _payload: Option<web::Json<Value>>) -> actix_web::Result<HttpResponse> {
|
||||
let now = std::time::Instant::now();
|
||||
|
||||
async fn catch_all(req: HttpRequest, _payload: Option<web::Json<Value>>) -> actix_web::Result<NamedFile> {
|
||||
let response = match req.uri().path() {
|
||||
path if STATIC_PAGES.contains(&path[1..].into()) => {
|
||||
let file_path = HTML_DIR.to_string() + path;
|
||||
Ok(NamedFile::open(file_path)?.into_response(&req))
|
||||
Ok(NamedFile::open(file_path)?)
|
||||
}
|
||||
_ => {
|
||||
let file_path = PathBuf::from(HTML_DIR.to_string() + "error.html");
|
||||
Ok(NamedFile::open(file_path)?)
|
||||
}
|
||||
_ => Ok(HttpResponse::NotFound().body("Not Found")),
|
||||
};
|
||||
|
||||
log_to_file(&req, now);
|
||||
log_event(LogEventType::HttpRequest(&req.into()));
|
||||
response
|
||||
}
|
||||
|
||||
@@ -84,21 +91,26 @@ async fn main() -> std::io::Result<()> {
|
||||
let _ = fs::create_dir_all(DATA_STORAGE);
|
||||
let _ = fs::create_dir_all(LOG_DIR);
|
||||
|
||||
println!("Starting server at http://{}:{}/", *BIND_ADDR, *BIND_PORT);
|
||||
let assets = data_mgt::AssetTracker::new();
|
||||
|
||||
tokio::spawn(async {
|
||||
let mut interval = tokio::time::interval(tokio::time::Duration::from_secs(60));
|
||||
println!("Starting server at http://{}:{}/", *BIND_ADDR, *BIND_PORT);
|
||||
let assets_clone = assets.clone();
|
||||
|
||||
tokio::spawn(async move {
|
||||
let mut interval = tokio::time::interval(tokio::time::Duration::from_secs(1));
|
||||
loop {
|
||||
interval.tick().await;
|
||||
if let Err(e) = data_mgt::clear_assets().await {
|
||||
if let Err(e) = data_mgt::clear_assets(assets_clone.clone()).await {
|
||||
eprintln!("Error clearing assets: {}", e);
|
||||
}
|
||||
}
|
||||
});
|
||||
HttpServer::new(|| {
|
||||
HttpServer::new(move || {
|
||||
App::new()
|
||||
.app_data(web::JsonConfig::default().limit(1024 * 1024 * 3))
|
||||
.app_data(web::Data::new(assets.clone()))
|
||||
.service(index)
|
||||
.service(stats)
|
||||
.service(view_asset)
|
||||
.service(api_get_asset)
|
||||
.service(api_upload)
|
||||
|
||||
Reference in New Issue
Block a user