Compare commits
2 Commits
1.3.13
...
2026-01-01
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
f354d89ece | ||
|
|
5247a37fac |
69
BUILDING.md
69
BUILDING.md
@@ -379,6 +379,50 @@ rsync -azvu -e "ssh -i ~/.ssh/..." dist ubuntutest@test.timesafari.app:time-safa
|
|||||||
- Record the new hash in the changelog. Edit package.json to increment version &
|
- Record the new hash in the changelog. Edit package.json to increment version &
|
||||||
add "-beta", `npm install`, commit, and push. Also record what version is on production.
|
add "-beta", `npm install`, commit, and push. Also record what version is on production.
|
||||||
|
|
||||||
|
## WebAuthn Verification Server
|
||||||
|
|
||||||
|
TimeSafari includes a server-side WebAuthn verification service for secure passkey registration and authentication. This server must be running for passkey features to work (unless offline mode is enabled).
|
||||||
|
|
||||||
|
### Quick Setup
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Navigate to server directory
|
||||||
|
cd server
|
||||||
|
|
||||||
|
# Install dependencies
|
||||||
|
npm install
|
||||||
|
|
||||||
|
# Copy and configure environment
|
||||||
|
cp .env.example .env
|
||||||
|
# Edit .env with your RP_ID, RP_NAME, RP_ORIGIN
|
||||||
|
|
||||||
|
# Start development server
|
||||||
|
npm run dev
|
||||||
|
```
|
||||||
|
|
||||||
|
The server runs on `http://localhost:3002` by default.
|
||||||
|
|
||||||
|
### Production Deployment
|
||||||
|
|
||||||
|
For production, you'll need to:
|
||||||
|
|
||||||
|
1. **Configure environment variables** in `.env`:
|
||||||
|
- `RP_ID`: Your domain (e.g., `timesafari.app`)
|
||||||
|
- `RP_NAME`: Application name
|
||||||
|
- `RP_ORIGIN`: Your app's origin URL
|
||||||
|
- `PORT`: Server port (default: 3002)
|
||||||
|
|
||||||
|
2. **Replace in-memory storage** with:
|
||||||
|
- Redis for challenge storage
|
||||||
|
- Database for credential persistence
|
||||||
|
- Session management for user binding
|
||||||
|
|
||||||
|
3. **Deploy the server** alongside your main application
|
||||||
|
|
||||||
|
4. **Configure client** via `VITE_WEBAUTHN_SERVER_URL` environment variable
|
||||||
|
|
||||||
|
See [server/README.md](server/README.md) for complete API documentation and deployment guide.
|
||||||
|
|
||||||
## Docker Deployment
|
## Docker Deployment
|
||||||
|
|
||||||
The application can be containerized using Docker for consistent deployment across
|
The application can be containerized using Docker for consistent deployment across
|
||||||
@@ -1534,6 +1578,7 @@ VITE_APP_SERVER=https://timesafari.app
|
|||||||
|
|
||||||
# Feature Flags
|
# Feature Flags
|
||||||
VITE_PASSKEYS_ENABLED=true
|
VITE_PASSKEYS_ENABLED=true
|
||||||
|
VITE_WEBAUTHN_SERVER_URL=http://localhost:3002
|
||||||
VITE_BVC_MEETUPS_PROJECT_CLAIM_ID=https://endorser.ch/entity/01HWE8FWHQ1YGP7GFZYYPS272F
|
VITE_BVC_MEETUPS_PROJECT_CLAIM_ID=https://endorser.ch/entity/01HWE8FWHQ1YGP7GFZYYPS272F
|
||||||
```
|
```
|
||||||
|
|
||||||
@@ -1547,6 +1592,9 @@ VITE_DEFAULT_ENDORSER_API_SERVER=http://localhost:3000
|
|||||||
VITE_DEFAULT_PARTNER_API_SERVER=http://localhost:3000
|
VITE_DEFAULT_PARTNER_API_SERVER=http://localhost:3000
|
||||||
VITE_DEFAULT_IMAGE_API_SERVER=https://test-image-api.timesafari.app
|
VITE_DEFAULT_IMAGE_API_SERVER=https://test-image-api.timesafari.app
|
||||||
VITE_APP_SERVER=http://localhost:8080
|
VITE_APP_SERVER=http://localhost:8080
|
||||||
|
|
||||||
|
# WebAuthn Server (for passkey verification)
|
||||||
|
VITE_WEBAUTHN_SERVER_URL=http://localhost:3002
|
||||||
```
|
```
|
||||||
|
|
||||||
**Test Environment** (`.env.test`):
|
**Test Environment** (`.env.test`):
|
||||||
@@ -1724,14 +1772,12 @@ npx prettier --write ./sw_scripts/
|
|||||||
The `prebuild` script automatically runs before any build:
|
The `prebuild` script automatically runs before any build:
|
||||||
|
|
||||||
```json
|
```json
|
||||||
"prebuild": "eslint --ext .js,.ts,.vue --ignore-path .gitignore src && node sw_combine.js && node scripts/copy-wasm.js"
|
"prebuild": "eslint --ext .js,.ts,.vue --ignore-path .gitignore src && node scripts/copy-wasm.js"
|
||||||
```
|
```
|
||||||
|
|
||||||
**What happens automatically:**
|
**What happens automatically:**
|
||||||
|
|
||||||
- **ESLint**: Checks and fixes code formatting in `src/`
|
- **ESLint**: Checks and fixes code formatting in `src/`
|
||||||
- **Script Combination**: `sw_combine.js` combines all `sw_scripts/*.js` files
|
|
||||||
into `sw_scripts-combined.js`
|
|
||||||
- **WASM Copy**: `copy-wasm.js` copies SQLite WASM files to `public/wasm/`
|
- **WASM Copy**: `copy-wasm.js` copies SQLite WASM files to `public/wasm/`
|
||||||
|
|
||||||
#### Build Process Architecture
|
#### Build Process Architecture
|
||||||
@@ -1739,10 +1785,10 @@ The `prebuild` script automatically runs before any build:
|
|||||||
**Web Build Process:**
|
**Web Build Process:**
|
||||||
|
|
||||||
```text
|
```text
|
||||||
1. Pre-Build: ESLint + Script Combination + WASM Copy
|
1. Pre-Build: ESLint + WASM Copy
|
||||||
2. Environment Setup: Load .env files, set NODE_ENV
|
2. Environment Setup: Load .env files, set NODE_ENV
|
||||||
3. Vite Build: Bundle web assets with PWA support
|
3. Vite Build: Bundle web assets with PWA support
|
||||||
4. Service Worker: Inject combined scripts into PWA
|
4. Service Worker: Inject service worker scripts into PWA
|
||||||
5. Output: Production-ready files in dist/
|
5. Output: Production-ready files in dist/
|
||||||
```
|
```
|
||||||
|
|
||||||
@@ -1770,10 +1816,8 @@ The `prebuild` script automatically runs before any build:
|
|||||||
|
|
||||||
**Script Organization:**
|
**Script Organization:**
|
||||||
|
|
||||||
- `sw_scripts/` - Individual third-party scripts
|
- `sw_scripts/` - Individual third-party scripts for service worker
|
||||||
- `sw_combine.js` - Combines scripts into single file
|
- `vite.config.utils.mts` - PWA configuration
|
||||||
- `sw_scripts-combined.js` - Combined service worker (317KB, 10K+ lines)
|
|
||||||
- `vite.config.utils.mts` - PWA configuration using combined script
|
|
||||||
|
|
||||||
**PWA Integration:**
|
**PWA Integration:**
|
||||||
|
|
||||||
@@ -1781,18 +1825,16 @@ The `prebuild` script automatically runs before any build:
|
|||||||
// vite.config.utils.mts
|
// vite.config.utils.mts
|
||||||
pwaConfig: {
|
pwaConfig: {
|
||||||
strategies: "injectManifest",
|
strategies: "injectManifest",
|
||||||
filename: "sw_scripts-combined.js", // Uses our combined script
|
filename: "sw_scripts-combined.js", // Service worker file
|
||||||
// ... manifest configuration
|
// ... manifest configuration
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
**What Gets Combined:**
|
**Service Worker Scripts:**
|
||||||
|
|
||||||
- `nacl.js` - NaCl cryptographic library
|
- `nacl.js` - NaCl cryptographic library
|
||||||
- `noble-curves.js` - Elliptic curve cryptography (177KB)
|
- `noble-curves.js` - Elliptic curve cryptography (177KB)
|
||||||
- `noble-hashes.js` - Cryptographic hash functions (91KB)
|
- `noble-hashes.js` - Cryptographic hash functions (91KB)
|
||||||
- `safari-notifications.js` - Safari-specific notifications
|
|
||||||
- `additional-scripts.js` - Additional service worker functionality
|
|
||||||
|
|
||||||
#### Process Environment Configuration
|
#### Process Environment Configuration
|
||||||
|
|
||||||
@@ -1828,6 +1870,7 @@ VITE_APP_SERVER=https://timesafari.app
|
|||||||
|
|
||||||
# Feature Flags
|
# Feature Flags
|
||||||
VITE_PASSKEYS_ENABLED=true
|
VITE_PASSKEYS_ENABLED=true
|
||||||
|
VITE_WEBAUTHN_SERVER_URL=http://localhost:3002
|
||||||
VITE_BVC_MEETUPS_PROJECT_CLAIM_ID=https://endorser.ch/entity/01HWE8FWHQ1YGP7GFZYYPS272F
|
VITE_BVC_MEETUPS_PROJECT_CLAIM_ID=https://endorser.ch/entity/01HWE8FWHQ1YGP7GFZYYPS272F
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|||||||
59
README.md
59
README.md
@@ -89,6 +89,65 @@ VITE_LOG_LEVEL=debug npm run build:web:dev
|
|||||||
|
|
||||||
See [Logging Configuration Guide](doc/logging-configuration.md) for complete details.
|
See [Logging Configuration Guide](doc/logging-configuration.md) for complete details.
|
||||||
|
|
||||||
|
## WebAuthn Verification Server
|
||||||
|
|
||||||
|
TimeSafari includes a server-side WebAuthn verification service for secure passkey registration and authentication.
|
||||||
|
|
||||||
|
### Why a Separate Server?
|
||||||
|
|
||||||
|
WebAuthn verification **must** be performed server-side for security. Client-side verification can be tampered with and should never be trusted. The server:
|
||||||
|
|
||||||
|
- Verifies attestation signatures during registration
|
||||||
|
- Validates authentication signatures during login
|
||||||
|
- Prevents replay attacks by tracking counters
|
||||||
|
- Stores credentials securely with proper user binding
|
||||||
|
- Enforces origin and RP ID validation
|
||||||
|
|
||||||
|
**Note**: The client includes an optional "offline mode" for development (`VITE_OFFLINE_WEBAUTHN_VERIFY=true`), but this is not recommended for production as it compromises security.
|
||||||
|
|
||||||
|
### Quick Start
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Navigate to server directory
|
||||||
|
cd server
|
||||||
|
|
||||||
|
# Install dependencies
|
||||||
|
npm install
|
||||||
|
|
||||||
|
# Copy environment template
|
||||||
|
cp .env.example .env
|
||||||
|
|
||||||
|
# Edit .env with your configuration
|
||||||
|
# RP_ID=your-domain.com
|
||||||
|
# RP_NAME=Time Safari
|
||||||
|
# RP_ORIGIN=https://your-app-url.com
|
||||||
|
|
||||||
|
# Start development server
|
||||||
|
npm run dev
|
||||||
|
```
|
||||||
|
|
||||||
|
The server runs on `http://localhost:3002` by default (configurable via `PORT` in `.env`).
|
||||||
|
|
||||||
|
### Documentation
|
||||||
|
|
||||||
|
See [server/README.md](server/README.md) for:
|
||||||
|
- Complete API documentation
|
||||||
|
- Endpoint specifications
|
||||||
|
- Production deployment guide
|
||||||
|
- Security considerations
|
||||||
|
|
||||||
|
### Client Configuration
|
||||||
|
|
||||||
|
The client automatically uses the server when `VITE_OFFLINE_WEBAUTHN_VERIFY` is not set to `true`. Configure the server URL via:
|
||||||
|
|
||||||
|
- Environment variable: `VITE_WEBAUTHN_SERVER_URL`
|
||||||
|
- Defaults to `http://localhost:3002` in development
|
||||||
|
- Defaults to same origin in production
|
||||||
|
|
||||||
|
### Development Database Clearing
|
||||||
|
|
||||||
|
TimeSafari provides a simple script-based approach to clear the local database (not the claim server) for development purposes.
|
||||||
|
|
||||||
### Quick Usage
|
### Quick Usage
|
||||||
```bash
|
```bash
|
||||||
# Run the database clearing script
|
# Run the database clearing script
|
||||||
|
|||||||
@@ -105,8 +105,7 @@ Build Scripts:
|
|||||||
├── electron/** # Electron build files
|
├── electron/** # Electron build files
|
||||||
├── android/** # Android build configuration
|
├── android/** # Android build configuration
|
||||||
├── ios/** # iOS build configuration
|
├── ios/** # iOS build configuration
|
||||||
├── sw_scripts/** # Service worker scripts
|
└── sw_scripts/** # Service worker scripts
|
||||||
└── sw_combine.js # Service worker combination
|
|
||||||
|
|
||||||
Deployment:
|
Deployment:
|
||||||
├── Dockerfile # Docker configuration
|
├── Dockerfile # Docker configuration
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ import electronIsDev from 'electron-is-dev';
|
|||||||
import unhandled from 'electron-unhandled';
|
import unhandled from 'electron-unhandled';
|
||||||
// import { autoUpdater } from 'electron-updater';
|
// import { autoUpdater } from 'electron-updater';
|
||||||
import { promises as fs } from 'fs';
|
import { promises as fs } from 'fs';
|
||||||
import { join } from 'path';
|
import { join, basename } from 'path';
|
||||||
|
|
||||||
import { ElectronCapacitorApp, setupContentSecurityPolicy, setupReloadWatcher } from './setup';
|
import { ElectronCapacitorApp, setupContentSecurityPolicy, setupReloadWatcher } from './setup';
|
||||||
|
|
||||||
@@ -151,15 +151,47 @@ app.on('activate', async function () {
|
|||||||
* This provides a secure, native way to save files directly to the Downloads
|
* This provides a secure, native way to save files directly to the Downloads
|
||||||
* directory using the main process's file system access.
|
* directory using the main process's file system access.
|
||||||
*
|
*
|
||||||
|
* Security: File names are sanitized to prevent path traversal attacks.
|
||||||
|
* Only safe file extensions are allowed (.json, .txt, .csv, .md).
|
||||||
|
*
|
||||||
* @param fileName - The name of the file to save (including extension)
|
* @param fileName - The name of the file to save (including extension)
|
||||||
* @param data - The data to write to the file (string or buffer)
|
* @param data - The data to write to the file (string or buffer)
|
||||||
* @returns Promise<{success: boolean, path?: string, error?: string}>
|
* @returns Promise<{success: boolean, path?: string, error?: string}>
|
||||||
*/
|
*/
|
||||||
ipcMain.handle('export-data-to-downloads', async (_event, fileName: string, data: string) => {
|
ipcMain.handle('export-data-to-downloads', async (_event, fileName: string, data: string) => {
|
||||||
try {
|
try {
|
||||||
|
// Security: Sanitize file name to prevent path traversal
|
||||||
|
// 1. Extract only the basename (removes any directory components)
|
||||||
|
const sanitizedBaseName = basename(fileName);
|
||||||
|
|
||||||
|
// 2. Reject if still contains path separators (shouldn't happen after basename, but double-check)
|
||||||
|
if (sanitizedBaseName.includes('/') || sanitizedBaseName.includes('\\')) {
|
||||||
|
throw new Error('Invalid file name: path separators not allowed');
|
||||||
|
}
|
||||||
|
|
||||||
|
// 3. Enforce allowed file extensions for security
|
||||||
|
const allowedExtensions = ['.json', '.txt', '.csv', '.md', '.log'];
|
||||||
|
const hasAllowedExtension = allowedExtensions.some(ext =>
|
||||||
|
sanitizedBaseName.toLowerCase().endsWith(ext.toLowerCase())
|
||||||
|
);
|
||||||
|
|
||||||
|
if (!hasAllowedExtension) {
|
||||||
|
throw new Error(`Invalid file extension. Allowed: ${allowedExtensions.join(', ')}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// 4. Additional validation: reject empty or suspicious names
|
||||||
|
if (!sanitizedBaseName || sanitizedBaseName.trim().length === 0) {
|
||||||
|
throw new Error('File name cannot be empty');
|
||||||
|
}
|
||||||
|
|
||||||
|
// 5. Reject names that are too long (prevent potential filesystem issues)
|
||||||
|
if (sanitizedBaseName.length > 255) {
|
||||||
|
throw new Error('File name too long (max 255 characters)');
|
||||||
|
}
|
||||||
|
|
||||||
// Get the user's Downloads directory path
|
// Get the user's Downloads directory path
|
||||||
const downloadsDir = app.getPath('downloads');
|
const downloadsDir = app.getPath('downloads');
|
||||||
const filePath = join(downloadsDir, fileName);
|
const filePath = join(downloadsDir, sanitizedBaseName);
|
||||||
|
|
||||||
// Write the file to the Downloads directory
|
// Write the file to the Downloads directory
|
||||||
await fs.writeFile(filePath, data, 'utf-8');
|
await fs.writeFile(filePath, data, 'utf-8');
|
||||||
|
|||||||
@@ -218,17 +218,50 @@ export class ElectronCapacitorApp {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set a CSP up for our application based on the custom scheme
|
/**
|
||||||
|
* Set up Content Security Policy for Electron application
|
||||||
|
*
|
||||||
|
* CSP is assembled from structured directives to prevent truncation/corruption.
|
||||||
|
* This ensures the CSP string is always complete and valid.
|
||||||
|
*
|
||||||
|
* @param customScheme - The custom URL scheme for the Electron app (e.g., 'capacitor-electron')
|
||||||
|
*/
|
||||||
export function setupContentSecurityPolicy(customScheme: string): void {
|
export function setupContentSecurityPolicy(customScheme: string): void {
|
||||||
|
// Build CSP from structured directives to prevent truncation issues
|
||||||
|
const buildCSP = (isDev: boolean): string => {
|
||||||
|
const directives: string[] = [];
|
||||||
|
|
||||||
|
// Default source: allow custom scheme, inline scripts (required for some libs), and data URIs
|
||||||
|
const defaultSrc = [
|
||||||
|
`${customScheme}://*`,
|
||||||
|
"'unsafe-inline'",
|
||||||
|
"data:",
|
||||||
|
"https:",
|
||||||
|
];
|
||||||
|
|
||||||
|
if (isDev) {
|
||||||
|
// Development: allow devtools and eval for debugging
|
||||||
|
defaultSrc.push("devtools://*", "'unsafe-eval'", "http:");
|
||||||
|
}
|
||||||
|
|
||||||
|
directives.push(`default-src ${defaultSrc.join(" ")}`);
|
||||||
|
|
||||||
|
// Style source: allow custom scheme and inline styles
|
||||||
|
directives.push(`style-src ${customScheme}://* 'unsafe-inline'`);
|
||||||
|
|
||||||
|
// Font source: allow custom scheme and data URIs
|
||||||
|
directives.push(`font-src ${customScheme}://* data:`);
|
||||||
|
|
||||||
|
return directives.join("; ");
|
||||||
|
};
|
||||||
|
|
||||||
session.defaultSession.webRequest.onHeadersReceived((details, callback) => {
|
session.defaultSession.webRequest.onHeadersReceived((details, callback) => {
|
||||||
|
const csp = buildCSP(electronIsDev);
|
||||||
|
|
||||||
callback({
|
callback({
|
||||||
responseHeaders: {
|
responseHeaders: {
|
||||||
...details.responseHeaders,
|
...details.responseHeaders,
|
||||||
'Content-Security-Policy': [
|
'Content-Security-Policy': [csp],
|
||||||
electronIsDev
|
|
||||||
? `default-src ${customScheme}://* 'unsafe-inline' devtools://* 'unsafe-eval' data: https: http:; style-src ${customScheme}://* 'unsafe-inline'; font-src ${customScheme}://* data:`
|
|
||||||
: `default-src ${customScheme}://* 'unsafe-inline' data: https:; style-src ${customScheme}://* 'unsafe-inline'; font-src ${customScheme}://* data:`,
|
|
||||||
],
|
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -10,10 +10,10 @@
|
|||||||
"lint-fix": "eslint --ext .js,.ts,.vue --ignore-path .gitignore --fix src",
|
"lint-fix": "eslint --ext .js,.ts,.vue --ignore-path .gitignore --fix src",
|
||||||
"type-safety-check": "./scripts/type-safety-check.sh",
|
"type-safety-check": "./scripts/type-safety-check.sh",
|
||||||
"type-check": "tsc --noEmit",
|
"type-check": "tsc --noEmit",
|
||||||
"prebuild": "eslint --ext .js,.ts,.vue --ignore-path .gitignore src && node sw_combine.js && node scripts/copy-wasm.js",
|
"prebuild": "eslint --ext .js,.ts,.vue --ignore-path .gitignore src && node scripts/copy-wasm.js",
|
||||||
"test:prerequisites": "node scripts/check-prerequisites.js",
|
"test:prerequisites": "node scripts/check-prerequisites.js",
|
||||||
"check:dependencies": "./scripts/check-dependencies.sh",
|
"check:dependencies": "./scripts/check-dependencies.sh",
|
||||||
"test:all": "npm run lint && tsc && npm run test:web && npm run test:mobile && ./scripts/test-safety-check.sh && echo '\n\n\nGotta add the performance tests'",
|
"test:all": "npm run lint && tsc && npm run test:web && npm run test:mobile && echo '\n\n\nGotta add the performance tests'",
|
||||||
"test:web": "npx playwright test -c playwright.config-local.ts --trace on",
|
"test:web": "npx playwright test -c playwright.config-local.ts --trace on",
|
||||||
"test:mobile": "./scripts/test-mobile.sh",
|
"test:mobile": "./scripts/test-mobile.sh",
|
||||||
"test:android": "node scripts/test-android.js",
|
"test:android": "node scripts/test-android.js",
|
||||||
@@ -64,6 +64,7 @@
|
|||||||
"build:web:serve:test": "./scripts/build-web.sh --serve --test",
|
"build:web:serve:test": "./scripts/build-web.sh --serve --test",
|
||||||
"build:web:serve:prod": "./scripts/build-web.sh --serve --prod",
|
"build:web:serve:prod": "./scripts/build-web.sh --serve --prod",
|
||||||
"docker:up": "docker-compose up",
|
"docker:up": "docker-compose up",
|
||||||
|
"build:web:build": "./scripts/build-web.sh",
|
||||||
"docker:up:test": "npm run build:web:build -- --mode test && docker-compose up test",
|
"docker:up:test": "npm run build:web:build -- --mode test && docker-compose up test",
|
||||||
"docker:up:prod": "npm run build:web:build -- --mode production && docker-compose up production",
|
"docker:up:prod": "npm run build:web:build -- --mode production && docker-compose up production",
|
||||||
"docker:down": "docker-compose down",
|
"docker:down": "docker-compose down",
|
||||||
|
|||||||
@@ -24,7 +24,6 @@ SENSITIVE=(
|
|||||||
"android/**"
|
"android/**"
|
||||||
"ios/**"
|
"ios/**"
|
||||||
"sw_scripts/**"
|
"sw_scripts/**"
|
||||||
"sw_combine.js"
|
|
||||||
"Dockerfile"
|
"Dockerfile"
|
||||||
"docker/**"
|
"docker/**"
|
||||||
"capacitor.config.ts"
|
"capacitor.config.ts"
|
||||||
|
|||||||
@@ -1,6 +1,11 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
#
|
||||||
|
# Critical Files Migration Validator
|
||||||
|
# Author: Matthew Raymer
|
||||||
|
# Description: Validates migration status of critical files
|
||||||
|
#
|
||||||
|
|
||||||
echo 🔍 Critical Files Migration Validator"
|
echo "🔍 Critical Files Migration Validator"
|
||||||
echo "====================================="
|
echo "====================================="
|
||||||
|
|
||||||
# Function to check actual usage (not comments)
|
# Function to check actual usage (not comments)
|
||||||
@@ -10,77 +15,87 @@ check_actual_usage() {
|
|||||||
local description="$3"
|
local description="$3"
|
||||||
|
|
||||||
# Remove comments and check for actual usage
|
# Remove comments and check for actual usage
|
||||||
local count=$(grep -v ^[[:space:]]*//\|^[[:space:]]*\*\|^[[:space:]]*<!--" "$file" | \
|
local count=$(grep -v "^[[:space:]]*//\|^[[:space:]]*\*\|^[[:space:]]*<!--" "$file" | \
|
||||||
grep -v TODO.*migration\|FIXME.*migration" | \
|
grep -v "TODO.*migration\|FIXME.*migration" | \
|
||||||
grep -v "Migration.*replaced\|migrated.*from" | \
|
grep -v "Migration.*replaced\|migrated.*from" | \
|
||||||
grep -c $pattern" || echo 0)
|
grep -c "$pattern" || echo 0)
|
||||||
|
|
||||||
if [$count" -gt0 then
|
if [ "$count" -gt 0 ]; then
|
||||||
echo ❌ $description: $count instances
|
echo " ❌ $description: $count instances"
|
||||||
return 1 else
|
return 1
|
||||||
echo ✅$description: None found
|
else
|
||||||
|
echo " ✅ $description: None found"
|
||||||
return 0
|
return 0
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
# Function to check notification migration
|
# Function to check notification migration
|
||||||
check_notifications() {
|
check_notifications() {
|
||||||
local file="$1
|
local file="$1"
|
||||||
# Check for notification helpers
|
# Check for notification helpers
|
||||||
local has_helpers=$(grep -c "createNotifyHelpers" $file" || echo "0")
|
local has_helpers=$(grep -c "createNotifyHelpers" "$file" || echo "0")
|
||||||
|
|
||||||
# Check for direct $notify calls (excluding helper setup)
|
# Check for direct $notify calls (excluding helper setup)
|
||||||
local direct_notify=$(grep -v "createNotifyHelpers" "$file" | \
|
local direct_notify=$(grep -v "createNotifyHelpers" "$file" | \
|
||||||
grep -v this\.notify\." | \
|
grep -v "this\.notify\." | \
|
||||||
grep -c "this\.\$notify" || echo 0)
|
grep -c "this\.\$notify" || echo 0)
|
||||||
|
|
||||||
if $has_helpers" -gt0 && $direct_notify" -eq0 then
|
if [ "$has_helpers" -gt 0 ] && [ "$direct_notify" -eq 0 ]; then
|
||||||
echo " ✅ Complete notification migration
|
echo " ✅ Complete notification migration"
|
||||||
return 0
|
return 0
|
||||||
elif $has_helpers" -gt0 && $direct_notify" -gt0 then
|
elif [ "$has_helpers" -gt 0 ] && [ "$direct_notify" -gt 0 ]; then
|
||||||
echo " ⚠️ Mixed pattern: $direct_notify direct calls
|
echo " ⚠️ Mixed pattern: $direct_notify direct calls"
|
||||||
return 1 else
|
return 1
|
||||||
echo " ❌ No notification migration
|
else
|
||||||
|
echo " ❌ No notification migration"
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
# Function to analyze a file
|
# Function to analyze a file
|
||||||
analyze_file() {
|
analyze_file() {
|
||||||
local file="$1 echo ""
|
local file="$1"
|
||||||
|
echo ""
|
||||||
echo "📄 Analyzing: $file"
|
echo "📄 Analyzing: $file"
|
||||||
echo "----------------------------------------"
|
echo "----------------------------------------"
|
||||||
|
|
||||||
local issues=0 # Check legacy patterns
|
local issues=0
|
||||||
echo "🔍 Legacy Patterns:
|
# Check legacy patterns
|
||||||
check_actual_usage$file aseUtil" "databaseUtil usage || ((issues++))
|
echo "🔍 Legacy Patterns:"
|
||||||
check_actual_usage "$filelogConsoleAndDb ConsoleAndDb usage || ((issues++))
|
check_actual_usage "$file" "databaseUtil" "databaseUtil usage" || ((issues++))
|
||||||
check_actual_usage$file formServiceFactory\.getInstance ct PlatformService usage ||((issues++))
|
check_actual_usage "$file" "logConsoleAndDb" "ConsoleAndDb usage" || ((issues++))
|
||||||
|
check_actual_usage "$file" "formServiceFactory\.getInstance" "PlatformService usage" || ((issues++))
|
||||||
|
|
||||||
# Check notifications
|
# Check notifications
|
||||||
echo 🔔 Notifications:"
|
echo "🔔 Notifications:"
|
||||||
check_notifications "$file ||((issues++))
|
check_notifications "$file" || ((issues++))
|
||||||
|
|
||||||
# Check PlatformServiceMixin
|
# Check PlatformServiceMixin
|
||||||
echo "🔧 PlatformServiceMixin:"
|
echo "🔧 PlatformServiceMixin:"
|
||||||
local has_mixin=$(grep -cPlatformServiceMixin" $file || echo 0)
|
local has_mixin=$(grep -c "PlatformServiceMixin" "$file" || echo 0)
|
||||||
local has_mixins=$(grep -cmixins.*PlatformServiceMixin\|mixins.*\[PlatformServiceMixin" $file" || echo 0)
|
local has_mixins=$(grep -c "mixins.*PlatformServiceMixin\|mixins.*\[PlatformServiceMixin" "$file" || echo 0)
|
||||||
|
|
||||||
if $has_mixin" -gt 0 && $has_mixins" -gt0 then
|
if [ "$has_mixin" -gt 0 ] && [ "$has_mixins" -gt 0 ]; then
|
||||||
echo " ✅ PlatformServiceMixin properly integrated elif $has_mixin" -gt 0 && $has_mixins" -eq0 then
|
echo " ✅ PlatformServiceMixin properly integrated"
|
||||||
echo " ⚠️ Imported but not used as mixin ((issues++))
|
elif [ "$has_mixin" -gt 0 ] && [ "$has_mixins" -eq 0 ]; then
|
||||||
|
echo " ⚠️ Imported but not used as mixin"
|
||||||
|
((issues++))
|
||||||
else
|
else
|
||||||
echo " ❌ No PlatformServiceMixin usage ((issues++))
|
echo " ❌ No PlatformServiceMixin usage"
|
||||||
|
((issues++))
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Check TODO comments
|
# Check TODO comments
|
||||||
local todo_count=$(grep -c TODO.*migration\|FIXME.*migration" $file || echo "0) if $todo_count" -gt0 then
|
local todo_count=$(grep -c "TODO.*migration\|FIXME.*migration" "$file" || echo "0")
|
||||||
echo ⚠️ TODO/FIXME comments: $todo_count ((issues++))
|
if [ "$todo_count" -gt 0 ]; then
|
||||||
|
echo " ⚠️ TODO/FIXME comments: $todo_count"
|
||||||
|
((issues++))
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if$issues" -eq0 then
|
if [ "$issues" -eq 0 ]; then
|
||||||
echo "✅ File is fully migrated else
|
echo "✅ File is fully migrated"
|
||||||
echo❌ $issues issues found"
|
else
|
||||||
|
echo "❌ $issues issues found"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
return $issues
|
return $issues
|
||||||
@@ -88,35 +103,39 @@ analyze_file() {
|
|||||||
|
|
||||||
# Main analysis
|
# Main analysis
|
||||||
echo ""
|
echo ""
|
||||||
echo 📊 Critical Files Analysis"
|
echo "📊 Critical Files Analysis"
|
||||||
echo "=========================="
|
echo "=========================="
|
||||||
|
|
||||||
# Critical files from our assessment
|
# Critical files from our assessment
|
||||||
files=(
|
files=(
|
||||||
src/components/MembersList.vue"
|
"src/components/MembersList.vue"
|
||||||
"src/views/ContactsView.vue"
|
"src/views/ContactsView.vue"
|
||||||
src/views/OnboardMeetingSetupView.vue"
|
"src/views/OnboardMeetingSetupView.vue"
|
||||||
src/db/databaseUtil.ts"
|
"src/db/databaseUtil.ts"
|
||||||
src/db/index.ts
|
"src/db/index.ts"
|
||||||
)
|
)
|
||||||
|
|
||||||
total_issues=0
|
total_issues=0
|
||||||
for file in ${files[@]}"; do
|
for file in "${files[@]}"; do
|
||||||
if [ -f "$file" ]; then
|
if [ -f "$file" ]; then
|
||||||
analyze_file "$file"
|
analyze_file "$file"
|
||||||
total_issues=$((total_issues + $?))
|
total_issues=$((total_issues + $?))
|
||||||
else
|
else
|
||||||
echo ❌ File not found: $file"
|
echo "❌ File not found: $file"
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
|
|
||||||
# Summary
|
# Summary
|
||||||
echo "echo📋 Summary"
|
echo ""
|
||||||
echo=========="
|
echo "📋 Summary"
|
||||||
|
echo "=========="
|
||||||
echo "Files analyzed: ${#files[@]}"
|
echo "Files analyzed: ${#files[@]}"
|
||||||
echo "Total issues found: $total_issues"
|
echo "Total issues found: $total_issues"
|
||||||
|
|
||||||
if$total_issues" -eq 0]; then
|
if [ "$total_issues" -eq 0 ]; then
|
||||||
echo "✅ All critical files are properly migrated exit 0 echo "❌ Migration issues require attention"
|
echo "✅ All critical files are properly migrated"
|
||||||
|
exit 0
|
||||||
|
else
|
||||||
|
echo "❌ Migration issues require attention"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|||||||
11
server/.env
Normal file
11
server/.env
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
# Relying Party Configuration
|
||||||
|
RP_ID=localhost
|
||||||
|
RP_NAME=Time Safari
|
||||||
|
RP_ORIGIN=http://localhost:8080
|
||||||
|
|
||||||
|
# Server Configuration
|
||||||
|
PORT=3002
|
||||||
|
HOST=0.0.0.0
|
||||||
|
|
||||||
|
# CORS (optional, defaults to RP_ORIGIN)
|
||||||
|
# CORS_ORIGIN=http://localhost:8080
|
||||||
11
server/.env.example
Normal file
11
server/.env.example
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
# Relying Party Configuration
|
||||||
|
RP_ID=localhost
|
||||||
|
RP_NAME=Time Safari
|
||||||
|
RP_ORIGIN=http://localhost:8080
|
||||||
|
|
||||||
|
# Server Configuration
|
||||||
|
PORT=3002
|
||||||
|
HOST=0.0.0.0
|
||||||
|
|
||||||
|
# CORS (optional, defaults to RP_ORIGIN)
|
||||||
|
# CORS_ORIGIN=http://localhost:8080
|
||||||
197
server/README.md
Normal file
197
server/README.md
Normal file
@@ -0,0 +1,197 @@
|
|||||||
|
# WebAuthn Verification Server
|
||||||
|
|
||||||
|
Server-side WebAuthn verification service for Time Safari.
|
||||||
|
|
||||||
|
## Why This Server Exists
|
||||||
|
|
||||||
|
WebAuthn verification **must** be performed server-side for security. Client-side verification can be tampered with and should never be trusted for security-critical operations.
|
||||||
|
|
||||||
|
### Security Rationale
|
||||||
|
|
||||||
|
1. **Trust Boundary**: The client bundle runs in an untrusted environment (user's browser). Any verification code in the client can be modified, bypassed, or replaced by an attacker.
|
||||||
|
|
||||||
|
2. **Attestation Verification**: During registration, the server must verify:
|
||||||
|
- The attestation signature is valid
|
||||||
|
- The authenticator is genuine (not a software emulator)
|
||||||
|
- The challenge matches what was issued
|
||||||
|
- The origin and RP ID are correct
|
||||||
|
|
||||||
|
3. **Authentication Verification**: During authentication, the server must verify:
|
||||||
|
- The signature is valid for the stored credential
|
||||||
|
- The challenge matches
|
||||||
|
- The counter has increased (replay attack prevention)
|
||||||
|
- The origin and RP ID are correct
|
||||||
|
|
||||||
|
4. **Credential Storage**: Credentials must be stored securely server-side with proper user binding to prevent unauthorized access.
|
||||||
|
|
||||||
|
### Offline Mode
|
||||||
|
|
||||||
|
The client includes an optional "offline mode" (`VITE_OFFLINE_WEBAUTHN_VERIFY=true`) that allows client-side verification, but this is:
|
||||||
|
- **Not recommended for production** - security can be compromised
|
||||||
|
- **Intended for development/testing** - when a server isn't available
|
||||||
|
- **Clearly documented** - with security warnings
|
||||||
|
|
||||||
|
### Architecture
|
||||||
|
|
||||||
|
```
|
||||||
|
┌─────────────┐ ┌──────────────┐ ┌─────────────┐
|
||||||
|
│ Client │────────▶│ WebAuthn │────────▶│ Authenticator
|
||||||
|
│ (Browser) │ │ Server │ │ (Passkey)
|
||||||
|
└─────────────┘ └──────────────┘ └─────────────┘
|
||||||
|
│ │
|
||||||
|
│ 1. Request options │
|
||||||
|
│◀─────────────────────────│
|
||||||
|
│ │
|
||||||
|
│ 2. Create credential │
|
||||||
|
│ (browser API) │
|
||||||
|
│ │
|
||||||
|
│ 3. Send attestation │
|
||||||
|
│────────────────────────▶│
|
||||||
|
│ │
|
||||||
|
│ 4. Verify & store │
|
||||||
|
│ (server-side only) │
|
||||||
|
│ │
|
||||||
|
│◀─────────────────────────│
|
||||||
|
│ 5. Return credential info│
|
||||||
|
```
|
||||||
|
|
||||||
|
The server acts as the **Relying Party (RP)** and performs all cryptographic verification that cannot be safely done client-side.
|
||||||
|
|
||||||
|
## Setup
|
||||||
|
|
||||||
|
1. Install dependencies:
|
||||||
|
```bash
|
||||||
|
npm install
|
||||||
|
```
|
||||||
|
|
||||||
|
2. Copy `.env.example` to `.env` and configure:
|
||||||
|
```bash
|
||||||
|
cp .env.example .env
|
||||||
|
```
|
||||||
|
|
||||||
|
3. Update `.env` with your Relying Party configuration:
|
||||||
|
```
|
||||||
|
RP_ID=your-domain.com
|
||||||
|
RP_NAME=Time Safari
|
||||||
|
RP_ORIGIN=https://your-app-url.com
|
||||||
|
```
|
||||||
|
|
||||||
|
## Development
|
||||||
|
|
||||||
|
Run in development mode with hot reload:
|
||||||
|
```bash
|
||||||
|
npm run dev
|
||||||
|
```
|
||||||
|
|
||||||
|
## Production
|
||||||
|
|
||||||
|
Build and run:
|
||||||
|
```bash
|
||||||
|
npm run build
|
||||||
|
npm start
|
||||||
|
```
|
||||||
|
|
||||||
|
## Endpoints
|
||||||
|
|
||||||
|
### POST /webauthn/registration/options
|
||||||
|
Generate registration options for a new passkey.
|
||||||
|
|
||||||
|
**Request:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"username": "User Name",
|
||||||
|
"userId": "optional-user-id"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Response:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"rp": { "name": "Time Safari", "id": "localhost" },
|
||||||
|
"user": { "id": "...", "name": "User Name", "displayName": "User Name" },
|
||||||
|
"challenge": "...",
|
||||||
|
"pubKeyCredParams": [...],
|
||||||
|
...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### POST /webauthn/registration/verify
|
||||||
|
Verify a registration response.
|
||||||
|
|
||||||
|
**Request:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"options": { ... },
|
||||||
|
"attestationResponse": { ... }
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Response:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"verified": true,
|
||||||
|
"credential": {
|
||||||
|
"credentialID": "...",
|
||||||
|
"credentialPublicKey": [...],
|
||||||
|
"counter": 0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### POST /webauthn/authentication/options
|
||||||
|
Generate authentication options.
|
||||||
|
|
||||||
|
**Request:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"credentialId": "...",
|
||||||
|
"userId": "optional-user-id"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Response:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"challenge": "...",
|
||||||
|
"rpId": "localhost",
|
||||||
|
"allowCredentials": [...],
|
||||||
|
...
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### POST /webauthn/authentication/verify
|
||||||
|
Verify an authentication response.
|
||||||
|
|
||||||
|
**Request:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"options": { ... },
|
||||||
|
"assertionResponse": { ... }
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Response:**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"verified": true,
|
||||||
|
"counter": 1
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## Storage
|
||||||
|
|
||||||
|
**Development:** Uses in-memory storage (challenges and credentials).
|
||||||
|
|
||||||
|
**Production:** Replace with:
|
||||||
|
- Redis for challenge storage
|
||||||
|
- Database for credential persistence
|
||||||
|
- Session management for user binding
|
||||||
|
|
||||||
|
## Security Notes
|
||||||
|
|
||||||
|
- Challenges expire after 5 minutes
|
||||||
|
- Credentials are stored in-memory (lost on restart)
|
||||||
|
- In production, implement proper credential persistence and user binding
|
||||||
|
- Use HTTPS in production
|
||||||
|
- Validate origin and RP ID strictly
|
||||||
|
|
||||||
1333
server/package-lock.json
generated
Normal file
1333
server/package-lock.json
generated
Normal file
File diff suppressed because it is too large
Load Diff
25
server/package.json
Normal file
25
server/package.json
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
{
|
||||||
|
"name": "timesafari-webauthn-server",
|
||||||
|
"version": "1.0.0",
|
||||||
|
"description": "WebAuthn verification server for Time Safari",
|
||||||
|
"type": "module",
|
||||||
|
"main": "dist/index.js",
|
||||||
|
"scripts": {
|
||||||
|
"dev": "tsx watch src/index.ts",
|
||||||
|
"build": "tsc",
|
||||||
|
"start": "node dist/index.js"
|
||||||
|
},
|
||||||
|
"dependencies": {
|
||||||
|
"@simplewebauthn/server": "^9.0.0",
|
||||||
|
"fastify": "^4.24.3",
|
||||||
|
"zod": "^3.22.4",
|
||||||
|
"@fastify/cors": "^8.4.0",
|
||||||
|
"dotenv": "^16.3.1"
|
||||||
|
},
|
||||||
|
"devDependencies": {
|
||||||
|
"@types/node": "^20.10.0",
|
||||||
|
"tsx": "^4.7.0",
|
||||||
|
"typescript": "^5.3.3"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
340
server/src/index.ts
Normal file
340
server/src/index.ts
Normal file
@@ -0,0 +1,340 @@
|
|||||||
|
/**
|
||||||
|
* WebAuthn Verification Server
|
||||||
|
*
|
||||||
|
* Fastify-based server for WebAuthn registration and authentication verification.
|
||||||
|
* This server handles the server-side verification of WebAuthn credentials.
|
||||||
|
*
|
||||||
|
* @author Matthew Raymer
|
||||||
|
*/
|
||||||
|
|
||||||
|
import Fastify from "fastify";
|
||||||
|
import cors from "@fastify/cors";
|
||||||
|
import dotenv from "dotenv";
|
||||||
|
import {
|
||||||
|
generateRegistrationOptions,
|
||||||
|
verifyRegistrationResponse,
|
||||||
|
generateAuthenticationOptions,
|
||||||
|
verifyAuthenticationResponse,
|
||||||
|
} from "@simplewebauthn/server";
|
||||||
|
import type {
|
||||||
|
PublicKeyCredentialCreationOptionsJSON,
|
||||||
|
PublicKeyCredentialRequestOptionsJSON,
|
||||||
|
VerifyRegistrationResponseOpts,
|
||||||
|
VerifyAuthenticationResponseOpts,
|
||||||
|
} from "@simplewebauthn/types";
|
||||||
|
|
||||||
|
// Load environment variables
|
||||||
|
dotenv.config();
|
||||||
|
|
||||||
|
const fastify = Fastify({
|
||||||
|
logger: true,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Register CORS
|
||||||
|
await fastify.register(cors, {
|
||||||
|
origin: process.env.RP_ORIGIN || "http://localhost:8080",
|
||||||
|
credentials: true,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Relying Party configuration from environment
|
||||||
|
const rpId = process.env.RP_ID || "localhost";
|
||||||
|
const rpName = process.env.RP_NAME || "Time Safari";
|
||||||
|
const rpOrigin = process.env.RP_ORIGIN || "http://localhost:8080";
|
||||||
|
|
||||||
|
// In-memory challenge storage (for development)
|
||||||
|
// In production, use Redis or a database
|
||||||
|
interface ChallengeStore {
|
||||||
|
challenge: string;
|
||||||
|
userId?: string;
|
||||||
|
expiresAt: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
const challengeStore = new Map<string, ChallengeStore>();
|
||||||
|
|
||||||
|
// Credential storage (in-memory for development)
|
||||||
|
// In production, use a database
|
||||||
|
interface StoredCredential {
|
||||||
|
credentialID: string;
|
||||||
|
credentialPublicKey: Uint8Array;
|
||||||
|
counter: number;
|
||||||
|
userId?: string;
|
||||||
|
createdAt: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
const credentialStore = new Map<string, StoredCredential>();
|
||||||
|
|
||||||
|
// Cleanup expired challenges every 5 minutes
|
||||||
|
setInterval(() => {
|
||||||
|
const now = Date.now();
|
||||||
|
for (const [key, value] of challengeStore.entries()) {
|
||||||
|
if (value.expiresAt < now) {
|
||||||
|
challengeStore.delete(key);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}, 5 * 60 * 1000);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /webauthn/registration/options
|
||||||
|
* Generate registration options for a new passkey
|
||||||
|
*/
|
||||||
|
fastify.post<{
|
||||||
|
Body: {
|
||||||
|
username?: string;
|
||||||
|
userId?: string;
|
||||||
|
};
|
||||||
|
}>("/webauthn/registration/options", async (request, reply) => {
|
||||||
|
try {
|
||||||
|
const { username, userId } = request.body;
|
||||||
|
|
||||||
|
const options = await generateRegistrationOptions({
|
||||||
|
rpName,
|
||||||
|
rpID: rpId,
|
||||||
|
userName: username || rpName + " User",
|
||||||
|
userID: userId || crypto.randomUUID(),
|
||||||
|
timeout: 60000,
|
||||||
|
attestationType: "none",
|
||||||
|
authenticatorSelection: {
|
||||||
|
residentKey: "preferred",
|
||||||
|
userVerification: "preferred",
|
||||||
|
authenticatorAttachment: "platform",
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// Store challenge for verification
|
||||||
|
const challengeKey = userId || options.user.id;
|
||||||
|
challengeStore.set(challengeKey, {
|
||||||
|
challenge: options.challenge,
|
||||||
|
userId: userId,
|
||||||
|
expiresAt: Date.now() + 5 * 60 * 1000, // 5 minutes
|
||||||
|
});
|
||||||
|
|
||||||
|
return options;
|
||||||
|
} catch (error) {
|
||||||
|
fastify.log.error(error);
|
||||||
|
reply.code(500).send({ error: "Failed to generate registration options" });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /webauthn/registration/verify
|
||||||
|
* Verify a registration response
|
||||||
|
*/
|
||||||
|
fastify.post<{
|
||||||
|
Body: {
|
||||||
|
options: PublicKeyCredentialCreationOptionsJSON;
|
||||||
|
attestationResponse: unknown;
|
||||||
|
};
|
||||||
|
}>("/webauthn/registration/verify", async (request, reply) => {
|
||||||
|
try {
|
||||||
|
const { options, attestationResponse } = request.body;
|
||||||
|
|
||||||
|
// Retrieve stored challenge
|
||||||
|
const challengeKey = options.user.id;
|
||||||
|
const storedChallenge = challengeStore.get(challengeKey);
|
||||||
|
|
||||||
|
if (!storedChallenge) {
|
||||||
|
reply.code(400).send({ error: "Challenge not found or expired" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (storedChallenge.expiresAt < Date.now()) {
|
||||||
|
challengeStore.delete(challengeKey);
|
||||||
|
reply.code(400).send({ error: "Challenge expired" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verify registration response
|
||||||
|
const verification = await verifyRegistrationResponse({
|
||||||
|
response: attestationResponse as any,
|
||||||
|
expectedChallenge: storedChallenge.challenge,
|
||||||
|
expectedOrigin: rpOrigin,
|
||||||
|
expectedRPID: rpId,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Clean up challenge
|
||||||
|
challengeStore.delete(challengeKey);
|
||||||
|
|
||||||
|
if (!verification.verified || !verification.registrationInfo) {
|
||||||
|
reply.code(400).send({ verified: false, error: "Verification failed" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Store credential
|
||||||
|
const credentialID = verification.registrationInfo.credentialID;
|
||||||
|
credentialStore.set(credentialID, {
|
||||||
|
credentialID: credentialID,
|
||||||
|
credentialPublicKey: verification.registrationInfo.credentialPublicKey,
|
||||||
|
counter: verification.registrationInfo.counter,
|
||||||
|
userId: storedChallenge.userId,
|
||||||
|
createdAt: Date.now(),
|
||||||
|
});
|
||||||
|
|
||||||
|
return {
|
||||||
|
verified: true,
|
||||||
|
credential: {
|
||||||
|
credentialID: credentialID,
|
||||||
|
credentialPublicKey: Array.from(verification.registrationInfo.credentialPublicKey),
|
||||||
|
counter: verification.registrationInfo.counter,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
} catch (error) {
|
||||||
|
fastify.log.error(error);
|
||||||
|
reply.code(500).send({ error: "Verification failed", details: String(error) });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /webauthn/authentication/options
|
||||||
|
* Generate authentication options for an existing passkey
|
||||||
|
*/
|
||||||
|
fastify.post<{
|
||||||
|
Body: {
|
||||||
|
credentialId?: string;
|
||||||
|
userId?: string;
|
||||||
|
};
|
||||||
|
}>("/webauthn/authentication/options", async (request, reply) => {
|
||||||
|
try {
|
||||||
|
const { credentialId, userId } = request.body;
|
||||||
|
|
||||||
|
// Find credential(s) for user
|
||||||
|
let credentials: StoredCredential[] = [];
|
||||||
|
if (credentialId) {
|
||||||
|
const cred = credentialStore.get(credentialId);
|
||||||
|
if (cred) {
|
||||||
|
credentials = [cred];
|
||||||
|
}
|
||||||
|
} else if (userId) {
|
||||||
|
credentials = Array.from(credentialStore.values()).filter(
|
||||||
|
(c) => c.userId === userId
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
reply.code(400).send({ error: "credentialId or userId required" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (credentials.length === 0) {
|
||||||
|
reply.code(404).send({ error: "Credential not found" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
const options = await generateAuthenticationOptions({
|
||||||
|
rpID: rpId,
|
||||||
|
allowCredentials: credentials.map((cred) => ({
|
||||||
|
id: cred.credentialID,
|
||||||
|
transports: ["internal"],
|
||||||
|
})),
|
||||||
|
userVerification: "preferred",
|
||||||
|
});
|
||||||
|
|
||||||
|
// Store challenge for verification
|
||||||
|
const challengeKey = credentialId || userId || options.challenge;
|
||||||
|
challengeStore.set(challengeKey, {
|
||||||
|
challenge: options.challenge,
|
||||||
|
userId: userId,
|
||||||
|
expiresAt: Date.now() + 5 * 60 * 1000, // 5 minutes
|
||||||
|
});
|
||||||
|
|
||||||
|
return options;
|
||||||
|
} catch (error) {
|
||||||
|
fastify.log.error(error);
|
||||||
|
reply.code(500).send({ error: "Failed to generate authentication options" });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* POST /webauthn/authentication/verify
|
||||||
|
* Verify an authentication response
|
||||||
|
*/
|
||||||
|
fastify.post<{
|
||||||
|
Body: {
|
||||||
|
options: PublicKeyCredentialRequestOptionsJSON;
|
||||||
|
assertionResponse: unknown;
|
||||||
|
};
|
||||||
|
}>("/webauthn/authentication/verify", async (request, reply) => {
|
||||||
|
try {
|
||||||
|
const { options, assertionResponse } = request.body;
|
||||||
|
|
||||||
|
// Find credential by ID
|
||||||
|
const credentialId = (assertionResponse as any).id;
|
||||||
|
const credential = credentialStore.get(credentialId);
|
||||||
|
|
||||||
|
if (!credential) {
|
||||||
|
reply.code(404).send({ error: "Credential not found" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Retrieve stored challenge
|
||||||
|
const challengeKey = credentialId;
|
||||||
|
const storedChallenge = challengeStore.get(challengeKey);
|
||||||
|
|
||||||
|
if (!storedChallenge) {
|
||||||
|
reply.code(400).send({ error: "Challenge not found or expired" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (storedChallenge.expiresAt < Date.now()) {
|
||||||
|
challengeStore.delete(challengeKey);
|
||||||
|
reply.code(400).send({ error: "Challenge expired" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verify authentication response
|
||||||
|
const verification = await verifyAuthenticationResponse({
|
||||||
|
response: assertionResponse as any,
|
||||||
|
expectedChallenge: storedChallenge.challenge,
|
||||||
|
expectedOrigin: rpOrigin,
|
||||||
|
expectedRPID: rpId,
|
||||||
|
authenticator: {
|
||||||
|
credentialID: credential.credentialID,
|
||||||
|
credentialPublicKey: credential.credentialPublicKey,
|
||||||
|
counter: credential.counter,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// Clean up challenge
|
||||||
|
challengeStore.delete(challengeKey);
|
||||||
|
|
||||||
|
if (!verification.verified) {
|
||||||
|
reply.code(400).send({ verified: false, error: "Verification failed" });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update counter
|
||||||
|
if (verification.authenticationInfo) {
|
||||||
|
credential.counter = verification.authenticationInfo.newCounter;
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
verified: true,
|
||||||
|
counter: credential.counter,
|
||||||
|
};
|
||||||
|
} catch (error) {
|
||||||
|
fastify.log.error(error);
|
||||||
|
reply.code(500).send({ error: "Verification failed", details: String(error) });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Health check endpoint
|
||||||
|
*/
|
||||||
|
fastify.get("/health", async () => {
|
||||||
|
return { status: "ok", timestamp: new Date().toISOString() };
|
||||||
|
});
|
||||||
|
|
||||||
|
// Start server
|
||||||
|
const start = async () => {
|
||||||
|
try {
|
||||||
|
const port = parseInt(process.env.PORT || "3002");
|
||||||
|
const host = process.env.HOST || "0.0.0.0";
|
||||||
|
|
||||||
|
await fastify.listen({ port, host });
|
||||||
|
fastify.log.info(`WebAuthn server listening on ${host}:${port}`);
|
||||||
|
fastify.log.info(`RP ID: ${rpId}, RP Name: ${rpName}, RP Origin: ${rpOrigin}`);
|
||||||
|
} catch (err) {
|
||||||
|
fastify.log.error(err);
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
start();
|
||||||
|
|
||||||
21
server/tsconfig.json
Normal file
21
server/tsconfig.json
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
{
|
||||||
|
"compilerOptions": {
|
||||||
|
"target": "ES2022",
|
||||||
|
"module": "ES2022",
|
||||||
|
"lib": ["ES2022"],
|
||||||
|
"moduleResolution": "node",
|
||||||
|
"outDir": "./dist",
|
||||||
|
"rootDir": "./src",
|
||||||
|
"strict": true,
|
||||||
|
"esModuleInterop": true,
|
||||||
|
"skipLibCheck": true,
|
||||||
|
"forceConsistentCasingInFileNames": true,
|
||||||
|
"resolveJsonModule": true,
|
||||||
|
"declaration": true,
|
||||||
|
"declarationMap": true,
|
||||||
|
"sourceMap": true
|
||||||
|
},
|
||||||
|
"include": ["src/**/*"],
|
||||||
|
"exclude": ["node_modules", "dist"]
|
||||||
|
}
|
||||||
|
|
||||||
@@ -47,6 +47,16 @@ export const DEFAULT_PARTNER_API_SERVER =
|
|||||||
export const DEFAULT_PUSH_SERVER =
|
export const DEFAULT_PUSH_SERVER =
|
||||||
import.meta.env.VITE_DEFAULT_PUSH_SERVER || AppString.PROD_PUSH_SERVER;
|
import.meta.env.VITE_DEFAULT_PUSH_SERVER || AppString.PROD_PUSH_SERVER;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* WebAuthn server endpoint URL
|
||||||
|
* Defaults to localhost:3002 for development, or can be set via VITE_WEBAUTHN_SERVER_URL
|
||||||
|
*/
|
||||||
|
export const DEFAULT_WEBAUTHN_SERVER =
|
||||||
|
import.meta.env.VITE_WEBAUTHN_SERVER_URL ||
|
||||||
|
(import.meta.env.DEV || window.location.hostname === "localhost"
|
||||||
|
? "http://localhost:3002"
|
||||||
|
: window.location.origin);
|
||||||
|
|
||||||
export const IMAGE_TYPE_PROFILE = "profile";
|
export const IMAGE_TYPE_PROFILE = "profile";
|
||||||
|
|
||||||
export const PASSKEYS_ENABLED =
|
export const PASSKEYS_ENABLED =
|
||||||
|
|||||||
@@ -94,7 +94,7 @@ const MIGRATIONS = [
|
|||||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
dateCreated TEXT NOT NULL,
|
dateCreated TEXT NOT NULL,
|
||||||
derivationPath TEXT,
|
derivationPath TEXT,
|
||||||
did TEXT NOT NULL,
|
did TEXT NOT NULL UNIQUE, -- UNIQUE constraint ensures no duplicate DIDs
|
||||||
identityEncrBase64 TEXT, -- encrypted & base64-encoded
|
identityEncrBase64 TEXT, -- encrypted & base64-encoded
|
||||||
mnemonicEncrBase64 TEXT, -- encrypted & base64-encoded
|
mnemonicEncrBase64 TEXT, -- encrypted & base64-encoded
|
||||||
passkeyCredIdHex TEXT,
|
passkeyCredIdHex TEXT,
|
||||||
|
|||||||
@@ -31,18 +31,7 @@ export async function updateDidSpecificSettings(
|
|||||||
|
|
||||||
const platform = PlatformServiceFactory.getInstance();
|
const platform = PlatformServiceFactory.getInstance();
|
||||||
|
|
||||||
// First, let's see what's currently in the database
|
// Generate and execute the update statement
|
||||||
const checkResult = await platform.dbQuery(
|
|
||||||
"SELECT * FROM settings WHERE accountDid = ?",
|
|
||||||
[accountDid],
|
|
||||||
);
|
|
||||||
|
|
||||||
// Get the current values for comparison
|
|
||||||
const currentRecord = checkResult?.values?.length
|
|
||||||
? mapColumnsToValues(checkResult.columns, checkResult.values)[0]
|
|
||||||
: null;
|
|
||||||
|
|
||||||
// First try to update existing record
|
|
||||||
const { sql: updateSql, params: updateParams } = generateUpdateStatement(
|
const { sql: updateSql, params: updateParams } = generateUpdateStatement(
|
||||||
settingsChanges,
|
settingsChanges,
|
||||||
"settings",
|
"settings",
|
||||||
@@ -50,66 +39,13 @@ export async function updateDidSpecificSettings(
|
|||||||
[accountDid],
|
[accountDid],
|
||||||
);
|
);
|
||||||
|
|
||||||
await platform.dbExec(updateSql, updateParams);
|
// dbExec() now returns reliable changes count across all platforms
|
||||||
|
// (normalized using SQLite's changes() function in Capacitor/Electron,
|
||||||
// **WORKAROUND**: AbsurdSQL doesn't return changes count correctly
|
// and reliable from AbsurdSQL in web platform)
|
||||||
// Instead, check if the record was actually updated
|
const result = await platform.dbExec(updateSql, updateParams);
|
||||||
const postUpdateResult = await platform.dbQuery(
|
|
||||||
"SELECT * FROM settings WHERE accountDid = ?",
|
// Return true if any rows were affected
|
||||||
[accountDid],
|
return result.changes > 0;
|
||||||
);
|
|
||||||
|
|
||||||
const updatedRecord = postUpdateResult?.values?.length
|
|
||||||
? mapColumnsToValues(postUpdateResult.columns, postUpdateResult.values)[0]
|
|
||||||
: null;
|
|
||||||
|
|
||||||
// Note that we want to eliminate this check (and fix the above if it doesn't work).
|
|
||||||
// Check if any of the target fields were actually changed
|
|
||||||
let actuallyUpdated = false;
|
|
||||||
if (currentRecord && updatedRecord) {
|
|
||||||
for (const key of Object.keys(settingsChanges)) {
|
|
||||||
if (key !== "accountDid" && currentRecord[key] !== updatedRecord[key]) {
|
|
||||||
actuallyUpdated = true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// If the standard update didn't work, try a different approach
|
|
||||||
if (
|
|
||||||
!actuallyUpdated &&
|
|
||||||
settingsChanges.firstName &&
|
|
||||||
settingsChanges.isRegistered !== undefined
|
|
||||||
) {
|
|
||||||
// Update firstName
|
|
||||||
await platform.dbExec(
|
|
||||||
"UPDATE settings SET firstName = ? WHERE accountDid = ?",
|
|
||||||
[settingsChanges.firstName, accountDid],
|
|
||||||
);
|
|
||||||
|
|
||||||
// Update isRegistered
|
|
||||||
await platform.dbExec(
|
|
||||||
"UPDATE settings SET isRegistered = ? WHERE accountDid = ?",
|
|
||||||
[settingsChanges.isRegistered ? 1 : 0, accountDid],
|
|
||||||
);
|
|
||||||
|
|
||||||
// Check if the individual updates worked
|
|
||||||
const finalCheckResult = await platform.dbQuery(
|
|
||||||
"SELECT * FROM settings WHERE accountDid = ?",
|
|
||||||
[accountDid],
|
|
||||||
);
|
|
||||||
|
|
||||||
const finalRecord = finalCheckResult?.values?.length
|
|
||||||
? mapColumnsToValues(finalCheckResult.columns, finalCheckResult.values)[0]
|
|
||||||
: null;
|
|
||||||
|
|
||||||
if (finalRecord) {
|
|
||||||
actuallyUpdated =
|
|
||||||
finalRecord.firstName === settingsChanges.firstName &&
|
|
||||||
finalRecord.isRegistered === (settingsChanges.isRegistered ? 1 : 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return actuallyUpdated;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const DEFAULT_SETTINGS: Settings = {
|
const DEFAULT_SETTINGS: Settings = {
|
||||||
|
|||||||
73
src/interfaces/diagnostics.ts
Normal file
73
src/interfaces/diagnostics.ts
Normal file
@@ -0,0 +1,73 @@
|
|||||||
|
/**
|
||||||
|
* Platform Diagnostics Interface
|
||||||
|
*
|
||||||
|
* Provides comprehensive diagnostic information about the current platform,
|
||||||
|
* database backend, worker status, and build information.
|
||||||
|
*
|
||||||
|
* @author Matthew Raymer
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { PlatformCapabilities } from "@/services/PlatformService";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Database backend information
|
||||||
|
*/
|
||||||
|
export interface DatabaseDiagnostics {
|
||||||
|
/** Type of database backend in use */
|
||||||
|
kind: "absurd-sql" | "capacitor-sqlite" | "electron-sqlite" | "unknown";
|
||||||
|
/** SharedArrayBuffer availability status (web platform only) */
|
||||||
|
sharedArrayBuffer?: "available" | "fallback" | "unknown";
|
||||||
|
/** Worker thread status (web platform only) */
|
||||||
|
worker?: {
|
||||||
|
/** Whether the worker is ready to process messages */
|
||||||
|
ready: boolean;
|
||||||
|
/** Number of pending messages */
|
||||||
|
pending: number;
|
||||||
|
/** Time since last ping in milliseconds */
|
||||||
|
lastPingMs?: number;
|
||||||
|
};
|
||||||
|
/** Operation queue status (Capacitor/Electron platforms) */
|
||||||
|
queue?: {
|
||||||
|
/** Current queue length */
|
||||||
|
current: number;
|
||||||
|
/** Peak queue size reached */
|
||||||
|
maxReached: number;
|
||||||
|
/** Maximum queue size limit */
|
||||||
|
limit: number;
|
||||||
|
/** Whether queue is currently processing */
|
||||||
|
isProcessing: boolean;
|
||||||
|
};
|
||||||
|
/** Database initialization status */
|
||||||
|
initialized: boolean;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Build information
|
||||||
|
*/
|
||||||
|
export interface BuildDiagnostics {
|
||||||
|
/** Application version from package.json */
|
||||||
|
version?: string;
|
||||||
|
/** Git commit hash */
|
||||||
|
commit?: string;
|
||||||
|
/** Build mode (development, test, production) */
|
||||||
|
mode?: string;
|
||||||
|
/** Build timestamp */
|
||||||
|
timestamp?: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Complete platform diagnostics
|
||||||
|
*/
|
||||||
|
export interface PlatformDiagnostics {
|
||||||
|
/** Detected platform */
|
||||||
|
platform: "web" | "capacitor" | "electron" | "development" | string;
|
||||||
|
/** Platform capabilities */
|
||||||
|
capabilities: PlatformCapabilities;
|
||||||
|
/** Database diagnostics */
|
||||||
|
db: DatabaseDiagnostics;
|
||||||
|
/** Build information */
|
||||||
|
build: BuildDiagnostics;
|
||||||
|
/** Additional platform-specific diagnostics */
|
||||||
|
metadata?: Record<string, unknown>;
|
||||||
|
}
|
||||||
|
|
||||||
219
src/libs/crypto/vc/passkeyDidPeer.client.ts
Normal file
219
src/libs/crypto/vc/passkeyDidPeer.client.ts
Normal file
@@ -0,0 +1,219 @@
|
|||||||
|
/**
|
||||||
|
* Client-side WebAuthn Passkey Functions
|
||||||
|
*
|
||||||
|
* This module provides client-side WebAuthn operations using @simplewebauthn/browser.
|
||||||
|
* All verification is performed server-side via API endpoints.
|
||||||
|
*
|
||||||
|
* @author Matthew Raymer
|
||||||
|
*/
|
||||||
|
|
||||||
|
import {
|
||||||
|
startAuthentication,
|
||||||
|
startRegistration,
|
||||||
|
} from "@simplewebauthn/browser";
|
||||||
|
import type {
|
||||||
|
PublicKeyCredentialCreationOptionsJSON,
|
||||||
|
PublicKeyCredentialRequestOptionsJSON,
|
||||||
|
PublicKeyCredentialJSON,
|
||||||
|
} from "@simplewebauthn/types";
|
||||||
|
import { AppString } from "../../../constants/app";
|
||||||
|
import { logger } from "../../../utils/logger";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* WebAuthn server endpoint configuration
|
||||||
|
*/
|
||||||
|
const getWebAuthnServerUrl = (): string => {
|
||||||
|
// Check for custom endpoint in settings/env
|
||||||
|
const customUrl = import.meta.env.VITE_WEBAUTHN_SERVER_URL;
|
||||||
|
if (customUrl) {
|
||||||
|
return customUrl;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Default to localhost:3002 for development (matches server default port)
|
||||||
|
// In production, this should point to your WebAuthn verification service
|
||||||
|
if (import.meta.env.DEV || window.location.hostname === "localhost") {
|
||||||
|
return "http://localhost:3002";
|
||||||
|
}
|
||||||
|
|
||||||
|
// Production: use same origin or configured endpoint
|
||||||
|
return window.location.origin;
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Registration result from server verification
|
||||||
|
*/
|
||||||
|
export interface RegistrationVerificationResult {
|
||||||
|
verified: boolean;
|
||||||
|
credential: {
|
||||||
|
credentialID: string;
|
||||||
|
credentialPublicKey: Uint8Array;
|
||||||
|
counter: number;
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Authentication result from server verification
|
||||||
|
*/
|
||||||
|
export interface AuthenticationVerificationResult {
|
||||||
|
verified: boolean;
|
||||||
|
counter?: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Register a new passkey credential
|
||||||
|
*
|
||||||
|
* Flow:
|
||||||
|
* 1. Request registration options from server
|
||||||
|
* 2. Start registration with browser API
|
||||||
|
* 3. Send attestation response to server for verification
|
||||||
|
* 4. Return verified credential info
|
||||||
|
*
|
||||||
|
* @param passkeyName - Optional name for the passkey
|
||||||
|
* @param userId - Optional user ID (if not provided, server generates)
|
||||||
|
* @returns Verified registration result with credential info
|
||||||
|
*/
|
||||||
|
export async function registerPasskey(
|
||||||
|
passkeyName?: string,
|
||||||
|
userId?: string
|
||||||
|
): Promise<RegistrationVerificationResult> {
|
||||||
|
const serverUrl = getWebAuthnServerUrl();
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Step 1: Request registration options from server
|
||||||
|
const optionsResponse = await fetch(`${serverUrl}/webauthn/registration/options`, {
|
||||||
|
method: "POST",
|
||||||
|
headers: {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
username: passkeyName || AppString.APP_NAME + " User",
|
||||||
|
userId: userId,
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!optionsResponse.ok) {
|
||||||
|
throw new Error(
|
||||||
|
`Failed to get registration options: ${optionsResponse.statusText}`
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const options: PublicKeyCredentialCreationOptionsJSON =
|
||||||
|
await optionsResponse.json();
|
||||||
|
|
||||||
|
// Step 2: Start registration with browser API
|
||||||
|
const attestationResponse = await startRegistration(options);
|
||||||
|
|
||||||
|
// Step 3: Send attestation response to server for verification
|
||||||
|
const verifyResponse = await fetch(`${serverUrl}/webauthn/registration/verify`, {
|
||||||
|
method: "POST",
|
||||||
|
headers: {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
options: options,
|
||||||
|
attestationResponse: attestationResponse,
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!verifyResponse.ok) {
|
||||||
|
throw new Error(
|
||||||
|
`Registration verification failed: ${verifyResponse.statusText}`
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const verification: RegistrationVerificationResult =
|
||||||
|
await verifyResponse.json();
|
||||||
|
|
||||||
|
if (!verification.verified) {
|
||||||
|
throw new Error("Registration verification failed on server");
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.debug("[passkeyDidPeer.client] Registration successful", {
|
||||||
|
credentialID: verification.credential.credentialID,
|
||||||
|
});
|
||||||
|
|
||||||
|
return verification;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error("[passkeyDidPeer.client] Registration failed:", error);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Authenticate with an existing passkey credential
|
||||||
|
*
|
||||||
|
* Flow:
|
||||||
|
* 1. Request authentication options from server
|
||||||
|
* 2. Start authentication with browser API
|
||||||
|
* 3. Send assertion response to server for verification
|
||||||
|
* 4. Return verification result
|
||||||
|
*
|
||||||
|
* @param credentialId - Base64URL encoded credential ID
|
||||||
|
* @param userId - Optional user ID (if not provided, server looks up by credential)
|
||||||
|
* @returns Verification result
|
||||||
|
*/
|
||||||
|
export async function authenticatePasskey(
|
||||||
|
credentialId: string,
|
||||||
|
userId?: string
|
||||||
|
): Promise<AuthenticationVerificationResult> {
|
||||||
|
const serverUrl = getWebAuthnServerUrl();
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Step 1: Request authentication options from server
|
||||||
|
const optionsResponse = await fetch(`${serverUrl}/webauthn/authentication/options`, {
|
||||||
|
method: "POST",
|
||||||
|
headers: {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
credentialId: credentialId,
|
||||||
|
userId: userId,
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!optionsResponse.ok) {
|
||||||
|
throw new Error(
|
||||||
|
`Failed to get authentication options: ${optionsResponse.statusText}`
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const options: PublicKeyCredentialRequestOptionsJSON =
|
||||||
|
await optionsResponse.json();
|
||||||
|
|
||||||
|
// Step 2: Start authentication with browser API
|
||||||
|
const assertionResponse = await startAuthentication(options);
|
||||||
|
|
||||||
|
// Step 3: Send assertion response to server for verification
|
||||||
|
const verifyResponse = await fetch(`${serverUrl}/webauthn/authentication/verify`, {
|
||||||
|
method: "POST",
|
||||||
|
headers: {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
options: options,
|
||||||
|
assertionResponse: assertionResponse,
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!verifyResponse.ok) {
|
||||||
|
throw new Error(
|
||||||
|
`Authentication verification failed: ${verifyResponse.statusText}`
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const verification: AuthenticationVerificationResult =
|
||||||
|
await verifyResponse.json();
|
||||||
|
|
||||||
|
if (!verification.verified) {
|
||||||
|
throw new Error("Authentication verification failed on server");
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.debug("[passkeyDidPeer.client] Authentication successful");
|
||||||
|
|
||||||
|
return verification;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error("[passkeyDidPeer.client] Authentication failed:", error);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
123
src/libs/crypto/vc/passkeyDidPeer.offlineVerify.ts
Normal file
123
src/libs/crypto/vc/passkeyDidPeer.offlineVerify.ts
Normal file
@@ -0,0 +1,123 @@
|
|||||||
|
/**
|
||||||
|
* Offline WebAuthn Verification Module
|
||||||
|
*
|
||||||
|
* This module contains server-side WebAuthn verification functions that are
|
||||||
|
* only available when VITE_OFFLINE_WEBAUTHN_VERIFY is enabled.
|
||||||
|
*
|
||||||
|
* SECURITY WARNING: Client-side verification can be tampered with and should
|
||||||
|
* not be trusted for security-critical operations. This module is intended
|
||||||
|
* for offline-only mode where server-side verification is not available.
|
||||||
|
*
|
||||||
|
* For production deployments, verification should be performed on a server.
|
||||||
|
*/
|
||||||
|
|
||||||
|
import type {
|
||||||
|
Base64URLString,
|
||||||
|
VerifyAuthenticationResponseOpts,
|
||||||
|
} from "@simplewebauthn/types";
|
||||||
|
import { logger } from "../../../utils/logger";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Dynamically import server-side verification functions
|
||||||
|
* This prevents bundling @simplewebauthn/server in normal builds
|
||||||
|
*/
|
||||||
|
async function getServerVerification() {
|
||||||
|
// Check if offline verification is enabled
|
||||||
|
const offlineVerifyEnabled =
|
||||||
|
import.meta.env.VITE_OFFLINE_WEBAUTHN_VERIFY === "true";
|
||||||
|
|
||||||
|
if (!offlineVerifyEnabled) {
|
||||||
|
throw new Error(
|
||||||
|
"Offline WebAuthn verification is disabled. " +
|
||||||
|
"Set VITE_OFFLINE_WEBAUTHN_VERIFY=true to enable offline mode. " +
|
||||||
|
"For production, use server-side verification instead."
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Dynamic import prevents bundling in normal builds
|
||||||
|
const serverModule = await import("@simplewebauthn/server");
|
||||||
|
return {
|
||||||
|
verifyRegistrationResponse: serverModule.verifyRegistrationResponse,
|
||||||
|
verifyAuthenticationResponse: serverModule.verifyAuthenticationResponse,
|
||||||
|
generateRegistrationOptions: serverModule.generateRegistrationOptions,
|
||||||
|
generateAuthenticationOptions: serverModule.generateAuthenticationOptions,
|
||||||
|
};
|
||||||
|
} catch (error) {
|
||||||
|
logger.error(
|
||||||
|
"[passkeyDidPeer.offlineVerify] Failed to load server verification module:",
|
||||||
|
error
|
||||||
|
);
|
||||||
|
throw new Error(
|
||||||
|
"Server-side WebAuthn verification module is not available. " +
|
||||||
|
"This feature requires VITE_OFFLINE_WEBAUTHN_VERIFY=true and @simplewebauthn/server to be installed."
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Verify registration response (offline mode only)
|
||||||
|
*
|
||||||
|
* @throws Error if offline verification is not enabled
|
||||||
|
*/
|
||||||
|
export async function verifyRegistrationResponseOffline(
|
||||||
|
response: unknown,
|
||||||
|
expectedChallenge: string,
|
||||||
|
expectedOrigin: string,
|
||||||
|
expectedRPID: string
|
||||||
|
) {
|
||||||
|
const { verifyRegistrationResponse } = await getServerVerification();
|
||||||
|
return verifyRegistrationResponse({
|
||||||
|
response: response as any,
|
||||||
|
expectedChallenge,
|
||||||
|
expectedOrigin,
|
||||||
|
expectedRPID,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Verify authentication response (offline mode only)
|
||||||
|
*
|
||||||
|
* @throws Error if offline verification is not enabled
|
||||||
|
*/
|
||||||
|
export async function verifyAuthenticationResponseOffline(
|
||||||
|
opts: VerifyAuthenticationResponseOpts
|
||||||
|
) {
|
||||||
|
const { verifyAuthenticationResponse } = await getServerVerification();
|
||||||
|
return verifyAuthenticationResponse(opts);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generate registration options (offline mode only)
|
||||||
|
*
|
||||||
|
* @throws Error if offline verification is not enabled
|
||||||
|
*/
|
||||||
|
export async function generateRegistrationOptionsOffline(opts: {
|
||||||
|
rpName: string;
|
||||||
|
rpID: string;
|
||||||
|
userName: string;
|
||||||
|
attestationType?: string;
|
||||||
|
authenticatorSelection?: {
|
||||||
|
residentKey?: string;
|
||||||
|
userVerification?: string;
|
||||||
|
authenticatorAttachment?: string;
|
||||||
|
};
|
||||||
|
}) {
|
||||||
|
const { generateRegistrationOptions } = await getServerVerification();
|
||||||
|
return generateRegistrationOptions(opts);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generate authentication options (offline mode only)
|
||||||
|
*
|
||||||
|
* @throws Error if offline verification is not enabled
|
||||||
|
*/
|
||||||
|
export async function generateAuthenticationOptionsOffline(opts: {
|
||||||
|
challenge: Uint8Array;
|
||||||
|
rpID: string;
|
||||||
|
allowCredentials?: Array<{ id: string }>;
|
||||||
|
}) {
|
||||||
|
const { generateAuthenticationOptions } = await getServerVerification();
|
||||||
|
return generateAuthenticationOptions(opts);
|
||||||
|
}
|
||||||
|
|
||||||
@@ -5,18 +5,21 @@ import {
|
|||||||
startAuthentication,
|
startAuthentication,
|
||||||
startRegistration,
|
startRegistration,
|
||||||
} from "@simplewebauthn/browser";
|
} from "@simplewebauthn/browser";
|
||||||
import {
|
/**
|
||||||
generateAuthenticationOptions,
|
* Client-side WebAuthn Passkey Functions
|
||||||
generateRegistrationOptions,
|
*
|
||||||
verifyAuthenticationResponse,
|
* This module provides client-side WebAuthn operations using @simplewebauthn/browser.
|
||||||
verifyRegistrationResponse,
|
* Server-side verification is isolated in passkeyDidPeer.offlineVerify.ts and only
|
||||||
VerifyAuthenticationResponseOpts,
|
* available when VITE_OFFLINE_WEBAUTHN_VERIFY=true.
|
||||||
} from "@simplewebauthn/server";
|
*
|
||||||
import {
|
* For production deployments, verification should be performed on a server endpoint.
|
||||||
|
*/
|
||||||
|
import type {
|
||||||
Base64URLString,
|
Base64URLString,
|
||||||
PublicKeyCredentialCreationOptionsJSON,
|
PublicKeyCredentialCreationOptionsJSON,
|
||||||
PublicKeyCredentialRequestOptionsJSON,
|
PublicKeyCredentialRequestOptionsJSON,
|
||||||
AuthenticatorAssertionResponse,
|
AuthenticatorAssertionResponse,
|
||||||
|
VerifyAuthenticationResponseOpts,
|
||||||
} from "@simplewebauthn/types";
|
} from "@simplewebauthn/types";
|
||||||
|
|
||||||
import { AppString } from "../../../constants/app";
|
import { AppString } from "../../../constants/app";
|
||||||
@@ -36,32 +39,103 @@ export interface JWK {
|
|||||||
y: string;
|
y: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Register a new passkey credential
|
||||||
|
*
|
||||||
|
* This is a facade that routes to either:
|
||||||
|
* - Client module (server-side verification) - default/production
|
||||||
|
* - Offline verification module (client-side) - only if VITE_OFFLINE_WEBAUTHN_VERIFY=true
|
||||||
|
*
|
||||||
|
* @param passkeyName - Optional name for the passkey
|
||||||
|
* @returns Registration result with credential info
|
||||||
|
*/
|
||||||
export async function registerCredential(passkeyName?: string) {
|
export async function registerCredential(passkeyName?: string) {
|
||||||
const options: PublicKeyCredentialCreationOptionsJSON =
|
const offlineVerifyEnabled =
|
||||||
await generateRegistrationOptions({
|
import.meta.env.VITE_OFFLINE_WEBAUTHN_VERIFY === "true";
|
||||||
|
|
||||||
|
if (offlineVerifyEnabled) {
|
||||||
|
// Offline mode: use dynamic import for client-side verification
|
||||||
|
const {
|
||||||
|
generateRegistrationOptionsOffline,
|
||||||
|
verifyRegistrationResponseOffline,
|
||||||
|
} = await import("./passkeyDidPeer.offlineVerify");
|
||||||
|
|
||||||
|
const options = await generateRegistrationOptionsOffline({
|
||||||
rpName: AppString.APP_NAME,
|
rpName: AppString.APP_NAME,
|
||||||
rpID: window.location.hostname,
|
rpID: window.location.hostname,
|
||||||
userName: passkeyName || AppString.APP_NAME + " User",
|
userName: passkeyName || AppString.APP_NAME + " User",
|
||||||
// Don't prompt users for additional information about the authenticator
|
|
||||||
// (Recommended for smoother UX)
|
|
||||||
attestationType: "none",
|
attestationType: "none",
|
||||||
authenticatorSelection: {
|
authenticatorSelection: {
|
||||||
// Defaults
|
|
||||||
residentKey: "preferred",
|
residentKey: "preferred",
|
||||||
userVerification: "preferred",
|
userVerification: "preferred",
|
||||||
// Optional
|
|
||||||
authenticatorAttachment: "platform",
|
authenticatorAttachment: "platform",
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
// someday, instead of simplwebauthn, we'll go direct: navigator.credentials.create with PublicKeyCredentialCreationOptions
|
|
||||||
// with pubKeyCredParams: { type: "public-key", alg: -7 }
|
const attResp = await startRegistration(options);
|
||||||
const attResp = await startRegistration(options);
|
const verification = await verifyRegistrationResponseOffline(
|
||||||
const verification = await verifyRegistrationResponse({
|
attResp,
|
||||||
response: attResp,
|
options.challenge,
|
||||||
expectedChallenge: options.challenge,
|
window.location.origin,
|
||||||
expectedOrigin: window.location.origin,
|
window.location.hostname
|
||||||
expectedRPID: window.location.hostname,
|
);
|
||||||
});
|
|
||||||
|
return extractCredentialInfo(attResp, verification);
|
||||||
|
} else {
|
||||||
|
// Production mode: use client module with server-side verification
|
||||||
|
const { registerPasskey } = await import("./passkeyDidPeer.client");
|
||||||
|
const verification = await registerPasskey(passkeyName);
|
||||||
|
|
||||||
|
// Convert server response to expected format
|
||||||
|
const credIdBase64Url = verification.credential.credentialID;
|
||||||
|
const credIdHex = Buffer.from(
|
||||||
|
base64URLStringToArrayBuffer(credIdBase64Url),
|
||||||
|
).toString("hex");
|
||||||
|
const { publicKeyJwk } = cborToKeys(
|
||||||
|
verification.credential.credentialPublicKey
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
authData: undefined, // Not available from server response
|
||||||
|
credIdHex: credIdHex,
|
||||||
|
publicKeyJwk: publicKeyJwk,
|
||||||
|
publicKeyBytes: verification.credential.credentialPublicKey,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Extract credential info from attestation response and verification result
|
||||||
|
* Used by offline mode
|
||||||
|
*/
|
||||||
|
function extractCredentialInfo(
|
||||||
|
attResp: any,
|
||||||
|
verification: any
|
||||||
|
): {
|
||||||
|
authData: unknown;
|
||||||
|
credIdHex: string;
|
||||||
|
publicKeyJwk: JWK;
|
||||||
|
publicKeyBytes: Uint8Array;
|
||||||
|
} {
|
||||||
|
const credIdBase64Url = verification.registrationInfo?.credentialID as string;
|
||||||
|
if (attResp.rawId !== credIdBase64Url) {
|
||||||
|
logger.warn("Warning! The raw ID does not match the credential ID.");
|
||||||
|
}
|
||||||
|
const credIdHex = Buffer.from(
|
||||||
|
base64URLStringToArrayBuffer(credIdBase64Url),
|
||||||
|
).toString("hex");
|
||||||
|
const { publicKeyJwk } = cborToKeys(
|
||||||
|
verification.registrationInfo?.credentialPublicKey as Uint8Array,
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
authData: verification.registrationInfo?.attestationObject,
|
||||||
|
credIdHex: credIdHex,
|
||||||
|
publicKeyJwk: publicKeyJwk,
|
||||||
|
publicKeyBytes: verification.registrationInfo
|
||||||
|
?.credentialPublicKey as Uint8Array,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
// references for parsing auth data and getting the public key
|
// references for parsing auth data and getting the public key
|
||||||
// https://github.com/MasterKale/SimpleWebAuthn/blob/master/packages/server/src/helpers/parseAuthenticatorData.ts#L11
|
// https://github.com/MasterKale/SimpleWebAuthn/blob/master/packages/server/src/helpers/parseAuthenticatorData.ts#L11
|
||||||
@@ -113,12 +187,32 @@ export class PeerSetup {
|
|||||||
};
|
};
|
||||||
this.challenge = new Uint8Array(Buffer.from(JSON.stringify(fullPayload)));
|
this.challenge = new Uint8Array(Buffer.from(JSON.stringify(fullPayload)));
|
||||||
// const payloadHash: Uint8Array = sha256(this.challenge);
|
// const payloadHash: Uint8Array = sha256(this.challenge);
|
||||||
const options: PublicKeyCredentialRequestOptionsJSON =
|
// Use offline verification if enabled
|
||||||
await generateAuthenticationOptions({
|
const offlineVerifyEnabled =
|
||||||
|
import.meta.env.VITE_OFFLINE_WEBAUTHN_VERIFY === "true";
|
||||||
|
|
||||||
|
let options: PublicKeyCredentialRequestOptionsJSON;
|
||||||
|
|
||||||
|
if (offlineVerifyEnabled) {
|
||||||
|
const { generateAuthenticationOptionsOffline } = await import(
|
||||||
|
"./passkeyDidPeer.offlineVerify"
|
||||||
|
);
|
||||||
|
options = await generateAuthenticationOptionsOffline({
|
||||||
challenge: this.challenge,
|
challenge: this.challenge,
|
||||||
rpID: window.location.hostname,
|
rpID: window.location.hostname,
|
||||||
allowCredentials: [{ id: credentialId }],
|
allowCredentials: [{ id: credentialId }],
|
||||||
});
|
});
|
||||||
|
} else {
|
||||||
|
// Production mode: should use server endpoint
|
||||||
|
// For now, fall back to direct navigator.credentials.get
|
||||||
|
// TODO: Implement server endpoint for authentication options
|
||||||
|
options = {
|
||||||
|
challenge: arrayBufferToBase64URLString(this.challenge.buffer),
|
||||||
|
rpId: window.location.hostname,
|
||||||
|
allowCredentials: [{ id: credentialId, type: "public-key" }],
|
||||||
|
userVerification: "preferred",
|
||||||
|
} as PublicKeyCredentialRequestOptionsJSON;
|
||||||
|
}
|
||||||
// console.log("simple authentication options", options);
|
// console.log("simple authentication options", options);
|
||||||
|
|
||||||
const clientAuth = await startAuthentication(options);
|
const clientAuth = await startAuthentication(options);
|
||||||
@@ -345,6 +439,22 @@ export async function verifyJwtSimplewebauthn(
|
|||||||
clientDataJsonBase64Url: Base64URLString,
|
clientDataJsonBase64Url: Base64URLString,
|
||||||
signature: Base64URLString,
|
signature: Base64URLString,
|
||||||
) {
|
) {
|
||||||
|
// Only allow offline verification if explicitly enabled
|
||||||
|
const offlineVerifyEnabled =
|
||||||
|
import.meta.env.VITE_OFFLINE_WEBAUTHN_VERIFY === "true";
|
||||||
|
|
||||||
|
if (!offlineVerifyEnabled) {
|
||||||
|
throw new Error(
|
||||||
|
"Client-side WebAuthn verification is disabled for security. " +
|
||||||
|
"Please use server-side verification endpoint or enable offline mode " +
|
||||||
|
"with VITE_OFFLINE_WEBAUTHN_VERIFY=true (not recommended for production)."
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const { verifyAuthenticationResponseOffline } = await import(
|
||||||
|
"./passkeyDidPeer.offlineVerify"
|
||||||
|
);
|
||||||
|
|
||||||
const authData = arrayToBase64Url(Buffer.from(authenticatorData));
|
const authData = arrayToBase64Url(Buffer.from(authenticatorData));
|
||||||
const publicKeyBytes = peerDidToPublicKeyBytes(issuerDid);
|
const publicKeyBytes = peerDidToPublicKeyBytes(issuerDid);
|
||||||
const credId = arrayBufferToBase64URLString(
|
const credId = arrayBufferToBase64URLString(
|
||||||
@@ -372,7 +482,7 @@ export async function verifyJwtSimplewebauthn(
|
|||||||
type: "public-key",
|
type: "public-key",
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
const verification = await verifyAuthenticationResponse(authOpts);
|
const verification = await verifyAuthenticationResponseOffline(authOpts);
|
||||||
return verification.verified;
|
return verification.verified;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -3,7 +3,8 @@ import { logger } from "./utils/logger";
|
|||||||
|
|
||||||
const platform = process.env.VITE_PLATFORM;
|
const platform = process.env.VITE_PLATFORM;
|
||||||
|
|
||||||
// PWA service worker is automatically registered by VitePWA plugin
|
// Note: PWA functionality is currently not implemented.
|
||||||
|
// Service worker registration would be handled here when PWA support is added.
|
||||||
|
|
||||||
const app = initializeApp();
|
const app = initializeApp();
|
||||||
|
|
||||||
|
|||||||
@@ -5,31 +5,21 @@ if (typeof window === "undefined") {
|
|||||||
globalThis.window = globalThis;
|
globalThis.window = globalThis;
|
||||||
|
|
||||||
// Enhanced crypto polyfill
|
// Enhanced crypto polyfill
|
||||||
|
// SECURITY: Never use Math.random() for cryptographic operations
|
||||||
|
// If crypto is missing, fail fast rather than silently using insecure randomness
|
||||||
if (typeof crypto === "undefined") {
|
if (typeof crypto === "undefined") {
|
||||||
globalThis.crypto = {
|
throw new Error(
|
||||||
getRandomValues: (array) => {
|
"[SQLWorker] crypto API is not available in worker context. " +
|
||||||
// Simple fallback for worker context
|
"This is required for secure database operations. " +
|
||||||
for (let i = 0; i < array.length; i++) {
|
"Please ensure the worker is running in a secure context with crypto support."
|
||||||
array[i] = Math.floor(Math.random() * 256);
|
);
|
||||||
}
|
|
||||||
return array;
|
|
||||||
},
|
|
||||||
subtle: {
|
|
||||||
generateKey: async () => ({ type: "secret" }),
|
|
||||||
sign: async () => new Uint8Array(32),
|
|
||||||
verify: async () => true,
|
|
||||||
digest: async () => new Uint8Array(32),
|
|
||||||
},
|
|
||||||
};
|
|
||||||
} else if (!crypto.getRandomValues) {
|
} else if (!crypto.getRandomValues) {
|
||||||
// Crypto exists but doesn't have getRandomValues - extend it
|
// Crypto exists but doesn't have getRandomValues - fail fast
|
||||||
crypto.getRandomValues = (array) => {
|
throw new Error(
|
||||||
// Simple fallback for worker context
|
"[SQLWorker] crypto.getRandomValues is not available. " +
|
||||||
for (let i = 0; i < array.length; i++) {
|
"This is required for secure database operations. " +
|
||||||
array[i] = Math.floor(Math.random() * 256);
|
"Please ensure the worker environment supports the Web Crypto API."
|
||||||
}
|
);
|
||||||
return array;
|
|
||||||
};
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -280,6 +280,17 @@ const routes: Array<RouteRecordRaw> = [
|
|||||||
name: "test",
|
name: "test",
|
||||||
component: () => import("../views/TestView.vue"),
|
component: () => import("../views/TestView.vue"),
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
path: "/debug/diagnostics",
|
||||||
|
name: "debug-diagnostics",
|
||||||
|
component: () => import("../views/debug/PlatformDiagnosticsView.vue"),
|
||||||
|
meta: {
|
||||||
|
title: "Platform Diagnostics",
|
||||||
|
requiresAuth: false,
|
||||||
|
// Only show in dev mode or if explicitly enabled in settings
|
||||||
|
devOnly: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
{
|
{
|
||||||
path: "/user-profile/:id?",
|
path: "/user-profile/:id?",
|
||||||
name: "user-profile",
|
name: "user-profile",
|
||||||
|
|||||||
@@ -1,17 +1,13 @@
|
|||||||
// **WORKER-COMPATIBLE CRYPTO POLYFILL**: Must be at the very top
|
// **SECURITY**: Crypto API is required for secure database operations
|
||||||
// This prevents "crypto is not defined" errors when running in worker context
|
// This service runs in a worker context where crypto should be available via Web Crypto API
|
||||||
|
// If crypto is missing, fail fast rather than silently using insecure Math.random()
|
||||||
|
// This matches the fail-fast approach in registerSQLWorker.js
|
||||||
if (typeof window === "undefined" && typeof crypto === "undefined") {
|
if (typeof window === "undefined" && typeof crypto === "undefined") {
|
||||||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
throw new Error(
|
||||||
(globalThis as any).crypto = {
|
"[AbsurdSqlDatabaseService] crypto API is not available. " +
|
||||||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
"This is required for secure database operations. " +
|
||||||
getRandomValues: (array: any) => {
|
"Please ensure the worker is running in a secure context with crypto support."
|
||||||
// Simple fallback for worker context
|
);
|
||||||
for (let i = 0; i < array.length; i++) {
|
|
||||||
array[i] = Math.floor(Math.random() * 256);
|
|
||||||
}
|
|
||||||
return array;
|
|
||||||
},
|
|
||||||
};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
import initSqlJs from "@jlongster/sql.js";
|
import initSqlJs from "@jlongster/sql.js";
|
||||||
|
|||||||
169
src/services/DiagnosticExportService.ts
Normal file
169
src/services/DiagnosticExportService.ts
Normal file
@@ -0,0 +1,169 @@
|
|||||||
|
/**
|
||||||
|
* Diagnostic Export Service
|
||||||
|
*
|
||||||
|
* Provides functionality to export comprehensive diagnostic information
|
||||||
|
* including platform diagnostics, settings, logs, and build information.
|
||||||
|
*
|
||||||
|
* @author Matthew Raymer
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { PlatformServiceFactory } from "./PlatformServiceFactory";
|
||||||
|
import { retrieveSettingsForActiveAccount } from "@/db/databaseUtil";
|
||||||
|
import type { PlatformDiagnostics } from "@/interfaces/diagnostics";
|
||||||
|
import { logger, getMemoryLogs } from "@/utils/logger";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Redacts sensitive information from diagnostic data
|
||||||
|
*/
|
||||||
|
function redactSensitive(data: unknown): unknown {
|
||||||
|
if (typeof data !== "object" || data === null) {
|
||||||
|
return data;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (Array.isArray(data)) {
|
||||||
|
return data.map(redactSensitive);
|
||||||
|
}
|
||||||
|
|
||||||
|
const redacted: Record<string, unknown> = {};
|
||||||
|
const sensitiveKeys = [
|
||||||
|
"privateKey",
|
||||||
|
"privateKeyHex",
|
||||||
|
"mnemonic",
|
||||||
|
"secret",
|
||||||
|
"password",
|
||||||
|
"token",
|
||||||
|
"apiKey",
|
||||||
|
"identityEncrBase64",
|
||||||
|
"mnemonicEncrBase64",
|
||||||
|
];
|
||||||
|
|
||||||
|
for (const [key, value] of Object.entries(data)) {
|
||||||
|
if (sensitiveKeys.some((sk) => key.toLowerCase().includes(sk.toLowerCase()))) {
|
||||||
|
redacted[key] = "[REDACTED]";
|
||||||
|
} else if (typeof value === "object" && value !== null) {
|
||||||
|
redacted[key] = redactSensitive(value);
|
||||||
|
} else {
|
||||||
|
redacted[key] = value;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return redacted;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Exports comprehensive diagnostic bundle
|
||||||
|
*
|
||||||
|
* @returns Promise resolving to diagnostic bundle as JSON string
|
||||||
|
*/
|
||||||
|
export async function exportDiagnostics(): Promise<string> {
|
||||||
|
const platform = PlatformServiceFactory.getInstance();
|
||||||
|
const timestamp = new Date().toISOString();
|
||||||
|
|
||||||
|
// Collect diagnostics
|
||||||
|
const diagnostics: PlatformDiagnostics | null = platform.getDiagnostics
|
||||||
|
? await platform.getDiagnostics()
|
||||||
|
: null;
|
||||||
|
|
||||||
|
// Collect settings
|
||||||
|
let settingsDefault = null;
|
||||||
|
let settingsActive = null;
|
||||||
|
try {
|
||||||
|
// Note: retrieveSettingsForDefaultAccount might not exist, handle gracefully
|
||||||
|
settingsActive = await retrieveSettingsForActiveAccount();
|
||||||
|
} catch (error) {
|
||||||
|
logger.debug("[DiagnosticExport] Could not retrieve settings:", error);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Collect recent logs from memory
|
||||||
|
let memoryLogs: string[] = [];
|
||||||
|
try {
|
||||||
|
memoryLogs = getMemoryLogs(1000); // Get last 1000 log entries
|
||||||
|
} catch (error) {
|
||||||
|
logger.debug("[DiagnosticExport] Could not retrieve memory logs:", error);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Collect recent logs from database (if logs table exists)
|
||||||
|
let dbLogs: unknown[] = [];
|
||||||
|
try {
|
||||||
|
const logsResult = await platform.dbQuery(
|
||||||
|
"SELECT * FROM logs ORDER BY date DESC LIMIT 100"
|
||||||
|
);
|
||||||
|
if (logsResult?.values) {
|
||||||
|
dbLogs = logsResult.values.map((row) => {
|
||||||
|
const logEntry: Record<string, unknown> = {};
|
||||||
|
if (logsResult.columns && row) {
|
||||||
|
logsResult.columns.forEach((col, idx) => {
|
||||||
|
logEntry[col] = row[idx];
|
||||||
|
});
|
||||||
|
}
|
||||||
|
return logEntry;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logger.debug("[DiagnosticExport] Could not retrieve DB logs:", error);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get build info
|
||||||
|
let packageJson: { version?: string } = {};
|
||||||
|
try {
|
||||||
|
packageJson = await import("../../../package.json");
|
||||||
|
} catch (error) {
|
||||||
|
logger.debug("[DiagnosticExport] Could not load package.json:", error);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get git commit hash if available
|
||||||
|
let commitHash: string | undefined;
|
||||||
|
try {
|
||||||
|
// This would need to be set at build time via Vite define
|
||||||
|
commitHash = import.meta.env.VITE_GIT_HASH;
|
||||||
|
} catch (error) {
|
||||||
|
// Ignore
|
||||||
|
}
|
||||||
|
|
||||||
|
// Assemble diagnostic bundle
|
||||||
|
const bundle = {
|
||||||
|
timestamp,
|
||||||
|
version: "1.0",
|
||||||
|
diagnostics: diagnostics ? redactSensitive(diagnostics) : null,
|
||||||
|
settings: {
|
||||||
|
active: redactSensitive(settingsActive),
|
||||||
|
default: redactSensitive(settingsDefault),
|
||||||
|
},
|
||||||
|
logs: {
|
||||||
|
memory: redactSensitive(memoryLogs),
|
||||||
|
database: redactSensitive(dbLogs),
|
||||||
|
},
|
||||||
|
build: {
|
||||||
|
version: packageJson.version,
|
||||||
|
commit: commitHash,
|
||||||
|
mode: import.meta.env.MODE,
|
||||||
|
platform: import.meta.env.VITE_PLATFORM,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
|
return JSON.stringify(bundle, null, 2);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Exports diagnostic bundle to file
|
||||||
|
*
|
||||||
|
* @param fileName - Optional custom filename (default: diagnostics-{timestamp}.json)
|
||||||
|
* @returns Promise that resolves when file is exported
|
||||||
|
*/
|
||||||
|
export async function exportDiagnosticsToFile(
|
||||||
|
fileName?: string
|
||||||
|
): Promise<void> {
|
||||||
|
const platform = PlatformServiceFactory.getInstance();
|
||||||
|
const timestamp = new Date().toISOString().replace(/[:.]/g, "-");
|
||||||
|
const defaultFileName = fileName || `diagnostics-${timestamp}.json`;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const bundle = await exportDiagnostics();
|
||||||
|
await platform.writeAndShareFile(defaultFileName, bundle);
|
||||||
|
logger.log(`[DiagnosticExport] Diagnostic bundle exported: ${defaultFileName}`);
|
||||||
|
} catch (error) {
|
||||||
|
logger.error("[DiagnosticExport] Failed to export diagnostics:", error);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@@ -215,7 +215,10 @@ export interface PlatformService {
|
|||||||
*/
|
*/
|
||||||
registerServiceWorker?(): void;
|
registerServiceWorker?(): void;
|
||||||
|
|
||||||
|
// --- Diagnostics (optional, for debugging) ---
|
||||||
/**
|
/**
|
||||||
* Returns true if PWA is enabled (web only)
|
* Gets comprehensive diagnostic information about the platform
|
||||||
|
* @returns Promise resolving to platform diagnostics
|
||||||
*/
|
*/
|
||||||
|
getDiagnostics?(): Promise<import("@/interfaces/diagnostics").PlatformDiagnostics>;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -39,7 +39,9 @@ export class PlatformServiceFactory {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Only log when actually creating the instance
|
// Only log when actually creating the instance
|
||||||
const platform = process.env.VITE_PLATFORM || "web";
|
// Use import.meta.env for Vite environment variables (standard Vite pattern)
|
||||||
|
// process.env.VITE_PLATFORM is defined via Vite's define config, but import.meta.env is preferred
|
||||||
|
const platform = (import.meta.env?.VITE_PLATFORM || process.env.VITE_PLATFORM || "web") as string;
|
||||||
|
|
||||||
if (!PlatformServiceFactory.creationLogged) {
|
if (!PlatformServiceFactory.creationLogged) {
|
||||||
// Use console for critical startup message to avoid circular dependency
|
// Use console for critical startup message to avoid circular dependency
|
||||||
|
|||||||
155
src/services/dbResultNormalizer.ts
Normal file
155
src/services/dbResultNormalizer.ts
Normal file
@@ -0,0 +1,155 @@
|
|||||||
|
/**
|
||||||
|
* Database Result Normalizer
|
||||||
|
*
|
||||||
|
* Provides shared logic to normalize database execution results across
|
||||||
|
* platforms, ensuring reliable changes count and last insert row ID.
|
||||||
|
*
|
||||||
|
* This addresses platform-specific inconsistencies where plugins may not
|
||||||
|
* return reliable change counts. The normalizer queries SQLite's connection
|
||||||
|
* state directly when plugin-provided values are missing or unreliable.
|
||||||
|
*
|
||||||
|
* @author Matthew Raymer
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { logger } from "@/utils/logger";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Result from a database run operation
|
||||||
|
*/
|
||||||
|
interface RunResult {
|
||||||
|
changes?: {
|
||||||
|
changes?: number;
|
||||||
|
lastId?: number;
|
||||||
|
};
|
||||||
|
lastId?: number;
|
||||||
|
changes?: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Normalized database execution result
|
||||||
|
*/
|
||||||
|
export interface NormalizedRunResult {
|
||||||
|
changes: number;
|
||||||
|
lastId?: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Query function type for fallback queries
|
||||||
|
* Must use the same database connection to ensure changes() is accurate
|
||||||
|
*/
|
||||||
|
type QueryFunction = (
|
||||||
|
sql: string,
|
||||||
|
params?: unknown[]
|
||||||
|
) => Promise<{
|
||||||
|
values?: Array<Record<string, unknown>>;
|
||||||
|
columns?: string[];
|
||||||
|
}>;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Normalizes a database run result to ensure reliable changes count
|
||||||
|
*
|
||||||
|
* Strategy:
|
||||||
|
* 1. Prefer plugin-provided values if present and numeric
|
||||||
|
* 2. Fall back to querying SQLite connection state (changes(), last_insert_rowid())
|
||||||
|
* 3. Return normalized result with guaranteed numeric changes count
|
||||||
|
*
|
||||||
|
* @param runResult - Raw result from database plugin
|
||||||
|
* @param queryFn - Optional query function for fallback (must use same connection)
|
||||||
|
* @returns Promise resolving to normalized result with reliable changes count
|
||||||
|
*/
|
||||||
|
export async function normalizeRunResult(
|
||||||
|
runResult: RunResult,
|
||||||
|
queryFn?: QueryFunction
|
||||||
|
): Promise<NormalizedRunResult> {
|
||||||
|
let changes = 0;
|
||||||
|
let lastId: number | undefined;
|
||||||
|
|
||||||
|
// Extract plugin-provided values (handle different plugin response shapes)
|
||||||
|
if (runResult.changes?.changes !== undefined) {
|
||||||
|
changes = Number(runResult.changes.changes) || 0;
|
||||||
|
lastId = runResult.changes.lastId
|
||||||
|
? Number(runResult.changes.lastId)
|
||||||
|
: undefined;
|
||||||
|
} else if (runResult.changes !== undefined) {
|
||||||
|
changes = Number(runResult.changes) || 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (runResult.lastId !== undefined && !lastId) {
|
||||||
|
lastId = Number(runResult.lastId);
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we have a query function and changes is 0 (or missing), query SQLite directly
|
||||||
|
// This ensures correctness even if plugin doesn't return reliable counts
|
||||||
|
if (queryFn && (changes === 0 || runResult.changes === undefined)) {
|
||||||
|
try {
|
||||||
|
// Query SQLite's changes() function for the actual number of rows affected
|
||||||
|
// This must use the same connection to get accurate results
|
||||||
|
const changesResult = await queryFn("SELECT changes() AS changes");
|
||||||
|
if (
|
||||||
|
changesResult.values &&
|
||||||
|
changesResult.values.length > 0 &&
|
||||||
|
changesResult.values[0]
|
||||||
|
) {
|
||||||
|
const changesValue = Object.values(changesResult.values[0])[0];
|
||||||
|
if (typeof changesValue === "number") {
|
||||||
|
changes = changesValue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Query last_insert_rowid() for INSERT statements
|
||||||
|
const lastIdResult = await queryFn("SELECT last_insert_rowid() AS lastId");
|
||||||
|
if (
|
||||||
|
lastIdResult.values &&
|
||||||
|
lastIdResult.values.length > 0 &&
|
||||||
|
lastIdResult.values[0]
|
||||||
|
) {
|
||||||
|
const lastIdValue = Object.values(lastIdResult.values[0])[0];
|
||||||
|
if (typeof lastIdValue === "number" && lastIdValue > 0) {
|
||||||
|
lastId = lastIdValue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
// If querying SQLite state fails, log but don't fail the operation
|
||||||
|
// Fall back to plugin-provided values (which may be 0)
|
||||||
|
logger.debug(
|
||||||
|
"[dbResultNormalizer] Failed to query SQLite state, using plugin values:",
|
||||||
|
error
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
changes: Math.max(0, changes), // Ensure non-negative
|
||||||
|
lastId: lastId && lastId > 0 ? lastId : undefined,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Synchronous version that uses provided values only
|
||||||
|
* Use this when query function is not available or not needed
|
||||||
|
*/
|
||||||
|
export function normalizeRunResultSync(
|
||||||
|
runResult: RunResult
|
||||||
|
): NormalizedRunResult {
|
||||||
|
let changes = 0;
|
||||||
|
let lastId: number | undefined;
|
||||||
|
|
||||||
|
if (runResult.changes?.changes !== undefined) {
|
||||||
|
changes = Number(runResult.changes.changes) || 0;
|
||||||
|
lastId = runResult.changes.lastId
|
||||||
|
? Number(runResult.changes.lastId)
|
||||||
|
: undefined;
|
||||||
|
} else if (runResult.changes !== undefined) {
|
||||||
|
changes = Number(runResult.changes) || 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (runResult.lastId !== undefined && !lastId) {
|
||||||
|
lastId = Number(runResult.lastId);
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
changes: Math.max(0, changes),
|
||||||
|
lastId: lastId && lastId > 0 ? lastId : undefined,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
@@ -23,6 +23,8 @@ import {
|
|||||||
} from "../PlatformService";
|
} from "../PlatformService";
|
||||||
import { logger } from "../../utils/logger";
|
import { logger } from "../../utils/logger";
|
||||||
import { BaseDatabaseService } from "./BaseDatabaseService";
|
import { BaseDatabaseService } from "./BaseDatabaseService";
|
||||||
|
import type { PlatformDiagnostics } from "@/interfaces/diagnostics";
|
||||||
|
import { normalizeRunResult } from "../dbResultNormalizer";
|
||||||
|
|
||||||
interface QueuedOperation {
|
interface QueuedOperation {
|
||||||
type: "run" | "query" | "rawQuery";
|
type: "run" | "query" | "rawQuery";
|
||||||
@@ -54,6 +56,8 @@ export class CapacitorPlatformService
|
|||||||
private initializationPromise: Promise<void> | null = null;
|
private initializationPromise: Promise<void> | null = null;
|
||||||
private operationQueue: Array<QueuedOperation> = [];
|
private operationQueue: Array<QueuedOperation> = [];
|
||||||
private isProcessingQueue: boolean = false;
|
private isProcessingQueue: boolean = false;
|
||||||
|
private readonly MAX_QUEUE_SIZE = 1000; // Maximum queue size before failing
|
||||||
|
private maxQueueSizeReached = 0; // Track peak queue size for telemetry
|
||||||
|
|
||||||
constructor() {
|
constructor() {
|
||||||
super();
|
super();
|
||||||
@@ -217,14 +221,34 @@ export class CapacitorPlatformService
|
|||||||
let result: unknown;
|
let result: unknown;
|
||||||
switch (operation.type) {
|
switch (operation.type) {
|
||||||
case "run": {
|
case "run": {
|
||||||
|
// Execute the statement
|
||||||
const runResult = await this.db.run(
|
const runResult = await this.db.run(
|
||||||
operation.sql,
|
operation.sql,
|
||||||
operation.params,
|
operation.params,
|
||||||
);
|
);
|
||||||
result = {
|
|
||||||
changes: runResult.changes?.changes || 0,
|
// Normalize using shared normalizer with query fallback
|
||||||
lastId: runResult.changes?.lastId,
|
// The query function uses the same connection to ensure changes() is accurate
|
||||||
};
|
const normalized = await normalizeRunResult(
|
||||||
|
runResult,
|
||||||
|
async (sql: string, params?: unknown[]) => {
|
||||||
|
const queryResult = await this.db.query(sql, params || []);
|
||||||
|
return {
|
||||||
|
values: queryResult.values?.map((row) => {
|
||||||
|
const obj: Record<string, unknown> = {};
|
||||||
|
if (queryResult.columns) {
|
||||||
|
queryResult.columns.forEach((col, idx) => {
|
||||||
|
obj[col] = row[idx];
|
||||||
|
});
|
||||||
|
}
|
||||||
|
return obj;
|
||||||
|
}),
|
||||||
|
columns: queryResult.columns,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
);
|
||||||
|
|
||||||
|
result = normalized;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case "query": {
|
case "query": {
|
||||||
@@ -371,6 +395,31 @@ export class CapacitorPlatformService
|
|||||||
});
|
});
|
||||||
|
|
||||||
return new Promise<R>((resolve, reject) => {
|
return new Promise<R>((resolve, reject) => {
|
||||||
|
// Queue size guard: prevent memory exhaustion from unbounded queue growth
|
||||||
|
if (this.operationQueue.length >= this.MAX_QUEUE_SIZE) {
|
||||||
|
const error = new Error(
|
||||||
|
`Database operation queue is full (${this.MAX_QUEUE_SIZE} operations). ` +
|
||||||
|
`This usually indicates the database is not initializing properly or operations are being queued too quickly.`
|
||||||
|
);
|
||||||
|
logger.error(
|
||||||
|
`[CapacitorPlatformService] Queue size limit reached: ${this.operationQueue.length}/${this.MAX_QUEUE_SIZE}`,
|
||||||
|
);
|
||||||
|
reject(error);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Track peak queue size for telemetry
|
||||||
|
if (this.operationQueue.length > this.maxQueueSizeReached) {
|
||||||
|
this.maxQueueSizeReached = this.operationQueue.length;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Log warning if queue is getting large (but not at limit yet)
|
||||||
|
if (this.operationQueue.length > this.MAX_QUEUE_SIZE * 0.8) {
|
||||||
|
logger.warn(
|
||||||
|
`[CapacitorPlatformService] Queue size is high: ${this.operationQueue.length}/${this.MAX_QUEUE_SIZE}`,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
// Create completely plain objects that Vue cannot make reactive
|
// Create completely plain objects that Vue cannot make reactive
|
||||||
// Step 1: Deep clone the converted params to ensure they're plain objects
|
// Step 1: Deep clone the converted params to ensure they're plain objects
|
||||||
const plainParams = JSON.parse(JSON.stringify(convertedParams));
|
const plainParams = JSON.parse(JSON.stringify(convertedParams));
|
||||||
@@ -865,6 +914,27 @@ export class CapacitorPlatformService
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets telemetry information about the database operation queue.
|
||||||
|
* Useful for debugging and monitoring queue health.
|
||||||
|
* @returns Queue telemetry data
|
||||||
|
*/
|
||||||
|
getQueueTelemetry(): {
|
||||||
|
currentSize: number;
|
||||||
|
maxSize: number;
|
||||||
|
peakSize: number;
|
||||||
|
isProcessing: boolean;
|
||||||
|
initialized: boolean;
|
||||||
|
} {
|
||||||
|
return {
|
||||||
|
currentSize: this.operationQueue.length,
|
||||||
|
maxSize: this.MAX_QUEUE_SIZE,
|
||||||
|
peakSize: this.maxQueueSizeReached,
|
||||||
|
isProcessing: this.isProcessingQueue,
|
||||||
|
initialized: this.initialized,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Checks and requests storage permissions if needed
|
* Checks and requests storage permissions if needed
|
||||||
* @returns Promise that resolves when permissions are granted
|
* @returns Promise that resolves when permissions are granted
|
||||||
@@ -1409,6 +1479,38 @@ export class CapacitorPlatformService
|
|||||||
// --- PWA/Web-only methods (no-op for Capacitor) ---
|
// --- PWA/Web-only methods (no-op for Capacitor) ---
|
||||||
public registerServiceWorker(): void {}
|
public registerServiceWorker(): void {}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets comprehensive diagnostic information about the Capacitor platform
|
||||||
|
* @returns Promise resolving to platform diagnostics
|
||||||
|
*/
|
||||||
|
async getDiagnostics(): Promise<PlatformDiagnostics> {
|
||||||
|
const platform = Capacitor.getPlatform();
|
||||||
|
const queueTelemetry = this.getQueueTelemetry();
|
||||||
|
|
||||||
|
return {
|
||||||
|
platform: "capacitor",
|
||||||
|
capabilities: this.getCapabilities(),
|
||||||
|
db: {
|
||||||
|
kind: "capacitor-sqlite",
|
||||||
|
queue: {
|
||||||
|
current: queueTelemetry.currentSize,
|
||||||
|
maxReached: queueTelemetry.peakSize,
|
||||||
|
limit: queueTelemetry.maxSize,
|
||||||
|
isProcessing: queueTelemetry.isProcessing,
|
||||||
|
},
|
||||||
|
initialized: this.initialized,
|
||||||
|
},
|
||||||
|
build: {
|
||||||
|
appVersion: import.meta.env.VITE_APP_VERSION,
|
||||||
|
mode: import.meta.env.MODE,
|
||||||
|
gitHash: import.meta.env.VITE_GIT_HASH,
|
||||||
|
},
|
||||||
|
metadata: {
|
||||||
|
nativePlatform: platform,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
// Database utility methods - inherited from BaseDatabaseService
|
// Database utility methods - inherited from BaseDatabaseService
|
||||||
// generateInsertStatement, updateDefaultSettings, updateActiveDid,
|
// generateInsertStatement, updateDefaultSettings, updateActiveDid,
|
||||||
// getActiveIdentity, insertNewDidIntoSettings, updateDidSpecificSettings,
|
// getActiveIdentity, insertNewDidIntoSettings, updateDidSpecificSettings,
|
||||||
|
|||||||
@@ -22,6 +22,7 @@
|
|||||||
|
|
||||||
import { CapacitorPlatformService } from "./CapacitorPlatformService";
|
import { CapacitorPlatformService } from "./CapacitorPlatformService";
|
||||||
import { logger } from "../../utils/logger";
|
import { logger } from "../../utils/logger";
|
||||||
|
import type { PlatformDiagnostics } from "@/interfaces/diagnostics";
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Electron-specific platform service implementation.
|
* Electron-specific platform service implementation.
|
||||||
@@ -164,6 +165,24 @@ export class ElectronPlatformService extends CapacitorPlatformService {
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets comprehensive diagnostic information about the Electron platform
|
||||||
|
* @returns Promise resolving to platform diagnostics
|
||||||
|
*/
|
||||||
|
async getDiagnostics(): Promise<PlatformDiagnostics> {
|
||||||
|
const baseDiagnostics = await super.getDiagnostics();
|
||||||
|
return {
|
||||||
|
...baseDiagnostics,
|
||||||
|
platform: "electron",
|
||||||
|
capabilities: this.getCapabilities(),
|
||||||
|
metadata: {
|
||||||
|
...baseDiagnostics.metadata,
|
||||||
|
electronVersion: process.versions?.electron,
|
||||||
|
nodeVersion: process.versions?.node,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
// --- PWA/Web-only methods (no-op for Electron) ---
|
// --- PWA/Web-only methods (no-op for Electron) ---
|
||||||
public registerServiceWorker(): void {}
|
public registerServiceWorker(): void {}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ import {
|
|||||||
import { logger } from "../../utils/logger";
|
import { logger } from "../../utils/logger";
|
||||||
import { QueryExecResult } from "@/interfaces/database";
|
import { QueryExecResult } from "@/interfaces/database";
|
||||||
import { BaseDatabaseService } from "./BaseDatabaseService";
|
import { BaseDatabaseService } from "./BaseDatabaseService";
|
||||||
|
import type { PlatformDiagnostics } from "@/interfaces/diagnostics";
|
||||||
// Dynamic import of initBackend to prevent worker context errors
|
// Dynamic import of initBackend to prevent worker context errors
|
||||||
import type {
|
import type {
|
||||||
WorkerRequest,
|
WorkerRequest,
|
||||||
@@ -673,6 +674,42 @@ export class WebPlatformService
|
|||||||
// SharedArrayBuffer initialization is handled by initBackend call in initializeWorker
|
// SharedArrayBuffer initialization is handled by initBackend call in initializeWorker
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets comprehensive diagnostic information about the web platform
|
||||||
|
* @returns Promise resolving to platform diagnostics
|
||||||
|
*/
|
||||||
|
async getDiagnostics(): Promise<PlatformDiagnostics> {
|
||||||
|
const platform = (import.meta.env?.VITE_PLATFORM || process.env.VITE_PLATFORM || "web") as string;
|
||||||
|
const sabAvailable = typeof SharedArrayBuffer !== "undefined";
|
||||||
|
|
||||||
|
// Get version from build-time env var if available
|
||||||
|
let version: string | undefined;
|
||||||
|
try {
|
||||||
|
version = import.meta.env.VITE_APP_VERSION;
|
||||||
|
} catch {
|
||||||
|
// Ignore
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
platform,
|
||||||
|
capabilities: this.getCapabilities(),
|
||||||
|
db: {
|
||||||
|
kind: "absurd-sql",
|
||||||
|
sharedArrayBuffer: sabAvailable ? "available" : "fallback",
|
||||||
|
worker: {
|
||||||
|
ready: this.workerReady,
|
||||||
|
pending: this.pendingMessages.size,
|
||||||
|
},
|
||||||
|
initialized: this.workerReady,
|
||||||
|
},
|
||||||
|
build: {
|
||||||
|
appVersion: version,
|
||||||
|
mode: import.meta.env.MODE,
|
||||||
|
gitHash: import.meta.env.VITE_GIT_HASH,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
// Database utility methods - inherited from BaseDatabaseService
|
// Database utility methods - inherited from BaseDatabaseService
|
||||||
// generateInsertStatement, updateDefaultSettings, updateActiveDid,
|
// generateInsertStatement, updateDefaultSettings, updateActiveDid,
|
||||||
// getActiveIdentity, insertNewDidIntoSettings, updateDidSpecificSettings,
|
// getActiveIdentity, insertNewDidIntoSettings, updateDidSpecificSettings,
|
||||||
|
|||||||
269
src/views/debug/PlatformDiagnosticsView.vue
Normal file
269
src/views/debug/PlatformDiagnosticsView.vue
Normal file
@@ -0,0 +1,269 @@
|
|||||||
|
<template>
|
||||||
|
<div class="platform-diagnostics-view p-6 max-w-6xl mx-auto">
|
||||||
|
<div class="mb-6">
|
||||||
|
<h1 class="text-3xl font-bold mb-2">Platform Diagnostics</h1>
|
||||||
|
<p class="text-gray-600">
|
||||||
|
Comprehensive diagnostic information about the current platform, database,
|
||||||
|
and build configuration.
|
||||||
|
</p>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="mb-4 flex gap-4">
|
||||||
|
<button
|
||||||
|
@click="refreshDiagnostics"
|
||||||
|
:disabled="loading"
|
||||||
|
class="px-4 py-2 bg-blue-600 text-white rounded hover:bg-blue-700 disabled:opacity-50"
|
||||||
|
>
|
||||||
|
{{ loading ? "Loading..." : "Refresh" }}
|
||||||
|
</button>
|
||||||
|
<button
|
||||||
|
@click="exportDiagnostics"
|
||||||
|
:disabled="loading || !diagnostics"
|
||||||
|
class="px-4 py-2 bg-green-600 text-white rounded hover:bg-green-700 disabled:opacity-50"
|
||||||
|
>
|
||||||
|
Export Diagnostics Bundle
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div v-if="error" class="mb-4 p-4 bg-red-100 border border-red-400 text-red-700 rounded">
|
||||||
|
{{ error }}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div v-if="diagnostics" class="space-y-6">
|
||||||
|
<!-- Platform Info -->
|
||||||
|
<div class="bg-white rounded-lg shadow p-6">
|
||||||
|
<h2 class="text-xl font-semibold mb-4">Platform Information</h2>
|
||||||
|
<dl class="grid grid-cols-2 gap-4">
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Platform</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.platform }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Build Mode</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.build.mode || "N/A" }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">App Version</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.build.appVersion || "N/A" }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Git Hash</dt>
|
||||||
|
<dd class="text-gray-900 font-mono text-sm">
|
||||||
|
{{ diagnostics.build.gitHash || "N/A" }}
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
</dl>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<!-- Capabilities -->
|
||||||
|
<div class="bg-white rounded-lg shadow p-6">
|
||||||
|
<h2 class="text-xl font-semibold mb-4">Platform Capabilities</h2>
|
||||||
|
<dl class="grid grid-cols-2 gap-4">
|
||||||
|
<div v-for="(value, key) in diagnostics.capabilities" :key="key">
|
||||||
|
<dt class="font-medium text-gray-700">{{ formatKey(key) }}</dt>
|
||||||
|
<dd class="text-gray-900">
|
||||||
|
<span
|
||||||
|
:class="
|
||||||
|
value
|
||||||
|
? 'text-green-600 font-semibold'
|
||||||
|
: 'text-gray-400'
|
||||||
|
"
|
||||||
|
>
|
||||||
|
{{ value ? "✓ Yes" : "✗ No" }}
|
||||||
|
</span>
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
</dl>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<!-- Database Info -->
|
||||||
|
<div class="bg-white rounded-lg shadow p-6">
|
||||||
|
<h2 class="text-xl font-semibold mb-4">Database Information</h2>
|
||||||
|
<dl class="grid grid-cols-2 gap-4">
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Backend Type</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.kind }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Initialized</dt>
|
||||||
|
<dd class="text-gray-900">
|
||||||
|
<span
|
||||||
|
:class="
|
||||||
|
diagnostics.db.initialized
|
||||||
|
? 'text-green-600 font-semibold'
|
||||||
|
: 'text-red-600 font-semibold'
|
||||||
|
"
|
||||||
|
>
|
||||||
|
{{ diagnostics.db.initialized ? "✓ Yes" : "✗ No" }}
|
||||||
|
</span>
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
<div v-if="diagnostics.db.sharedArrayBuffer">
|
||||||
|
<dt class="font-medium text-gray-700">SharedArrayBuffer</dt>
|
||||||
|
<dd class="text-gray-900">
|
||||||
|
<span
|
||||||
|
:class="
|
||||||
|
diagnostics.db.sharedArrayBuffer === 'available'
|
||||||
|
? 'text-green-600 font-semibold'
|
||||||
|
: 'text-yellow-600 font-semibold'
|
||||||
|
"
|
||||||
|
>
|
||||||
|
{{ diagnostics.db.sharedArrayBuffer }}
|
||||||
|
</span>
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
</dl>
|
||||||
|
|
||||||
|
<!-- Worker Status (Web Platform) -->
|
||||||
|
<div v-if="diagnostics.db.worker" class="mt-4 pt-4 border-t">
|
||||||
|
<h3 class="font-semibold mb-2">Worker Status</h3>
|
||||||
|
<dl class="grid grid-cols-2 gap-4">
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Ready</dt>
|
||||||
|
<dd class="text-gray-900">
|
||||||
|
<span
|
||||||
|
:class="
|
||||||
|
diagnostics.db.worker.ready
|
||||||
|
? 'text-green-600 font-semibold'
|
||||||
|
: 'text-red-600 font-semibold'
|
||||||
|
"
|
||||||
|
>
|
||||||
|
{{ diagnostics.db.worker.ready ? "✓ Yes" : "✗ No" }}
|
||||||
|
</span>
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Pending Messages</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.worker.pending }}</dd>
|
||||||
|
</div>
|
||||||
|
<div v-if="diagnostics.db.worker.lastPingMs">
|
||||||
|
<dt class="font-medium text-gray-700">Last Ping</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.worker.lastPingMs }}ms ago</dd>
|
||||||
|
</div>
|
||||||
|
</dl>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<!-- Queue Status (Capacitor/Electron) -->
|
||||||
|
<div v-if="diagnostics.db.queue" class="mt-4 pt-4 border-t">
|
||||||
|
<h3 class="font-semibold mb-2">Operation Queue</h3>
|
||||||
|
<dl class="grid grid-cols-2 gap-4">
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Current Size</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.queue.current }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Limit</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.queue.limit }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Peak Reached</dt>
|
||||||
|
<dd class="text-gray-900">{{ diagnostics.db.queue.maxReached }}</dd>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<dt class="font-medium text-gray-700">Processing</dt>
|
||||||
|
<dd class="text-gray-900">
|
||||||
|
<span
|
||||||
|
:class="
|
||||||
|
diagnostics.db.queue.isProcessing
|
||||||
|
? 'text-green-600 font-semibold'
|
||||||
|
: 'text-gray-400'
|
||||||
|
"
|
||||||
|
>
|
||||||
|
{{ diagnostics.db.queue.isProcessing ? "✓ Yes" : "✗ No" }}
|
||||||
|
</span>
|
||||||
|
</dd>
|
||||||
|
</div>
|
||||||
|
</dl>
|
||||||
|
<div class="mt-2">
|
||||||
|
<div class="w-full bg-gray-200 rounded-full h-2">
|
||||||
|
<div
|
||||||
|
class="bg-blue-600 h-2 rounded-full transition-all"
|
||||||
|
:style="{
|
||||||
|
width: `${Math.min(
|
||||||
|
(diagnostics.db.queue.current / diagnostics.db.queue.limit) * 100,
|
||||||
|
100
|
||||||
|
)}%`,
|
||||||
|
}"
|
||||||
|
></div>
|
||||||
|
</div>
|
||||||
|
<p class="text-xs text-gray-500 mt-1">
|
||||||
|
{{ Math.round((diagnostics.db.queue.current / diagnostics.db.queue.limit) * 100) }}%
|
||||||
|
capacity
|
||||||
|
</p>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<!-- Metadata -->
|
||||||
|
<div v-if="diagnostics.metadata && Object.keys(diagnostics.metadata).length > 0" class="bg-white rounded-lg shadow p-6">
|
||||||
|
<h2 class="text-xl font-semibold mb-4">Additional Metadata</h2>
|
||||||
|
<pre class="bg-gray-100 p-4 rounded text-sm overflow-auto">{{ JSON.stringify(diagnostics.metadata, null, 2) }}</pre>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div v-else-if="!loading" class="text-center py-12 text-gray-500">
|
||||||
|
No diagnostics available. Click "Refresh" to load.
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</template>
|
||||||
|
|
||||||
|
<script setup lang="ts">
|
||||||
|
import { ref, onMounted } from "vue";
|
||||||
|
import { PlatformServiceFactory } from "@/services/PlatformServiceFactory";
|
||||||
|
import { exportDiagnosticsToFile } from "@/services/DiagnosticExportService";
|
||||||
|
import type { PlatformDiagnostics } from "@/interfaces/diagnostics";
|
||||||
|
import { logger } from "@/utils/logger";
|
||||||
|
|
||||||
|
const diagnostics = ref<PlatformDiagnostics | null>(null);
|
||||||
|
const loading = ref(false);
|
||||||
|
const error = ref<string | null>(null);
|
||||||
|
|
||||||
|
const formatKey = (key: string): string => {
|
||||||
|
return key
|
||||||
|
.replace(/([A-Z])/g, " $1")
|
||||||
|
.replace(/^./, (str) => str.toUpperCase())
|
||||||
|
.trim();
|
||||||
|
};
|
||||||
|
|
||||||
|
const refreshDiagnostics = async () => {
|
||||||
|
loading.value = true;
|
||||||
|
error.value = null;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const platform = PlatformServiceFactory.getInstance();
|
||||||
|
if (platform.getDiagnostics) {
|
||||||
|
diagnostics.value = await platform.getDiagnostics();
|
||||||
|
logger.debug("[PlatformDiagnosticsView] Diagnostics refreshed", diagnostics.value);
|
||||||
|
} else {
|
||||||
|
error.value = "Diagnostics not available on this platform";
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
error.value = `Failed to load diagnostics: ${err}`;
|
||||||
|
logger.error("[PlatformDiagnosticsView] Failed to refresh diagnostics:", err);
|
||||||
|
} finally {
|
||||||
|
loading.value = false;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
const exportDiagnostics = async () => {
|
||||||
|
try {
|
||||||
|
await exportDiagnosticsToFile();
|
||||||
|
logger.log("[PlatformDiagnosticsView] Diagnostics exported successfully");
|
||||||
|
} catch (err) {
|
||||||
|
error.value = `Failed to export diagnostics: ${err}`;
|
||||||
|
logger.error("[PlatformDiagnosticsView] Failed to export diagnostics:", err);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
onMounted(() => {
|
||||||
|
refreshDiagnostics();
|
||||||
|
});
|
||||||
|
</script>
|
||||||
|
|
||||||
|
<style scoped>
|
||||||
|
.platform-diagnostics-view {
|
||||||
|
min-height: 100vh;
|
||||||
|
background-color: #f5f5f5;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
|
||||||
@@ -4,6 +4,7 @@ import dotenv from "dotenv";
|
|||||||
import { loadAppConfig } from "./vite.config.utils.mts";
|
import { loadAppConfig } from "./vite.config.utils.mts";
|
||||||
import path from "path";
|
import path from "path";
|
||||||
import { fileURLToPath } from 'url';
|
import { fileURLToPath } from 'url';
|
||||||
|
import { execSync } from "child_process";
|
||||||
|
|
||||||
// Load environment variables
|
// Load environment variables
|
||||||
dotenv.config({ path: `.env.${process.env.NODE_ENV}` })
|
dotenv.config({ path: `.env.${process.env.NODE_ENV}` })
|
||||||
@@ -20,6 +21,15 @@ export async function createBuildConfig(platform: string): Promise<UserConfig> {
|
|||||||
// Set platform - PWA is always enabled for web platforms
|
// Set platform - PWA is always enabled for web platforms
|
||||||
process.env.VITE_PLATFORM = platform;
|
process.env.VITE_PLATFORM = platform;
|
||||||
|
|
||||||
|
// Get git commit hash for build info (fallback to empty if git not available)
|
||||||
|
let gitHash = "";
|
||||||
|
try {
|
||||||
|
gitHash = execSync("git rev-parse --short HEAD", { encoding: "utf-8" }).trim();
|
||||||
|
} catch (error) {
|
||||||
|
// Git not available or not a git repo - use empty string
|
||||||
|
gitHash = "";
|
||||||
|
}
|
||||||
|
|
||||||
// Environment variables are loaded from .env files via dotenv.config() above
|
// Environment variables are loaded from .env files via dotenv.config() above
|
||||||
|
|
||||||
return {
|
return {
|
||||||
@@ -71,6 +81,8 @@ export async function createBuildConfig(platform: string): Promise<UserConfig> {
|
|||||||
'process.env.NODE_ENV': JSON.stringify(process.env.NODE_ENV),
|
'process.env.NODE_ENV': JSON.stringify(process.env.NODE_ENV),
|
||||||
'process.env.VITE_PLATFORM': JSON.stringify(platform),
|
'process.env.VITE_PLATFORM': JSON.stringify(platform),
|
||||||
'process.env.VITE_LOG_LEVEL': JSON.stringify(process.env.VITE_LOG_LEVEL),
|
'process.env.VITE_LOG_LEVEL': JSON.stringify(process.env.VITE_LOG_LEVEL),
|
||||||
|
'import.meta.env.VITE_GIT_HASH': JSON.stringify(gitHash),
|
||||||
|
'process.env.VITE_GIT_HASH': JSON.stringify(gitHash),
|
||||||
// PWA is always enabled for web platforms
|
// PWA is always enabled for web platforms
|
||||||
__dirname: JSON.stringify(process.cwd()),
|
__dirname: JSON.stringify(process.cwd()),
|
||||||
__IS_MOBILE__: JSON.stringify(isCapacitor),
|
__IS_MOBILE__: JSON.stringify(isCapacitor),
|
||||||
|
|||||||
Reference in New Issue
Block a user