52 Commits
1.1.5 ... dev

Author SHA1 Message Date
William Oldham
f5cec7ba24 Merge pull request #40 from qtchaos/index-OK
Add a success message to the base path
2024-02-26 12:41:12 +00:00
qtchaos
db00030df0 feat: add a success message to the base path 2024-02-26 14:13:59 +02:00
mrjvs
2fe48d24d7 Merge branch 'master' into dev 2024-01-25 22:33:24 +01:00
William Oldham
b0f5b28084 Merge pull request #38 from movie-web/scrapetools
Add tool metrics
2024-01-25 21:29:24 +00:00
mrjvs
b5ee9f1f7a Merge branch 'dev' into scrapetools 2024-01-25 22:28:33 +01:00
mrjvs
eb0f47ba53 Bump version 2024-01-25 22:28:08 +01:00
mrjvs
26b46876c5 fix linting 2024-01-25 22:27:54 +01:00
mrjvs
f58c2c86e5 Merge pull request #37 from gsi-kevincarrera/fix/backend-always-return-true
Fix: Implement function to correctly handle boolean strings with zod
2024-01-25 22:27:15 +01:00
mrjvs
4af2d32b72 Add tool metrics 2024-01-25 22:24:58 +01:00
Kevin Carrera Calzado
a911e52ddb Add default false value to boolean schema 2024-01-24 12:54:24 -05:00
Kevin Carrera Calzado
7b26b029de Fix:Implement function to handle boolean strings 2024-01-24 12:34:01 -05:00
William Oldham
06ad2249d6 Merge pull request #35 from movie-web/remove-arm
Remove ARM 32-bit from Docker Build
2024-01-09 21:36:01 +00:00
William Oldham
14601b9010 Remove ARM 32-bit from Docker Build 2024-01-09 21:01:52 +00:00
William Oldham
b1eeb21ba4 Merge pull request #34 from weeryan17/feature/arm-build
Update workflows to support arm
2024-01-07 17:58:15 +00:00
weeryan17
5855943a56 Update .github/workflows/linting_testing.yml
Co-authored-by: William Oldham <github@binaryoverload.co.uk>
2024-01-07 12:55:30 -05:00
weeryan17
9073010f4c Update .github/workflows/linting_testing.yml
Co-authored-by: William Oldham <github@binaryoverload.co.uk>
2024-01-07 12:55:24 -05:00
weeryan17
78e6b6443f Update CODEOWNERS file 2024-01-07 12:51:09 -05:00
weeryan17
e7c5d93cc3 Update workflows to support arm 2024-01-07 12:42:45 -05:00
mrjvs
9fe8cb4877 Merge pull request #31 from Caio-Nogueira/fix-backend-captcha
update deprecated composer syntax; change body format in API request
2024-01-05 21:04:32 +01:00
mrjvs
66840d7894 Merge branch 'dev' into fix-backend-captcha 2024-01-05 20:19:26 +01:00
William Oldham
befa173445 Merge pull request #32 from movie-web/dev
Version 1.3.0 backend
2024-01-02 21:45:53 +00:00
mrjvs
1c1d70fa05 Merge branch 'master' into dev 2024-01-02 22:41:31 +01:00
mrjvs
6b83fcd158 Merge branch 'dev' of https://github.com/movie-web/backend into dev 2024-01-02 22:40:32 +01:00
mrjvs
f8870b5929 Bump version 2024-01-02 22:40:15 +01:00
mrjvs
727a9cdd43 Merge pull request #28 from qtchaos/proxy-syncing
Add proxyUrls column to UserSettings model
2024-01-01 19:04:34 +01:00
mrjvs
206844e758 Merge branch 'dev' into proxy-syncing 2024-01-01 19:03:43 +01:00
qtchaos
b2e0841d0b Swap to using ArrayType 2024-01-01 19:46:46 +02:00
mrjvs
886df2ffb4 Merge pull request #29 from qtchaos/ssl
Add config support for SSL postgres connections.
2024-01-01 18:09:39 +01:00
mrjvs
6ea07ea685 Merge branch 'dev' into ssl 2024-01-01 18:06:13 +01:00
mrjvs
1010eaf71e Merge branch 'dev' into proxy-syncing 2024-01-01 18:04:59 +01:00
mrjvs
1632f138b0 Merge branch 'dev' into fix-backend-captcha 2023-12-31 14:10:18 +01:00
caio_nogueira_27
0c3de831fb update deprecated composer syntax; change body format in API request 2023-12-31 13:03:32 +00:00
mrjvs
ebe00ecbd6 Merge pull request #30 from qtchaos/pnpm
Swap to using `pnpm exec` instead of `npx` in scripts
2023-12-31 12:59:16 +01:00
qtchaos
3bb427401f Update package.json scripts to use pnpm 2023-12-30 14:07:38 +02:00
qtchaos
b2598e3d82 Add SSL support for PostgreSQL connection 2023-12-30 01:09:04 +02:00
qtchaos
bf55be2978 Add proxyUrls column to UserSettings model 2023-12-29 23:56:25 +02:00
William Oldham
8f52dad296 Merge pull request #27 from movie-web/dev
Register Captcha Solves Metric
2023-12-21 21:06:09 +00:00
William Oldham
9a27b0b0b4 Merge branch 'master' into dev 2023-12-21 21:04:00 +00:00
William Oldham
46baac466d Register Captcha Solves Metric 2023-12-21 21:03:18 +00:00
William Oldham
26316c7755 Merge pull request #26 from movie-web/dev
Backend v1.2.0
2023-12-21 20:21:44 +00:00
mrjvs
d5851b522a Merge branch 'master' into dev 2023-12-21 21:19:26 +01:00
mrjvs
7145d6c246 Bump version 2023-12-21 21:17:32 +01:00
William Oldham
961ac7eeca Merge pull request #25 from movie-web/fix-stuff
Ratelimit fix & optimize metrics
2023-12-21 20:11:10 +00:00
mrjvs
59ab9b48bd Implement trusted cloudflare ips
Co-authored-by: William Oldham <github@binaryoverload.co.uk>
2023-12-21 20:38:51 +01:00
mrjvs
07ecd445f9 Add captcha solves metric 2023-12-21 20:11:17 +01:00
mrjvs
5ebecd1476 Remove provider metric table and replace prometheus metrics with something more efficient
Co-authored-by: William Oldham <github@binaryoverload.co.uk>
2023-12-21 20:01:27 +01:00
William Oldham
f4e9a96666 Merge pull request #24 from movie-web/dev
Update bugfix
2023-12-14 20:48:14 +00:00
William Oldham
60dda8ac57 Merge branch 'master' into dev 2023-12-14 20:48:04 +00:00
William Oldham
d84cdc4239 Merge pull request #23 from movie-web/fix-show-delete
Fix show deletion
2023-12-14 20:47:40 +00:00
mrjvs
c2cba27e68 remove unused import 2023-12-14 21:10:48 +01:00
mrjvs
c3259156ac bump version 2023-12-14 21:07:38 +01:00
mrjvs
9ef12d1c0f Delete all entries of a show 2023-12-14 21:07:24 +01:00
29 changed files with 277 additions and 330 deletions

View File

@@ -12,6 +12,8 @@ services:
POSTGRES_PASSWORD: postgres
volumes:
- 'postgres_data:/var/lib/postgresql/data'
redis:
image: redis
# custom services
backend:
@@ -39,7 +41,7 @@ services:
links:
- postgres:postgres
environment:
- DATABASE_URL=postgres://postgres:postgres@postgres:5432/postgres?sslmode=disable
- PGWEB_DATABASE_URL=postgres://postgres:postgres@postgres:5432/postgres?sslmode=disable
depends_on:
- postgres

4
.github/CODEOWNERS vendored
View File

@@ -1,3 +1 @@
* @movie-web/core
.github @binaryoverload
* @movie-web/project-leads

View File

@@ -14,16 +14,16 @@ jobs:
steps:
- name: Checkout code
uses: actions/checkout@v3
uses: actions/checkout@v4
- uses: pnpm/action-setup@v2
with:
version: 8
- name: Install Node.js
uses: actions/setup-node@v3
uses: actions/setup-node@v4
with:
node-version: 18
node-version: 20
cache: 'pnpm'
- name: Install packages
@@ -38,16 +38,16 @@ jobs:
steps:
- name: Checkout code
uses: actions/checkout@v3
uses: actions/checkout@v4
- uses: pnpm/action-setup@v2
with:
version: 8
- name: Install Node.js
uses: actions/setup-node@v3
uses: actions/setup-node@v4
with:
node-version: 18
node-version: 20
cache: 'pnpm'
- name: Install packages
@@ -62,10 +62,10 @@ jobs:
steps:
- name: Checkout code
uses: actions/checkout@v3
uses: actions/checkout@v4
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v2
uses: docker/setup-buildx-action@v3
- name: Build
uses: docker/build-push-action@v4
uses: docker/build-push-action@v5

View File

@@ -12,7 +12,7 @@ jobs:
steps:
- name: Checkout code
uses: actions/checkout@v3
uses: actions/checkout@v4
- name: Get version
id: package-version
@@ -42,10 +42,10 @@ jobs:
steps:
- name: Checkout repository
uses: actions/checkout@v3
uses: actions/checkout@v4
- name: Setup Docker buildx
uses: docker/setup-buildx-action@v2
uses: docker/setup-buildx-action@v3
- name: Get version
id: package-version
@@ -70,9 +70,12 @@ jobs:
- name: Build and push Docker image
id: build-and-push
uses: docker/build-push-action@v4
uses: docker/build-push-action@v5
with:
push: true
platforms: linux/amd64,linux/arm64
context: .
labels: ${{ steps.meta.outputs.labels }}
tags: ${{ steps.meta.outputs.tags }}
cache-from: type=gha
cache-to: type=gha,mode=max

View File

@@ -1,6 +1,6 @@
{
"name": "backend",
"version": "1.1.5",
"version": "1.3.1",
"private": true,
"homepage": "https://github.com/movie-web/backend",
"engines": {
@@ -8,16 +8,16 @@
},
"scripts": {
"dev": "nodemon -r tsconfig-paths/register src/main.ts",
"build": "npm run build:pre && npm run build:compile",
"build": "pnpm run build:pre && pnpm run build:compile",
"start": "node dist/main.js",
"lint": "eslint --ext .ts,.js,.json,.tsx src/",
"lint:fix": "eslint --fix --ext .ts,.js,.json,.tsx src/",
"build:pre": "rimraf dist/",
"build:compile": "tsc && tsc-alias",
"preinstall": "npx -y only-allow pnpm",
"migration:create": "npx -y mikro-orm migration:create",
"migration:up": "npx -y mikro-orm migration:up",
"migration:down": "npx -y mikro-orm migration:down"
"migration:create": "pnpm exec mikro-orm migration:create",
"migration:up": "pnpm exec mikro-orm migration:up",
"migration:down": "pnpm exec mikro-orm migration:down"
},
"mikro-orm": {
"useTsNode": true,

View File

@@ -4,4 +4,8 @@ export const dockerFragment: FragmentSchema = {
postgres: {
connection: 'postgres://postgres:postgres@postgres:5432/postgres',
},
ratelimits: {
enabled: true,
redisUrl: 'redis://redis:6379',
},
};

View File

@@ -1,10 +1,20 @@
import { devFragment } from '@/config/fragments/dev';
import { dockerFragment } from '@/config/fragments/docker';
import { createConfigLoader } from 'neat-config';
import { z } from 'zod';
import { booleanSchema } from './schema';
const fragments = {
dev: devFragment,
dockerdev: dockerFragment,
};
export const ormConfigSchema = z.object({
postgres: z.object({
// connection URL for postgres database
connection: z.string(),
// whether to use SSL for the connection
ssl: booleanSchema.default(false),
}),
});
@@ -15,6 +25,8 @@ export const ormConf = createConfigLoader()
prefix: 'MWB_',
})
.addFromFile('config.json')
.setFragmentKey('usePresets')
.addConfigFragments(fragments)
.addZodSchema(ormConfigSchema)
.freeze()
.load();

View File

@@ -1,5 +1,7 @@
import { z } from 'zod';
export const booleanSchema = z.preprocess((val) => val === 'true', z.boolean());
export const configSchema = z.object({
server: z
.object({
@@ -11,10 +13,13 @@ export const configSchema = z.object({
// disable cross origin restrictions, allow any site.
// overwrites the cors option above
allowAnySite: z.coerce.boolean().default(false),
allowAnySite: booleanSchema.default(false),
// should it trust reverse proxy headers? (for ip gathering)
trustProxy: z.coerce.boolean().default(false),
trustProxy: booleanSchema.default(false),
// should it trust cloudflare headers? (for ip gathering, cloudflare has priority)
trustCloudflare: booleanSchema.default(false),
// prefix for where the instance is run on. for example set it to /backend if you're hosting it on example.com/backend
// if this is set, do not apply url rewriting before proxing
@@ -27,7 +32,7 @@ export const configSchema = z.object({
format: z.enum(['json', 'pretty']).default('pretty'),
// show debug logs?
debug: z.coerce.boolean().default(false),
debug: booleanSchema.default(false),
})
.default({}),
postgres: z.object({
@@ -35,16 +40,19 @@ export const configSchema = z.object({
connection: z.string(),
// run all migrations on boot of the application
migrateOnBoot: z.coerce.boolean().default(false),
migrateOnBoot: booleanSchema.default(false),
// try to sync the schema on boot, useful for development
// will always keep the database schema in sync with the connected database
// it is extremely destructive, do not use it EVER in production
syncSchema: z.coerce.boolean().default(false),
syncSchema: booleanSchema.default(false),
// Enable debug logging for MikroORM - Outputs queries and entity management logs
// Do NOT use in production, leaks all sensitive data
debugLogging: z.coerce.boolean().default(false),
debugLogging: booleanSchema.default(false),
// Enable SSL for the postgres connection
ssl: booleanSchema.default(false),
}),
crypto: z.object({
// session secret. used for signing session tokens
@@ -59,7 +67,7 @@ export const configSchema = z.object({
captcha: z
.object({
// enabled captchas on register
enabled: z.coerce.boolean().default(false),
enabled: booleanSchema.default(false),
// captcha secret
secret: z.string().min(1).optional(),
@@ -70,7 +78,7 @@ export const configSchema = z.object({
ratelimits: z
.object({
// enabled captchas on register
enabled: z.coerce.boolean().default(false),
enabled: booleanSchema.default(false),
redisUrl: z.string().optional(),
})
.default({}),

View File

@@ -268,143 +268,6 @@
"checks": [],
"foreignKeys": {}
},
{
"columns": {
"id": {
"name": "id",
"type": "uuid",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "uuid"
},
"tmdb_id": {
"name": "tmdb_id",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
},
"type": {
"name": "type",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
},
"title": {
"name": "title",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
},
"season_id": {
"name": "season_id",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "string"
},
"episode_id": {
"name": "episode_id",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "string"
},
"created_at": {
"name": "created_at",
"type": "timestamptz(0)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"length": 0,
"mappedType": "datetime"
},
"status": {
"name": "status",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
},
"provider_id": {
"name": "provider_id",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
},
"embed_id": {
"name": "embed_id",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "string"
},
"error_message": {
"name": "error_message",
"type": "text",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "text"
},
"full_error": {
"name": "full_error",
"type": "text",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "text"
},
"hostname": {
"name": "hostname",
"type": "varchar(255)",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": false,
"mappedType": "string"
}
},
"name": "provider_metrics",
"schema": "public",
"indexes": [
{
"keyName": "provider_metrics_pkey",
"columnNames": [
"id"
],
"composite": false,
"primary": true,
"unique": true
}
],
"checks": [],
"foreignKeys": {}
},
{
"columns": {
"id": {
@@ -620,6 +483,15 @@
"primary": false,
"nullable": true,
"mappedType": "string"
},
"proxy_urls": {
"name": "proxy_urls",
"type": "text[]",
"unsigned": false,
"autoincrement": false,
"primary": false,
"nullable": true,
"mappedType": "array"
}
},
"name": "user_settings",

View File

@@ -0,0 +1,13 @@
import { Migration } from '@mikro-orm/migrations';
export class Migration20231221185725 extends Migration {
async up(): Promise<void> {
this.addSql('drop table if exists "provider_metrics" cascade;');
}
async down(): Promise<void> {
this.addSql('create table "provider_metrics" ("id" uuid not null default null, "tmdb_id" varchar not null default null, "type" varchar not null default null, "title" varchar not null default null, "season_id" varchar null default null, "episode_id" varchar null default null, "created_at" timestamptz not null default null, "status" varchar not null default null, "provider_id" varchar not null default null, "embed_id" varchar null default null, "error_message" text null default null, "full_error" text null default null, "hostname" varchar not null default null, constraint "provider_metrics_pkey" primary key ("id"));');
}
}

View File

@@ -0,0 +1,13 @@
import { Migration } from '@mikro-orm/migrations';
export class Migration20231229214215 extends Migration {
async up(): Promise<void> {
this.addSql('alter table "user_settings" add column "proxy_urls" text[] null;');
}
async down(): Promise<void> {
this.addSql('alter table "user_settings" drop column "proxy_urls";');
}
}

View File

@@ -1,51 +0,0 @@
import { Entity, PrimaryKey, Property } from '@mikro-orm/core';
import { randomUUID } from 'crypto';
export const status = {
failed: 'failed',
notfound: 'notfound',
success: 'success',
} as const;
type Status = keyof typeof status;
@Entity({ tableName: 'provider_metrics' })
export class ProviderMetric {
@PrimaryKey({ name: 'id', type: 'uuid' })
id: string = randomUUID();
@Property({ name: 'tmdb_id' })
tmdbId!: string;
@Property({ name: 'type' })
type!: string;
@Property({ name: 'title' })
title!: string;
@Property({ name: 'season_id', nullable: true })
seasonId?: string;
@Property({ name: 'episode_id', nullable: true })
episodeId?: string;
@Property({ name: 'created_at', type: 'date' })
createdAt = new Date();
@Property({ name: 'status' })
status!: Status;
@Property({ name: 'provider_id' })
providerId!: string;
@Property({ name: 'embed_id', nullable: true })
embedId?: string;
@Property({ name: 'error_message', nullable: true, type: 'text' })
errorMessage?: string;
@Property({ name: 'full_error', nullable: true, type: 'text' })
fullError?: string;
@Property({ name: 'hostname' })
hostname!: string;
}

View File

@@ -1,4 +1,4 @@
import { Entity, PrimaryKey, Property } from '@mikro-orm/core';
import { ArrayType, Entity, PrimaryKey, Property } from '@mikro-orm/core';
@Entity({ tableName: 'user_settings' })
export class UserSettings {
@@ -13,6 +13,9 @@ export class UserSettings {
@Property({ name: 'default_subtitle_language', nullable: true })
defaultSubtitleLanguage?: string | null;
@Property({ name: 'proxy_urls', type: ArrayType, nullable: true })
proxyUrls?: string[] | null;
}
export interface UserSettingsDTO {
@@ -20,6 +23,7 @@ export interface UserSettingsDTO {
applicationTheme?: string | null;
applicationLanguage?: string | null;
defaultSubtitleLanguage?: string | null;
proxyUrls?: string[] | null;
}
export function formatUserSettings(
@@ -30,5 +34,6 @@ export function formatUserSettings(
applicationTheme: userSettings.applicationTheme,
applicationLanguage: userSettings.applicationLanguage,
defaultSubtitleLanguage: userSettings.defaultSubtitleLanguage,
proxyUrls: userSettings.proxyUrls,
};
}

View File

@@ -1,4 +1,4 @@
import { ormConf } from '@/config/orm';
import { makeOrmConfig } from '@/modules/mikro/orm';
export default makeOrmConfig(ormConf.postgres.connection);
export default makeOrmConfig(ormConf.postgres.connection, ormConf.postgres.ssl);

View File

@@ -1,3 +1,4 @@
import { indexRouter } from '@/routes';
import { loginAuthRouter } from '@/routes/auth/login';
import { manageAuthRouter } from '@/routes/auth/manage';
import { metaRouter } from '@/routes/meta';
@@ -25,4 +26,5 @@ export async function setupRoutes(app: FastifyInstance) {
await app.register(userSettingsRouter.register);
await app.register(userGetRouter.register);
await app.register(metricsRouter.register);
await app.register(indexRouter.register);
}

View File

@@ -1,11 +1,9 @@
import { challengeCodeJob } from '@/modules/jobs/list/challengeCode';
import { sessionExpiryJob } from '@/modules/jobs/list/sessionExpiry';
import { userDeletionJob } from '@/modules/jobs/list/userDeletion';
import { providerMetricCleanupJob } from '@/modules/jobs/list/providerMetricCleanup';
export async function setupJobs() {
challengeCodeJob.start();
sessionExpiryJob.start();
userDeletionJob.start();
providerMetricCleanupJob.start();
}

View File

@@ -1,27 +0,0 @@
import { ProviderMetric } from '@/db/models/ProviderMetrics';
import { job } from '@/modules/jobs/job';
import ms from 'ms';
// every day at 12:00:00
export const providerMetricCleanupJob = job(
'provider-metric-cleanup',
'0 12 * * *',
async ({ em, log }) => {
const now = new Date();
const thirtyDaysAgo = new Date(now.getTime() - ms('30d'));
const deletedMetrics = await em
.createQueryBuilder(ProviderMetric)
.delete()
.where({
createdAt: {
$lt: thirtyDaysAgo,
},
})
.execute<{ affectedRows: number }>('run');
log.info(
`Removed ${deletedMetrics.affectedRows} metrics that were older than 30 days`,
);
},
);

View File

@@ -9,17 +9,11 @@ const log = scopedLogger('metrics');
export type Metrics = {
user: Counter<'namespace'>;
providerMetrics: Counter<
| 'title'
| 'tmdb_id'
| 'season_id'
| 'episode_id'
| 'status'
| 'type'
| 'provider_id'
| 'embed_id'
| 'hostname'
>;
captchaSolves: Counter<'success'>;
providerHostnames: Counter<'hostname'>;
providerStatuses: Counter<'provider_id' | 'status'>;
watchMetrics: Counter<'title' | 'tmdb_full_id' | 'provider_id' | 'success'>;
toolMetrics: Counter<'tool'>;
};
let metrics: null | Metrics = null;
@@ -42,31 +36,45 @@ export async function setupMetrics(app: FastifyInstance) {
metrics = {
user: new Counter({
name: 'user_count',
help: 'user_help',
name: 'mw_user_count',
help: 'mw_user_help',
labelNames: ['namespace'],
}),
providerMetrics: new Counter({
name: 'provider_metrics',
help: 'provider_metrics',
labelNames: [
'episode_id',
'provider_id',
'season_id',
'status',
'title',
'tmdb_id',
'type',
'embed_id',
'hostname',
],
captchaSolves: new Counter({
name: 'mw_captcha_solves',
help: 'mw_captcha_solves',
labelNames: ['success'],
}),
providerHostnames: new Counter({
name: 'mw_provider_hostname_count',
help: 'mw_provider_hostname_count',
labelNames: ['hostname'],
}),
providerStatuses: new Counter({
name: 'mw_provider_status_count',
help: 'mw_provider_status_count',
labelNames: ['provider_id', 'status'],
}),
watchMetrics: new Counter({
name: 'mw_media_watch_count',
help: 'mw_media_watch_count',
labelNames: ['title', 'tmdb_full_id', 'provider_id', 'success'],
}),
toolMetrics: new Counter({
name: 'mw_provider_tool_count',
help: 'mw_provider_tool_count',
labelNames: ['tool'],
}),
};
const promClient = app.metrics.client;
promClient.register.registerMetric(metrics.user);
promClient.register.registerMetric(metrics.providerMetrics);
promClient.register.registerMetric(metrics.providerHostnames);
promClient.register.registerMetric(metrics.providerStatuses);
promClient.register.registerMetric(metrics.watchMetrics);
promClient.register.registerMetric(metrics.captchaSolves);
promClient.register.registerMetric(metrics.toolMetrics);
const orm = getORM();
const em = orm.em.fork();

View File

@@ -18,6 +18,7 @@ export async function setupMikroORM() {
conf.postgres.connection,
conf.postgres.debugLogging,
(msg) => log.info(msg),
conf.postgres.ssl,
);
if (conf.postgres.syncSchema) {

View File

@@ -2,7 +2,10 @@ import { Options } from '@mikro-orm/core';
import { MikroORM, PostgreSqlDriver } from '@mikro-orm/postgresql';
import path from 'path';
export function makeOrmConfig(url: string): Options<PostgreSqlDriver> {
export function makeOrmConfig(
url: string,
ssl: boolean,
): Options<PostgreSqlDriver> {
return {
type: 'postgresql',
clientUrl: url,
@@ -13,6 +16,11 @@ export function makeOrmConfig(url: string): Options<PostgreSqlDriver> {
pathTs: './migrations',
path: './migrations',
},
driverOptions: {
connection: {
ssl,
},
},
};
}
@@ -20,9 +28,10 @@ export async function createORM(
url: string,
debug: boolean,
log: (msg: string) => void,
ssl: boolean,
) {
return await MikroORM.init<PostgreSqlDriver>({
...makeOrmConfig(url),
...makeOrmConfig(url, ssl),
logger: log,
debug,
});

View File

@@ -2,6 +2,7 @@ import Redis from 'ioredis';
import RateLimiter from 'async-ratelimiter';
import ms from 'ms';
import { StatusError } from '@/services/error';
import { IpReq, getIp } from '@/services/ip';
export interface LimiterOptions {
redis: Redis;
@@ -26,8 +27,8 @@ export class Limiter {
this.redis = ops.redis;
}
async bump(req: { ip: string }, ops: BucketOptions) {
const ip = req.ip;
async bump(req: IpReq, ops: BucketOptions) {
const ip = getIp(req);
if (!this.buckets[ops.id]) {
this.buckets[ops.id] = {
limiter: new RateLimiter({
@@ -54,7 +55,7 @@ export class Limiter {
};
}
async assertAndBump(req: { ip: string }, ops: BucketOptions) {
async assertAndBump(req: IpReq, ops: BucketOptions) {
const { hasBeenLimited } = await this.bump(req, ops);
if (hasBeenLimited) {
throw new StatusError('Ratelimited', 429);

14
src/routes/index.ts Normal file
View File

@@ -0,0 +1,14 @@
import { version } from '@/config';
import { handle } from '@/services/handler';
import { makeRouter } from '@/services/router';
export const indexRouter = makeRouter((app) => {
app.get(
'/',
handle(async () => {
return {
message: `Backend is working as expected (${version})`,
};
}),
);
});

View File

@@ -1,8 +1,8 @@
import { handle } from '@/services/handler';
import { makeRouter } from '@/services/router';
import { z } from 'zod';
import { ProviderMetric, status } from '@/db/models/ProviderMetrics';
import { getMetrics } from '@/modules/metrics';
import { status } from '@/routes/statuses';
const metricsProviderSchema = z.object({
tmdbId: z.string(),
@@ -19,6 +19,7 @@ const metricsProviderSchema = z.object({
const metricsProviderInputSchema = z.object({
items: z.array(metricsProviderSchema).max(10).min(1),
tool: z.string().optional(),
});
export const metricsRouter = makeRouter((app) => {
@@ -29,7 +30,7 @@ export const metricsRouter = makeRouter((app) => {
body: metricsProviderInputSchema,
},
},
handle(async ({ em, body, req, limiter }) => {
handle(async ({ body, req, limiter }) => {
await limiter?.assertAndBump(req, {
id: 'provider_metrics',
max: 300,
@@ -37,43 +38,65 @@ export const metricsRouter = makeRouter((app) => {
window: '30m',
});
const hostname = req.headers.origin?.slice(0, 255) ?? 'unknown origin';
const entities = body.items.map((v) => {
const errorMessage = v.errorMessage?.slice(0, 200);
const truncatedFullError = v.fullError?.slice(0, 2000);
const metric = new ProviderMetric();
em.assign(metric, {
providerId: v.providerId,
embedId: v.embedId,
fullError: truncatedFullError,
errorMessage: errorMessage,
episodeId: v.episodeId,
seasonId: v.seasonId,
status: v.status,
title: v.title,
tmdbId: v.tmdbId,
type: v.type,
hostname,
});
return metric;
const hostname = req.headers.origin?.slice(0, 255) ?? '<UNKNOWN>';
getMetrics().providerHostnames.inc({
hostname,
});
entities.forEach((entity) => {
getMetrics().providerMetrics.inc({
episode_id: entity.episodeId,
provider_id: entity.providerId,
season_id: entity.seasonId,
status: entity.status,
title: entity.title,
tmdb_id: entity.tmdbId,
type: entity.type,
hostname,
body.items.forEach((item) => {
getMetrics().providerStatuses.inc({
provider_id: item.embedId ?? item.providerId,
status: item.status,
});
});
await em.persistAndFlush(entities);
const itemList = [...body.items];
itemList.reverse();
const lastSuccessfulItem = body.items.find(
(v) => v.status === status.success,
);
const lastItem = itemList[0];
if (lastItem) {
getMetrics().watchMetrics.inc({
tmdb_full_id: lastItem.type + '-' + lastItem.tmdbId,
provider_id: lastSuccessfulItem?.providerId ?? lastItem.providerId,
title: lastItem.title,
success: (!!lastSuccessfulItem).toString(),
});
}
if (body.tool) {
getMetrics().toolMetrics.inc({
tool: body.tool,
});
}
return true;
}),
);
app.post(
'/metrics/captcha',
{
schema: {
body: z.object({
success: z.boolean(),
}),
},
},
handle(async ({ body, req, limiter }) => {
await limiter?.assertAndBump(req, {
id: 'captcha_solves',
max: 300,
inc: 1,
window: '30m',
});
getMetrics().captchaSolves.inc({
success: body.success.toString(),
});
return true;
}),
);

6
src/routes/statuses.ts Normal file
View File

@@ -0,0 +1,6 @@
export const status = {
failed: 'failed',
notfound: 'notfound',
success: 'success',
} as const;
export type Status = keyof typeof status;

View File

@@ -6,7 +6,6 @@ import {
import { StatusError } from '@/services/error';
import { handle } from '@/services/handler';
import { makeRouter } from '@/services/router';
import { randomUUID } from 'crypto';
import { z } from 'zod';
const bookmarkDataSchema = z.object({

View File

@@ -6,6 +6,7 @@ import {
import { StatusError } from '@/services/error';
import { handle } from '@/services/handler';
import { makeRouter } from '@/services/router';
import { FilterQuery } from '@mikro-orm/core';
import { randomUUID } from 'crypto';
import { z } from 'zod';
@@ -164,22 +165,28 @@ export const userProgressRouter = makeRouter((app) => {
if (auth.user.id !== params.uid)
throw new StatusError('Cannot modify user other than yourself', 403);
const progressItem = await em.findOne(ProgressItem, {
const query: FilterQuery<ProgressItem> = {
userId: params.uid,
tmdbId: params.tmdbid,
episodeId: body.episodeId,
seasonId: body.seasonId,
});
if (!progressItem) {
};
if (body.seasonId) query.seasonId = body.seasonId;
if (body.episodeId) query.episodeId = body.episodeId;
const progressItems = await em.find(ProgressItem, query);
if (progressItems.length === 0) {
return {
count: 0,
tmdbId: params.tmdbid,
episodeId: body.episodeId,
seasonId: body.seasonId,
};
}
await em.removeAndFlush(progressItem);
progressItems.forEach((v) => em.remove(v));
await em.flush();
return {
count: progressItems.length,
tmdbId: params.tmdbid,
episodeId: body.episodeId,
seasonId: body.seasonId,

View File

@@ -41,6 +41,7 @@ export const userSettingsRouter = makeRouter((app) => {
applicationLanguage: z.string().nullable().optional(),
applicationTheme: z.string().nullable().optional(),
defaultSubtitleLanguage: z.string().nullable().optional(),
proxyUrls: z.string().array().nullable().optional(),
}),
},
},
@@ -64,6 +65,7 @@ export const userSettingsRouter = makeRouter((app) => {
settings.defaultSubtitleLanguage = body.defaultSubtitleLanguage;
if (body.applicationTheme !== undefined)
settings.applicationTheme = body.applicationTheme;
if (body.proxyUrls !== undefined) settings.proxyUrls = body.proxyUrls;
await em.persistAndFlush(settings);
return formatUserSettings(settings);

View File

@@ -4,16 +4,14 @@ import { StatusError } from '@/services/error';
export async function isValidCaptcha(token: string): Promise<boolean> {
if (!conf.captcha.secret)
throw new Error('isValidCaptcha() is called but no secret set');
const formData = new URLSearchParams();
formData.append('secret', conf.captcha.secret);
formData.append('response', token);
const res = await fetch('https://www.google.com/recaptcha/api/siteverify', {
method: 'POST',
body: JSON.stringify({
secret: conf.captcha.secret,
response: token,
}),
headers: {
'content-type': 'application/json',
},
body: formData,
});
const json = await res.json();
return !!json.success;
}

27
src/services/ip.ts Normal file
View File

@@ -0,0 +1,27 @@
import { conf } from '@/config';
import { IncomingHttpHeaders } from 'http';
export type IpReq = {
ip: string;
headers: IncomingHttpHeaders;
};
const trustCloudflare = conf.server.trustCloudflare;
function getSingleHeader(
headers: IncomingHttpHeaders,
key: string,
): string | undefined {
const header = headers[key];
if (Array.isArray(header)) return header[0];
return header;
}
export function getIp(req: IpReq) {
const cfIp = getSingleHeader(req.headers, 'cf-connecting-ip');
if (trustCloudflare && cfIp) {
return cfIp;
}
return req.ip;
}