added all files lol

This commit is contained in:
J-onasJones 2024-06-24 20:14:21 +02:00
parent b5e5bafd97
commit 547b32592a
10 changed files with 3615 additions and 0 deletions

12
.editorconfig Normal file
View file

@ -0,0 +1,12 @@
# http://editorconfig.org
root = true
[*]
indent_style = tab
end_of_line = lf
charset = utf-8
trim_trailing_whitespace = true
insert_final_newline = true
[*.yml]
indent_style = space

172
.gitignore vendored Normal file
View file

@ -0,0 +1,172 @@
# Logs
logs
_.log
npm-debug.log_
yarn-debug.log*
yarn-error.log*
lerna-debug.log*
.pnpm-debug.log*
# Diagnostic reports (https://nodejs.org/api/report.html)
report.[0-9]_.[0-9]_.[0-9]_.[0-9]_.json
# Runtime data
pids
_.pid
_.seed
\*.pid.lock
# Directory for instrumented libs generated by jscoverage/JSCover
lib-cov
# Coverage directory used by tools like istanbul
coverage
\*.lcov
# nyc test coverage
.nyc_output
# Grunt intermediate storage (https://gruntjs.com/creating-plugins#storing-task-files)
.grunt
# Bower dependency directory (https://bower.io/)
bower_components
# node-waf configuration
.lock-wscript
# Compiled binary addons (https://nodejs.org/api/addons.html)
build/Release
# Dependency directories
node_modules/
jspm_packages/
# Snowpack dependency directory (https://snowpack.dev/)
web_modules/
# TypeScript cache
\*.tsbuildinfo
# Optional npm cache directory
.npm
# Optional eslint cache
.eslintcache
# Optional stylelint cache
.stylelintcache
# Microbundle cache
.rpt2_cache/
.rts2_cache_cjs/
.rts2_cache_es/
.rts2_cache_umd/
# Optional REPL history
.node_repl_history
# Output of 'npm pack'
\*.tgz
# Yarn Integrity file
.yarn-integrity
# dotenv environment variable files
.env
.env.development.local
.env.test.local
.env.production.local
.env.local
# parcel-bundler cache (https://parceljs.org/)
.cache
.parcel-cache
# Next.js build output
.next
out
# Nuxt.js build / generate output
.nuxt
dist
# Gatsby files
.cache/
# Comment in the public line in if your project uses Gatsby and not Next.js
# https://nextjs.org/blog/next-9-1#public-directory-support
# public
# vuepress build output
.vuepress/dist
# vuepress v2.x temp and cache directory
.temp
.cache
# Docusaurus cache and generated files
.docusaurus
# Serverless directories
.serverless/
# FuseBox cache
.fusebox/
# DynamoDB Local files
.dynamodb/
# TernJS port file
.tern-port
# Stores VSCode versions used for testing VSCode extensions
.vscode-test
# yarn v2
.yarn/cache
.yarn/unplugged
.yarn/build-state.yml
.yarn/install-state.gz
.pnp.\*
# wrangler project
.dev.vars
.wrangler/

6
.prettierrc Normal file
View file

@ -0,0 +1,6 @@
{
"printWidth": 140,
"singleQuote": true,
"semi": true,
"useTabs": true
}

1
example.txt Normal file
View file

@ -0,0 +1 @@
some example text

3110
package-lock.json generated Normal file

File diff suppressed because it is too large Load diff

16
package.json Normal file
View file

@ -0,0 +1,16 @@
{
"name": "bold-cell-47f0",
"version": "0.0.0",
"private": true,
"scripts": {
"deploy": "wrangler deploy",
"dev": "wrangler dev",
"start": "wrangler dev",
"test": "vitest"
},
"devDependencies": {
"@cloudflare/vitest-pool-workers": "^0.4.5",
"wrangler": "^3.60.3",
"vitest": "1.5.0"
}
}

151
src/index.js Normal file
View file

@ -0,0 +1,151 @@
import { nanoid } from "nanoid";
addEventListener("fetch", event => {
event.respondWith(handleRequest(event.request));
});
async function handleRequest(request) {
const url = new URL(request.url);
const { pathname } = url;
await cleanUpExpiredFiles();
if (request.method === 'POST' && pathname === '/upload') {
return handleUpload(request);
}
if (request.method === 'GET' && pathname.startsWith('/download/')) {
const filename = pathname.replace('/download/', '');
return handleDownload(filename);
}
if (request.method === 'DELETE' && pathname.startsWith('/delete/')) {
const filename = pathname.replace('/delete/', '');
return handleDelete(filename);
}
return new Response('Not Found', { status: 404 });
}
async function cleanUpExpiredFiles() {
const indexData = await CDN_BUCKET.get('/tempupload/index.json');
if (!indexData) {
return;
}
let index = await indexData.json();
const now = Date.now();
const updatedIndex = [];
for (const fileRecord of index) {
if (fileRecord.deletionTimestamp <= now) {
await CDN_BUCKET.delete(`/tempupload/content/${fileRecord.hash}`);
} else {
updatedIndex.push(fileRecord);
}
}
if (updatedIndex.length !== index.length) {
await CDN_BUCKET.put('/tempupload/index.json', JSON.stringify(updatedIndex), {
httpMetadata: { contentType: 'application/json' },
});
}
}
async function handleUpload(request) {
// Password protection
const password = request.headers.get('X-Custom-Auth-Key');
if (password !== `Bearer ${AUTH_KEY_SECRET}`) {
return new Response('Unauthorized', { status: 401 });
}
const formData = await request.formData();
const file = formData.get('file');
if (!file) {
return new Response('Bad Request', { status: 400 });
}
const hash = nanoid();
const contentType = file.type;
const originalName = file.name;
const deletionTimestamp = Date.now() + 24 * 60 * 60 * 1000; // 24 hours
await CDN_BUCKET.put(`tempupload/content/${hash}`, file.stream(), {
httpMetadata: { contentType },
});
const indexData = await CDN_BUCKET.get('tempupload/index.json');
const index = indexData ? await indexData.json() : [];
index.push({
hash,
contentType,
originalName,
deletionTimestamp,
});
await CDN_BUCKET.put('tempupload/index.json', JSON.stringify(index), {
httpMetadata: { contentType: 'application/json' },
});
return new Response(JSON.stringify({ downloadLink: `https://fileshare.jonasjones.dev/download/${hash}` }), {
headers: { 'Content-Type': 'application/json' },
});
}
async function handleDownload(hash) {
const indexData = await CDN_BUCKET.get('tempupload/index.json');
if (!indexData) {
return new Response('Not Found', { status: 404 });
}
const index = await indexData.json();
const fileRecord = index.find(file => file.hash === hash);
if (!fileRecord) {
return new Response('Not Found', { status: 404 });
}
const file = await CDN_BUCKET.get(`tempupload/content/${hash}`);
if (!file) {
return new Response('Not Found', { status: 404 });
}
const response = new Response(file.body, {
headers: {
'Content-Type': fileRecord.contentType,
'Content-Disposition': `inline; filename="${fileRecord.originalName}"`,
},
});
return response;
}
async function handleDelete(filename) {
// Password protection
const password = request.headers.get('X-Custom-Auth-Key');
if (password !== `Bearer ${AUTH_KEY_SECRET}`) {
return new Response('Unauthorized', { status: 401 });
}
const indexData = await CDN_BUCKET.get('tempupload/index.json');
if (!indexData) {
return new Response('Not Found', { status: 404 });
}
let index = await indexData.json();
const fileRecord = index.find(file => file.originalName === filename);
if (!fileRecord) {
return new Response('Not Found', { status: 404 });
}
await CDN_BUCKET.delete(`tempupload/content/${fileRecord.hash}`);
index = index.filter(file => file.originalName !== filename);
await CDN_BUCKET.put('tempupload/index.json', JSON.stringify(index), {
httpMetadata: { contentType: 'application/json' },
});
return new Response('File deleted', { status: 200 });
}

20
test/index.spec.js Normal file
View file

@ -0,0 +1,20 @@
import { env, createExecutionContext, waitOnExecutionContext, SELF } from 'cloudflare:test';
import { describe, it, expect } from 'vitest';
import worker from '../src';
describe('Hello World worker', () => {
it('responds with Hello World! (unit style)', async () => {
const request = new Request('http://example.com');
// Create an empty context to pass to `worker.fetch()`.
const ctx = createExecutionContext();
const response = await worker.fetch(request, env, ctx);
// Wait for all `Promise`s passed to `ctx.waitUntil()` to settle before running test assertions
await waitOnExecutionContext(ctx);
expect(await response.text()).toMatchInlineSnapshot(`"Hello World!"`);
});
it('responds with Hello World! (integration style)', async () => {
const response = await SELF.fetch(request, env, ctx);
expect(await response.text()).toMatchInlineSnapshot(`"Hello World!"`);
});
});

11
vitest.config.js Normal file
View file

@ -0,0 +1,11 @@
import { defineWorkersConfig } from '@cloudflare/vitest-pool-workers/config';
export default defineWorkersConfig({
test: {
poolOptions: {
workers: {
wrangler: { configPath: './wrangler.toml' },
},
},
},
});

116
wrangler.toml Normal file
View file

@ -0,0 +1,116 @@
#:schema node_modules/wrangler/config-schema.json
name = "fileshare"
main = "src/index.js"
compatibility_date = "2024-06-20"
compatibility_flags = ["nodejs_compat"]
account_id = "d9e259f328167af6c866c4e624f12c0d"
workers_dev = true
[[r2_buckets]]
binding = 'CDN_BUCKET'
bucket_name = 'cdn'
# Automatically place your workloads in an optimal location to minimize latency.
# If you are running back-end logic in a Worker, running it closer to your back-end infrastructure
# rather than the end user may result in better performance.
# Docs: https://developers.cloudflare.com/workers/configuration/smart-placement/#smart-placement
# [placement]
# mode = "smart"
# Variable bindings. These are arbitrary, plaintext strings (similar to environment variables)
# Docs:
# - https://developers.cloudflare.com/workers/wrangler/configuration/#environment-variables
# Note: Use secrets to store sensitive data.
# - https://developers.cloudflare.com/workers/configuration/secrets/
# [vars]
# MY_VARIABLE = "production_value"
# Bind the Workers AI model catalog. Run machine learning models, powered by serverless GPUs, on Cloudflares global network
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#workers-ai
# [ai]
# binding = "AI"
# Bind an Analytics Engine dataset. Use Analytics Engine to write analytics within your Pages Function.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#analytics-engine-datasets
# [[analytics_engine_datasets]]
# binding = "MY_DATASET"
# Bind a headless browser instance running on Cloudflare's global network.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#browser-rendering
# [browser]
# binding = "MY_BROWSER"
# Bind a D1 database. D1 is Cloudflares native serverless SQL database.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#d1-databases
# [[d1_databases]]
# binding = "MY_DB"
# database_name = "my-database"
# database_id = "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx"
# Bind a dispatch namespace. Use Workers for Platforms to deploy serverless functions programmatically on behalf of your customers.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#dispatch-namespace-bindings-workers-for-platforms
# [[dispatch_namespaces]]
# binding = "MY_DISPATCHER"
# namespace = "my-namespace"
# Bind a Durable Object. Durable objects are a scale-to-zero compute primitive based on the actor model.
# Durable Objects can live for as long as needed. Use these when you need a long-running "server", such as in realtime apps.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#durable-objects
# [[durable_objects.bindings]]
# name = "MY_DURABLE_OBJECT"
# class_name = "MyDurableObject"
# Durable Object migrations.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#migrations
# [[migrations]]
# tag = "v1"
# new_classes = ["MyDurableObject"]
# Bind a Hyperdrive configuration. Use to accelerate access to your existing databases from Cloudflare Workers.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#hyperdrive
# [[hyperdrive]]
# binding = "MY_HYPERDRIVE"
# id = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# Bind a KV Namespace. Use KV as persistent storage for small key-value pairs.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#kv-namespaces
# [[kv_namespaces]]
# binding = "MY_KV_NAMESPACE"
# id = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# Bind an mTLS certificate. Use to present a client certificate when communicating with another service.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#mtls-certificates
# [[mtls_certificates]]
# binding = "MY_CERTIFICATE"
# certificate_id = "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx"
# Bind a Queue producer. Use this binding to schedule an arbitrary task that may be processed later by a Queue consumer.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#queues
# [[queues.producers]]
# binding = "MY_QUEUE"
# queue = "my-queue"
# Bind a Queue consumer. Queue Consumers can retrieve tasks scheduled by Producers to act on them.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#queues
# [[queues.consumers]]
# queue = "my-queue"
# Bind an R2 Bucket. Use R2 to store arbitrarily large blobs of data, such as files.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#r2-buckets
# [[r2_buckets]]
# binding = "MY_BUCKET"
# bucket_name = "my-bucket"
# Bind another Worker service. Use this binding to call another Worker without network overhead.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#service-bindings
# [[services]]
# binding = "MY_SERVICE"
# service = "my-service"
# Bind a Vectorize index. Use to store and query vector embeddings for semantic search, classification and other vector search use-cases.
# Docs: https://developers.cloudflare.com/workers/wrangler/configuration/#vectorize-indexes
# [[vectorize]]
# binding = "MY_INDEX"
# index_name = "my-index"