summary refs log tree commit diff
path: root/src/routes/external.ts
blob: be680003ca71243c74fca72b5621f8c516d8b8ec (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
import bodyParser from "body-parser";
import { Router } from "express";
import fetch from "node-fetch";
import cheerio from "cheerio";
import btoa from "btoa";
import { URL } from "url";

const router = Router();

type crawled = {
	id: string;
	ogTitle: string;
	ogType: string;
	ogDescription: string;
	ogUrl: string;
	cachedImage: string;
};

const DEFAULT_FETCH_OPTIONS: any = {
	redirect: "follow",
	follow: 1,
	headers: {
		"user-agent": "Mozilla/5.0 (compatible; Discordbot/2.0; +https://discordapp.com)",
	},
	size: 1024 * 1024 * 8,
	compress: true,
	method: "GET",
};

router.post("/", bodyParser.json(), async (req, res) => {
	if (!req.body) throw new Error("Invalid Body (url missing) \nExample: url:https://discord.com");

	const { db } = req.server;
	const { url } = req.body;

	const ID = btoa(url);

	const cache = await db.data.crawler({ id: ID }).get();
	if (cache) return res.send(cache);

	try {
		const request = await fetch(url, DEFAULT_FETCH_OPTIONS);

		const text = await request.text();
		const : any = cheerio.load(text);

		const ogTitle = ('meta[property="og:title"]').attr("content");
		const ogDescription = ('meta[property="og:description"]').attr("content");
		const ogImage = ('meta[property="og:image"]').attr("content");
		const ogUrl = ('meta[property="og:url"]').attr("content");
		const ogType = ('meta[property="og:type"]').attr("content");

		const filename = new URL(url).host.split(".")[0];

		const ImageResponse = await fetch(ogImage, DEFAULT_FETCH_OPTIONS);
		const ImageType = ImageResponse.headers.get("content-type");
		const ImageExtension = ImageType?.split("/")[1];
		const ImageResponseBuffer = (await ImageResponse.buffer()).toString("base64");
		const cachedImage = `/external/${ID}/${filename}.${ImageExtension}`;

		await db.data.externals.push({ image: ImageResponseBuffer, id: ID, type: ImageType });

		const new_cache_entry: crawled = { id: ID, ogTitle, ogDescription, cachedImage, ogUrl, ogType };
		await db.data.crawler.push(new_cache_entry);

		res.send(new_cache_entry);
	} catch (error) {
		console.log(error);

		throw new Error("Couldn't fetch website");
	}
});

router.get("/:id/:filename", async (req, res) => {
	const { db } = req.server;
	const { id, filename } = req.params;
	const { image, type } = await db.data.externals({ id: id }).get();
	const imageBuffer = Buffer.from(image, "base64");

	res.set("Content-Type", type);
	res.send(imageBuffer);
});

export default router;