1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
|
import bodyParser from "body-parser";
import { Router } from "express";
import fetch from "node-fetch";
import cheerio from "cheerio";
import btoa from "btoa";
import { URL } from "url";
const router = Router();
type crawled = {
id: string;
ogTitle: string;
ogType: string;
ogDescription: string;
ogUrl: string;
cachedImage: string;
};
const DEFAULT_FETCH_OPTIONS: any = {
redirect: "follow",
follow: 1,
headers: {
"user-agent": "Mozilla/5.0 (compatible; Discordbot/2.0; +https://discordapp.com)",
},
size: 1024 * 1024 * 8,
compress: true,
method: "GET",
};
router.post("/", bodyParser.json(), async (req, res) => {
if (!req.body) throw new Error("Invalid Body (url missing) \nExample: url:https://discord.com");
const { db } = req.server;
const { url } = req.body;
const ID = btoa(url);
const cache = await db.data.crawler({ id: ID }).get();
if (cache) return res.send(cache);
try {
const request = await fetch(url, DEFAULT_FETCH_OPTIONS);
const text = await request.text();
const ツ: any = cheerio.load(text);
const ogTitle = ツ('meta[property="og:title"]').attr("content");
const ogDescription = ツ('meta[property="og:description"]').attr("content");
const ogImage = ツ('meta[property="og:image"]').attr("content");
const ogUrl = ツ('meta[property="og:url"]').attr("content");
const ogType = ツ('meta[property="og:type"]').attr("content");
const filename = new URL(url).host.split(".")[0];
const ImageResponse = await fetch(ogImage, DEFAULT_FETCH_OPTIONS);
const ImageType = ImageResponse.headers.get("content-type");
const ImageExtension = ImageType?.split("/")[1];
const ImageResponseBuffer = (await ImageResponse.buffer()).toString("base64");
const cachedImage = `/external/${ID}/${filename}.${ImageExtension}`;
await db.data.externals.push({ image: ImageResponseBuffer, id: ID, type: ImageType });
const new_cache_entry: crawled = { id: ID, ogTitle, ogDescription, cachedImage, ogUrl, ogType };
await db.data.crawler.push(new_cache_entry);
res.send(new_cache_entry);
} catch (error) {
console.log(error);
throw new Error("Couldn't fetch website");
}
});
router.get("/:id/:filename", async (req, res) => {
const { db } = req.server;
const { id, filename } = req.params;
const { image, type } = await db.data.externals({ id: id }).get();
const imageBuffer = Buffer.from(image, "base64");
res.set("Content-Type", type);
res.send(imageBuffer);
});
export default router;
|