@nire0510/wmcli
Version:
Useful tools for webmasters
355 lines (299 loc) • 11.3 kB
text/typescript
import open from 'open';
import ora from 'ora';
import puppeteer from 'puppeteer';
import Crawler from './crawler';
import * as utils from './utils';
const spinner = ora('wait...');
function browse(url: string | any, options?: any): void {
if (typeof url === 'string') {
open(url);
}
else {
open(url[Object.keys(options)
.find((key) => Object.keys(url).includes(key)) || Object.keys(url)[0]]);
}
process.exit(0);
}
async function execute(command: string): Promise<void> {
await genericCommand(null, async () => {
const info = await utils.execute(command);
spinner.succeed(info);
});
}
async function genericCommand(crawler: Crawler | null, fnc: any): Promise<void> {
spinner.start();
try {
await fnc();
}
catch (error) {
// console.error(error);
spinner.fail('An error has occurred. Is URL correct?');
process.exit(0);
}
finally {
spinner.stop();
crawler && await crawler.deconstructor();
}
}
export async function archive(url: string): Promise<void> {
browse(`https://web.archive.org/web/*/${url}`);
}
export function audit(url: string, options: any): void {
browse({
pagespeed: `https://developers.google.com/speed/pagespeed/insights/?url=${url}`,
seoptimer: `https://www.seoptimer.com/${url}`,
yellowlab: `https://yellowlab.tools/?url=${encodeURIComponent(url)}`,
}, options);
}
export async function coverage(url: string, options: any): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'html');
await genericCommand(crawler, async () => {
const coverage: any[] = await crawler.coverage(url, options?.stylesheet ? 'css' : 'js');
const items = coverage.map((entry) => {
const item = {
'URL': entry.url,
'Total Bytes': entry.text.length,
'Used Bytes': entry.ranges.reduce((a: number, c: any) => a + c.end - c.start - 1, 0),
'Usage': 'N/A',
};
item['Usage'] = `${(item['Used Bytes'] / item['Total Bytes'] * 100).toFixed(2)}%`;
return item;
});
await utils.writeFile(file, await utils.generateFileFromTemplate('table', items));
open (file);
});
}
export async function extract(url: string, options: any): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'html');
await genericCommand(crawler, async () => {
if (options?.headers) {
const headers: HTMLElement[] = await crawler.querySelectorAll(url,
[
`${options && options.selector || ''} h1`.trim(),
`${options && options.selector || ''} h2`.trim(),
`${options && options.selector || ''} h3`.trim(),
`${options && options.selector || ''} h4`.trim(),
`${options && options.selector || ''} h5`.trim(),
`${options && options.selector || ''} h6`.trim(),
].join(','),
(elements: Element[]) => elements.map((element) => ({
'Tag Name': element.tagName,
'Content': element.textContent,
})));
await utils.writeFile(file, await utils.generateFileFromTemplate('table', headers));
}
else if (options?.links) {
const links: HTMLElement[] = await crawler.querySelectorAll(url,
`${options && options.selector || ''} a`.trim(),
(elements: Element[]) => elements.map((element) => ({
'Href': decodeURI(element.getAttribute('href') || ''),
'Text': element.textContent,
'Title': element.getAttribute('title'),
})));
await utils.writeFile(file, await utils.generateFileFromTemplate('table', links));
}
else if (options?.images || options?.imagesGallery) {
const images: HTMLElement[] = await crawler.querySelectorAll(url,
`${options && options.selector || ''} img`.trim(),
(elements: Element[]) => elements.map((element) => ({
'Title': element.getAttribute('title'),
'Alternate Text': element.getAttribute('alt'),
'Source': element.getAttribute('src'),
})));
if (options?.imagesGallery) {
const html = images
.map((image: any) => `<img src="${image['Source'].startsWith('http') ? '' : url}${image['Source']}" title="${image['Title'] || image['Alternate Text']}">`).join('');
await utils.writeFile(file, html);
}
else {
await utils.writeFile(file, await utils.generateFileFromTemplate('table', images));
}
}
// text
else {
const text: string = await crawler.querySelector(url,
`${options && options.selector || 'body'}`.trim(),
(element: any) => element.innerText);
if (options?.textCloud) {
const wordCount = utils.countWords(text);
const wordCountArray = Object.keys(wordCount)
.map((word: string) => ({
word,
count: wordCount[word],
}));
await utils.writeFile(file, await utils.generateFileFromTemplate('cloud', wordCountArray));
}
else {
await utils.writeFile(file, text.replace(/\n/g, '<br>'));
}
}
open (file);
});
}
export function info(domain: string, options: any): void {
browse({
wmtips: `https://www.wmtips.com/tools/info/s/${domain}`,
alexa: `https://www.alexa.com/siteinfo/${domain}`,
similarweb: `https://www.similarweb.com/website/${domain}`,
w3techs: `https://w3techs.com/sites/info/${domain}`,
webtechsurvey: `https://webtechsurvey.com/website/${domain}`,
}, options);
}
export async function ip(domain: string): Promise<void> {
await execute(`dig +short ${domain} A`);
}
export async function log(url: string, options: any): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'html');
await genericCommand(crawler, async () => {
const items = await crawler.intercept(url, options?.responses ? 'response' : 'request');
const urlObject = new URL(url);
await utils.writeFile(file, await utils.generateFileFromTemplate('table', items.map((item) => {
if (options?.responses) {
const response = (item as puppeteer.HTTPResponse);
return {
'Url': response.url(),
'Status': response.status(),
'Status Text': response.statusText(),
'Content Type': response.headers()['content-type'],
'Cache': response.fromCache(),
'Service Worker': response.fromServiceWorker(),
'Success': response.ok(),
'Remote Address': `${response.remoteAddress().ip}${response.remoteAddress().port ? `:${response.remoteAddress().port}` : ''}`,
'External': !response.url().startsWith(urlObject.origin),
};
}
else {
const request = (item as puppeteer.HTTPRequest);
return {
'Url': request.url(),
'Method': request.method(),
'Resource Type': request.resourceType(),
'Post Data': JSON.stringify(request.postData()),
'External': !request.url().startsWith(urlObject.origin),
};
}
})));
open (file);
});
}
export async function pdf(url: string): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'pdf');
await genericCommand(crawler, async () => {
await crawler.pdf(url, file);
open(file);
});
}
export async function robots(domain: string): Promise<void> {
await genericCommand(null, async () => {
const url = `https://${domain}/robots.txt`;
try {
const isUrlExists = await utils.isUrlExists(url);
if (isUrlExists) {
spinner.succeed(`robots.txt found: ${url}`);
return browse(url);
}
spinner.warn(`robots.txt file could not be found for domain ${domain}`);
}
catch (error) {
spinner.fail(`Domain name ${domain} does not exists`);
process.exit(0);
}
});
}
export async function rss(domain: string): Promise<void> {
await genericCommand(null, async () => {
const crawler = new Crawler();
const origin = `https://${domain}`;
const rss = await crawler.querySelectorAll(origin, 'head > link[type="application/rss+xml"]', ([link]: HTMLLinkElement[]) => link ? link.getAttribute('href') : null);
if (rss && Array.isArray(rss) && rss.length) {
spinner.succeed(`RSS found: ${rss[0]}`);
browse(rss[0]);
}
else {
[
`${origin}/feed`,
`${origin}/feeds/posts/default`,
].find(async (url) => {
const isUrlExists = await utils.isUrlExists(url);
if (isUrlExists) {
spinner.succeed(`RSS found: ${url}`);
browse(url);
return true;
}
return false;
});
}
});
}
export async function screenshot(url: string, options: any): Promise<void> {
const crawler = new Crawler({
defaultViewport: {
width: 1920,
height: 1080,
},
} as puppeteer.LaunchOptions);
const file = utils.generateTempFilePath(url, 'png');
await genericCommand(crawler, async () => {
await crawler.screenshot(url, file, {
fullPage: options?.fullPage,
omitBackground: options?.omitBackground,
});
open(file);
});
}
export async function security(domain: string): Promise<void> {
const crawler = new Crawler();
const origin = `https://${domain}`;
const file = utils.generateTempFilePath(origin, 'json');
await genericCommand(crawler, async () => {
const securityDetails: any = await crawler.security(origin);
await utils.writeFile(file, JSON.stringify(securityDetails, null, 2));
open (file);
});
}
export async function source(url: string): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'txt');
await genericCommand(crawler, async () => {
const html: string = await crawler.querySelector(url,
'html',
(element: Element) => element.outerHTML);
await utils.writeFile(file, html);
open (file);
});
}
export function stack(domain: string, options: any): void {
browse({
builtwith: `https://builtwith.com/${domain}`,
netcraft: `https://sitereport.netcraft.com/?url=${domain}`,
similartech: `https://www.similartech.com/websites/${domain}`,
wappalyzer: `https://www.wappalyzer.com/lookup/${domain}`,
}, options);
}
export async function trace(url: string): Promise<void> {
const crawler = new Crawler();
const file = utils.generateTempFilePath(url, 'json');
await genericCommand(crawler, async () => {
await crawler.trace(url, file);
open(file);
});
}
export function translate(url: string): void {
browse(`https://translate.google.com/translate?sl=autoS&u=${url}`);
}
export function validate(url: string, options: { [key: string]: string }): void {
browse({
html: `https://validator.w3.org/nu/?doc=${url}`,
css: `https://jigsaw.w3.org/css-validator/validator?uri=${url}`,
i18n: `https://validator.w3.org/i18n-checker/check?uri=${url}`,
links: `https://validator.w3.org/checklink?uri=${url}`,
structured: `https://search.google.com/test/rich-results?utm_campaign=sdtt&utm_medium=message&url=${url}&user_agent=1`,
}, options);
}
export async function whois(domain: string): Promise<void> {
await execute(`whois ${domain}`);
}