-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathcrawl.js
84 lines (77 loc) · 2.67 KB
/
crawl.js
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
const {JSDOM} = require('jsdom')
// crawlPage is a recursive function...the for loop give it the nexturls to crawl again
async function crawlPage(baseURL,currentURL,pages){
const baseURLObj = new URL(baseURL)
const currentURLObj = new URL(currentURL)
if(baseURLObj.hostname !== currentURLObj.hostname){
return pages
}
const normalizedCurrentURL = normalizeURL(currentURL)
if(pages[normalizedCurrentURL] > 0){
pages[normalizedCurrentURL]++
return pages
}
// crawling a new page
pages[normalizedCurrentURL]=1
console.log(`actively crawling: ${currentURL}`)
try{
const resp = await fetch(currentURL)
if(resp.status>399){
console.log(`error in fetch with status code: ${resp.status} on page:${currentURL}`)
return pages
}
const contentType = resp.headers.get("content-type")
if(!contentType.includes("text/html")){// here we can us if(contentType!=="text/html"){}
console.log(`non html reponse,content type: ${contentType}, on page:${currentURL}`)
return pages
}
const htmlBody = await resp.text()
const nextURLS = getURLsFromHTML(htmlBody,baseURL)
for( const nextURL of nextURLS){
//updating the pages object for the newurl by calling the crawlPage function
pages = await crawlPage(baseURL,nextURL,pages)
}
}catch(err){
console.log(`error in fetch: ${err.message}, on page:${currentURL}`)
}
return pages
}
//extracting all the links from the html
function getURLsFromHTML(htmlBody,baseURL) {
const urls =[]
const dom = new JSDOM(htmlBody)
const linkElements = dom.window.document.querySelectorAll('a')
for (const linkElement of linkElements){
if (linkElement.href.slice(0,1) === '/'){
//relative
try{
const urlObj = new URL(`${baseURL}${linkElement.href}`)
urls.push(urlObj.href)}
catch (err){
console.log(`error with relative url: ${err.message}`)
}
} else{
// absolute
try{
const urlObj = new URL(linkElement.href)
urls.push(urlObj.href)}
catch (err){
console.log(`error with absolute url: ${err.message}`)
}
}
}
return urls
}
function normalizeURL(urlString){
const urlObj = new URL(urlString)
const hostpath = `${urlObj.hostname}${urlObj.pathname}`
if (hostpath.length > 0 && hostpath.slice(-1) === '/'){
return hostpath.slice(0,-1)
}
return hostpath
}
module.exports = {
normalizeURL,
getURLsFromHTML,
crawlPage
}