@@ -7,9 +7,8 @@ const visitedLinks = new Set<string>();
7
7
8
8
async function main ( ) {
9
9
const browserInstance = await chromium . launch ( ) ;
10
- const baseUrl = 'https://www.docs.sei.io/' ;
10
+ const baseUrl = process . env . BASE_URL || 'https://www.docs.sei.io/' ;
11
11
await crawlPages ( baseUrl , browserInstance , 'main' ) ;
12
-
13
12
fs . writeFileSync ( 'brokenLinks.json' , JSON . stringify ( [ ...brokenLinks ] , null , 2 ) ) ;
14
13
15
14
if ( brokenLinks . size > 0 ) {
@@ -31,7 +30,7 @@ async function crawlPages(url: string, browser: Browser, path: string) {
31
30
}
32
31
33
32
function isInternal ( url : string ) {
34
- return url . includes ( 'docs.sei' ) ;
33
+ return url . includes ( 'docs.sei' ) || url . includes ( 'localhost:3000' ) ;
35
34
}
36
35
37
36
async function checkInternalLinks ( url : string , page : Page , path : string , browser : Browser ) {
@@ -58,7 +57,7 @@ async function checkExternalLinks(url: string, page: Page, path: string, browser
58
57
59
58
async function isLinkBroken ( page : Page , url : string , path : string ) {
60
59
if ( visitedLinks . has ( url ) ) return false ;
61
- if ( url . includes ( 'localhost' ) || url . includes ( '.tar.gz' ) ) return false ;
60
+ if ( ( url . includes ( 'localhost' ) && ! url . includes ( ':3000' ) ) || url . includes ( '.tar.gz' ) ) return false ;
62
61
63
62
let pageResponse : Response ;
64
63
try {
0 commit comments