build-reject-domainset.js 8.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278
  1. const { promises: fsPromises } = require('fs');
  2. const fse = require('fs-extra');
  3. const { resolve: pathResolve } = require('path');
  4. const Piscina = require('piscina');
  5. const { processHosts, processFilterRules, preprocessFullDomainSetBeforeUsedAsWorkerData } = require('./lib/parse-filter');
  6. const cpuCount = require('os').cpus().length;
  7. const { isCI } = require('ci-info');
  8. const threads = isCI ? cpuCount : cpuCount / 2;
  9. const { HOSTS, ADGUARD_FILTERS, PREDEFINED_WHITELIST } = require('./lib/reject-data-source');
  10. const { withBannerArray } = require('./lib/with-banner');
  11. const { compareAndWriteFile } = require('./lib/string-array-compare');
  12. const filterRuleWhitelistDomainSets = new Set(PREDEFINED_WHITELIST);
  13. (async () => {
  14. console.time('Total Time - build-reject-domain-set');
  15. /** @type Set<string> */
  16. const domainSets = new Set();
  17. console.log('Downloading hosts file...');
  18. console.time('* Download and process Hosts');
  19. // Parse from remote hosts & domain lists
  20. (await Promise.all(
  21. HOSTS.map(entry => processHosts(entry[0], entry[1]))
  22. )).forEach(hosts => {
  23. hosts.forEach(host => {
  24. if (host) {
  25. domainSets.add(host);
  26. }
  27. });
  28. });
  29. console.timeEnd('* Download and process Hosts');
  30. let previousSize = domainSets.size;
  31. console.log(`Import ${previousSize} rules from hosts files!`);
  32. // Parse from AdGuard Filters
  33. console.time('* Download and process AdBlock Filter Rules');
  34. let shouldStop = false;
  35. await Promise.all(ADGUARD_FILTERS.map(input => {
  36. const promise = typeof input === 'string'
  37. ? processFilterRules(input, undefined, false)
  38. : processFilterRules(input[0], input[1] ?? undefined, input[2] ?? false)
  39. return promise.then((i) => {
  40. if (i) {
  41. const { white, black, foundDebugDomain } = i;
  42. if (foundDebugDomain) {
  43. shouldStop = true;
  44. }
  45. white.forEach(i => filterRuleWhitelistDomainSets.add(i));
  46. black.forEach(i => domainSets.add(i));
  47. } else {
  48. process.exit(1);
  49. }
  50. });
  51. }));
  52. await Promise.all([
  53. 'https://raw.githubusercontent.com/AdguardTeam/AdGuardSDNSFilter/master/Filters/exceptions.txt',
  54. 'https://raw.githubusercontent.com/AdguardTeam/AdGuardSDNSFilter/master/Filters/exclusions.txt'
  55. ].map(
  56. input => processFilterRules(input).then((i) => {
  57. if (i) {
  58. const { white, black } = i;
  59. white.forEach(i => filterRuleWhitelistDomainSets.add(i));
  60. black.forEach(i => filterRuleWhitelistDomainSets.add(i));
  61. } else {
  62. process.exit(1);
  63. }
  64. })
  65. ));
  66. console.timeEnd('* Download and process AdBlock Filter Rules');
  67. if (shouldStop) {
  68. process.exit(1);
  69. }
  70. previousSize = domainSets.size - previousSize;
  71. console.log(`Import ${previousSize} rules from adguard filters!`);
  72. await fsPromises.readFile(pathResolve(__dirname, '../Source/domainset/reject_sukka.conf'), { encoding: 'utf-8' }).then(data => {
  73. data.split('\n').forEach(line => {
  74. const trimmed = line.trim();
  75. if (
  76. line.startsWith('#')
  77. || line.startsWith(' ')
  78. || line.startsWith('\r')
  79. || line.startsWith('\n')
  80. || trimmed === ''
  81. ) {
  82. return;
  83. }
  84. domainSets.add(trimmed);
  85. });
  86. });
  87. // Copy reject_sukka.conf for backward compatibility
  88. await fse.copy(pathResolve(__dirname, '../Source/domainset/reject_sukka.conf'), pathResolve(__dirname, '../List/domainset/reject_sukka.conf'))
  89. previousSize = domainSets.size - previousSize;
  90. console.log(`Import ${previousSize} rules from reject_sukka.conf!`);
  91. // Read DOMAIN Keyword
  92. const domainKeywordsSet = new Set();
  93. const domainSuffixSet = new Set();
  94. await fsPromises.readFile(pathResolve(__dirname, '../List/non_ip/reject.conf'), { encoding: 'utf-8' }).then(data => {
  95. data.split('\n').forEach(line => {
  96. if (line.startsWith('DOMAIN-KEYWORD')) {
  97. const [, ...keywords] = line.split(',');
  98. domainKeywordsSet.add(keywords.join(',').trim());
  99. } else if (line.startsWith('DOMAIN-SUFFIX')) {
  100. const [, ...keywords] = line.split(',');
  101. domainSuffixSet.add(keywords.join(',').trim());
  102. }
  103. });
  104. });
  105. // Read Special Phishing Suffix list
  106. await fsPromises.readFile(pathResolve(__dirname, '../List/domainset/reject_phishing.conf'), { encoding: 'utf-8' }).then(data => {
  107. data.split('\n').forEach(line => {
  108. const trimmed = line.trim();
  109. if (
  110. line.startsWith('#')
  111. || line.startsWith(' ')
  112. || line.startsWith('\r')
  113. || line.startsWith('\n')
  114. || trimmed === ''
  115. ) {
  116. return;
  117. }
  118. domainSuffixSet.add(trimmed);
  119. });
  120. });
  121. console.log(`Import ${domainKeywordsSet.size} black keywords and ${domainSuffixSet.size} black suffixes!`);
  122. previousSize = domainSets.size;
  123. // Dedupe domainSets
  124. console.log(`Start deduping from black keywords/suffixes! (${previousSize})`);
  125. console.time(`* Dedupe from black keywords/suffixes`);
  126. for (const domain of domainSets) {
  127. let isTobeRemoved = false;
  128. for (const suffix of domainSuffixSet) {
  129. if (domain.endsWith(suffix)) {
  130. isTobeRemoved = true;
  131. break;
  132. }
  133. }
  134. if (!isTobeRemoved) {
  135. for (const keyword of domainKeywordsSet) {
  136. if (domain.includes(keyword)) {
  137. isTobeRemoved = true;
  138. break;
  139. }
  140. }
  141. }
  142. if (!isTobeRemoved) {
  143. if (isInWhiteList(domain)) {
  144. isTobeRemoved = true;
  145. }
  146. }
  147. if (isTobeRemoved) {
  148. domainSets.delete(domain);
  149. }
  150. }
  151. console.timeEnd(`* Dedupe from black keywords/suffixes`);
  152. console.log(`Deduped ${previousSize} - ${domainSets.size} = ${previousSize - domainSets.size} from black keywords and suffixes!`);
  153. previousSize = domainSets.size;
  154. // Dedupe domainSets
  155. console.log(`Start deduping! (${previousSize})`);
  156. const START_TIME = Date.now();
  157. const piscina = new Piscina({
  158. filename: pathResolve(__dirname, 'worker/build-reject-domainset-worker.js'),
  159. workerData: preprocessFullDomainSetBeforeUsedAsWorkerData([...domainSets]),
  160. idleTimeout: 50,
  161. minThreads: threads,
  162. maxThreads: threads
  163. });
  164. console.log(`Launching ${threads} threads...`)
  165. const tasksArray = Array.from(domainSets)
  166. .reduce((result, element, index) => {
  167. const chunk = index % threads;
  168. result[chunk] ??= [];
  169. result[chunk].push(element);
  170. return result;
  171. }, []);
  172. (
  173. await Promise.all(
  174. Array.from(domainSets)
  175. .reduce((result, element, index) => {
  176. const chunk = index % threads;
  177. result[chunk] ??= [];
  178. result[chunk].push(element);
  179. return result;
  180. }, [])
  181. .map(chunk => piscina.run({ chunk }, { name: 'dedupe' }))
  182. )
  183. ).forEach((result, taskIndex) => {
  184. const chunk = tasksArray[taskIndex];
  185. for (let i = 0, len = result.length; i < len; i++) {
  186. if (result[i]) {
  187. domainSets.delete(chunk[i]);
  188. }
  189. }
  190. });
  191. console.log(`* Dedupe from covered subdomain - ${(Date.now() - START_TIME) / 1000}s`);
  192. console.log(`Deduped ${previousSize - domainSets.size} rules!`);
  193. await piscina.destroy();
  194. console.time('* Write reject.conf');
  195. await compareAndWriteFile(
  196. withBannerArray(
  197. 'Sukka\'s Surge Rules - Reject Base',
  198. [
  199. 'License: AGPL 3.0',
  200. 'Homepage: https://ruleset.skk.moe',
  201. 'GitHub: https://github.com/SukkaW/Surge',
  202. '',
  203. 'The domainset supports AD blocking, tracking protection, privacy protection, anti-phishing, anti-mining',
  204. '',
  205. 'Build from:',
  206. ...HOSTS.map(host => ` - ${host[0]}`),
  207. ...ADGUARD_FILTERS.map(filter => ` - ${Array.isArray(filter) ? filter[0] : filter}`),
  208. ],
  209. new Date(),
  210. [...domainSets].sort()
  211. ),
  212. pathResolve(__dirname, '../List/domainset/reject.conf')
  213. );
  214. console.timeEnd('* Write reject.conf');
  215. console.timeEnd('Total Time - build-reject-domain-set');
  216. if (piscina.queueSize === 0) {
  217. process.exit(0);
  218. }
  219. })();
  220. function isInWhiteList (domain) {
  221. for (const white of filterRuleWhitelistDomainSets) {
  222. if (domain === white || domain.endsWith(white)) {
  223. return true;
  224. }
  225. if (white.endsWith(domain)) {
  226. // If a whole domain is in blacklist but a subdomain is in whitelist
  227. // We have no choice but to remove the whole domain from blacklist
  228. return true;
  229. }
  230. }
  231. return false;
  232. }