Online Website Downloader. Static HTML to CMS Converter

archivarix

Пользователь
Регистрация
30.09.2017
Сообщения
32
Благодарностей
20
Баллы
8
In addition to our Wayback Machine Downloader we have launched Online Website Downloader service/HTML-to-CMS corverter -
Unlike other programs such as HTTrack or Teleport, our Downloader does not just copy the site, but integrates files into CMS and creates a fully functional, valid copy of the website with the ability to edit and update. As well as for sites restored from Archive.org, all Archivarix functions are available for downloaded sites — such as optimizing graphics and HTML code, clearing external links, deleting counters, analytics, and others.

Archivarix online downloader - https://en.archivarix.com/
 

archivarix

Пользователь
Регистрация
30.09.2017
Сообщения
32
Благодарностей
20
Баллы
8
Recently our system has been updated and now we have two new options.
First - you can download Darknet .onion sites. Just enter .onion website address in the "Domain" field here https://en.archivarix.com/website-downloader-cms-converter/ and our system will download it from the Tor network just like a regular website.
And the second - Archivarix can not only download existing sites or restore them from the Web Archive but also extract content from them. Here https://en.archivarix.com/restore/ in the "Advanced options" field you need to select "Extract structured content".
After that you will recieve a complete archive of the entire site, and an archive of articles in xml, csv, wxr and json format.
When creating an archive of articles our parser takes into account only meaningful content excluding duplicate articles, control elements and service pages.
 

archivarix

Пользователь
Регистрация
30.09.2017
Сообщения
32
Благодарностей
20
Баллы
8
In the near future, our team plans to launch a unique service https://archivarix.net/ that combines the capabilities of the Internet Archive (archive.org) and a search engine.
We plan to index the main pages of all sites that have ever been saved in the web archive. Our website database will also contain an archive of various metrics such as Alexa, Ahrefs, Majestic, keywords, WHOIS domain and other historical data from the entire lifetime of the website. Thus, it will be possible to search for the desired site to restore drops or extract deleted content by a huge number of parameters, such as keywords in text and tags, traffic on a specific date, the presence of historical links, name server, and so on. Using this system, it will be possible to make selections of any complexity, for example, find all domains deleted in 2018 containing the words "webmaster analytics", which then had an Alexa rating of less than 300k and which had more than 30 unique visitors per day for a specific keyword in 2016 year. Our service will have a convenient interface for finding the required data and will contain screenshots of websites with all parameters in the form of graphs and tables.
The next stage in the development of the system will be the indexing of live sites and the expansion of the indexed content base. For search, media files and some internal pages of the site will be available, selected according to an algorithm that takes into account the importance of this page. As a result, the system will use mainly its own database of archived sites and become independent of Archive.org.
 

Кто просматривает тему: (Всего: 1, Пользователи: 0, Гости: 1)