WebScrapBook에 대한 리뷰
WebScrapBook 제작자: Danny Lin
14802265 님의 리뷰
I tried this app to save webpages completely and accurately. It works on some pages like ghacks.net perfectly with scripted single html . On other pages like nytimes.com it captures the page out of sync even though all of the content seems to be there (large gap spaces, enlarged photos, etc.) Save Page WE has the same issue. On Washingtonpost.com WebScrapbook was almost perfect but there is a bug that will add incorrect characters if there is an apostrophe in the text(which in a news article there will undoubtedly be). I used scripted single html option on this also. I do have specific scripts for the Times and WPost running, but they are not the issue since Mozilla Archive Format and SingleFile always works perfectly on the same sites with the same scripts running. But since MAF doesnt work for current browsers and SingleFile works somewhat inconsistently (it stalls a lot), I was hoping WebScrapbook would work but no go.
Also, I havent seen an option to save the original page url either in the title or in the .html file for reference like MAF, Singlefile, or SavePage WE can.
I noticed the saved webpage nytimes.com icon was used in the tab, but Webscrapbook couldnt find the icon for washingtonpost.com tab. If the developer wants to see the output files, just tell me where to forward them.
This app might be able to save websites but if it cant do it accurately what's the point of using it.
Also, I havent seen an option to save the original page url either in the title or in the .html file for reference like MAF, Singlefile, or SavePage WE can.
I noticed the saved webpage nytimes.com icon was used in the tab, but Webscrapbook couldnt find the icon for washingtonpost.com tab. If the developer wants to see the output files, just tell me where to forward them.
This app might be able to save websites but if it cant do it accurately what's the point of using it.
개발자 답글
6년 전에 게시됨Thank you for the feedback.
The issue on nytimes.com is same as the one with styled components and we are working on it (https://github.com/danny0838/webscrapbook/issues/109). It's a complicated issue as there are many things behind the scene to deal with. We almost have the solution but still need sometime to implement it, maybe next one or two revision.
I can't see an issue for washingtonpost.com, maybe it's really related with the scripts you've mentioned. Could you confirm it (by disabling your scripts and see if the issue's still there) and provide the scripts you are using, for further investigation?
The source page URL is recorded in the source code of the saved page but not shown directly. You'll be able to see it from the metadata if the backend server is used; otherwise you can see it from the source code. We are still investigating an appropriate way to present such metadata without altering the document too explictly.
As this addon site doesn't allow discussion, you can report issues to the source code repo (like the link provided above) so that we can discuss and trace them better:)
The issue on nytimes.com is same as the one with styled components and we are working on it (https://github.com/danny0838/webscrapbook/issues/109). It's a complicated issue as there are many things behind the scene to deal with. We almost have the solution but still need sometime to implement it, maybe next one or two revision.
I can't see an issue for washingtonpost.com, maybe it's really related with the scripts you've mentioned. Could you confirm it (by disabling your scripts and see if the issue's still there) and provide the scripts you are using, for further investigation?
The source page URL is recorded in the source code of the saved page but not shown directly. You'll be able to see it from the metadata if the backend server is used; otherwise you can see it from the source code. We are still investigating an appropriate way to present such metadata without altering the document too explictly.
As this addon site doesn't allow discussion, you can report issues to the source code repo (like the link provided above) so that we can discuss and trace them better:)
리뷰 138개
- 5점 만점에 5점Firefox 사용자 18834571 님, 3달 전Удалось в первозданном виде сохранить страницу, которую не сохранял ни один браузер. Автору большое спасибо и успехов в работе!
- Unable to get this to work on Firefox 133 and Fedora 40. Installed the backend and the browser extension. Could not find any configuration for the backend that would work. There needs to be a more specific set of instructions written by someone who has successfully installed this extension in Linux, unless this only works in Windows. Perhaps this is a great extension but it is worthless if it cannot be installed.
개발자 답글
5달 전에 게시됨Have you read the documentation: https://github.com/danny0838/webscrapbook/wiki/Basic#3-browser-sidebar-approach ? - 5점 만점에 5점Firefox 사용자 18235051 님, 일 년 전I needed to backup a website that used form login, making a simple scraping not possible. This extension worked like a charm after figuring out some of the configuration options.
- 5점 만점에 5점Firefox 사용자 14643647 님, 2년 전
- 5점 만점에 1점Firefox 사용자 13058149 님, 3년 전I choose the WebScrapBook/data option but it wants me to configure Backend server... ? ? ?
Useless.
I miss the old Scrapbook Extension.개발자 답글
3년 전에 게시됨Please consult the documentation about different approaches to capture a page: https://github.com/danny0838/webscrapbook/wiki/Basic. Raise an issue with more details (e.g. a screenshot illustrating where you are asked for backend server configuration) in the source repository if you still don't get it. - 5점 만점에 5점Firefox 사용자 12472805 님, 3년 전
- 5점 만점에 1점Firefox 사용자 13474132 님, 3년 전Tanto utile prima ora completamente inusabile perchè oltremodo macchinoso.
- 5점 만점에 5점Firefox 사용자 15902721 님, 3년 전