如何保存包含所有媒体的公共 html 页面并保留结构
正在寻找一个 Linux 应用程序
(或 Firefox 扩展),它可以让我抓取 HTML 模型并保持页面的完整性。
Firefox 做得几乎完美,但没有抓取 CSS 中引用的图像。
Firefox 的剪贴簿扩展可以获取所有内容,但会扁平化目录结构。
如果所有文件夹都成为 index
页面的子级,我不会非常介意。
Looking for a Linux application
(or Firefox extension) that will allow me to scrape an HTML mockup and keep the page's integrity.
Firefox does an almost perfect job but doesn't grab images referenced in the CSS.
The Scrapbook extension for Firefox gets everything, but flattens the directory structure.
I wouldn't terribly mind if all folders became children of the index
page.
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。
绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(4)
请参阅使用 wget 进行网站镜像
See Website Mirroring With wget
您尝试过 wget 吗?
Have you tried wget?
wget -r
可以做你想要的事情,如果没有,有很多标志可以配置它。 请参阅man wget
。另一个选项是
curl
,它的功能更强大。 请参阅http://curl.haxx.se/。wget -r
does what you want, and if not, there are plenty of flags to configure it. Seeman wget
.Another option is
curl
, which is even more powerful. See http://curl.haxx.se/.Teleport Pro 非常适合此类事情。 您可以将其指向完整的网站,它会在本地下载一个副本,维护目录结构,并根据需要用相对链接替换绝对链接。 您还可以指定是否希望将其他第三方网站的内容链接到原始网站。
Teleport Pro is great for this sort of thing. You can point it at complete websites and it will download a copy locally maintaining directory structure, and replacing absolute links with relative ones as necessary. You can also specify whether you want content from other third-party websites linked to from the original site.