This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 2 years ago.
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
接受
或继续使用网站,即表示您同意使用 Cookies 和您的相关数据。
发布评论
评论(2)
ppspy做以下内容。
它读取您的Shopify站点地图以在商店中找到产品。
.../stitemap_products_1.xml
作为后备,它解析了网址:
.../collections/all?sort_by =最畅销 - 和尝试
在那里找到产品。
接下来,它使用Shopify的JSON URL。再次试图找到
所有产品。一个示例URL:
.../products.json?page = 1& limit = 250-大多数商店所有者甚至都不知道存在。
之后,它调用了每种产品的JSON URL。你可以得到这个
只需打开产品页面并写作,在您的在线商店中的URL
“ .json”在URL中。示例URL:
.../products/your-productName.json。
在此JSON中,有一个“ Updated_at”字段。每次进行更改时,都会更新此字段。另外,发生订单时(股票更改)。
因此,可以(大约)跟踪销售。
PPSPY does the following.
It reads your Shopify sitemap to find the products in the store.
.../sitemap_products_1.xml
As fallback, it parses the URL:
.../collections/all?sort_by=best-selling - and tries
to find the products there.
Next, it uses the JSON URL from Shopify. There again it tries to find
all products. An example URL:
.../products.json?page=1&limit=250 - most store owners don't even know this exists.
After that, it calls the JSON URL for each product. You can get this
URL in your online store by simply opening a product page and writing
".json" after it in the URL. Example URL:
.../products/your-productname.json.
In this JSON there is a field "updated_at". This field is updated every time a change is made. Also, when an order take place (the stock is changed).
And with this, it is possible to track the sales (approximately).
它们称为Web刮板或爬网。他们转到您的产品页面(浏览您的电子商务中的所有链接)并了解页面的内容。他们提取产品名称和产品价格。他们将每X小时或X天做一次,并收集信息,因此他们不需要任何网络钩。
从理论上讲,您可以使您的页面变得足够复杂,以至于不容易爬网,例如,您可以用JavaScript显示价格(爬行者通常不启用JavaScript)。但这将使您的网站访问降低,尤其是对于Google而言,这实际上是另一个爬网。
They are called web scraper or crawlers. They go to your product page (going through all the links in your ecommerce) and understand the content of the page. They extract the product name and the product price. They will do that every X hours or X days and collect the information, so they don't need any webhook.
In theory you could make your page complicated enough to not make it easy to crawl, for example you could show the price with Javascript (the crawlers typically have javascript not enabled). But that would make your website less accessible, especially to Google, which is in fact another crawler.