Releases: crwlrsoft/crawler-ext-browser
Releases · crwlrsoft/crawler-ext-browser
v2.0.0
v1.4.0
Added
Screenshot::take()
in addition toScreenshot::loadAndTake()
, allowing to take a screenshot of an already opened page, in a separate step. This way you can even add this step after anHttp::crawl()
step and get screenshots of all the crawled pages.
Fixed
- Change calls to
HttpLoader::browserHelper()
toHttpLoader::browser()
and requirecrwlr/crawler
with constraint^1.9.3
to make sure the method exists.
v1.3.0
Added
- Merge things from the
crwlr/crwl-ext-browser
package to this one, because they are too tightly coupled. The other package will be abandoned. - Add
timeout
config param toScreenshotBuilder
, therefore also requirecrwlr/crawler
v1.9.0 (or greater), with the new functionality to configure timeouts for the headless browser.
Fixed
- Prepare for
crwlr/crawler
v2.0.
v1.2.1
v1.2.0
v1.1.0
v1.0.0
Changed
- Change the output of the
Screenshot
step, from an array['response' => RespondedRequest, 'screenshotPath' => string]
to aRespondedRequestWithScreenshot
object, that has ascreenshotPath
property. The problem with the previous solution was: when using the response cache, the step failed, because it gets a cached response from the loader that was not actually loaded in the headless browser. When the step afterwards tries to take a screenshot from the page that is still open in the browser, it just fails because there is no open page. Now, with the newRespondedRequestWithScreenshot
object, thescreenshotPath
is also saved in the cached response.
v0.1.2
Fixed
- Upgrade to
crwlr/crawler
v1.5.3 and remove the separateHeadlessBrowserLoader
andHeadlessBrowserCrawler
. The steps shall simply use the normalHttpLoader
and automatically switch to use the headless browser for loading and switch back afterwards if the loader was configured to use the HTTP client.