Skip to content

Releases: crwlrsoft/robots-txt

v1.1.1

08 Nov 12:26
Compare
Choose a tag to compare

Fixed

The Parser now also trims hidden whitespace characters that aren't covered by PHP's trim() function by default. Such characters at the beginning of a line can cause parsing to fail, because it's important that user-agent and rule lines actually start with the corresponding keywords.

v1.1.0

23 Sep 00:22
Compare
Choose a tag to compare

Added

Method isExplicitlyNotAllowedFor() that ignores rules for wildcard user-agent (*) and checks if some path is explicitly not allowed for a certain user-agent.

v1.0.0

22 Sep 20:52
Compare
Choose a tag to compare

Changed

  • Required PHP version is now 8.0.

Added

  • It now also parses Sitemap: lines. You can get all referenced sitemaps via the sitemaps() method of the RobotsTxt class.

v0.1.2

16 Sep 00:16
Compare
Choose a tag to compare

Fixed

  • Also allow usage of crwlr/url 1.0 as it's not a problem at all and the PHP version requirement of this package is still ^7.4|^8.0.

v0.1.1

16 Sep 00:06
Compare
Choose a tag to compare

Changed

  • Upgraded crwlr/url package version constraint.

v0.1.0

07 Nov 21:42
Compare
Choose a tag to compare
v0.1.0 Pre-release
Pre-release

[0.1.0] - 2021-10-07

Added

  • Parsing robots.txt files
    • user-agent groups
    • allow and disallow rule patterns
  • Check if a certain uri/path is allowed for a certain bot/user-agent according to the definitions in robots.txt
  • Implemented as described in https://datatracker.ietf.org/doc/html/draft-koster-rep