- 05 Oct, 2015 1 commit
-
-
Ben Bodenmiller authored
Commit page has valuable information that search engines should be allowed to crawl however the .patch and .diff pages have no new information that is not on commit page
-
- 18 Aug, 2015 1 commit
-
-
Ben Bodenmiller authored
Update default robots.txt rules to disallow irrelevant pages that search engines should not care about. This will still allow important pages like the files, commit details, merge requests, issues, comments, etc. to be crawled.
-
- 13 Oct, 2011 2 commits
-
-
Dmitriy Zaporozhets authored
-
gitlabhq authored
-
- 08 Oct, 2011 1 commit
-
-
gitlabhq authored
-