File: help.ubuntu.com

package info (click to toggle)
python-protego 0.5.0%2Bdfsg-1
  • links: PTS, VCS
  • area: main
  • in suites: forky, sid
  • size: 30,052 kB
  • sloc: python: 1,579; perl: 190; cpp: 33; sh: 4; makefile: 3
file content (44 lines) | stat: -rw-r--r-- 1,402 bytes parent folder | download | duplicates (4)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
# https://help.ubuntu.com/robots.txt
#
# Notes:
#
# 2016.11.23: We, the docs team, do not actually have access to the
#	web server access logs, which would help in improving this
#	robots.txt file. Some things we do herein are based on experiences
#	from other web servers, where we do have access to the access logs.
#	Always keep in mind that crawlers, and google in particular, seem
#	to keep looking for deleted content for a very very long time after
#	it is gone.
#
# 2016.11.23: Canonical is moving the CloudDocs elsewhere.
#	A disallow lts/clouddocs is being added in order to assist
#	web crawlers to realize that they should delete that content.
#	It should be left for at least a year, probably two.
#	For this type of server, a crawl delay of 20 is too long,
#	changing to 5, which might still be a little long.

User-agent: *
Crawl-delay: 5
Disallow: /img/
Disallow: /libs/
Disallow: /14.04/
Disallow: /16.04/
Disallow: /16.10/
Disallow: /17.04/
Disallow: /17.10/
Disallow: /18.04/
Disallow: /18.10/
Disallow: /19.04/
Disallow: /19.10/
Disallow: /20.04/
Disallow: /20.10/
Disallow: /21.04/
Disallow: /21.10/
Disallow: /lts/ubuntu-help/
Disallow: /stable/installation-guide/
Disallow: /stable/serverguide/
Disallow: /dev/
# Leave the following two lines at least until 2017.12.01, perferrably 2018.12.01
Disallow: /stable/clouddocs/
Disallow: /lts/clouddocs/
Disallow: /community/*?action=