webui : robots.txt, prevent search-engine indexing
This commit is contained in:
parent
07939d46b2
commit
596e8bbf21
@ -91,7 +91,8 @@ urls = (
|
|||||||
#"/downloads/(.*)","downloads" disabled until it can handle large downloads
|
#"/downloads/(.*)","downloads" disabled until it can handle large downloads
|
||||||
#default-pages
|
#default-pages
|
||||||
"/", "home",
|
"/", "home",
|
||||||
"", "home"
|
"", "home",
|
||||||
|
"/robots.txt","robots"
|
||||||
)
|
)
|
||||||
#/routing
|
#/routing
|
||||||
|
|
||||||
@ -345,5 +346,12 @@ class downloads(static_handler):
|
|||||||
if not ws.config.get('share_downloads'):
|
if not ws.config.get('share_downloads'):
|
||||||
raise Exception('Access to downloads is forbidden.')
|
raise Exception('Access to downloads is forbidden.')
|
||||||
return static_handler.GET(self, name)
|
return static_handler.GET(self, name)
|
||||||
|
|
||||||
|
class robots:
|
||||||
|
def GET(self):
|
||||||
|
"no robots/prevent searchengines from indexing"
|
||||||
|
web.header("Content-Type", "text/plain")
|
||||||
|
print "User-agent: *\nDisallow:\n"
|
||||||
|
|
||||||
#/pages
|
#/pages
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user