zvi: self-portrait: short, fat, black dyke in bunny slippers (Default)
still kind of a stealthy love ninja ([personal profile] zvi) wrote in [site community profile] dw_suggestions2009-08-17 12:49 pm

Finer grained control of robots.txt

Title:
Finer grained control of robots.txt

Area:
privacy, account settings

Summary:
Allow users to upload their own robots.txt file

Description:
The current robots.txt file generated when you check "Minimize my journal's inclusion in search engine results" forbids all automated robots from indexing your journal.

However, not all robots originate with search engines. For instance, someone who did not want their journal to be googlable might still want the backup provided by archive.org. (Or, perhaps it's the other way around, you do want Google to search, but you don't want a permanent record.) Browsershots.org uses a robot to display a web page in a variety of different browsers. I'm sure there are other webservices out there (translation? found art?) which use robots and obey robots.txt directives.

I suggest permitting users to upload a robots.txt file for their journal, so they can have finer grained control of their privacy.

If everybody having a robots.txt file is too expensive, let people add individual robots to allow and have the service automatically generate a robots.txt which allows those user-agents on a person's journal.

Poll #1022 Finer grained control of robots.txt
Open to: Registered Users, detailed results viewable to: All, participants: 40


This suggestion:

View Answers

Should be implemented as-is.
24 (60.0%)

Should be implemented with changes.
6 (15.0%)

Shouldn't be implemented.
1 (2.5%)

(I have no opinion)
8 (20.0%)

(Other: please comment)
1 (2.5%)


Post a comment in response:

If you don't have an account you can create one now.
HTML doesn't work in the subject.
More info about formatting

If you are unable to use this captcha for any reason, please contact us by email at support@dreamwidth.org