Updated robots.txt handler on broker

Updated robots.txt file to disallow crawling as in ticket #29565.
This commit is contained in:
Cecylia Bocovich 2019-04-15 16:43:21 -04:00
parent 6399ef9d4f
commit d4c5d4e65b

View file

@ -218,7 +218,7 @@ func debugHandler(ctx *BrokerContext, w http.ResponseWriter, r *http.Request) {
func robotsTxtHandler(w http.ResponseWriter, r *http.Request) { func robotsTxtHandler(w http.ResponseWriter, r *http.Request) {
w.Header().Set("Content-Type", "text/plain; charset=utf-8") w.Header().Set("Content-Type", "text/plain; charset=utf-8")
w.Write([]byte("User-agent: *\nDisallow:\n")) w.Write([]byte("User-agent: *\nDisallow: /\n"))
} }
func main() { func main() {