X-Git-Url: http://git.squeep.com/?a=blobdiff_plain;f=docs%2Fadministration%2FCLI_tasks%2Frobots_txt.md;h=7eeedf57119d2d41ece8ba973985d32e4c248b63;hb=8871ca5aa35e9533e57b4a15420687869378a981;hp=b1de0981b8fe1436804f527ab867dbf6b0609633;hpb=e0dee833f2b192e07cd00cc4fd78f646a3cd21d9;p=akkoma diff --git a/docs/administration/CLI_tasks/robots_txt.md b/docs/administration/CLI_tasks/robots_txt.md index b1de0981b..7eeedf571 100644 --- a/docs/administration/CLI_tasks/robots_txt.md +++ b/docs/administration/CLI_tasks/robots_txt.md @@ -1,17 +1,21 @@ -# Managing robot.txt +# Managing robots.txt {! backend/administration/CLI_tasks/general_cli_task_info.include !} -## Generate a new robot.txt file and add it to the static directory +## Generate a new robots.txt file and add it to the static directory The `robots.txt` that ships by default is permissive. It allows well-behaved search engines to index all of your instance's URIs. If you want to generate a restrictive `robots.txt`, you can run the following mix task. The generated `robots.txt` will be written in your instance [static directory](../../../configuration/static_dir/). -```elixir tab="OTP" -./bin/pleroma_ctl robots_txt disallow_all -``` +=== "OTP" -```elixir tab="From Source" -mix pleroma.robots_txt disallow_all -``` + ```sh + ./bin/pleroma_ctl robots_txt disallow_all + ``` + +=== "From Source" + + ```sh + mix pleroma.robots_txt disallow_all + ```