0
0
Fork 0
mirror of https://codeberg.org/forgejo/docs.git synced 2024-12-10 20:44:45 -05:00
forgejo-docs/docs/admin/search-engines-indexation.md
Earl Warren ae97171f57 sync with the latest changes from Gitea
git log --patch faa28b5a44912f1c63afddab9396bae9e6fe061c..8d9e2d07f3f84a86265fdbe0ab7fcf63cc34ddbd . ':(glob)docs/**/*.en-us.md'

(cherry picked from commit b77359bfb1)
2023-09-22 17:36:27 +00:00

42 lines
1.4 KiB
Markdown

---
title: 'Search Engines Indexation'
license: 'Apache-2.0'
origin_url: 'https://github.com/go-gitea/gitea/blob/8d9e2d07f3f84a86265fdbe0ab7fcf63cc34ddbd/docs/content/administration/search-engines-indexation.en-us.md'
---
## Search engines indexation of your Forgejo installation
By default your Forgejo installation will be indexed by search engines.
If you don't want your repository to be visible for search engines read further.
### Block search engines indexation using robots.txt
To make Forgejo serve a custom `robots.txt` (default: empty 404) for top level installations,
create a file called `robots.txt` at the root of the `CustomPath` as displayed in the `/admin` page.
Examples on how to configure the `robots.txt` can be found at [https://moz.com/learn/seo/robotstxt](https://moz.com/learn/seo/robotstxt).
```txt
User-agent: *
Disallow: /
```
If you installed Forgejo in a subdirectory, you will need to create or edit the `robots.txt` in the top level directory.
```txt
User-agent: *
Disallow: /forgejo/
```
### Disallow crawling archives to save disk space
If the archive files are crawled, they will be generated dynamically
and kept around which can amount to a lot of disk. To prevent that
from happening, add the following to the `robots.txt` file:
```txt
User-agent: *
Disallow: /*/*/archive/
```
See also a more complete example [at Codeberg](https://codeberg.org/robots.txt).