mirror of
https://codeberg.org/forgejo/docs.git
synced 2024-12-24 22:32:43 -05:00
b77359bfb1
git log --patch faa28b5a44912f1c63afddab9396bae9e6fe061c..8d9e2d07f3f84a86265fdbe0ab7fcf63cc34ddbd . ':(glob)docs/**/*.en-us.md'
42 lines
1.4 KiB
Markdown
42 lines
1.4 KiB
Markdown
---
|
|
title: 'Search Engines Indexation'
|
|
license: 'Apache-2.0'
|
|
origin_url: 'https://github.com/go-gitea/gitea/blob/8d9e2d07f3f84a86265fdbe0ab7fcf63cc34ddbd/docs/content/administration/search-engines-indexation.en-us.md'
|
|
---
|
|
|
|
## Search engines indexation of your Forgejo installation
|
|
|
|
By default your Forgejo installation will be indexed by search engines.
|
|
If you don't want your repository to be visible for search engines read further.
|
|
|
|
### Block search engines indexation using robots.txt
|
|
|
|
To make Forgejo serve a custom `robots.txt` (default: empty 404) for top level installations,
|
|
create a file called `robots.txt` at the root of the `CustomPath` as displayed in the `/admin` page.
|
|
|
|
Examples on how to configure the `robots.txt` can be found at [https://moz.com/learn/seo/robotstxt](https://moz.com/learn/seo/robotstxt).
|
|
|
|
```txt
|
|
User-agent: *
|
|
Disallow: /
|
|
```
|
|
|
|
If you installed Forgejo in a subdirectory, you will need to create or edit the `robots.txt` in the top level directory.
|
|
|
|
```txt
|
|
User-agent: *
|
|
Disallow: /forgejo/
|
|
```
|
|
|
|
### Disallow crawling archives to save disk space
|
|
|
|
If the archive files are crawled, they will be generated dynamically
|
|
and kept around which can amount to a lot of disk. To prevent that
|
|
from happening, add the following to the `robots.txt` file:
|
|
|
|
```txt
|
|
User-agent: *
|
|
Disallow: /*/*/archive/
|
|
```
|
|
|
|
See also a more complete example [at Codeberg](https://codeberg.org/robots.txt).
|