0
0
Fork 0
mirror of https://codeberg.org/forgejo/docs.git synced 2024-12-26 22:51:22 -05:00
forgejo-docs/admin/search-engines-indexation.md
Caesar Schinas e688f3324b
fix: remove h1-level headings in docs
Increases depth of markdown headings to avoid 1st-level headings

# Conflicts:
#	v1.19/user/project.md
#	v1.19/user/protection.md
#	v1.19/user/push-to-create.md
2023-08-13 17:08:33 +01:00

42 lines
1.4 KiB
Markdown

---
title: 'Search Engines Indexation'
license: 'Apache-2.0'
origin_url: 'https://github.com/go-gitea/gitea/blob/faa28b5a44912f1c63afddab9396bae9e6fe061c/docs/content/doc/administration/search-engines-indexation.en-us.md'
---
## Search engines indexation of your Forgejo installation
By default your Forgejo installation will be indexed by search engines.
If you don't want your repository to be visible for search engines read further.
### Block search engines indexation using robots.txt
To make Forgejo serve a custom `robots.txt` (default: empty 404) for top level installations,
create a file called `robots.txt` at the root of the `CustomPath` as displayed in the `/admin` page.
Examples on how to configure the `robots.txt` can be found at [https://moz.com/learn/seo/robotstxt](https://moz.com/learn/seo/robotstxt).
```txt
User-agent: *
Disallow: /
```
If you installed Forgejo in a subdirectory, you will need to create or edit the `robots.txt` in the top level directory.
```txt
User-agent: *
Disallow: /forgejo/
```
### Disallow crawling archives to save disk space
If the archive files are crawled, they will be generated dynamically
and kept around which can amount to a lot of disk. To prevent that
from happening, add the following to the `robots.txt` file:
```txt
User-agent: *
Disallow: /*/*/archive/
```
See also a more complete example [at Codeberg](https://codeberg.org/robots.txt).