Robots

The Robots.txt file is used to help prevent certain URLs from being found in external search results.

Robots Governance

Each robots will contain this core set of URLs:

Disallow: /stage.digital2.aveva.com
Disallow: /stage.digital3.aveva.com
Disallow: /search-results
Disallow: /search-results*
Disallow: /404
Disallow: /500
Disallow: /search
Disallow: /search/*
Disallow: Search.aspx
Disallow: Search?search*
Disallow: search?q=*
Disallow: admin
Disallow: admin/*
Disallow: assets/*
Disallow: /styles
Disallow: /styles/*

Only en will include these:

Sitemap: https://www.aveva.com/sitemap.xml
Sitemap: https://www.aveva.com/sitemap_es.xml
Sitemap: https://www.aveva.com/sitemap_de.xml
Sitemap: https://www.aveva.com/sitemap_pt.xml
Sitemap: https://www.aveva.com/sitemap_ko.xml
Sitemap: https://www.aveva.com/sitemap_ja.xml
Sitemap: https://www.aveva.com/sitemap_cn.xml
Sitemap: https://www.aveva.com/sitemap_fr.xml

Only cn will include this:

Sitemap: https://www.aveva.cn/sitemap.xml

Live Site

The correct version of the robots.txt file must show on these URLs.

If updates are reflecting at the PROD URLs, but they are not showing on these live site URLs, then ask Adobe to cache the URLs.

PROD

STAGE, QA, DEV

A recurring problem that AVEVA experiences is the URLs from the lower environments (STAGE, QA, DEV) getting into external search results.

To minimize or prevent this issue from occuring, the robots file on the lower environments should contain the following additional rule at the top of the file:

  • Disallow: /