Config Using Route Rules
Learn how to configure robots through route rules.
If you prefer, you can use route rules to configure how your routes are indexed by search engines.
You can provide the following rules:
{ robots: false } - Will disable the route from being indexed using the robotsDisabledValue config.{ robots: '<rule>' } - Will add the provided string as the robots rule
The rules are applied using the following logic:
X-Robots-Tag header - SSR only,<meta name="robots"> - When using thedefineRobotMeta orRobotMeta composable or component/robots.txt disallow entry - When disallowNonIndexableRoutes is enabled
Inline Route Rules
Requires enabling the experimental
<script lang="ts" setup>
defineRouteRules({
robots: false,
})
</script>
Nuxt Config
nuxt.config.ts
export default defineNuxtConfig({
routeRules: {
// use the `index` shortcut for simple rules
'/secret/**': { robots: false },
// add exceptions for individual routes
'/secret/visible': { robots: true },
// use the `robots` rule if you need finer control
'/custom-robots': { robots: 'index, follow' },
}
})
Did this page help you?