Install Nuxt Robots

Get started with Nuxt Robots by installing the dependency to your project.

Setup Module

Want to know why you need this module? Check out the introduction.

To get started with Nuxt Robots, you need to install the dependency and add it to your Nuxt config.

bash
npx nuxi module add robots

Verifying Installation

To ensure the module is behaving as expected, you should first check /robots.txt is being generated.

It should show that the site is disallowed from indexing, this is good as development environments should not be indexed by search engines.

However, we want to see what a production environment would look like.

For this, it's recommended to use the Nuxt DevTools Robots tab to see the current configuration and how it's being applied.

The DevTools will show you that in production we're just serving a minimal robots.txt file.

robots.txt
User-agent: *
Disallow:

This allows all search engines to index the site.

Configuration

Every site is different and will require their own unique configuration, to give you a head start you may consider the following areas to configure.

  • Disallow Pages - You should consider excluding pages that are not useful to search engines, for example any routes which require authentication should be ignored.
  • Disabling Site Indexing - If you have non-production environments you should disable indexing for these environments, while this works out-of-the-box for most providers, it's good to verify this is working as expected.

Make sure you understand the differences between robots.txt vs robots meta tag with the Conquering Web Crawlers guide.

Next Steps

You've successfully installed Nuxt Robots and configured it for your project.

Documentation is provided for module integrations, check them out if you're using them.

  • Nuxt I18n - Disallows are automatically expanded to your configured locales.
  • Nuxt Content - Configure robots from your markdown files.

Otherwise, just learn more about how the module works.

Did this page help you?