Robots.txt File and Indexing
Manage site indexing with a robots.txt file.
This section provides information on how to use a robots.txt
file to manage indexing for Front-End Sites.
Robots.txt File
A robots.txt
file instructs search engine crawlers on which URLs and files can be accessed on your domain. This allows you to prevent crawling and indexing to specified areas of your website.
Install and Configure the Gatsby robots.txt Plugin
You can install the gatsby-plugin-robots-txt
plugin to if you are using the Gatsby WordPress framework. The steps below are intended as a guide to help you get started. Refer to the Gatsby robots-txt-plugin documentation for more information.
Install
gatsby-plugin-robots-txt
with yarn or npm:yarn:
yarn add gatsby-plugin-robots-txt
npm:
npm install --save gatsby-plugin-robots-txt
Open your
gatsby-config.js
file and add the URLs you want to block or allow crawler access to.//gatsby-config.js module.exports = { siteMetadata: { siteUrl: 'https://www.example.com' }, plugins: ['gatsby-plugin-robots-txt'] };
Example configuration:
//gatsby-config.js module.exports = { plugins: [ { resolve: 'gatsby-plugin-robots-txt', options: { host: 'https://www.mygatsbysite.com', sitemap: 'https://www.mygatsbysite.com/sitemap.xml', policy: [{userAgent: '*', allow: '/'}] } } ] };