robots.txt defaults to disallow
This commit is contained in:
parent
4b98cf45f3
commit
9268fa45e7
2 changed files with 7 additions and 0 deletions
2
static/robots.txt
Normal file
2
static/robots.txt
Normal file
|
@ -0,0 +1,2 @@
|
|||
User-agent: *
|
||||
Disallow: /
|
|
@ -37,6 +37,11 @@ const copyWebpackCommands = [
|
|||
},
|
||||
force: true,
|
||||
},
|
||||
{
|
||||
from: `${STATIC_ROOT}/robots.txt`,
|
||||
to: `${DIST_ROOT}/robots.txt`,
|
||||
force: true,
|
||||
},
|
||||
{
|
||||
from: `${STATIC_ROOT}/img/favicon.png`,
|
||||
to: `${DIST_ROOT}/public/favicon.png`,
|
||||
|
|
Loading…
Reference in a new issue