robots.txt defaults to disallow
This commit is contained in:
parent
4b98cf45f3
commit
9268fa45e7
2 changed files with 7 additions and 0 deletions
2
static/robots.txt
Normal file
2
static/robots.txt
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
User-agent: *
|
||||||
|
Disallow: /
|
|
@ -37,6 +37,11 @@ const copyWebpackCommands = [
|
||||||
},
|
},
|
||||||
force: true,
|
force: true,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
from: `${STATIC_ROOT}/robots.txt`,
|
||||||
|
to: `${DIST_ROOT}/robots.txt`,
|
||||||
|
force: true,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
from: `${STATIC_ROOT}/img/favicon.png`,
|
from: `${STATIC_ROOT}/img/favicon.png`,
|
||||||
to: `${DIST_ROOT}/public/favicon.png`,
|
to: `${DIST_ROOT}/public/favicon.png`,
|
||||||
|
|
Loading…
Reference in a new issue