This will useful when web site has more than 2 domains. For an example webs site is configured to use one domain as real/actual domain and other domain use for testing purposes/temporary. But if you use more than 1 domain for web site, web crawlers such as Google bot will penalize the site as it contains duplicate content.
Simplest solution is, use Nginx managed robots.txt file for a different sub domain by checking the domain name.
server {
## some code ....
listen 80;
root /path/to/webroot/folder;
index index.html index.htm index.php;
server_name realdomain.com testdomain.com;
location /robots.txt
{
if ($host = 'testdomain.com') {
return 200 "User-agent: *\n Disallow: /";
}
}
## some code ....
}