How to Add Robots.txt in Django Easy Way
Hi Everyone, In today’s tutorial I am sharing a simple and quick way to add robots.txt file in your django project. Before diving further, you must have an idea why robots.txt is needed and what exactly a robots.txt file looks like. I will provide a sample of robots.txt file in this tutorial though.
Creating Robots app is an easy way to add robots.txt in django website
Step 1:
Create an app named robots from django shell
python manage.py startapp robots
add this app to INSTALLED_APPS in settings.py
INSTALLED_APPS = [ ..... , 'robots', ]
Step 2:
Create file name urls.py inside the robots folder (this folder will be created inside your project automatically when you create an app in step1.)
Add below lines to urls.py:
from django.urls import path from . import views urlpatterns = [ path('', views.robots, name='robots'), ]
Step 3:
Open view.py from robots folder and add following code (change khalsalabs.com to your website name in the following code):
from django.shortcuts import render, HttpResponse # Create your views here. robots_file = '''User-Agent: * Disallow: Sitemap: https://khalsalabs.com/sitemap.xml ''' def robots(request): content = robots_file return HttpResponse(content, content_type='text/plain')
Step 4:
In your projects urls.py (urls file of main project)
Add single line in the end under urlpatterns followed by ,
path('robots.txt', include('robots.urls'))
TADA !! your are done
Now check your file at http://localhost:8000/robots.txt or http://