Optimization of Robots.txt File In Opencart Marketplace

What is a Robots.txt file?

Robots.txt is a standard file that the websites use to instruct the spiders or web crawlers such as Googlebot when they are visiting your websites or marketplaces.

So basically, you are telling Google and Bing or MSN or all other basic search engines what they can and can not see on your websites. More importantly, what they are allowed to see and where they are allowed to go on your websites or marketplaces.

Why do I need a robot.txt file?

If you are not using a robot.txt file then a lot of things that can go wrong with your marketplace or a website like –

How to create a robots.txt file?

First, you need to check whether your website is having a robots.txt file, you can check it by going to your website URL and adding the text robots.txt at the end – www.example.com/robots.txt. If you are having this file then you will see it or else if it’s not then you can create one easily.

You can create a robots.txt file by using a whole bunch of text editors that are available in the market. Open your notepad to start writing the robots.txt file.

After you have created this file with what you want to allow and disallow, you have to upload it to your website’s root folder.

A screenshot has been shared below to let you know how the Opencart robots.txt file appears visually.

For example –

User-agent : *
Allow: / 

This says that any web crawlers coming to the website can crawl everything. If you want to block your website from all crawlers then just replace Allow: / by Disallow: / in the above example.

Certain people specify specific user agents like the Googlebot or other random ones that people typically do really focus on and will start the robots.txt with User-agent: Googlebot and then they will disallow certain parts of their websites.

User agent : *
Disallow: /staging/   

This is just a suggestion to any of the crawlers coming to the site not to crawl this folder. But if any malicious crawler comes it can ignore the robots.txt file and hence is not a good security mechanism.

There are advantages and disadvantages to this. This is not a way to hide your sensitive data as it’s just an instruction to crawl or not.

<html>
<head>
<title>My Website</title >
<meta name= “robots” content = “noindex,  nofollow
</head>
</html>

This code will instruct “Do not index this content“, hence the crawlers will not be able to index this website page and provides a good mechanism in hiding the sensitive data on your website or marketplaces from the crawlers.

Thus making the best or most effective use of the robots.txt file you will be able to make a set of rules for the search crawlers. The rules will guide them on what they are allowed to see and where they are allowed to go on your websites. For more information click here robots.txt specifications.

Hope this was somewhat helpful. Stay put! for the next article that’s on Search Engine Friendly URLs.

Support

That’s all for Optimization of Robots.txt File In Opencart Marketplace still have any issues feel free to add a ticket and let us know your views to make the module better contact us.

 

Categories: blog
Leave a comment
Exit mobile version