Discuz! Board

 找回密碼
 立即註冊
搜索
熱搜: 活動 交友 discuz
查看: 3|回復: 0

Then robots. Txt can be accesse through domain

[複製鏈接]

1

主題

1

帖子

5

積分

新手上路

Rank: 1

積分
5
發表於 2024-3-5 18:04:13 | 顯示全部樓層 |閱讀模式
The disadvantage is that these tools are not freely customizable to some extent. Where to place the robots. Txt file? Place the robots. Txt file in the root directory folder of the corresponding domain name/subdomain name. For example. If.Com/robots.Txt . If you want to control the access restrictions of the second-level domain name. Such as blog. Domain.Com . Then its robots.Txt nees to be accesse through blog.Domain.Com/robots.Txt . Best practices for robots.Txt keep the following tips in mind to avoid unnecessary mistakes: each new instruction requires a new line each command requires a new line.


Otherwise. Search engines will misunderstand: error  France WhatsApp Number Data example: user-agent: * disallow: /directory/ disallow: /another-directory/ standard example: user-agent: * disallow: /directory/ disallow: /another-directory/ use wildcards to simplify directives not only can you use the wildcard character (*) to apply the directive to all user agents. But you can also use it to match identical urls when declaring the directive. For example. If you wante to prevent search engines from accessing parameterize product category urls on your site. You could list them like this: user-agent: * disallow: /products/t-shirts? Disallow: /products/hoodies? Disallow: /products/jackets? … but this is not concise. You can use wildcards. Abbreviate as follows: user-agent: * disallow: /products/*? This example blocks all search engine users from crawling all links containing question marks (?) in the /product/ directory.



In other words. All product links with parameters are blocke. Use the dollar sign ($) to annotate urls ending with specific characters add $ at the end of the command. For example. If you want to block all links ending in .Pdf. You can set your robots. Txt like this: user-agent: * disallow: /*. Pdf$ in this example. Search engines cannot crawl any links ending with .Pdf. Which means search engines cannot crawl /file. Pdf. But search engines can crawl this /file.Pdf?Id=68937586 because it does not end with . Pdf end. The same user agent is declare only once if you declare the same user agent multiple times.
回復

使用道具 舉報

您需要登錄後才可以回帖 登錄 | 立即註冊

本版積分規則

Archiver|手機版|自動贊助|GameHost抗攻擊論壇

GMT+8, 2024-11-23 10:21 , Processed in 0.034565 second(s), 18 queries .

抗攻擊 by GameHost X3.4

Copyright © 2001-2021, Tencent Cloud.

快速回復 返回頂部 返回列表
一粒米 | 中興米 | 論壇美工 | 設計 抗ddos | 天堂私服 | ddos | ddos | 防ddos | 防禦ddos | 防ddos主機 | 天堂美工 | 設計 防ddos主機 | 抗ddos主機 | 抗ddos | 抗ddos主機 | 抗攻擊論壇 | 天堂自動贊助 | 免費論壇 | 天堂私服 | 天堂123 | 台南清潔 | 天堂 | 天堂私服 | 免費論壇申請 | 抗ddos | 虛擬主機 | 實體主機 | vps | 網域註冊 | 抗攻擊遊戲主機 | ddos |