mod_rewrite规则不起作用

时间:2012-06-16 21:36:18

标签: apache .htaccess mod-rewrite

我已经接触过我的虚拟主机,但他们有点不太乐于助人,所以我来这里是天才。

我无法通过.htaccess文件重写我网站上的任何网址。

我只有一个.htaccess文件,它位于我的主目录的根目录中。这是该文件:

RewriteEngine On

RewriteCond %{REQUEST_FILENAME} !-d
RewriteCond %{REQUEST_FILENAME} !-f

# The support guys thought it was in issue with the L flag below, so I commented out my original implementation and used their supplied one

#RewriteCond %{HTTP_HOST} ^www.mythofechelon.co.uk$ [NC]
#RewriteRule ^(.*)$ http://mythofechelon.co.uk/$1 [R=301,L]

RewriteCond %{HTTP_HOST} ^www\.(([a-z0-9_]+\.)?mythofechelon\.co.uk)$ [NC]
RewriteRule .? http://%1%{REQUEST_URI} [R=301,L]


RewriteRule ^/$ /main/pages/index.php?home
RewriteRule ^/home(.*)?$ /main/pages/index.php?home
RewriteRule ^/404(.*)?$ /main/pages/index.php?404

#I will eventually change the following commands to link to the rewritten URLs when this all eventually works
DirectoryIndex /main/pages/index.php?home
ErrorDocument 404 /main/pages/index.php?404

AddType application/x-shockwave-flash swf

Options All -Indexes

#Protect .htaccess
<files .htaccess>
    order allow,deny
    deny from all
</files>

<Files *.reg>
    ForceType application/pdf
    Header set Content-Disposition attachment
</Files>

#Block bots
<limit GET POST HEAD>
    Order Allow,Deny
    Allow from all
    Deny from env=bad_bot
</limit>

RewriteRule ^.* - [F,L]
RewriteCond %{HTTP_USER_AGENT} ^Anarchie [OR]
RewriteCond %{HTTP_USER_AGENT} ^ASPSeek [OR]
RewriteCond %{HTTP_USER_AGENT} ^attach [OR]
RewriteCond %{HTTP_USER_AGENT} ^autoemailspider [OR]
RewriteCond %{HTTP_USER_AGENT} ^BlackWidow [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Bot\ mailto:craftbot@yahoo.com [OR] 
RewriteCond %{HTTP_USER_AGENT} ^ChinaClaw [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Custo [OR] 
RewriteCond %{HTTP_USER_AGENT} ^DISCo [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Download\ Demon [OR] 
RewriteCond %{HTTP_USER_AGENT} ^eCatch [OR] 
RewriteCond %{HTTP_USER_AGENT} ^EirGrabber [OR] 
RewriteCond %{HTTP_USER_AGENT} ^EmailSiphon [OR] 
RewriteCond %{HTTP_USER_AGENT} ^EmailWolf [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Express\ WebPictures [OR] 
RewriteCond %{HTTP_USER_AGENT} ^ExtractorPro [OR] 
RewriteCond %{HTTP_USER_AGENT} ^EyeNetIE [OR] 
RewriteCond %{HTTP_USER_AGENT} ^FlashGet [OR] 
RewriteCond %{HTTP_USER_AGENT} ^GetRight [OR] 
RewriteCond %{HTTP_USER_AGENT} ^GetWeb! [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Go!Zilla [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Go-Ahead-Got-It [OR] 
RewriteCond %{HTTP_USER_AGENT} ^GrabNet [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Grafula [OR] 
RewriteCond %{HTTP_USER_AGENT} ^HMView [OR] 
RewriteCond %{HTTP_USER_AGENT} HTTrack [NC,OR] 
RewriteCond %{HTTP_USER_AGENT} ^Image\ Stripper [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Image\ Sucker [OR] 
RewriteCond %{HTTP_USER_AGENT} Indy\ Library [NC,OR] 
RewriteCond %{HTTP_USER_AGENT} ^InterGET [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Internet\ Ninja [OR] 
RewriteCond %{HTTP_USER_AGENT} ^JetCar [OR] 
RewriteCond %{HTTP_USER_AGENT} ^JOC\ Web\ Spider [OR] 
RewriteCond %{HTTP_USER_AGENT} ^larbin [OR] 
RewriteCond %{HTTP_USER_AGENT} ^LeechFTP [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Mass\ Downloader [OR] 
RewriteCond %{HTTP_USER_AGENT} ^MIDown\ tool [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Mister\ PiX [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Navroad [OR] 
RewriteCond %{HTTP_USER_AGENT} ^NearSite [OR] 
RewriteCond %{HTTP_USER_AGENT} ^NetAnts [OR] 
RewriteCond %{HTTP_USER_AGENT} ^NetSpider [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Net\ Vampire [OR] 
RewriteCond %{HTTP_USER_AGENT} ^NetZIP [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Octopus [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Offline\ Explorer [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Offline\ Navigator [OR] 
RewriteCond %{HTTP_USER_AGENT} ^PageGrabber [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Papa\ Foto [OR] 
RewriteCond %{HTTP_USER_AGENT} ^pavuk [OR] 
RewriteCond %{HTTP_USER_AGENT} ^pcBrowser [OR] 
RewriteCond %{HTTP_USER_AGENT} ^RealDownload [OR] 
RewriteCond %{HTTP_USER_AGENT} ^ReGet [OR] 
RewriteCond %{HTTP_USER_AGENT} ^SiteSnagger [OR] 
RewriteCond %{HTTP_USER_AGENT} ^SmartDownload [OR] 
RewriteCond %{HTTP_USER_AGENT} ^SuperBot [OR] 
RewriteCond %{HTTP_USER_AGENT} ^SuperHTTP [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Surfbot [OR] 
RewriteCond %{HTTP_USER_AGENT} ^tAkeOut [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Teleport\ Pro [OR] 
RewriteCond %{HTTP_USER_AGENT} ^VoidEYE [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Web\ Image\ Collector [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Web\ Sucker [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebAuto [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebCopier [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebFetch [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebGo\ IS [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebLeacher [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebReaper [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebSauger [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Website\ eXtractor [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Website\ Quester [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebStripper [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebWhacker [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WebZIP [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Wget [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Widow [OR] 
RewriteCond %{HTTP_USER_AGENT} ^WWWOFFLE [OR] 
RewriteCond %{HTTP_USER_AGENT} ^Xaldon\ WebSpider [OR]
RewriteCond %{HTTP_USER_AGENT} ^Xenu [OR]
RewriteCond %{HTTP_USER_AGENT} ^Zeus.*Webster [OR]
RewriteCond %{HTTP_USER_AGENT} ^Zeus 

SetEnvIfNoCase user-Agent ^FrontPage [NC,OR]
SetEnvIfNoCase user-Agent ^Java.* [NC,OR]
SetEnvIfNoCase user-Agent ^Microsoft.URL [NC,OR]
SetEnvIfNoCase user-Agent ^MSFrontPage [NC,OR]
SetEnvIfNoCase user-Agent ^Offline.Explorer [NC,OR]
SetEnvIfNoCase user-Agent ^[Ww]eb[Bb]andit [NC,OR]
SetEnvIfNoCase user-Agent ^Zeus [NC]

(伙计,你们需要改变它,这样你就不必手动缩进每一行。)

我已经独立确认:

  • DirectoryIndex命令正常工作,DirectoryIndex /main/pages/index.php正常工作。
  • 显然ErrorDocument 404命令正常工作。
  • 使用PHP变量链接到文件,因为DirectoryIndexErrorDocument 404命令的当前实现可以正常工作。
  • 任何文件保护或僵尸程序拦截命令都不是问题。
  • 这不是&#34; www。&#34;的问题。删除命令,因为我已经注释并完全删除了所有尝试的实现,但仍然存在相同的问题。

问题似乎完全在于RewriteRule命令。至少在.htaccess中启用了RewriteEngine,并且mod_rewrite几天前正在重新启动我的网站。

我认为这可能是因为RewriteRule没有RewriteCond s,但这些确切的命令几天前就开始工作了。

1 个答案:

答案 0 :(得分:1)

在上面发布的.htaccess中,没有立即遵循这些规则的RewriteRule:

RewriteCond %{REQUEST_FILENAME} !-d
RewriteCond %{REQUEST_FILENAME} !-f

因此,它们将与下一个执行重定向的未注释规则相结合,但不会定义应处理这些请求的文件:

RewriteCond %{HTTP_HOST} ^www\.(([a-z0-9_]+\.)?mythofechelon\.co.uk)$ [NC]
RewriteRule .? http://%1%{REQUEST_URI} [R=301,L]

你会想要这样的东西:

# Strip www. from domain name 
RewriteCond %{HTTP_HOST} ^www\.(.+)$ [NC]
RewriteRule ^(.*)$ http://%1/$1 [R=301,L]

# Send requests for non-existent files and directories to index.php
RewriteCond %{REQUEST_FILENAME} !-d
RewriteCond %{REQUEST_FILENAME} !-f
RewriteRule ^home/(.*)$ /main/pages/index.php?$1 [L]

规则顺序对于mod_rewrite至关重要。在您的文件中,事情发生故障的另一个例子是您试图阻止坏机器人的部分。 RewriteRule必须遵循RewriteCond规则。此外,您的限制部分实际上没有做任何事情,因为关于机器人的规则实际上都没有设置环境变量。

实际上,您可以使用另一个指令来查看用户代理和设置环境变量:BrowserMatchBrowserMatchNoCase - http://httpd.apache.org/docs/2.2/mod/mod_setenvif.html#browsermatchnocase

我用这样的东西替换坏机器人的行:

BrowserMatchNoCase Anarchie bad_bot

然后将限制部分移到BrowserMatchNoCase条目下方 - 否则可能尚未设置环境变量。

此外,mod_rewrite标志对SetEnvIfNoCase条目无效。

<强>更新

要处理404,您可以添加以下内容:

RewriteCond %{REQUEST_FILENAME} !-d
RewriteCond %{REQUEST_FILENAME} !-f
RewriteRule ^!home/(.*)$ /main/pages/index.php?404 [L]

或(这就是我建议的)你可以将主页更改为。*然后更新你的php脚本以在适当时发送404。

相关问题