<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: Battling the &#039;bots? Try bringing out the big guns</title>
	<atom:link href="http://www.ojr.org/battling-the-bots-try-bringing-out-the-big-guns/feed/" rel="self" type="application/rss+xml" />
	<link>http://www.ojr.org/battling-the-bots-try-bringing-out-the-big-guns/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=battling-the-bots-try-bringing-out-the-big-guns</link>
	<description>Focusing on the future of digital journalism</description>
	<lastBuildDate>Sun, 07 Apr 2013 15:02:13 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.5.1</generator>
	<item>
		<title>By: Robert Niles</title>
		<link>http://www.ojr.org/battling-the-bots-try-bringing-out-the-big-guns/#comment-428</link>
		<dc:creator>Robert Niles</dc:creator>
		<pubDate>Wed, 14 Dec 2005 11:55:57 +0000</pubDate>
		<guid isPermaLink="false">http://www.ojr.org/?p=947#comment-428</guid>
		<description><![CDATA[Metrics becomes a huge issue here. Webmasters routinely knock survey-based readership numbers, often with comment along the lines of &quot;I know what my traffic is -- just look at my log files!&quot; But if you are not factoring bot traffic out of your log files, you are almost certainly overstating your human traffic, and possibly by an order of magnitude, if your site is getting hit like WebmasterWorld&#039;s.

When I took over OJR, and spent some time with our log files, I was surprised to find that almost three-fouths of the site&#039;s traffic was bots. I&#039;d had sites get hit hard by bots before, but never as much as OJR was. Now, serving that traffic wasn&#039;t a problem for us, and we&#039;ve taken only a few, standard steps toward keeping out the bad guys. But it did require my doing quite a bit of work setting up new log file filters when we switched ISPs, in order to get more accurate traffic numbers.]]></description>
		<content:encoded><![CDATA[<p>Metrics becomes a huge issue here. Webmasters routinely knock survey-based readership numbers, often with comment along the lines of &#8220;I know what my traffic is &#8212; just look at my log files!&#8221; But if you are not factoring bot traffic out of your log files, you are almost certainly overstating your human traffic, and possibly by an order of magnitude, if your site is getting hit like WebmasterWorld&#8217;s.</p>
<p>When I took over OJR, and spent some time with our log files, I was surprised to find that almost three-fouths of the site&#8217;s traffic was bots. I&#8217;d had sites get hit hard by bots before, but never as much as OJR was. Now, serving that traffic wasn&#8217;t a problem for us, and we&#8217;ve taken only a few, standard steps toward keeping out the bad guys. But it did require my doing quite a bit of work setting up new log file filters when we switched ISPs, in order to get more accurate traffic numbers.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Barry Parr</title>
		<link>http://www.ojr.org/battling-the-bots-try-bringing-out-the-big-guns/#comment-427</link>
		<dc:creator>Barry Parr</dc:creator>
		<pubDate>Wed, 14 Dec 2005 09:46:19 +0000</pubDate>
		<guid isPermaLink="false">http://www.ojr.org/?p=947#comment-427</guid>
		<description><![CDATA[Coastsider was nearly booted by my hosting service because of robot-related traffic. They were also wildly inflating my traffic numbers, making it difficult to set realistic ad rates.  I was able to get my cpu load down to a reasonable level by caching some content, excluding bots from certain directories, and excluding certain bots altogether.  It was a painful and time-consuming process and one I doubt is complete. But I could not afford to exclude robots altogether.

This also raises the question of how much robot traffic advertisers are paying for.]]></description>
		<content:encoded><![CDATA[<p>Coastsider was nearly booted by my hosting service because of robot-related traffic. They were also wildly inflating my traffic numbers, making it difficult to set realistic ad rates.  I was able to get my cpu load down to a reasonable level by caching some content, excluding bots from certain directories, and excluding certain bots altogether.  It was a painful and time-consuming process and one I doubt is complete. But I could not afford to exclude robots altogether.</p>
<p>This also raises the question of how much robot traffic advertisers are paying for.</p>
]]></content:encoded>
	</item>
</channel>
</rss>