So here it is, the GeekFlirt Lolcat Maker. Its got some cool image generation and ajaxy goodness, have fun.
Enter your domain name, select which version is your primary domain (with the WWW or without it) and submit. The tool will compare the number of pages indexed for a set of searches in Google to check common duplicate content problem areas. If it finds any it will tell you what they are, how bad they are, and how to fix them. So give it a whirl and tell me how you like it, just please don’t try to break it too hard.
I talk more about these duplicate content problems in my upcoming book, too.
Building on yesterday’s post about tips to make your blog rank better, here’s the plugins you’ll need to impliment that functionality. If I missed anything or there are better alternatives, please let me know.
- Optimal Title make sure your post titles occur at the begining of each permalink page’s title tag.
- Related Entries Deep linking baby.
- Google Sitemap Generator probably won’t make you rank better, but does provide good info.
- Sociable leverage social networks and parasite SEO, in a non-spammy way of course.
- yes-WWW add the WWW to all requests (using an SEO friendly 301 redirect) I use no-WWW (see below).
- no-WWW you know the drill.
- Top Posts intelligent deep linking, baby (thanks Derek).
- Permalink Redirect use handy 301’s to prevent clean permalink URLs don’t cause duplicate content issues.
Yet another great post today, this one from SEOmoz is about the problems and choices presented to SEOs by search engines because of their API and automated scraping policies.
I admit it. SEOmoz is a search engine scraper – we do it for our free public tools, for our internal research and we’ve even considered doing it for clients (though I’m seriously concerned about charging for data that’s obtained outside TOS). Many hundreds of large firms in the search space (including a few that are 10-20X our size) do it, too. Why? Because search engine APIs aren’t accurate.
I’m right there with randfish on this. I’ve developed some tools that scrape Google SERP data and return some awesome stuff, but I’m worried about publishing them for public consumption because of course, scraping is against TOS, and the APIs aren’t accurate. I really wish I could get access to real SERP data without pissing off the big G. I’m pretty sure there is some worry about reverse engineering or something that prevents them from allowing us access to this.
Who knows, maybe I’ll take my chances and release the tools, they are pretty sweet.
A while back I linked to a demo of a script I wrote implementing the unix command tail (Like for watching the data being appended to a file), so I could tail a log file. I finally got around to posting the source code.
You’ll need saja the secure ajax for PHP framework and my saja.functions.php file as well as the actual output page, tail.php.
As with most of my code its icky and hackish, but it works. For me at least.
I just whipped up a little PHP script that impliments the *nix command tail via AJAX in a browser, using any file available through a URL. This is useful for watching log files, because it will reverse the order of the file, that is, print what is being written to the bottom of the log file to the top of the browser window for easy viewing. The script uses AJAX to push the updates to the file through to the browser every 10 seconds. To use your file, just subsitutute your file ‘s URL for the text.txt URL given into the following URL: