I finished this a little while ago, but I’ve decided to try a soft launch for some of the sweet GeekFlirt stuff I’ve got queued up.
So here it is, the GeekFlirt Lolcat Maker. Its got some cool image generation and ajaxy goodness, have fun.
I finished this a little while ago, but I’ve decided to try a soft launch for some of the sweet GeekFlirt stuff I’ve got queued up.
So here it is, the GeekFlirt Lolcat Maker. Its got some cool image generation and ajaxy goodness, have fun.
Following my own advice (which some seem to have liked) I went and created a Duplicate Content Detector tool. Its still very beta (like all the other tools I’ve posted) but its simple and it works.
Enter your domain name, select which version is your primary domain (with the WWW or without it) and submit. The tool will compare the number of pages indexed for a set of searches in Google to check common duplicate content problem areas. If it finds any it will tell you what they are, how bad they are, and how to fix them. So give it a whirl and tell me how you like it, just please don’t try to break it too hard.
I talk more about these duplicate content problems in my upcoming book, too.
Building on yesterday’s post about tips to make your blog rank better, here’s the plugins you’ll need to impliment that functionality. If I missed anything or there are better alternatives, please let me know.
Yet another great post today, this one from SEOmoz is about the problems and choices presented to SEOs by search engines because of their API and automated scraping policies.
I admit it. SEOmoz is a search engine scraper – we do it for our free public tools, for our internal research and we’ve even considered doing it for clients (though I’m seriously concerned about charging for data that’s obtained outside TOS). Many hundreds of large firms in the search space (including a few that are 10-20X our size) do it, too. Why? Because search engine APIs aren’t accurate.
I’m right there with randfish on this. I’ve developed some tools that scrape Google SERP data and return some awesome stuff, but I’m worried about publishing them for public consumption because of course, scraping is against TOS, and the APIs aren’t accurate. I really wish I could get access to real SERP data without pissing off the big G. I’m pretty sure there is some worry about reverse engineering or something that prevents them from allowing us access to this.
Who knows, maybe I’ll take my chances and release the tools, they are pretty sweet.
A while back I linked to a demo of a script I wrote implementing the unix command tail (Like for watching the data being appended to a file), so I could tail a log file. I finally got around to posting the source code.
You’ll need saja the secure ajax for PHP framework and my saja.functions.php file as well as the actual output page, tail.php.
As with most of my code its icky and hackish, but it works. For me at least.
I ported Jason Wiener‘s python POS tagger into a PHP part of speech tagger for use in some stuff I’m working on. Its rough around the edges still, but check it out.
I just whipped up a little PHP script that impliments the *nix command tail via AJAX in a browser, using any file available through a URL. This is useful for watching log files, because it will reverse the order of the file, that is, print what is being written to the bottom of the log file to the top of the browser window for easy viewing. The script uses AJAX to push the updates to the file through to the browser every 10 seconds. To use your file, just subsitutute your file ‘s URL for the text.txt URL given into the following URL:
http://www.danzarrella.com/tail.php?file=http://www.danzarrella.com/test.txt