May 2007

Blue Hat Techniques20 May 2007 08:54 pm

Alrighty I’m moving this post up a bit to answer a few questions. In my Real Life SEO Example post I talked a bit about the technique of Log Link Matching. It’s an awesome technique that deserves a bit of attention. So here we go. :)

The reality of the search engines are that they only have a certain percentage of the documents on the web indexed. This is apparent by looking at your own saturation levels with your own sites. Often you’re very lucky if you get 80% of a large site indexed. Unfortunately this means that tons upon tons of the natural links out there aren’t counting and giving proper credit to you and their respective targets. This is a two edged sword. This means your competitors actually have quite a bit more links than it appears, and more than likely so do you. Naturally you can guess what has to be done. :)

Saturation usually refers to how many pages you have in the index in comparison to the total number of actual pages on your site. For instance if you have a 100 page site and 44 pages are indexed than you have 44% saturation. Since this is a topic that never really gets talked about, for the sake of making it easy on ourselves I’m going to refer to our goal as “link saturation.” In other words the number of links you have showing in the index in comparison to your total actual inbound links. So if you have 100 links in the index but you really have 200 actual links that are identifiable than you have 50% link saturation. That aside, our object is to use methods of early detection to quickly identify inbound links to our sites, get them indexed, and if possible give them a bit of link power so the link to our site will count for more. This will have an ultimate ending result of huge efficiency in our link building campaign. It also will more than likely stir up a large percentage of long dormant links on our older sites that are yet to use the Log Link Matching technique. First let’s focus on links we’ve already missed by taking a look at our log files.

Methodology #1 - The Log Files
Our site’s common log files are a great indicator of a new and old inbound links that the search engines may have missed. Most log files are usually located below the root of of the public html folder. If you’re on a standard CPanel setup the path to the log file can be easily found by downloading and viewing your Awstats config file, which is usually located in /tmp/awstats/ Around line 35 it’ll tell you the path of the log file: LogFile=”/usr/local/apache/domlogs/”. Typically your site as a Linux user has access to this file and can read it through a script. If not than contact your hosting provider and ask for read access to the log.

1) Open up the log file in a text editor and identify where all the referrers are then parse them out so you have a nice list of all the sites that link to you. If you use Textpad you can click Tools - Sort - Delete Duplicate Lines - OK. That will clean up the huge list and organize it into a manageable size.

2) Once you have your list of links there’s several routes you can take to get them indexed. These include but not limited to creating a third party rolling site map, roll over sites, or even distributing the links through blogrolls within your network. Those of course are the more complicated ways of doing it and also the most work intensive, but they’re by far the most effective simply because they involve using direct static links. The simplest of course would be to simply ping Blog Aggregators like the ones listed on Pingomatic or Pingoat. My recommendation is, if you are only getting a couple dozen links/day or are getting a huge volume of links (200+/day) than use the static link methods because they are more efficient and can be monitored more closely. If you’re somewhere in between than there’s no reason you can’t just keep it simple and continuously ping Blog Aggregators and hope a high percentage eventually will get indexed. After so many pings they will all eventually get in anyways. It may just take awhile and is harder to monitor (one of the biggest hatreds in my life..hehe).

There are several Windows applications that can help you mass ping this list of referral URLS. Since I use custom scripts instead of a single Windows app myself I have no strong recommendations for one, but feel free to browse around and find one you like. Another suggestion I have to help clean up your list a bit is to clean the list of any common referrers such as Google, MSN, and Yahoo referrals. That’ll at least save you a ton of wasted CPU time. Once you’ve gotten this taken care of you’ll want to start considering an automated way of doing this for any new links as they come in. I got a few suggestions for this as well.

Methodology #2 - Direct Referrals
Of course you can continue to do the method above to monitor for new referrals as long as you keep the list clean of duplicates. However it doesn’t hurt to consider accomplishing the task upon arrival. I talked a little bit about this last year with my Blog Ping Hack post, and the same principle applies except instead of pinging the current page we’ll ping the referral if it exists.

1) First check to see if a referral exists when the user display the page. If it does exist than have it open up the form submit for a place such as Pingomatic to automatically ping all the services using the users browser. Here’s a few examples of how to do it in various languages.

if(($ENV{'HTTP_REFERER'} ne "") || ($ENV{'HTTP_REFERER'} =~ m/http:\/\/(www\.)?$mydomain\//)) {
print qq~<iframe src="$title&blogurl=$ENV{'HTTP_REFERER'}&rssurl=$ENV{'HTTP_REFERER'}&chk_weblogscom=on&chk_blogs=on&chk_technorati=on&chk_feedburner=on&chk_syndic8=on&chk_newsgator=on&chk_feedster=on&chk_myyahoo=on&chk_pubsubcom=on&chk_blogdigger=on&chk_blogrolling=on&chk_blogstreet=on&chk_moreover=on&chk_weblogalot=on&chk_icerocket=on&chk_audioweblogs=on&chk_rubhub=on&chk_geourl=on&chk_a2b=on&chk_blogshares=on" border="0" width="1" height="1"></iframe>~;

if($_SERVER['HTTP_REFERER'] != "" || preg_match("/http:\/\/(www\.)?$mydomain\///i",$_SERVER['HTTP_REFERER'] > 0) {
echo "<iframe src="$title&blogurl=$_SERVER['HTTP_REFERER']&rssurl=$_SERVER['HTTP_REFERER']&chk_weblogscom=on&chk_blogs=on&chk_technorati=on&chk_feedburner=on&chk_syndic8=on&chk_newsgator=on&chk_feedster=on&chk_myyahoo=on&chk_pubsubcom=on&chk_blogdigger=on&chk_blogrolling=on&chk_blogstreet=on&chk_moreover=on&chk_weblogalot=on&chk_icerocket=on&chk_audioweblogs=on&chk_rubhub=on&chk_geourl=on&chk_a2b=on&chk_blogshares=on" border="0" width="1" height="1"></iframe>";

I really don’t know. :) Can someone fill this in for me? It’s entirely possible I just don’t know Javascript regex well enough.

This will check to see if the referrer exists. If it does and its not a referrer from within your domain than it’ll display an invisible IFRAME that automatically submits the referrer to PingOMatic. If you wanted to get a bit advanced with it you could also check for Google, MSN, and Yahoo referrers or any other unclean referrers you may get on a regular basis.

If you have an older site and you use this technique you’ll probably be shocked as hell about how many actual links you already had. Like I mentioned in the other post, at first you’ll start seeing your links tripling and even quadrupling but as also mentioned its just an illusion. You’ve had those links all along they just didn’t count since they weren’t indexed in the engines. After that starts to plateau, as long as you keep it up you’ll notice considerable difference in the efficiency and accuracy of your link saturation campaigns. I really believe this technique should be done on almost every site you use to target search traffic. Link Saturation is just too damn fundamental to be ignored. Yet, at the same time, its very good for those of us who are aware that it is not a common practice. Just the difference between your link saturation percentage and your competitors could be the difference between who outranks who.

Any other ideas for methods of early detection you can use to identify new inbound links? Technorati perhaps? How about ideas for ways to not only get the inbound links indexed but boost their creditability in an automated and efficient way? I didn’t mention this but when you’re pinging or rolling the pages through your indexing sites it doesn’t hurt to use YOUR anchor text, it won’t help much but it never hurts to help push the relevancy factor of your own site to their pages while you’re at it.

General Articles10 May 2007 11:49 pm

Every so often I get an Email from someone who instead of having a question or comment they give me a url and ask me to give them SEO advice on it. I normally don’t respond to these emails other than maybe a quick “did you have a specific question” type response. It’s not because I don’t want to or I take any offense to it, in fact its the complete opposite. Sorry to say I just don’t have the time to do full blown SEO analysis jobs for people. It’s not that I don’t want to, trust me I do, its just that I’m not one of those SEO bloggers whose career is blogging about SEO. I’m in the thick of it just like you guys are; day in and day out. I’m out there doing the techniques I talk about on this site everyday. Thats my job, and there’s no shortage of it. However, every once in awhile I’ll hit a gem. One of those unique situations that I can’t help but mole it around in my mind. One such example came to me last week by a reader here who was referred to by Jon Waraas(excellent blog; check out), and I think it could apply to a lot of other people in the same situation. He was kind and patient enough to allow me to publicize my response in an effort to help out others. So out of the usual context I’m going to take this example and show you all exactly what I would do in his situation.

The Situation
His site is going for the term “Myspace Layouts.” He already used to rank #1 for a solid amount of time. However after some time he admittedly got lazy and lost his ranking. He now still ranks in the top 10 steadily, but not quite top 5. He wants to rank number one again and was wondering if implementing my Keyword Fluffing and SERP Domination strategies would help.

I consider “Myspace Layouts” as a highly competitive term so this will be a great example. It’s also has the added difficulty factor of being an extremely fast growing niche; especially in just the last year. However, our mission isn’t just to take the number one rank, but to keep it intimidating for the others thinking about trying to take it themselves. So first we’ll look at what we got and then we’ll analyze what we’re up against and see if we can spot any weaknesses we can use to our advantage.

What We Got
Without even having to look at his site I can assume his on-site optimization is near perfect. He used to rank #1 steadily and where he’s at now he’s holding firm. Obviously, if he ranked at one point there’s no reason why he can’t rank again. So my first suggestion would be to not do a damn thing to the site itself. In my SERP Domination post I talked about how to break into the top 10 on a highly competitive niche by splitting up the site into a network of smaller content sites. In this situation he’s already in and has everything he needs to rank in the top position, so I would definitely advice against breaking up the site or making any drastic on-site SEO changes. Instead we’ll focus on off-site optimization and getting what links we need to earn that coveted spot again. Right now he has about 320,000 inbound links according to the Yahoo linkdomain: command. About 303,000 go directly to the main page. ~86,000 of his links come directly from Myspace Profiles. He also has a PR6.

What They Got
His top competitor has about 551,000 inbound links. 344,000 come directly from Myspace Profiles. Taking a quick look at the second and third placed competitors they are slightly less than the number one and theres nothing too notable about them. So for now we’re going to say fuck ‘em and not concern ourselves with what they’re doing because we know if we take the #1 site we’ll beat them as well. That after all is our main target and without extenuating circumstances they can be ignored in this case. So with that information out of the way for now, lets look at some strong determining factors and some weaknesses of both sites in respect to their rankings.

Spotting The Weaknesses and Loopholes
Lets analyze the math real quick. Without Myspace profile links he has 234,000 links on his own from other sites. His competitor has about 207,000. He is the clear winner in this instance which gives us a huge advantage. However, his 86,000 links directly from people’s Myspace profiles account for only about 27% of his inbound links. While his competitor clearly dominates by having his 344,000 Myspace profile links account for a whopping 62% of his total inbound links. We now have both a strength and a weakness for each of the sites. More importantly we now know why we’re loosing. Thus, we know what loopholes we’ll need to exploit in order to win.

What Do We know?
By looking at both our site and the competition we know that if we can increase our links from people’s Myspace profiles in the index to at least 62% than we will win. That means we’ll have to gain about 112,000 links on Myspace Profiles. We also know that our competitor is getting these profile links the same way we are. By giving out free Myspace layouts that include a link to our sites. There is a big problem standing our way though. Since they are ranked #1 and we’re ranked lower than we can assume they are getting more traffic than we are. Therefore, they are gaining these profile links faster than we are by giving out more layouts. It’s quite the pickle. We’ll have to do one of two things. We’ll either have to increase the value of each profile link or we’ll have to increase our volume at a much much faster rate. Neither sound like a fun solution, so we’re going to have to take a few shortcuts. One of which includes a twist on a technique I’ve already talked about, the other is an upcoming Blue Hat Technique I’m yet to discuss. First and foremost we need the proper mindset. We can’t just try to beat this guy, we need to brutally destroy him and get him so far below us that he’ll be stuck in the same situation we are in now. It’s the only way to maintain our competitive edge. So get any thoughts of playing nice out of your head right now. They aren’t going to do you any good.

Establishing A Plan Of Attack
First we’ll need to do an interesting twist on my Link Laundering technique and merge it with our SERP Domination tactic. Remember when I mentioned in one of the comments that all the techniques I’ve talked about on this site fit together perfectly like a puzzle to create an ultimate ranking strategy? I wasn’t fucking around. There’s no reason why you can’t create a network of sites within subniches that launder targeted links to your main site. He already mentioned in his original email that he wanted to create a few sites based on Myspace subniches to help build link popularity to his main site. Except, knowing what we know now, he doesn’t need more links from outside sites. He needs more links from Myspace Profiles. So I’d recommend doing just that. Build several sites on Myspace subniches such as image uploaders, profile editors, webproxies..etc. However, anytime they have an opportunity to get a link on the persons profile, instead put in the url to your main site. Looking at the current scenario he’s going to need quite a few of these sites in order to catch up, I’d say about 15 that perform on the average or even a little below average. It also wouldn’t hurt to sponsor a few of these types of sites in exchange for them laundering out links to you. I know places like Myspace Web Proxies have a hard time finding and keeping converting ads on their sites, they could be a very cheap solution. Just as long as they are consistently getting your main site links on peoples profiles thats our ultimate goal.

You got to understand, this guy only has 344,000 links on profiles. Thats not a big deal. If his site has been up for the last two years thats only about 15,000/day, not counting a sharp increase from when he started ranking of course. Getting 15 subniche sites within your network getting about 1,000 uses/day isn’t too incredibly hard. You can get that within a month or two. Once you accomplish that, you are at least matching him which is progress in the right direction.

Matching him? I’m sorry I fucked up. At that rate you’re no where NEAR matching him. There is the inevitable law of diminishing returns standing in your way. There are over 100 million Myspace profiles, only about 10 million exist in Google right now. Therefore less than 10% actually gets indexed. Since it’s only the ones that get indexed that matters to your rank than that means even if you match the 15,000/day; You’re really only getting 1,500 that matter. That’s not nearly enough. Consider your ass kicked. :) Or is it? We may need the help of a good Ol’ Blue Hat Technique to bridge the gap and turn that 10% into nearly 100%. The technique is called Log Link Matching. I’m not going to go into intricate detail about it until the post comes out of course, but I will hint to it and explain as it applies to this case.

Blue Hat Technique #XX - Log Link Matching
Anyone with experience can attest that not nearly all of the true links to your sites actually gets indexed and count in the search engines. It’s unfortunate but true. Not every page is indexed by the engines such as Google and therefore even if they link to you, they can’t pass a value. This is especially true for social sites like Myspace where only about 10% of the profiles get indexed. Log Link Matching is a technique used to ensure that nearly 100% of your real inbound links gets actually indexed by the search engines, thus
dramatically increasing your visible inbound links and bumping your rank in an indescribable way. Here’s how it works.

First create an automated Roll Over Site using my Abandoned Wordpress Accounts Part 2 post. Did I mention all my techniques merge together? Well they do. :) Do this by first building up several Wordpress blogs into high PR blogs by doing the method as described. This will get plenty of search engine spiders to the site on a steady and fast stream as well as give plenty of authority to the site(to boost each site). Lastly, make a system that allows you to quickly add links to your Blogrolls within these accounts (quit being lazy and learn how to code!).

After you got your blogs setup start parsing through your site’s log files. The log files are the first indicator of someone linking to your site. Everytime someone pulls an image on their Myspace account from your server it will show up in there, regardless of whether or not their profile actually shows up in the engines. Where there is a profile pulling your images there is more than likely a link to your site. So you must get that profile page indexed in the search engines in order for the link to count. So parse through your log file for any referrers coming from or*. Remember is a page on its own just like Therefore BOTH count. So there’s no reason why our site can’t instantly almost double all of his links from Myspace profiles at the exact same time as getting nearly 100% of them indexed. All he has to do is start adding each of these referrers to his blogrolls. After so many will automatically limit how many links actually show up and start randomizing which ones are shown on each and every pageview(or spider view). I think that limit is 15 currently but I’m not entirely sure. It seems to change every so often. Either way every time Googlebot visits the rollover sites it’ll be greeted by XX amount of Myspace profiles to index. Eventually it’ll get almost all of them. Thus the illusion of our inbound links tripling and even quadrupling daily will start happening. We won’t actually be performing nearly to the degree it appears, but we definitely won’t be short any needed momentum either. It’ll just appear that way because many links we’ve had for a long time will finally start showing up and giving us proper credit.

If we keep parsing the log files and checking for new people using our templates and ensuring they all get indexed eventually we will catch up to that #1 site. Coincide that with our link laundering sites within our network we should have no problem overtaking him and holding our ground. Once we are at that point, its check and mate for the time being. He’s either going to have to top that performance or back off and accept his second place trophy. If he does manage to pull something sneaky and come back, no big deal, persistence is worth more than gold.

I hope this little real life example helped a few people in the same situation, and I’ll move up the date on that Blue Hat Log Link Matching Technique so you can get the details of it and really learn how to utilize it in some very powerful ways. It never hurts to have more in our arsenal. :) Just remember, the fact that you already have more inbound links from other Myspace layout related sites than him. If you can just match him on his strong points you can beat him.

Go Get ‘Em Tiger!

Black Hole SEO06 May 2007 01:23 pm

This is a follow post to my 100’s Of Links/Hour Automated - Black Hole SEO post.

I’m going to concede on this one. I admittedly missed a few explanations of some fundamentals that I think left a lot of people out of all the fun. After reading a few comments, emails and inbound links (thanks Cucirca & YellowHouse for good measure) I realize that unless you already have adequate experience building RSS Scraper Sites than its very tough to fully understand my explanation of how to exploit them. So I’m going to do a complete re-explanation and keep it completely nontechnical. This post will become the post to explain how it works, and the other will be the one to explain how to do it. Fair enough? Good, lets get started with an explanation of exactly what a RSS scraper site is. So once again, this time with a MGD in my hand, cheers to Seis De Mayo!

Fundamentals Of Scraping RSS
Most blogs automatically publish an RSS feed in either a XML or ATOM format. Here’s mine for an example. These feeds basically consist of a small snipplet of your post(usually the first 500 or so characters) as well as the Title of the post and the Source URL. This is so people can add your blog into their Feed Readers and be updated when new posts arrive. Sometimes people like to be notified on a global scale of posts related to a specific topic. So there are blog search engines that are a compilation of all the RSS feeds they know about through either their own scrapings of the web or people submitting them through their submission forms. They allow you to search through millions of RSS feeds by simply entering a keyword or two. An example of this might be to use Google Blog Search’s direct XML search for the word puppy. Here’s the link. See how it resulted in a bunch of recent posts that included the word Puppy in either the title or the post content snippet (description). These are known as RSS Aggregators. The most popular of which would be, Google Blog Search, Yahoo News Search, & Daypop.

So when a black hatter in an attempt to create a massive site based on a set of keywords needs lots and lots of content one of the easiest ways would be to scrape these RSS Aggregators and use the Post Titles and Descriptions as actual pages of content. This however is a defacto form of copyright infringement since they are taking little bits of random people’s posts. The post Title’s don’t matter because they can’t be copyrighted but the actual written text can be if the person chose to have description within their feed include the entire post rather than just a snippet of it. I know it’s bullshit how Google is allowed to republish the information but no one else is, but like i said its defacto. It only matters to the beholder(which is usually a bunch of idiotic bloggers who don’t know better). So to keep in the up and up the Black Hatters always be sure to include proper credit to the original source of the post by linking to the original post as indicated in the RSS feed they grabbed. This backlink slows down the amount of complaints they have to deal with and makes their operation legitimate enough to continue stress free. At this point they are actually helping the original bloggers by not only driving traffic to their sites but giving them a free backlink, Google becomes the only real victim(boohoo). So when the many many people who use public RSS Scraper scripts such as Blog Solution and RSSGM on a mass scale start producing these sites they mass scrape thousands of posts from typically the three major RSS Aggregators listed above. They just insert their keywords in place of my “puppy” and automatically publish all the posts that result.

After that they need to get those individual pages indexed by the search engines. This is important because they want to start ranking for all these subkeywords that result from the post titles and within the post content. This results in huge traffic. Well not huge, but a small amount per RSS Scraper site they put up. This is usually done in mass scale over thousands of sites (also known as Splogs, spam blogs) which results in lots and lots of search engine traffic. They fill each page with ads (MFA, Made For Adsense Sites) and convert the click through rate on that traffic into money in their pockets. Some Black Hatters make this their entire profession. Some even create in the upwards of 5 figures worth of sites, each targeting different niches and keywords. One of the techniques they do to get these pages indexed quickly is to “ping” Blog Aggregators. Blog aggregators are nothing more than a rolling list of “recently updated blogs.” So they send a quick notification to these places by automatically filing out and submitting a form with the post title, and url to their new scraped page. A good example of the most common places they ping can be found in mass ping programs such as Ping-O-Matic. The biggest of those would probably include Weblogs. They also will do things such as comment spam on blogs and other link bombing techniques to generate lots of deep inbound links to these sites so they can outrank all the other sites going for the niche the original posts included. This is a good explanation of why is so worthless now. Black Hatters can supply these sites and generate thousands of RSS Scraped posts daily. Where legitimate bloggers can only do about one post every day or so. So these Blog Aggregator sites quickly get overrun and it can easily be assumed that about 90% of the posts that show up on there are actually pointed to and from RSS Scraper Sites. This is known as the Blog N’ Ping method.

I’m going to stop the explanation right there, because I keep saying “they” and it’s starting to bug me. Fuck guys I do this to! Haha. In fact most of the readers here do it as well. We already know tens of thousands, if not more, of these posts go up everyday and give links to whatever original source is specified in the RSS Aggregators. So all we got to do is figure out how to turn those links into OUR links. Now that you know what it is at least, lets learn how to exploit it to gain hundreds of automated links an hour.

What Do We Know So Far?
1) We know where these Splogs (RSS Scraper sites) get their content. They get them from RSS Aggregators such as Google Blog Search.

2) We know they post up the Title, Description (snippet of the original post) and a link to the Source URL on each individual page they make.

3) We know the majority of these new posts will eventually show up on popular Blog Aggregators such as We know these Blog Aggregators will post up the Title of the post and a link to the place it’s located on the Splogs.

4) We also know that somewhere within these post titles and/or descriptions are the real keywords they are targeting for their Splog.

5) Also, we know that if we republish these fake posts using these titles to the same RSS Aggregators the Black Hatters use eventually (usually within the same day) these Splogs will grab and republish our post on their sites.

6) Lastly, we know that if we put in our URL as the link to the original post the Splogs, once updated, will give us a backlink and probably post up just about any text we want them to.

We now have the makings of some serious inbound link gathering. :)

How To Get These Links
1) First we’ll go to the Blog Aggregators and make a note of all the post titles they provide us. This is done through our own little scraper.

2) We take all these post titles and store them in a database for use later.

3) Next we’ll need to create our own custom XML feed. So we’ll take 100 or so random post topics from our database and use a script to generate a .xml RSS or ATOM file. Inside that RSS file we’ll include each individual Title as our Post Title. We’ll put in our own custom description (could be a selling point for our site). Then we’ll put our actual site’s address as the Source URL. So that the RSS Scraper sites will link to us instead of someone else.

4) After that we’ll need to let the three popular RSS Aggregators listed above (Google,Yahoo,Daypop) know that our xml file exists. So, using a third script, we’ll go to their submission forms and automatically fill and submit each form with the URL to our RSS feed file( Here are the forms:

Google Blog Search
Yahoo News Search
Daypop RSS Search

Once the form is submitted than you are done! Your fake posts will now be included in the RSS Aggregators search results. Then all future Splog updates that use the RSS Aggregators to find their content will automatically pickup your fake posts and publish them. They will give you a link and drive traffic to whatever URL you specify. Want it to go to direct affiliate offers? Sure! Want your money making site to get tens of thousands of inbound links? Sure! It’s all possible from there, its just how do you want to twist it to your advantage.

I hope this cleared up the subject. Now that you know what you’re doing you are welcome to read the original post and figure out how to actually accomplish it from the technical view.

100’s Of Automated Links/Hour

Black Hole SEO05 May 2007 12:06 am

I really am holding a glass of Guinness right now so in all the authority it holds…Cheers! I’m kind of excited about this post because frankly it’s been a long time coming. For the last 7-9 months or so I’ve been hinting and hinting that there is more to Black Hat than people are willing to talk about. As “swell” as IP delivery and blog spam are there’s an awesome subculture of Black Hats that takes the rabbit hole quite a bit deeper than you can probably imagine. This is called Black Hole SEO. By no means am I an expert on it, but over the last few years I’ve been getting in quite a bit of practice and starting to really kick some ass with it. In the gist, Black Hole SEO is the deeper darker version of black hat. It’s the kind of stuff that makes those pioneering Black Hat Bloggers who dispel secrets like parasite hosting and link injection techniques look like pussies. Without getting into straight up hacking its the stuff black hatters dream about pulling off, and I am strangely comfortable with kicking in some doors on the subject. However lets start small and simple for now. Than if it takes well we’ll work our way up to some shit that’ll just make you laugh its so off the wall. Admit it, at one point you didn’t even think Advanced SEO existed. :)

In my White & Black Hat Parable post I subtly introduced this technique as well as the whole Black Hole SEO concept. It doesn’t really have a name but basically it follows all the rules of Black Hole SEO. It targets sites on a mass scale, particularly scraper sites. It tricks them into giving you legitimate and targeted links and it grabs its content on an authoritative scale (will be explained in a later related post). So lets begin our Black Hole SEO lesson by learning how to grab hundreds of links an hour in a completely automated and consenting method.

We will attempt to get black hat or scraper sites to mass grab our generated content and link to us. It’ll target just about every RSS scraper site out there, including Blog Solution and RSSGM installs including many private scrapers and Splogs.

1) First we’ll look at niche and target sources. Everyone knows the top technique for an RSS scraper is the classic Blog N’ Ping method. It’s basically where you create a scraped blog post from a search made on a popular Blog Aggregator like Google Blog Search or Yahoo Blog Search. Then they ping popular blog update services to get the post indexed by the engines. For a solid list of these checkout Something to chew on, how many of you actually go to to look for new interesting blog posts? Haha yeah thats what I thought. 90% of the posts there are pinged from spam RSS scraper blogs. On top of that there’s hundreds going in an hour. Kinda funny, but a great place to find targets for our link injections none the less.

2) We’ll take as an example. We know that at least 90% of those updates will be from RSS scrapers that will eventually update and grab more RSS content based upon their specified keywords. We know that the posts they make already contain the keywords they are looking for, otherwise they wouldn’t of scraped them in the first place. We also have a good idea of where they are getting their RSS content. So all we got to do is find what they want, where they are getting it from, change it up to benefit us, and give it back. :)

3) Write a simple script to to scrape all the post titles within the td class=”blogname” located between the !– START - WEBLOGS PING ROLLER — comments within the html. Once you got a list of all the titles store it in a database and keep doing it infinitely. Check for duplicates and continuously remove them.

4) Once you got all the titles steadily coming in write a small script on your site that outputs the titles into a rolling XML feed. I know I’m going to get questions about what a “rolling XML feed” is so I’ll just go ahead and answer them. It’s nothing more than an xml feed that basically updates in real time. You just keep adding posts to it as they come in and removing the previous ones. If the delay is too heavy you can always either make the feed larger (up to about 100 posts is usually fine) or you can create multiple XML feeds to accommodate the inevitably tremendous volume. I personally like the multiple feed idea.

5) Give each post within the feed the same title as you scraped from Weblogs. Then change the URL output field to your website address. Not the original! Haha that would do no good obviously. Then create a nice little sales post for your site. Don’t forget to include some html links inside your post content just in case their software forgets to remove it.

6) Ping a bunch of popular RSS blog search sites. The top 3 you should go for are:
Google Blog Search
Yahoo News Search
Daypop RSS Search

This will republish your changed up content so the RSS scrapers and all the sites you scraped the titles from will grab and republish your content once again. However, this time with your link. This won’t have any affect on legitimate sites or services so there really are no worries. Fair warning: be sure to make the link you want to inject into all these Splogs and scraped sites as a quickly changed and updated variable because this will gain you links VERY quickly. Lets just say I wasn’t exaggerating in the title :) A good idea would be to put the link in the database, and every time the XML publishing script loops through have it query it from the database. That way you can change it on the fly as it continuously runs.

As you’ve probably started to realize this technique doesn’t just stop at gaining links quickly, it’s also a VERY powerful affiliate marketing tool. I started playing around with this technique before last June and it still works amazingly. The switch to direct affiliate marketing is easy. Instead of putting in your URL, grab related affiliate offers and once you got a big enough list start matching for related keywords before you republish the XML feed. If a match is made, put in the affiliate link instead of your link and instead of the bullshit post content put in a quick prewritten sales post for that particular offer. The Black Hat sites will work hard to drive the traffic to the post and rank for the terms and you’ll be the one to benefit. :)

Each individual site may not give you much but when you scale it to several thousands of sites a day it starts really adding up quickly. By quickly I mean watch out. By no means is that a joke. It is quick. There are more RSS scraped pages and sites that go up everyday than any of us could possibly monetize no matter how fast you think your servers are.