Home / Tutorials / Internet Marketing Tools / GSA Toolset / How To Filter Your GSA Search Engine Ranker Lists For A Potential 963% Increase In LPM!

How To Filter Your GSA Search Engine Ranker Lists For A Potential 963% Increase In LPM!

As A reader of my blog, you are able to save 15% on the retail price of any GSA tool! To get this discount use any of the links below or in my posts then use the coupon code “shaunmarrs” on the checkout page.

GSA Search Engine Ranker
GSA Captcha Breaker
GSA Content Generator
GSA Platform Identifier
GSA Proxy Scraper

This post is part of my GSA Search Engine Ranker list filtering series and should be read in the following order.

List Filtering Theory

This Post

Contextual Filtered List Case StudyNon-Contextual Filtered List Case Study

In this post, I will explaining how you actually go about filtering your lists correctly.

Breaking The Mold With Your Folders

Now, I touched on this next part in my ultimate guide to GSA Search Engine Ranker and although I highly recommend you read it I will cover the basics quickly as it is almost 20,000 words long.

By default SER comes with four folders you are able to store links in, these are the identified, submitted, verified and failed folders and you can find the settings for them by clicking the options button and selecting the advanced tab. A common mistake many people make is to think these folders can only be used to store targets relevant to their names, for example only identified targets in the identified folder but this is incorrect.

These folders are set up with these names as they are essential for GSA SER to scrape targets for itself. They have a tick box to the left of their paths on the advanced tab that if ticked will trigger some hardcoded actions in SER meaning the tool will automatically write targets to these folders, as we want to keep these folders clean and only hold the targets we want them to we will not tick any of these tick boxes.

Now this next part is 100% theory on my part yet an essential step if it is correct and you are using a premium list. It is my belief that when you make a project and tell it to pull its targets from either of these four folders in the project options tab, SER will pull every single target in there no matter what you have selected in your project options. I have seen people on the GSA forum saying this is not what happens but due to my personal observations, I feel that my theory is correct.

Now, people who sell premium lists provide targets for all of the engines that SER supports including the ones you are unlikely to use in live projects such as the exploit and RSS platforms amongst a few others.

Let’s say that you have a project that has nothing but the article platforms selected and you tell the project to pull its targets from a folder. It is my belief that when this project goes active SER will pull the article targets from that folder and submit them but once it is out of targets it then starts to crawl all of the other platforms .txt in that folder checking if their footprints match with an article engine for it to post to.

As you can imagine this has the potential to slow SER down massively and when you are running filter projects you want SER running as fast as humanly possible.

[update] Although this is still technically only my theory I have since published this case study. In it, I show how the example list made in this very tutorial is able to produce an increase of 963.33% to project links per minute as well as an increase of 1380.48% in link submissions per hour when compared to an unfiltered list premium list.

Upgrading To Folders 2.0

Personally, I map all four of my folder locations to my dropbox account so they can automatically be synced between every VPS I need them on but you can map them to wherever you see fit. For this example, we are going to presume you have all of your verified targets that you wish to filter saved to the verified folder.

Now there are a few ways you can set your folders up but this is how I am currently using mine although I do sometimes put a second premium list service into the submitted folder rather than targets to re-verify.

  • Identified – Targets to filter from my premium list service.
  • Submitted – Targets from my verified and failed folder to re-filter.
  • Verified – Do Follow article, social network and wiki targets.
  • Failed – Do Follow blog comment, guestbook and image comment targets.

Now, because of the theory I explained earlier I only move the specific platform .txt files into my identified folder from the ones provided by the premium list service that I plan to use in future live projects. In my case, that’s Articles, Blog Comments, Guestbooks, Image Comments, Social Networks and Wikis. To keep this process quick I open the folder provided by the premium list service, click the top entry for a platform in the folder, scroll to the bottom one and shift click it before copying and pasting it into the relevant mapped SER folder.

Filtering Out Inefficient Engines

Normally once these platforms have been copied over I open the folder they are in and arrange it by size in descending order so the largest files are at the top. Due to the theory, I explained earlier I delete the following engines as well as a few others as their file is usually huge.

  • PHPWeb – Asian anchor with a large amount of domains having adult banner ads.
  • MediaWiki – Almost 100% No Follow
  • Gallary2 – Almost 100% No Follow

Now this next engine is down to personal choice, I currently process the Joomla K2 engine file as I am running it in some tests but not on live projects yet. I know a fair few people who do not use the Joomla K2 engine at all though so would delete its targets file.

Also, I know that the General Blogs engine file is huge and the majority of the targets are no follow but I still choose to process it as it sheer size still provides enough do follow targets for it to be my highest target count engine at the end of the process.

However, in this example, the only one of the above engine files I am going to remove from the process at this stage is PHPWeb. I am keeping the rest of the engines that I would normally remove for this example for two reasons.

Firstly, it is only down to personal preference that I remove them, I know a few people who use no follow links in their pyramid.

Secondly, I want to have the list stats available at the end of this tutorial when my live run is complete to show you the link count and the associated data so you can be better informed on what you will get from the various engines.

Click Here To Display My List Stats At This Stage

Article-AltoCMS-LiveStreet….: 36
Article-Article Directory Pro.: 3
Article-Article Friendly……: 1
Article-Article Script……..: 58
Article-Askbot…………….: 17
Article-AskIt……………..: 1
Article-BuddyPress…………: 1968
Article-ClassiPress………..: 405
Article-Drupal – Blog………: 244
Article-EasyBlog…………..: 230
Article-esoTalk……………: 30
Article-Joomla – Blog………: 5
Article-Joomla K2………….: 5836
Article-Moodle…………….: 1
Article-OSclass……………: 505
Article-OSQA………………: 2
Article-PHPMotion………….: 3
Article-Press Release Script..: 2
Article-Question2Answer…….: 133
Article-UCenter……………: 22
Article-vBulletin – Blog……: 29
Article-WeBid……………..: 17
Article-Wordpress Article…..: 127
Article-XpressEngine……….: 1081
blog comment-blogengine…….: 270
Blog Comment-Blogspot………: 13115
Blog Comment-Bravenet Comment.: 32
Blog Comment-Drupal – Comment.: 96
Blog Comment-General Blogs….: 139713
Blog Comment-HubSpot……….: 1060
Blog Comment-JComments……..: 758
Blog Comment-KeywordLuv…….: 575
Blog Comment-Lokomedia CMS….: 6828
Blog Comment-OrdaSoft Review..: 982
Blog Comment-PHP Fusion Comment: 269
Blog Comment-PrismoTube…….: 32
Blog Comment-Rocketeer CMS….: 1
Blog Comment-ShowNews………: 32019
Blog Comment-SPIP………….: 2824
Blog Comment-Storytlr………: 7
Blog Comment-Textcube………: 267
Guestbook-AchimWinkler Guestbook: 10
Guestbook-Advanced Guestbook..: 698
Guestbook-AkoBook………….: 39
Guestbook-Ard Guestbook…….: 7
Guestbook-aSgbookPHP……….: 23
Guestbook-Basti Guestbook…..: 3
Guestbook-BellaBook………..: 54
Guestbook-Burning Book……..: 20
Guestbook-Coder-World Gaestebuch: 1
Guestbook-DedeEIMS…………: 1405
Guestbook-DMS Guestbook…….: 84
Guestbook-Donation Guestbook..: 776
Guestbook-DRBGuestbook……..: 45
Guestbook-E-Guest Guestbook…: 19
Guestbook-Easy Guestbook……: 46
Guestbook-Easybook Reloaded…: 214
Guestbook-Firebook Guestbook..: 2
Guestbook-flat PHPbook……..: 2
Guestbook-Free Perl Guestbook.: 2
Guestbook-GA Gaestebuch…….: 7
Guestbook-GA Gaestebuch2……: 1
Guestbook-Gaijin.at Guestbook.: 16
Guestbook-Guestbook………..: 315
Guestbook-GuestWho…………: 3
Guestbook-g_book…………..: 23
Guestbook-Icybook………….: 1
Guestbook-Jambook Guestbook…: 1
Guestbook-Jax Guestbook…….: 12
Guestbook-KideShoutbox……..: 583
Guestbook-Lazarus Guestbook…: 2
Guestbook-Maian Guestbook…..: 2
Guestbook-Megabook Guestbook..: 2
Guestbook-MGB Guestbook…….: 20
Guestbook-Michatronic Guestbook: 2
Guestbook-Phoca Guestbook…..: 938
Guestbook-PHP Fusion Guestbook: 2
Guestbook-PJBlog Guestbook….: 34
Guestbook-Purple Yin Guestbook: 9
Guestbook-Ricar Guestbook…..: 27
Guestbook-ScarBook…………: 1
Guestbook-Shoutbox…………: 12337
Guestbook-Silentum Guestbook..: 5
guestbook-sr+denied………..: 71
Guestbook-TPK Guestbook…….: 1
Guestbook-Ultimate Guestbook..: 1
Guestbook-Unknown Guestbook…: 16
Guestbook-Unknown Guestbook 3.: 48
Guestbook-Unknown Guestbook 4.: 4
Guestbook-ViperGB………….: 25
Guestbook-Visitors Book…….: 2
Guestbook-vlBook…………..: 9
Guestbook-Webgarden………..: 34
Guestbook-write2me.nl………: 11
Guestbook-YapGB……………: 2
Image Comment-4image……….: 19
Image Comment-Coppermine Photo Gallery: 146
Image Comment-Coppermine Photo Gallery-login: 38
Image Comment-DatsoGallery….: 309
Image Comment-Gallery2……..: 11307
Image Comment-Piwigo……….: 68
Image Comment-Pixelpost…….: 149
Image Comment-PixelPost-2…..: 8
Image Comment-plogger………: 271
Image Comment-shutter………: 2
Image Comment-YaPig………..: 30
Image Comment-ZenPhoto……..: 281
Social Network-AvArcade…….: 246
Social Network-BoINC……….: 8
Social Network-Chameleon……: 8
Social Network-ClanSphere…..: 93
Social Network-DataLife CMS…: 2
Social Network-Dolphin……..: 2
Social Network-DotNetNuke…..: 59
Social Network-DZOIC Handshakes: 1
Social Network-Elgg………..: 119
Social Network-eMeeting…….: 2
Social Network-GeekLog……..: 2
Social Network-JCow………..: 32
Social Network-JomSocial……: 5
Social Network-MediaGoblin….: 4
Social Network-Oxwall………: 1
Social Network-PeoplePods…..: 20
Social Network-phpDolphin…..: 46
Social Network-PHPFox………: 257
Social Network-PJBlog………: 309
Social Network-Plone……….: 4
Social Network-SocialEngine…: 4
Social Network-WallpaperSiteScript: 19
Wiki-DokuWiki……………..: 44
Wiki-MacOSWiki…………….: 19
Wiki-MediaWiki…………….: 989
Wiki-MoinMoin……………..: 10
Wiki-TikiWiki……………..: 23
Wiki-Trac…………………: 8
Wiki-WikkaWiki…………….: 26
——————————-
Category – Article…………: 10756
Category – blog comment…….: 198848
Category – Guestbook……….: 18017
Category – Image Comment……: 12628
Category – Social Network…..: 1243
Category – Wiki……………: 1119
——————————-
Total…………………….: 242611

Sculpting Your Project Settings To Reach Your Goals

The platforms and engine selection you choose to use in your projects is down to personal preference and can change depending on what you are planning to do with your GSA Search Engine Ranker instance. As I have already mentioned I currently focus on the contextual article platforms of Article, Social Networks and Wikis along with the non-contextual platforms of Blog Comments, Guestbooks and Image Comments.

[Update] I have released this post where I explain how using a catch-all email service for your contextual projects can reduced your monthly email account costs while increasing your daily link output.

To enable me to have a better visual representation of how the premium list I am using for my projects is doing for each link type I usually use one project to filter contextual links and one project to filter non-contextual links. This way, I can quickly see the total link count coming from the different link types.

Ordinarily, I previously broke down the contextual filter projects down into two projects as contextual links from SER can be defined as either an article or a profile. One of these projects would be set up to post articles and one would be set up to only post profiles. This is so I can see the amount of the contextual links the projects are producing for each type. Currently, I only use the contextual article type links so don’t run any profile filtering projects in my live process. If you want to filter your projects for articles or profiles then go into your project options tab and select your link types as shown in the images below.

GSA Search Engine Ranking Contextual Article Settings GSA SER Contextual Profile Settings.

That being said, in the past, I have changed it up to meet different requirements I had at the time. For example, If I am running tests with JoomlaK2 links then I will run a project that will only filter Joomla K2 links and remove that engine from my generic contextual one. This helps to see how many contextual article links I was getting from my process from other engine types as well as how many Joomla K2 links I am getting for testing.

Additionally, I have done similar with the general blogs engine due to its pure size you may wish to have a separate general blogs project so you can easily see your verified link yield from that engine oppose to having it mixed in with your other non-contextual engines.

Another option that I have used in the past is rather than having projects broken down at link type level for contextual and non-contextual targets, I would break them down to platform level such as having separate projects for articles, blog comments, guestbooks, image comments, social network and wikis. This way I can instantly see how I am doing at a platform level.

In the past when trying to work out what engines in GSA SER were worth my time I have taken this even further and broken it down to engine level. This means I actually have a single project for each engine so I could see my verified link yield at engine level rather than platform or type level but in my opinion this is taking it too far.

To try and keep things as simple as possible I am only going to have two filtering projects for this example, one for all contextual links and one for all non-contextual links. I have set these projects up to pull targets from the identified folder where I hold my premium list targets. In my live filtering projects I would also copy and paste the targets currently held in my verified folder into my submitted folder to re-filter them but for this example, I won’t be doing this, if you are using a second premium link list then you could add its targets to your submitted folder if you chose to.

Additionally, I recommend that you set your filter projects up to build links to a fake URL such as http://filterproject1.com for example sake and I will explain the reason behind this later in the tutorial.

The Importance Of Limiting Your Target Retries!

For as long as I can remember users have has the ability to tell GSA Search Engine Ranker to retry a submission to a target where it previously had failed. Although this is an excellent feature, especially for the link filtering process it had one main problem. You were unable to set a limit for how many times SER would retry a site before giving up meaning, in theory, you could have SER running in a constant loop of targets that have been offline permanently producing no verified targets.

To improve this feature I made this suggestion and thankfully it has been implemented into the tool giving us an element of control that we previously did not have. Essentially it lets you tell the project how many times it is allowed to retry to post to a target.

Now you have to keep in mind that this is on a target basis, not a domain basis. This means that on a platform such as blog comments where the same domain can hold multiple targets you can still loop on a domain that has been taken offline to some extent but setting a limit massively helps you. One day I will suggest the option to be able to filter at domain level but Sven has recently publically confirmed he is working on a new tool keeping him even busier than he usually is.

There are a number of reasons a submission can fail and trigger this setting such as an incorrect captcha, the target domain being offline, the target domain timing out or the content management system being updated meaning the script is unable to post to the target anymore. Depending on how important you rate a target engine will depend on how many retries the project targeting that particular link type, platform or engine will have before leaving the target and using your active threads for potential new targets.

For example, I personally rate contextual article targets as the highest possible platforms that GSA Search Engine Ranker has the ability to produce without the SEREngines plugin. This means that I want my filtering projects to do their absolute best to try and verify these targets for me so I may have the project dealing with these requests set to 100 retries in an attempt to try and get as many verified targets from the project as possible.

Out of all of the contextual article platforms, I rate Drupal, WordPress and BuddyPress as the best engines available. In the past, I have set up a separate project just to focus on these engine types and removed them from the generic contextual article filtering project. There is a high chance that I will implement this strategy into my list filtering procedure again soon and because I value the targets for these engines so highly I plan to have their retries set higher than that of the regular contextual article project so perhaps 250.

Although I currently feel that the non-contextual link types such as blog comment, guestbook and image comment are still useful for indexing when used as a third tier I don’t rate them anywhere near as important to contextual articles. These platforms have plenty of targets for SER to try to submit to but they also consist of a majority of no follow links. Due to this, I don’t want SER to be wasting too many resources on this link type so I may only have the retry limit set to 25 or perhaps even less depending on the other stuff I have that GSA SER instance doing at the time.

Some people choose to use a human solved captcha service to try and get verified targets from domains running things such as ReCaptcha, solve and kcaptcha. I personally flip flop in and out of doing this depending what I am doing at the time but I currently do not permanently maintain a high-quality domain list for SER but it has recently became something I am seriously considering doing.

If you are willing to spend money to have a captcha solved then you are going to want to only attempt to post to the best such as either the contextual article link type or as I previously discussed only the Drupal, WordPress and BuddyPress article engines with profile posting turned off. Now previously your only two options were to either try to submit to each target once time meaning if for whatever reason the submission failed then the target was lost forever or, you could enable continue to post to targets that previously failed and risk being stuck in a loop of offline targets wasting all your human solved captcha credits.

Now you are able to set a limit to the number of times you are willing to have SER post to the target before dropping it. Although I have not implemented this into my permanent filtering process yet I am considering a retry of between 5 and 25 to try and keep the higher quality targets in my verified folder while limiting the potential human solved captcha cost for the project.

The Importance Of Turning Reverification Off!

I can’t stress this enough guys! You have to turn the re-verification of links off for your filter projects!

Imagine this, you are running your filter projects, you have the projects set up to initially verify the links automatically and you activate the projects. They do their thing and are verifying links like there’s no tomorrow, the project quickly builds up 1000 verified targets for you but then it kicks into re-verification mode and tries to reverify these links for you. For a number of reasons, 500 of them are no longer found so SER removes them from your total verified targets for the project leaving you with only 500 verified links.

When filtering you are trying to get as many domains that you are able to post to so you rather have the 1000 verified targets than the 500. There are a number of reasons the other 500 links could not be verified, for example, a platform like blog comments has many engines with moving pages where once other people comment your link rolls of the page. The link is still on the domain and the Google spider can crawl it but SER can’t see it so it thinks it has failed.

I know there is an option in the project options tab to enable SER to reverify links on engines with moving pages but this takes up threads and system resources that can be put to better use filtering more targets when there is a solution as simple as just turning re-verification off for the project.

The Potential Of Duplicate Projects

I have no idea why this happens exactly but I predict it has something to do with the captchas presented during submission or just good old RNGesus but I have seen this happen almost every time I run my filtering process. Although I won’t be doing it in this example to keep it as simple as possible you can see it for yourself by duplicating your filtering projects.

On my live filtering projects, I usually build the projects out as I want them and then will duplicate each of them an additional two times giving me a total of three identical projects per project type. Now you activate these projects and let them run.

The next day you come back and one of three things usually happen. Firstly, they all have a similar verified target count as well as unique domain count. Secondly, they have totally different verified target count but their unique domain count is similar.

Secondly, they have totally different verified target count but their unique domain count is similar. This could happen if the projects are running platforms such as blog comments where a single domain can host multiple targets. One of your project duplicates picks up many more verified targets from a single domain giving a much higher verified target count but due to them being from the same domain the unique domain count is similar between all projects.

Thirdly, your verified target count is similar but your unique domain count is total different between the projects. This usually happens on my contextual article filter projects when they have verified URLs for the same amount of time meaning they have roughly the same number of verified targets per project but the actual targets are hosted on different unique domains giving you a wider verified link spread on the domain level.

It is easy enough to view the difference in verified link targets just by comparing what is located in the verified column for the project. When it comes to checking the unique domains for the project it requires a little more work.

Initially, you will select the first of the project types so for example sake say your first contextual filter project and display chart as shown in the image below.

GSA Search Engine Ranker Display Chart

You are then presented with the link breakdown chart and in the top right you can see the unique domains for that project, remember that number. Next, you will select all of the contextual article projects by clicking the first one and then holding the Ctrl key while you click the others. You then repeat the above navigation and it will show you the unique domain count for all three projects and you will be able to see your domain spread on the backend.

GSA SER Show Diagram Window

Remember, filtering projects can take up a lot of active threads and system resources so if you are using a lower end VPS then it might be a good idea not to duplicate them.

[Update] I have covered the new captcha settings in GSA Search Engine Ranker in this post and covered how you are able to use it to increase the quality of the domains on your verified list to help build a higher quality list.

Unleashing The Beast

So, you have set your projects up as you see fit and duplicated them if you chose to. Now all you have to do is set them to active, press start and leave them to do their thing.

The time required for the projects to give you a decent filter depends on a number of things such as the specs of your VPS (my guide on how to choose, set up and optimise a VPS), other projects in your SER instance, how long it has been since you previously filtered your list and how many target retries you have set in your project options. This means there is no one size fits all answer for how long you should leave you filter projects before moving onto the next stage.

For this example, I have only let my example filter projects run for 12 hours to enable me to complete this post quicker and free up some resources for some up coming case studies.

Click Here To Display My List Stats At This Stage

Article-AltoCMS-LiveStreet….: 1
Article-BuddyPress…………: 95
Article-ClassiPress………..: 35
Article-Drupal – Blog………: 17
Article-EasyBlog…………..: 1
Article-esoTalk……………: 4
Article-Joomla – Blog………: 1
Article-Joomla K2………….: 823
Article-OSclass……………: 46
Article-Question2Answer…….: 16
Article-UCenter……………: 1
Article-Wordpress Article…..: 3
Article-XpressEngine……….: 53
Blog Comment-BlogEngine…….: 5
Blog Comment-Blogspot………: 55
Blog Comment-Bravenet Comment.: 1
Blog Comment-Drupal – Comment.: 7
Blog Comment-General Blogs….: 1777
Blog Comment-JComments……..: 28
Blog Comment-KeywordLuv…….: 2
Blog Comment-Lokomedia CMS….: 43
Blog Comment-OrdaSoft Review..: 13
Blog Comment-PHP Fusion Comment: 1
Blog Comment-ShowNews………: 60
Blog Comment-SPIP………….: 27
Blog Comment-Textcube………: 18
Guestbook-Advanced Guestbook..: 2
Guestbook-AkoBook………….: 1
Guestbook-aSgbookPHP……….: 5
Guestbook-Basti Guestbook…..: 1
Guestbook-BellaBook………..: 24
Guestbook-Burning Book……..: 5
Guestbook-Donation Guestbook..: 1
Guestbook-E-Guest Guestbook…: 5
Guestbook-Easy Guestbook……: 21
Guestbook-Easybook Reloaded…: 69
Guestbook-Guestbook………..: 8
Guestbook-g_book…………..: 1
Guestbook-Jax Guestbook…….: 9
Guestbook-KideShoutbox……..: 275
Guestbook-MGB Guestbook…….: 6
Guestbook-Michatronic Guestbook: 2
Guestbook-Phoca Guestbook…..: 149
Guestbook-PJBlog Guestbook….: 2
Guestbook-Purple Yin Guestbook: 1
Guestbook-Shoutbox…………: 110
Guestbook-Silentum Guestbook..: 4
Guestbook-SR+Denied………..: 1
Guestbook-Unknown Guestbook…: 1
Guestbook-Unknown Guestbook 4.: 1
Guestbook-ViperGB………….: 4
Guestbook-vlBook…………..: 2
Guestbook-Webgarden………..: 1
Image Comment-4image……….: 2
Image Comment-Coppermine Photo Gallery: 2
Image Comment-Gallery2……..: 79
Image Comment-Piwigo……….: 6
Image Comment-Pixelpost…….: 1
Image Comment-ZenPhoto……..: 1
Social Network-Chameleon……: 1
Social Network-ClanSphere…..: 4
Social Network-DotNetNuke…..: 2
Social Network-Elgg………..: 14
Social Network-JCow………..: 2
Social Network-PeoplePods…..: 2
Social Network-phpDolphin…..: 7
Social Network-PHPFox………: 18
Wiki-DokuWiki……………..: 2
Wiki-MediaWiki…………….: 148
Wiki-TikiWiki……………..: 7
Wiki-WikkaWiki…………….: 1
——————————-
Category – Article…………: 1096
Category – Blog Comment…….: 2037
Category – Guestbook……….: 711
Category – Image Comment……: 91
Category – Social Network…..: 50
Category – Wiki……………: 158
——————————-
Total…………………….: 4143

The Art Of Purging Unwanted Targets

This next step is by far my favorite thing of filtering your verified GSA SER list in this way as you have direct control over what you are pushing to your verified folder for your live projects to pull from and to my knowledge there is no other way to do this with 100% accuracy.

For this next stage, you can either have all of your filter projects selected or do them by type. Personally, I do them by link type as my keep criteria for contextual article targets is different to that of my non-contextual link targets. That being said, for this example, I will be selecting both the contextual and non-contextual projects for ease of explanation. Once you have selected the project you wish to purge complete the following navigation shown in the screenshot below.

GSA Search Engine Ranker Show Verified URLs

You will now be presented with the following window.

GSA Search Engine Ranker Verified Links Window

As you can see, this screen displays all targets that have been verified by your selected project. More importantly, it presents a number of filtering options for us along the top of the window such as…

  • URL – The verified target URL
  • Anchor Text – The anchor text of the URL
  • D – Is the link do follow or not, do follow links will have an x in the field.
  • Links (int) – The number of links on the page that link to internal pages on the domain.
  • Links (ext) – The number of links on the page that link to an external domain from the page.

All these filtering options are down to personal preference and to organize the window by the specific data option you simply click it once for ascending data and again to for descending.

Starting with URL, I organize the data firstly in ascending order to bring anything that starts with a number to the top of the window as these are usually domains with the raw server IP as their domain name. If you chose to purge these targets from your list you simply click the first entry, scroll down the list to the last entry with an IP as its name, shift click it and select delete selected.

GSA SER delete domains from verified targets with an ip as their name.

Next, we move onto the Anchor Text option, with this option I will arrange it by both ascending data and then descending as sometimes there are random anchors that I want to purge at both ends of the list. If I see any anchors I don’t want to go live on my projects then I do as above and shift click to select them all and delete them.

As you can see in the screenshot below I have chosen anchors using non-English characters. This is purely down to personal preference and as yet I have no data to back up if using these could affect your link pyramid.

GSA Search Engine Ranker removing non-English character anchor texts.

Personally, I currently choose not to filter my lists by the link count on the page but I can see the logic why some users may do this. If you do choose to do this then you would repeat the process above of organizing the links in descending order, selecting and removing them.

Although this next option is not currently available, I feel it would be a massive help if the country of the links hosting server is also displayed in this window. When a link is verified SER displays the flag of the links country in the verified link pane so the data is available.

In my opinion, this feature would be a good option for users to better purge engines such as PHPWeb or DedeEIMS where many of target domains are hosted in Asia wish Asian articles that I can’t read to find out if they are adult based or not and use Asian character anchor text. I suggested this feature here and Sven has confirmed that it is on his to-do list.

I know that there is an option in the project options tab to prevent posting by country so this offers a potential work around but in all honesty, I haven’t had the time to test it yet but it would still be nice to have the data in the verified window.

Click Here To Display My List Stats At This Stage

Article-AltoCMS-LiveStreet….: 1
Article-BuddyPress…………: 95
Article-ClassiPress………..: 35
Article-Drupal – Blog………: 17
Article-EasyBlog…………..: 1
Article-esoTalk……………: 4
Article-Joomla – Blog………: 1
Article-Joomla K2………….: 815
Article-OSclass……………: 46
Article-Question2Answer…….: 16
Article-UCenter……………: 1
Article-Wordpress Article…..: 3
Article-XpressEngine……….: 49
Blog Comment-BlogEngine…….: 5
Blog Comment-Blogspot………: 55
Blog Comment-Bravenet Comment.: 1
Blog Comment-Drupal – Comment.: 7
Blog Comment-General Blogs….: 1790
Blog Comment-JComments……..: 29
Blog Comment-KeywordLuv…….: 2
Blog Comment-Lokomedia CMS….: 43
Blog Comment-OrdaSoft Review..: 13
Blog Comment-PHP Fusion Comment: 1
Blog Comment-ShowNews………: 23
Guestbook-Advanced Guestbook..: 2
Guestbook-AkoBook………….: 1
Guestbook-aSgbookPHP……….: 5
Guestbook-Basti Guestbook…..: 1
Guestbook-BellaBook………..: 24
Guestbook-Burning Book……..: 5
Guestbook-Donation Guestbook..: 1
Guestbook-E-Guest Guestbook…: 5
Guestbook-Easy Guestbook……: 21
Guestbook-Easybook Reloaded…: 69
Guestbook-Guestbook………..: 8
Guestbook-g_book…………..: 1
Guestbook-Jax Guestbook…….: 9
Guestbook-KideShoutbox……..: 273
Guestbook-MGB Guestbook…….: 6
Guestbook-Michatronic Guestbook: 2
Guestbook-Phoca Guestbook…..: 149
Guestbook-PJBlog Guestbook….: 2
Guestbook-Purple Yin Guestbook: 1
Guestbook-Shoutbox…………: 112
Image Comment-4image……….: 2
Image Comment-Coppermine Photo Gallery: 2
Image Comment-Gallery2……..: 80
Image Comment-Piwigo……….: 5
Image Comment-Pixelpost…….: 1
Social Network-Chameleon……: 1
Social Network-ClanSphere…..: 4
Social Network-DotNetNuke…..: 2
Social Network-Elgg………..: 13
Social Network-JCow………..: 2
Social Network-PeoplePods…..: 2
Social Network-phpDolphin…..: 7
Social Network-PHPFox………: 17
Wiki-DokuWiki……………..: 2
Wiki-MediaWiki…………….: 143
Wiki-TikiWiki……………..: 7
Wiki-WikkaWiki…………….: 1
——————————-
Category – Article…………: 1084
Category – Blog Comment…….: 1969
Category – Guestbook……….: 697
Category – Image Comment……: 90
Category – Social Network…..: 48
Category – Wiki……………: 153
——————————-
Total…………………….: 4041

You also have the option to filter by do follow and no follow link status if you like. You can use the selecting method shown above but there is a hardcoded quick method in the tool as shown in the below screenshot to speed up the process of filtering links by this option.

GSA Search Engine Ranker hardcoded do follow or no follow link selection option.

Link Segmentation To Your Verified Folders

It is important that you delete all of the current targets sitting in the folders you send links to at this stage as pushing filtered links into a folder containing pre-filtered links defeats the whole point of the process as the unusable links will remain. I highly recommend you back up the targets sitting in the folder before you delete them so you always have something to restore to even though regular backups should be part of your SER maintenance. You can quickly and easily backup your folders with the inbuilt tool by following the following navigation in the screenshot below.

GSA Search Engine Ranker Export Site Lists Tool

 

Earlier in the tutorial, I explained how I currently use my verified folder to hold my do follow article, social network and wiki targets and my failed folder to hold my do follow blog comment, guestbook and image comment targets. This is only one option to hold these targets and depending what I am doing I do change this a fair bit.

The easiest method I have found to send these specific links to these specific folders like this, select the projects you wish to filter, in this example my contextual article filter project, bring up its verified window. Select all targets via the below navigation.

Right Click the verified target window – Select – All

Side note – If you have chosen to keep your no follow links from the purging step above then the navigation path would be.

Right Click the verified target window – Select – Do-Follow

Now with the relevant links selected you complete the navigation below and will be presented with the screenshot show.

Right Click the verified target window – Add to site lists

GSA SER send selected verified urls to specific site lists.

As you can see you are now presented with the ability to send your selected links to any folder you wish. I now send these targets to the verified folder and repeat the process with my non-contextual filter project selected but send its do follow targets to the failed folder.

Before getting to the final link counts after the process I want to state that this is not a fair representation of the link counts you may get as it was used for an example with the projects only running for 12 hours.

Click Here To Display My Do Follow Contextual List Stats With Joomla K2

Article-BuddyPress…………: 37
Article-ClassiPress………..: 31
Article-Drupal – Blog………: 11
Article-EasyBlog…………..: 1
Article-Joomla – Blog………: 1
Article-Joomla K2………….: 813
Article-OSclass……………: 29
Article-Question2Answer…….: 1
Article-Wordpress Article…..: 3
Article-XpressEngine……….: 49
Social Network-Chameleon……: 1
Social Network-ClanSphere…..: 4
Social Network-DotNetNuke…..: 2
Social Network-Elgg………..: 1
Social Network-PeoplePods…..: 2
Social Network-phpDolphin…..: 7
Social Network-PHPFox………: 2
Wiki-MediaWiki…………….: 2
Wiki-WikkaWiki…………….: 1
——————————-
Category – Article…………: 976
Category – Social Network…..: 19
Category – Wiki……………: 3
——————————-
Total…………………….: 998

Click Here To Display My Do Follow Contextual List Stats Without Joomla K2

Article-BuddyPress…………: 37
Article-ClassiPress………..: 31
Article-Drupal – Blog………: 11
Article-EasyBlog…………..: 1
Article-Joomla – Blog………: 1
Article-OSclass……………: 29
Article-Question2Answer…….: 1
Article-Wordpress Article…..: 3
Article-XpressEngine……….: 49
Social Network-Chameleon……: 1
Social Network-ClanSphere…..: 4
Social Network-DotNetNuke…..: 2
Social Network-Elgg………..: 1
Social Network-PeoplePods…..: 2
Social Network-phpDolphin…..: 7
Social Network-PHPFox………: 2
Wiki-MediaWiki…………….: 2
Wiki-WikkaWiki…………….: 1
——————————-
Category – Article…………: 163
Category – Social Network…..: 19
Category – Wiki……………: 3
——————————-
Total…………………….: 185

Click Here To Display My Do Follow Blow, Guestbook And Image List Stats

Blog Comment-BlogEngine…….: 5
Blog Comment-Blogspot………: 4
Blog Comment-Drupal – Comment.: 6
Blog Comment-General Blogs….: 75
Blog Comment-OrdaSoft Review..: 13
Guestbook-aSgbookPHP……….: 5
Guestbook-BellaBook………..: 24
Guestbook-Burning Book……..: 4
Guestbook-Easy Guestbook……: 21
Guestbook-Guestbook………..: 7
Guestbook-Phoca Guestbook…..: 133
Image Comment-Coppermine Photo Gallery: 1
Image Comment-Gallery2……..: 18
——————————-
Category – Blog Comment…….: 103
Category – Guestbook……….: 194
Category – Image Comment……: 19
——————————-
Total…………………….: 316

Some people may choose to keep their no follow targets to use on their live projects. In this case, the user may choose to send all of their do follow targets to their verified folder and all of their no follow targets to their failed folder. I have cleared my folders from the list stats above and re-sent the verified targets from my filter projects as explained in this paragraph and added a do follow v no follow comparison below.

The do follow link count is displayed on the left with the no follow link count on the right followed by a percentage, in all honesty, I have never paid much attention to the displayed percentage as its meaning seems to change per target engine.

Click Here To Display The Do Follow V No Follow Stats

Article-BuddyPress…………: 0000000037/0000000058 => 63.79%
Article-ClassiPress………..: 0000000031/0000000004 => 12.90%
Article-Drupal – Blog………: 0000000011/0000000006 => 54.55%
Article-EasyBlog…………..: 0000000001/———-
Article-Joomla – Blog………: 0000000001/———-
Article-Joomla K2………….: 0000000813/0000000002 => 0.25%
Article-OSclass……………: 0000000029/0000000017 => 58.62%
Article-Question2Answer…….: 0000000001/0000000015 => 6.67%
Article-Wordpress Article…..: 0000000003/———-
Article-XpressEngine……….: 0000000049/———-
Blog Comment-BlogEngine…….: 0000000005/———-
Blog Comment-Blogspot………: 0000000004/0000000051 => 7.84%
Blog Comment-Drupal – Comment.: 0000000006/0000000001 => 16.67%
Blog Comment-General Blogs….: 0000000075/0000001715 => 4.37%
Blog Comment-OrdaSoft Review..: 0000000013/———-
Guestbook-aSgbookPHP……….: 0000000005/———-
Guestbook-BellaBook………..: 0000000024/———-
Guestbook-Burning Book……..: 0000000004/0000000001 => 25.00%
Guestbook-Easy Guestbook……: 0000000021/———-
Guestbook-Guestbook………..: 0000000007/0000000001 => 14.29%
Guestbook-Phoca Guestbook…..: 0000000094/———-
Image Comment-Coppermine Photo Gallery: 0000000001/0000000001 =>
100.00%
Image Comment-Gallery2……..: 0000000018/0000000062 => 29.03%
Social Network-Chameleon……: 0000000001/———-
Social Network-ClanSphere…..: 0000000004/———-
Social Network-DotNetNuke…..: 0000000002/———-
Social Network-Elgg………..: 0000000001/0000000012 => 8.33%
Social Network-PeoplePods…..: 0000000002/———-
Social Network-phpDolphin…..: 0000000007/———-
Social Network-PHPFox………: 0000000002/0000000015 => 13.33%
Wiki-MediaWiki…………….: 0000000002/0000000141 => 1.42%
Wiki-WikkaWiki…………….: 0000000001/———-
Total…………………….: 0000000128/0000000196 => 65.31%

If you remember earlier in the tutorial, I said how the general blog engine is massive and mostly do follow but I choose to process it because its total link volume makes up for this still making producing more do follow links than any other blog platform and you can see this has turned out to remain true.

There is actually one more step I carry out at this stage for my contextual targets but I plan to do a separate blog post on it in the future so I won’t go into much detail on it here. The short version is I import and merge all of my verified contextuals into Scrapebox, trim them to root and index check them. I then export the domains that are not indexed to a text file and use the remove URLs from file option in the tools window to remove the domains from my saved files.

Now that you have segmented your links once, this set of filter projects are essentially useless. This is because GSA Search Engine Ranker automatically logs the targets a project has already processed for its particular projects for a particular URL. Remember earlier in the tutorial where I told you to set the project up to build links to http://filterproject1.com?  To get around this hard coded feature the next time you run your filtering projects you will change the URL to http://filterproject2.com and then each time thereafter increase the number by 1.

As the point of filtering is to reprocess everything from scratch this is a pain but I suggested this feature for SER and it has been implemented. Essentially it lets you run multiple reset options on a selected SER project at once rather than doing them all one at a time via the following navigation path once the project or projects are selected.

GSA Search Engine Ranker Reset Project Data Tool

Here is a screenshot of the options I have selected in the reset data window.

GSA Search Engine Ranker Reset Project Data Options.

Running this effectively resets the selected projects while leaving the email account details (if you are not using a catch-all email service then you would tick this and put new emails into the projects) and the project’s articles. With a filter project, the engines require article data but you don’t have to care about duplicate content as you won’t be bothered about indexing these links.

These removes the target URL history among other things so the next time you run your filter projects they are essentially brand new and with the new URL such as http://filterproject1.com they will be as good as new the next time you need them.

The Ideal Set Up

This next part totally depends on how you are using GSA Search Engine Ranker as well the scale you are using the tool. Due to a number of reasons I have scaled down my GSA SER usage recently but in the past I have ran the setup below and have been able to push out 1.5 million do follow links per day per VPS.

 

GSA Search Engine Ranker premium verified list filtering set up

As you can see, the list filtering process explained in this tutorial ran on its own VPS 24/7. I would then manually backup the verified target folders on the list filtering VPS, clear them of files and then send the new target list to the various folders and use dropbox to sync the new list to all of my live instances running on various VPS’s and dedicated servers.

Another advantage of using this method is that I was able to map my folders as required on the list filtering VPS  hot swapping them for different tasks while leaving the folders of my live instances of GSA Search Engine Ranker fixed to set dropbox files.

For example, when building the new list I would have the mapped folder patchs of the list filtering VPS as follows.

  • Identified – Locally saved premium list service 1
  • Submitted – Locally saved premium list service 2
  • Verified – Current verified list being refiltered
  • Failed – Locally saved manually scraped or link extracted targets to be processed.

The projects on this VPS would then pull their targets from these folders and process them throughout the day for me. Once I wanted to push a new list I would set all of its projects to inactive, wait for the thread count to decrease to 0 in the bottom left of the SER instance and press stop. This minimizes the risk of SER losing some of the link data when it is reset for the remapped folders as I had this happen a few times before I worked out it was linked to the tool still having active threads when telling it to close.

With all of the projects and threads stopped I would then remap this instances folders as so.

  • Identified – Dropbox do follow contextual article targets.
  • Submitted – Dropbox do follow forums.
  • Verified – Dropbox do follow blog, guestbook, image and trackbacks.
  • Failed – Dropbox human solved captcha contextual targets.

I would then restart SER so it knows its folders have been remapped, go through the various projects and push the required links to the required dropbox folders that my live VPS’ and dedicated servers were using to build links on live projects.

I would then merge all of these folders into a local backup on my filtering VPS and remap the folders to the original set and delete everything from them. I would then re-add the target files from the premium list services, restore the merged verified folder backup I made to the verified folder to re-filter it. The set up I had at the time would automatically keep the folder mapped to failed updated via GSA Platform Identifier so I had no need to take any action on it at this stage.

Then I would reset the local GSA Search Engine Ranker instance again to make sure it was aware its folders had been remapped. Now

Now just to clarify as people seem that link count and don’t understand what I was doing with it but these links WERE NOT pointed directly to my money site! I was using them as highly efficient indexing method that would get around 80% of my links indexed at the time. Here is an extremely simplified pyramid diagram to visually show the representative link counts on each tier and why I needed so many links per day but as I said this indexing system seems to have been patched out.

Simplified black hat search engine optimisation tiered link building diagram for GSA Search Engine Ranker.

All that being said, most users will be able to run the filtering projects on the same VPS as their live projects and still see a massive jump in links per minute with a properly filtered list and that brings me to the end of this tutorial. I hope it will help as many readers as possible improve both their link per minute and verified per minute link counts and help you in your black hat search engine optimization journey.

Check Also

How To Build Your Own Auto Accept List With GSA Search Engine Ranker!

As A reader of my blog, you are able to save 15% on the retail …

Join The Mailing List!