Hinkys' How To Easily Build HUGE Sites Lists for GSA SER
2013-10-19, 04:29 PM, (This post was last modified: 2013-10-19, 04:30 PM by boboMarley.)
#1
Hey guys, this is something I started doing a while back and wanted to share it with you guys cause it's THAT good.

Basically, with this trick you will increase your site lists EXPONENTIALLY.

[HIDE]I'll show you to use a seed list of as little as 100 AA blogs to create a list of 100,000 sites (after deduping both URLs and domains) that you can import to GSA SER and get a generous percentage of them on successful and verified. This process shouldn't take more than a few hours.

This list should then be imported in to a dedicated project in GSA (which I will explain how to setup later in the tutorial).

Note that you will need Scrapebox for this.

Last time I did this (which was a few days ago) I started with 2000k AA blogs taken from one of my projects and ended up with 2mil+ DIFFERENT DOMAINS to import in to SER. This resulted in 30k verified links. And that's with using out-of-the-box Captcha Breaker (See why I prefer CB to SC) and highly spun content. If I was to use more readable content and a more optimized version of CB, the submitted / verified numbers would be MUCH higher.
Idea behind this is the following:
If you managed to post a comment to a particular blog using GSA SER, there's a good chance that a lot of other GSA SER users have managed to post there as well. This means that there are a lot of other people building their Tier 2 / 3 in those same blogs.

By scraping all internal links on those blogs, we get a huge list of blog posts that have been violated by SER

Then by extracting all external links from the huge list of blog posts, we get a massive list of potential targets for SER

Detailed Version

Step 1 - Getting a seed list

First you need a list to start with. When you're doing this for the first time, you can go to your verified folder and copy all URLs from blog comments. Don't worry if you get very low number of URLs this way (I got around 500), you will see it's still as effective.

For every other time you're doing this, you can just use a list of verified blog comments you created from the last time you did this whole process. Other than that, you can simply harvest a list of blogs and check which ones are auto-approve by posting there with SB / SER.

When you found a list that you can use, import it to SB, trim to root, dedup URLs, save the list as "Step 1 - seed list" and proceed to step 2.

Step 2 - Extracting internal links

Fire up the "Link Extractor" plugin in Scrapebox and load the "Step 1 - seed list" file. Set the mode to "internal", use as much connections as your box can handle and start.

After it's done, import that list to scrapebox and close the link extractor. Scroll through the list and see if there's many comment (usually ending with /#comment), categories, tag links, etc. Use the "Remove URLs containing" to try and get rid of as many of those as possible. Ideally you'd want a list consisting of nothing but blog posts.

This filtering isn't necessary but depending on the size of your initial list, the next step could take considerably longer if you skip this.

After you're finished, save the list and name it "Step 2 - all internal links".


OPTIONALLY
Setup a dedicated project in SER and feed it those lists to filter out the junk from the actual blog posts
As a bonus, this is also a good way to build your AA list.



Step 3 - Extracting external links


Before starting, you should split "Step 2 - all internal links" file in to smaller ones, no more than 10k per file and then process them one at a time. The reason behind this is that Step 3 usually produces up to 200x more URLs than the number of links you start this step with. (For example, I usually use batches of 5k links which result in list of 400-900k de-duped URLs)

If you use large lists in this step, you will end up with couple of million of URLs and as far as my experience goes, Scrapebox doesn't handle more than 1mil URLs all that well.



You can use the "Dup Remover" Scrapebox plugin for splitting the files.

So open the link extractor again and load the first batch of the "Step 2 - all internal links" file. Set the mode to "external" and hit start. Go make yourself a coffee and once the link extract has finished, transfer the list to Scrapebox, dedup if needed and save the file as "Step 3 - Needs sorting".

Repeat the process until you have gone through all the smaller batches of the "Step 2 - all internal links" file.

You can run multiple instances of the link extractor simultaneously as long as your box can handle that if you want to speed this up (tho I wouldn't recommend more than 2-3 per SB instance since it will most likely crash).



Step 4 - Sorting the list

Now you should have quite large lists of sites that need to be sorted somehow. Luckily, GSA SER will do this automatically without too much hassle.

Setup a new project with the following settings:

[Image: brcf.jpg]

[Image: 1o5t.jpg]

Note that this setting will filter out all unindexed sites. If you don't care about PR and just want as many links as possible, just untick all PR filters, it should skip PR checking for sites, probably making your project faster.

Also make sure to untick all search engines and site lists, you want this project to ONLY post to target URL lists you import.


Double check that you're only using Captcha Sniper / Captcha Breaker for this project as anything else will very quickly deplete your balance.

Now just generate some relevant spun content (the type you would be using in all your other projects) and your new "Sitecheck" project is good to go.


After you got it setup, import the "Step 3 - Needs sorting" as target URLs for that project. (I don't know the limitations of GSA SER but I split everything above 1mil URLs in to smaller files and then let GSA go through the files one at a time.)

Now this project will now go through the list and fill up your identified / successful / verified lists!



Double check that you're saving Identified, Submitted and Verified sites in SER.


Now wait till it's finished, grab a list of verified blog comments / image comments you just created and repeat the process!

Obviously GSA SER won't be able to post to all of these sites but you will be surprised at how much it will!

Short Version

1. Get a list of verified blog comments from GSA
2. Extract all internal links
3. Extract all external links (from all internal)
4. Import the list to a project in GSA and let it sort it for you
5. Repeat from 1. with the verified links you just created

[/HIDE]

Thank is not my GUID , its from Platform provides competing CPA content sections ad such. Do not post this on CPA Elites please. , This guid help me a lot , and I think this will help you a lot.
Reply
Share This Thread :

2013-10-20, 01:20 PM,
#2
thank you man!! keep post usefull stuff like this if you found!
bookmarked!!
Reply
2013-10-20, 03:07 PM,
#3
(2013-10-20, 01:20 PM)alexparris Wrote: thank you man!! keep post usefull stuff like this if you found!
bookmarked!!


thanks i glad you like it :D , i will keep making good posts
Reply
2013-10-24, 09:20 PM,
#4
Time to test this out, just got Scrapebox :D
Reply
2014-03-19, 07:02 PM,
#5
Copied from Forum discussions are not allowed here.,this will hurt cpaelites ranking on google.Stop doing this and make a unique one.
Reply
2014-03-19, 07:16 PM,
#6
(2014-03-19, 07:02 PM)Khantahir Wrote: Copied from Forum discussions are not allowed here.,this will hurt cpaelites ranking on google.Stop doing this and make a unique one.

I just want help other member who cant see vip section on Platform provides competing CPA content sections ad such. Do not post this on CPA Elites please.

its not from Forum discussions are not allowed here.
Reply
2014-03-19, 07:35 PM,
#7
(2014-03-19, 07:16 PM)bobMarley519 Wrote:
(2014-03-19, 07:02 PM)Khantahir Wrote: Copied from Forum discussions are not allowed here.,this will hurt cpaelites ranking on google.Stop doing this and make a unique one.

I just want help other member who cant see vip section on Platform provides competing CPA content sections ad such. Do not post this on CPA Elites please.

its not from Forum discussions are not allowed here.

For your information you cant copy from other forums.
It will make this forum shitty and why whould people be here if they can get it from the forum u r copying !
Reply
2014-03-20, 03:04 AM,
#8
Because not everyone spends time browsing other forums. While I don't agree with copying it word for word, it's good to share information. That's how we learn and grow successful together.
Reply
2014-03-21, 02:22 AM,
#9
thanks i glad you like it :D , i will keep making good posts
Reply
2014-03-22, 11:16 AM,
#10
Thanks for the share.. That's a great information
Reply


Related Threads
Thread Author Replies Views Last Post
  Guest Blogging Sites List insomniac 60 4,862 Yesterday, 01:50 PM
Last Post: andrew43
  Quick GSA question direcT 3 198 2018-06-11, 12:28 AM
Last Post: CharlieHarper
  Which is the best and cheap GSA SER bot? dailyviral 1 247 2018-05-05, 05:15 PM
Last Post: DanaCr
  Guest Blogging Sites List insomniac 37 4,308 2018-05-03, 02:00 PM
Last Post: Rachitkhatri
  Time To Build Backlink? 5light 8 1,099 2018-03-23, 01:39 PM
Last Post: affmarketer101





About Us | Contact Us | CPA Elites | Advertise | Stats

© 2013-2018 CPA Elites Ltd
Enhanced by MyBB and WallBB
Return to top