Frequently Asked Questions

Most of the questions and answers have been copied from the forum thread. I will however update them over the next days and move everything here, since it's easier to maintain.

You don’t NEED them, but it makes things a lot faster and some people want to stay anonymous. Also, it is highly recommended to use proxies for everyone who has a VPS. Your account could get banned after your provider received too many complaints about the IP. However if you get complains is up to the way you use the tool. It's after all still a tool that can be used for good or bad. We are not responsible for what you do with it but strongly suggest you not to spam with it of course!

No! But note, usually you want to run the software 24/7 with a fast internet connection so you might consider getting one, especially if you have a slow internet connection.

A lot of people are using a software like Captcha Breaker and an online service like 2captcha.com. If you use Captcha Breaker and don't know what settings you should use please check our forum about that.

Make sure the program can access the pop3 server for checking emails. Also check that you don't delete emails in your email client once it logs into the server. All emails should be in your inbox and none should be saved in “Spam/Junk”. You should check from time to time, if you’re email has been blacklisted and if that’s the case exchange it with a new one.

Uncheck “Also analyse and post to competitors backlinks” and “Use URLs from global sites lists if enabled”.

The software can handle all UTF-8 characters (which includes nearly all known characters). Before submission the content is encoded to a format which works for the site. Sometimes you will still see display errors, but that is a problem of the site that can’t handle encoding correctly

How can I post to .gov, .edu, .de, .es etc. only?

Go to “Skip sites where the following words appear in the URL/domain” and add an exclamation mark followed by the domain ending. So for posting only to .edu sites you’d add !.edu Please note to enter all “white listed” domains in one line like “!.edu !.gov !.gsa” and not one in each line.

You can put a badwords list together yourself using sites like Swearword Dictionary or you use lists posted by users, e.g. s4nt0s.

  1. Green = submitting
  2. Blue= verification
  3. Orange = searching for new targets
  4. Red = something is wrong (not used for now)

Right-click on a project, select Show URLs → Submitted or Verified → Stats → Diagram.
Or with Right-click on a project → Show Diagram/Chart

There are 2 Options that are causing popups. Follow these steps:

Project Options → If a form field can't be filled (like category) → Choose Random
Options → Captcha → Uncheck “Finally ask user if everything else fails (else the submission is skipped)”

Insert <a href="%url%">%anchor_text%</a> at the place you want to have your URL + Anchortext or disable the option to place the link at the bottom (below article input field).

Why doesn’t GSA post articles on wiki sites?

According to Sven wiki article links are gone way sooner than links in the about section. He said there might be an option for changing that later.

Leave the respective fields blank and check “Uncheck Engines with empty fields”, or right-click on the field and choose “Disable Engines that use this”.

Edit/Double Click Project → Tools → Export → Account Data

Delete Cache - Click this only if you have e.g. imported a wrong target URL list or think you want to clear the URLs that a previous search task found.

Delete History - Click this if you..

  • a) want to submit to sites that you previously skipped (e.g. you didn't want to enter captchas back than but want to do now)
  • b) think it found sites in the past that it skipped but are now supported by e.g. new added engines
  • c) think it found sites in the past and somehow your internet connection blocked it to get sent to (proxy issue e.g.)

There are a lot of different opinions on that topic. If you set up the program correctly with strict filters, good content (best manual spun) and only for certain types of platforms, there shouldn’t be problems. Keep in mind that it always depends on the person who uses such a powerful tool like GSA.

Project Options → Pause the project after XX submissions or verifications for XX minutes/days.
So you could setup to post 10 links a day with:
Pause project after 10 verifications for 1440 minutes/1 day.

It can take up to 5 days until a submission is checked. Be patient.

If a submitted URL timed out (to many unsuccessful checks) it will be removed, so the number of submitted links can decrease and is sometimes lower than the number of verified links.

GSA suddenly finds a lot less new links to post to than the days before. Why is that?

Wait a few days to see if this is just a temporary problem. After that it most likely ran out of keywords. Try adding more keywords and choose some new search engines to scrape from.

Ozz:

  How should we know? When you are saying that you are 10 mph fast and asking if thats normal how should anyone know if you are driving a car, bicycle or a skateboard for
  instance.

There are so many factors that play a role. It’s impossible to give an answer to that question with so little information. We need to know the

Number of

  • Projects
  • URLs you are processing
  • Search engines you selected
  • Keywords you entered

Also, what

  • types of sites you selected
  • kind of filters you use (OBL, PR,…)
  • is the Processor power of PC you are running on
  • is the RAM of PC you are running on
  • speed your internet connection has

And how long you already have been running the project.

Check if GSA is blocked by a Firewall or Antivirus software. Another reason can be bad proxies, often public ones. Try without or choose other/private proxies to check if it works better then.

  • %NAME% - will be replaced with the a random users name
  • %EMAIL% - will be replaced with the a random email (auto generated)
  • %WEBSITE% - will be replaced with the users website url
  • %BLOGURL% - will be replaced with the target url you are about to submit to
  • %BLOGTITLE% - will be replaced with the page title of the target site
  • %keyword% - uses random keyword from project settings
  • %spinfile-% - takes a random line in that file
  • %source_url% - referrer page
  • %targeturl% - same as %blogurl%
  • %targethost% - just the http://www.domain.com without a path
  • %random_email% - same as %email%
  • %random-[min]-[max]% - a random number between min and max
  • %spinfolder-% - takes content of a random file in that folder
  • %file-<[ile]% - takes content of that file
  • %url_domain% - just the domain name of your url
  • %url% - your url
  • %?????% anything from the project input boxes e.g. %anchor_text% or %website_title% or %description_250% …
  • %random_url% - inserts another URL from your project
  • %random_anchor_text% - inserts another random anchor text.
  • %random_sanchor_text% - inserts another random secondary anchor text
  • %url1%, %url2% … - inserts the exact URL from your project.

special variables for blogs:

  • %meta_keyword% - takes a keyword from target meta tags
  • %blog_title% - same as %blogtitle%

special variables for image commenting:

  • %image_title% - the name of the image you comment on (sometimes it is sadly the filename only as some webmasters don't set a title)

A more complete list can be found in the macro guide.

Right-click on a running Project → Modify Project → Use New Engines

Select the Projects you want to backup → Right-click → Modify → Backup

No. Please support the developers and buy another version. It is already very cheap, some would say too cheap, so don’t try to do some shady tricks.

This question can’t be answered, because both tools are different in most parts. Scrapebox is basically a Link Harvester + Blog Commenter and GSA Search Engine Ranker is an Automatic Backlinking tool with lots of different backlink sources. A good idea is to combine the strengths of both tools.

You can use Scrapebox for harvesting URLs with GSA footprints. I extracted the footprints with the tool posted by s4nt0s. Put in your keyword list into Scrapebox → click on the “M” → load the list with GSA footprints. Now you can harvest the URLs with Scrapebox. After that you can use the Scrapebox filter functions, like “Remove duplicate URLs” and filter for PR. Now you have your own list, that you can use with GSA. See next Q&A.

Addition: You can scrape for a specific platform by:

  • Using the platform footprint in GSA SER:
  • Options > Advanced > Tools > Search Online for URLs > [Select Platform]
  • Add the footprints into Scrapbox, and scrape. Then just import those target URLs into GSA SER, and it will use these to post to.

Uncheck all search engines, uncheck “Also analyse and post to competitors backlinks”, uncheck “Use URLs from global sites lists if enabled”. Right-click on the project and click “Import Target URLs”.

“If you want to get some good feedback and assistance in this forum you need to have a detailed post with your issues. First you need to search this forum and ensure your question is not answered. Go to Google and type in: site:forum.gsa-online.de keyword If that doesn't help: Take screenshots, list all your settings. Experienced users will take a quick look at your screenshots and give you some basic ideas where you're going wrong. Copy this sample post: https://forum.gsa-online.de/discussion/1440/identify-low-performance-factor#Item_5

How to post Full Article on WIKI?

edit the project→click Options→go to Filter URLs→Type of Backlinks to create→enable Article-Wiki. However this is not recommended and considered as spam. Don't use it except you know what you are doing.

  • a) You use generic anchors in project data → uncheck it
  • b) You use domain anchors in project data → uncheck it
  • c) Some engines simply use there own anchors like “Homepage of Author” → right click on engines→uncheck engines not using your anchor texts
  • d) Some engines use Names instead of Anchor Texts or Generic Anchors like Blog Comments (see a or disable Blog Comments/Image Comments)

The Trackback Format2 engine is producing an anchor text like ”Anchor Text - www.yoursite.com“. Just disable this one if you do not want it.

  1. Right click on your project → Show URLs → Verified URLs
  2. Right click on the list → Select → By Date
  3. Click on the EXPORT button or simply right click again and choose Copy URL

This dialog is reachable via right click on a project → Show URLs → Verified URLs. Here you can reverify links or check the index status which then result in different colors. *

  • green background → successfully verified
  • red background → verification failed
  • yellow background → verification failed but with a timeout, means the server might be down temorarry
  • green text color → index check successfully finished and URL is indexed
  • red text color → index check successfully finished and URL is NOT indexed
  • yellow text color → index check failed due to a possible ban on the search engine with that IP/Proxy
  • Simple words like string and the selection has to have this word included, but will not match e.g. stringlist
  • If you want to match also parts of words, use *string*. The sign “*” stands for any or none chars.
  • If you want to match against multiple masks use “|” as a separator such as string1|string2|…. The “|” sign works as an “OR” operator.

It means that an account was created on that site with an email address. That however needs to be verified by a link that was sent to you via email. The project is now checking your email in intervals and hopes to see such email message to continue with the verification followed with login and link placement.

One of the following might have happened:

  • The website is simply no longer working and the email was not sent
  • Your email account does no longer work (e.g. you need to manually verify your email account → check the accounts in project options)
  • The message sent was tagged as spam and didn't reach your inbox. Manually login to that email account via web and check)
  • The message simply delayed and comes a lot later (can happen for manually verification by site admins but is not very common)
  • The previous registration attempt on that site actually failed but the program was unable to see that error and thought it was successful

If there have been a lot of tries to verify that account it will simply ignore that site and remove it from the submitted URLs list. It will decrease without further tries to submit.

All verified links of a project get checked for:

  • the webpage is is still valid (HTTP 200 as reply code)
  • the placed backlink to the URL of that project is still valid and present
  • for a tier project it is also checking if the placed backlink URL is actually still part of the main project's verified URLs

What should I do on a message such as "No targets to post to (...)"?

First, this is not an error, just a warning that at that current state of the project running, it is not getting enough URLs to place links on (Target URLs). The reason of this can be many and in brackets of that message are some hints that might help to find the issue. Most likely it's about proxies not being configured correctly or email accounts not working or to strict filters.

If you still can't figure it out, please use pastebin.com to upload some log lines from that project and discuss this on our forum.

That depends on your setup. If you are using “Wrodpress” and “Drupal” engines, then it is indeed kind of normal. There are a lot sites based on these two engines and registration is usually easy and allowed. So each registration results in one submitted entry in your project. Later if it is able to verify the account, it tries to login and even that usually works very well. But the submission of articles is not something that many sites will allow. In the end just a few verified URLs will get build from all the submitted. This is something that can not be easily seen when a site is identified and the account registration starts. So either you have a list of sites from somewhere else where this is pre-checked or have to live with this submitted vs verified amount of sites.