• Mirroring Adult Sites – Stage Three

    Date: 2011.02.23 | Category: Promotion, Tutorials | Response: 0

    So we hit stage three of our tutorial in this stage we are going to make some pages for our top 5 AVS systems. However, as we did in the last tutorial with the new TGP galleries we created we are going to use the same content and same pages that we have already built. On with the tutorial…

    Anyone who build AVS sites will know that you need to have a standard site layout. This usually consists of an entry page (or a warning page), a navigational page (or in our case our menu.html page) and some gallery pages.

    So the first thing we need to do is open up the warning page. Once this page has been opened we now need to do some editing of the page, what I would suggest is moving the warning text to the lower half of the page and at the top making some ‘juicy’ sales text, depending on the niche you are targeting with this site you may also like to add a couple of the original thumbnail images at the VERY top of your page however, these should NOT be linked to the full size image.

    Once you have edited your page to a reasonable standard for the AVS you now need to save this page in the /FreeSite/AVS/ folder that you originally created on your HD. I would suggest saving the page as something like avs1.html or the name of the AVS you will be using this page for.

    Once you have done this you now need to duplicate this page but save it as a different name in again, in the /FreeSite/AVS/ folder so you will end up with avs1.html, avs2.html, etc.

    At the moment these AVS pages have no AVS script on them so, we need to visit the AVS system we are going to be using and fill out the details to have the AVS generate an AVS signup page script for us. Once we have the coding that the AVS provides we now have to place this on our avs1.html page. You should repeat this for each of the AVS’s you wish to use and save each instance of the generated AVS script to a separate AVS*.html page.

    be aware, that when you are putting your information into the AVS to generate the script, you should enter the ‘members area url’ as the page we created for our navigation so, the AVS script location will be /FreeSite/AVS/avs1.html and, the members area page will be, /FreeSite/Menu/menu.html.

    Once you have placed the script on our avs.html pages you should now have the following sites ready:

    1) 50 Pic Free Site With Pics On HTML Pages.
    2) 5 TGP Galleries With Pics On HTML Pages.
    3) 5 AVS Entry Pages.

    Now we are starting to see how using one set of content can amount to a mass of sites all of which can be sending traffic to our sponsors and, to our surfer trap.

    This is where stage three of our tutorial ends however, in our next tutorial we will continue to make some more sites and in an effort to get some traffic we start to play with the ones we have currently created.

    Article written by Lee

  • Everything You Need To Know About Meta Tags And More!

    Date: 2011.02.22 | Category: Search Engine Optimization, WebDesign | Response: 0

    Everything You Need To Know About Meta Tags And More!

    There’s a plethora of different META Tags that you can make use of on your site. Because there are so many, it’s impossible for me to cover all of them in this article. However, I will describe the most common ones. But let’s start from the beginning, shall we?

    What Are META Tags?

    META tags are similar to standard HTML tags. However, there is one big difference: You have to insert all META tags between the <head> </head> tags on your page(s). META tags are used primarily to include information about a document. The META tags will be invisible to your site’s visitors, but will be seen by browsers and search engines.

    For The Search Engines.

    Several of the major search engines make use of the META Keyword tag, and virtually all of them make use of the Description tag. These tags help the search engine spider determine the content of your web site so that it can be indexed properly.

    This is what they look like:

    <META name=”keywords” content=”Webmaster Resource Site”>
    <META name=”description” content=”Online Webmaster Resource Site”>

    You can also instruct the search engine robot/spider how to index your site using the robots META Tag. This is what it looks like:

    <META name=”robots” content=”noindex,nofollow”>

    What the tag above does is tell the spider not to index the page that it appears on, and not to follow the links on that page. Here’s a complete list of attributes for the robots tag:

    index – the default, the page is added to the search engine database

    noindex – the page is not added to the search engine database

    nofollow – the spider doesn’t follow the links on that page

    none – same as “noindex, no follow”

    To use any of these attributes simply replace the “noindex,nofollow” text in the example above with whatever you want to use. If you need to use more than one attribute, separate them with a comma.

    Client Pull.

    Have you ever seen a page that automatically refreshes to another URL after a few seconds? Did you ever wondered how it was done? I’ll tell you. The page made use of one of the http-equiv META Tags to automatically “pull” you to a different page after a few seconds. Here’s what the code looks like:

    <META http-equiv=”refresh” content=”1; url=newpage.htm”>

    The value of content denotes how many seconds will pass before the new page is called for. If you want it to happen as soon as a person hits that page, then set the value to “0”.

    Prevent Caching.

    If you want to prevent a page being saved in your visitor’s cache you can do so by inserting the following three tags:

    <META HTTP-EQUIV=”expires” CONTENT=”0″>
    <META HTTP-EQUIV=”Pragma” CONTENT=”no-cache”>
    <META HTTP-EQUIV=”Cache-Control” CONTENT=”no-cache”>

    Why would you want a page not to be cached? If your site is updated frequently and you always want your visitors to see the newest content/changes, or if you want to ensure that a new banner is loaded each time from your server when a page is accessed, you’ll want to use the tags above.

    A quick note here on banner caching: In addition to using the tags above, you’ll want to append random numbers at the end of the tag calling the image.

    Rating Your Content.

    By using the rating META tag on your site, you can specify the appropriateness of your web site for kids. The tag looks like this:

    <META name=”rating” content=”general”>
    In addition to the general rating, you can use three others. They are:

    mature

    restricted

    14 years

    Misc. Tags

    Below I’ve listed several other popular tags that you may have seen around the web.

    The generator META tag is used to specify what program was used to create your web site. Many HTML editing tools automatically insert this so that a company can gauge their market penetration. The tag looks like this:

    <META name=”generator” content=”program name and version”>

    The author tag is used to identify the author of a page. Simply replace “author’s name” with your name or email address.

    <META name=”author” content=”author’s name”>
    The copyright tag identifies the individual or company that holds the copyright to a particular page. This is what it looks like:

    <META name=”copyright” content=”This page and all its contents are copyright 2003 by Lee Windsor. All Rights Reserved.”>

    I hope this article gave you some insight into the usage of Meta Tags and how they can benefit your day to day business.

    Article written by Lee

  • Formatting HTML Text Using Tags

    Date: 2011.02.22 | Category: WebDesign | Response: 0

    Formatting HTML Text Using Tags.

    As we know there are many elements to a website from graphics to text. In this article we are going to have a closer look at text and, more importantly how we format the text to look like we want it to.

    The text of the HTML pages we create, as you already know goes inside the <body> tag of our HTML pages but, unless you are using a WYSIWYG editor you can not just click a button and make it appear bold or italic therefore we need to get to know some of the text property tags that can be used to enhance our HTML text. Lets take a look at some of these now.

    <b> Any text inside these two tags will appear bold on our page</b>
    <i> Any text inside these two tags will appear in italic on our page</i>
    <u> Any text inside these two tags will appear to be underlined on our page</u>
    <big> Any text inside these two tags will appear BIG on our page</big>
    <small> Any text inside these two tags will appear small on our page</small>
    <sub> Any text inside these two tags will appear subscript on our page</sub>
    <sup> Any text inside these two tags will appear as superscript on our page</sup>

    There are also things called ‘heading tags’ these will work the same as the formatting tags mentioned above however there are only six of them and they look like this <hx> with the x being replaced by a number from 1-6 the lower this number is in the tag the LARGER our text will become so for example:

    <h1> Will be the largest heading text</h1>
    <h2> Will be the next smallest heading tag</h2>
    <h3> Will be smaller again </h3>
    <h4> Will be one size smaller again</h4>
    <h5> Will be the second to smallest heading size</h5>
    <h6> Is the smallest of the heading tags</h6>

    Also, you should remember that it *IS* possible to use more than one of the text tags in any single line of portion of text on our web page so for example, if i wanted to have bold underlined italic text my tags for the text would look like this:

    <b><i><u>This text is bold, in italics, and underlined</u></i></b>

    You will notice from the above example that the tags were opened and closed in the same order they were created this doesn’t have to be done like this but, in the long run, it is easier for you as a webmaster to code your pages this way.

    Hopefully this article has given you a further understanding on how we can format out text and you will be bale to put this into practice on the next site you build.

    Article written by Lee

  • Adult Employment Vacancies

    Date: 2011.02.21 | Category: Employment | Response: 0

    For some the prospect of being able to work for another company online is something that they may ultimately want to pursue however, just how do you go about finding work for one of the many adult companies and, more importantly, how can you ensure that you have a better chance of being hired for the job once you find the position you would like to have at the company?

    Online Adult Employment – Job Types.

    Working in the adult industry can take many forms however, the most commonly perceived adult industry job is that of a model but, what about other online job vacancies such as graphics designer, marketing manager, sales director, and webmaster support representative, just how do you find out what companies are hiring and more importantly, what adult employment vacancies they have.

    Adult Employment Agencies.

    For many webmasters the chances of finding employment in the adult industry at a ‘regular’ online employment agency such as monster.com may seem impossible however, that said, many of the larger companies have realized the value of the mainstream employment agencies and, as such are now advertising on these sites for their staff but where else can you look? Well there are actually several adult industry employment sites that adult webmasters can look at to find their ideal job for example Adult Staffing, Sexy Jobs and, Adult Help Wanted in addition to several smaller sites operated by companies such as Cybererotica.

    Improving Your Employment Chances.

    So now you know where to look for adult employment opportunities what can you do to ensure you have a greater chance of being hired? Well as with the mainstream bricks and mortar job market having a good curriculum vitae (CV) can be a good start, ensure you keep your CV regularly updated and that it contains pertinent information for the type of job vacancy you are looking for in addition, often the first method of contact you will make with these potential employers will be via email or cover letter, make sure that you spend some time to write a letter that is not only noticeable but, interesting and, more importantly, grammatically correct. Demonstrate in this first stage of contact with your potential future adult employer that you take pride in your work and, you might just get bumped to the top of the ‘must see’ applicants.

    Adult Jobs – Overview.

    Many webmasters forget that there is more to the online adult industry than being able to party with the best of us and, whilst this does form a chunk of the industry overall it is not conducive of you further enhancing your online career, take the time to investigate all of the avenues available to you and, most importantly, make sure you choose a job that you not only want to do but one that you will enjoy, your reputation is the one thing that a potential adult employer will take into consideration, if you have had several previous jobs all lasting for short periods of time, it wont look as good to them as if you have only had a couple of jobs for a lengthier time frame. Most of all remember that whenever you contact a company looking for employers, make sure you are prepared and professional even if it is just sending them a quick email asking for further information.

    Article written by Lee.

  • What’s A CHMOD?

    Date: 2011.02.24 | Category: Hosting, WebDesign | Response: 0

    CHMOD is another name for setting the permissions of files on your server.

    All UNIX based web servers understand the concept of permissions. these are broken down into three distinct areas.

    Permissions for the person who OWNS the file.
    Permissions for the group (all users are members of one or more groups).
    Permissions for everyone else.

    Permissions are listed one of two ways either using numbers such as 755 or, using letters such as rwx-xr-x however they both mean the same thing.

    The first number or first group of three letters stands for the permissions of the owner (IE 7 or rwx).
    The second number or second group of letters stands for the permissions of the group (IE 5 or xr).
    The third number or third group of letters stands for the permissions of everyone else (IE 5 or x).

    The way the system works is as follows:

    1 = x = you are able to execute (run) this program.
    2 = w = you are able to write to this file.
    4 = r = you are able to read the contents of this file.

    so if you add it up rwx = 7.

    So, in the example given above (755 or rwx-xr-x) the owner is allowed to read and write the file AND execute it.
    The group is allowed to execute and write to the file.
    Everyone else can execute the file.

    We hope this gives you a little more insight to how CHMOD’ing or, setting the permissions on your server can affect how files act.

    Article written by Lee

  • Redirecting Questionable Adult Traffic

    Date: 2011.02.24 | Category: 2257, Scripts | Response: 0

    The one thing almost all reputable adult webmasters agree on is that one way or the other, we want to rid the net of those webmasters that profit from traffic primarily gained from either the direct promotion of or, targeting of, keywords relating to child pornography however, how can you tell what traffic you are being sent and, more importantly, how can you do something about the type of traffic that you receive to your site in order to filter out this unwanted traffic? The answer is simple, use a script to redirect the traffic elsewhere before it even hits your site.

    Child Porn Redirection Php Script.

    The following php script when used on your server will enable you to send unwanted traffic gained by the promotion of using ‘illegal’ keywords in the search engines and sites which link to your own.

    <!– Start Copy Here –>

    <?

    // Redirect “Lolita” traffic

    $refer_full_path = “$HTTP_REFERER”.”$PATH_INFO”;

    if(( preg_match(“/lolita/i”, $refer_full_path)) ||
    ( preg_match(“/child/i”, $refer_full_path)) ||
    ( preg_match(“/preteen/i”, $refer_full_path)) ||
    ( preg_match(“/pre-teen/i”, $refer_full_path)) ||
    ( preg_match(“/pedo/i”, $refer_full_path)) ||
    ( preg_match(“/underage/i”, $refer_full_path)) ||
    ( preg_match(“/beast/i”, $refer_full_path)) ||
    ( preg_match(“/rape/i”, $refer_full_path)) ||
    ( preg_match(“/kinder/i”, $refer_full_path)) ||
    ( preg_match(“/incest/i”, $refer_full_path)) ||
    ( preg_match(“/kiddie/i”, $refer_full_path))) {

    header(“Location: $refer_full_path”);
    exit;
    }

    ?>

    <?
    $words=array(“childporn”,”underage”,”beast”,”interracial”,”lolita”,”preteen”);
    for($i=0;$i<count($words);$i++){
    if(eregi($words[$i],$HTTP_REFERER)){
    header(“Location: http://www.fbi.gov/?CHILD_PORN_ON_DISK_LOGGED_AND_REPORTED”);
    }
    }
    ?>

    <!– End Copy Here –>

    In order to use this script, simply add additional keywords or partial word matches to the top part of the script and, include the bottom half of the script at the top of your HTML coding.

    Any traffic being sent to your site via keywords which you have specifically told the script not to allow access to your web site will be forwarded to the url in the bottom part of the php script which again, can be changed to wherever you like.

    Article written by Lee

  • Geo Targeting – What Exactly Is It?

    Date: 2011.02.22 | Category: General, Promotion | Response: 0

    As the adult industry reaches further and further across the online global community we are beginning to realize the importance of our foreign surfers, gone are the days of shifting them off to a dialer site instead, we find more and more companies actually using geo-targeting techniques to filter and monetize this profitable traffic more but, how does geotargeting work and, more importantly, is it any good? They are the questions we will look at in this article.

    GEO::IP – The Basics.

    The Geo::ip module is a relatively simplistic database of IP addresses and, the matching country to which the IP address belongs. This means that the GEO::IP module can be used for a variety of purposes including automatically selecting the language and countries of your visitors, for credit card fraud detection, and for software export controls.

    Geo Targeting – The Basics.

    As described above, the GEO::IP module is what drives any and all forms of geographical location targeting system (geo targeting). Geo targeting is the method most commonly used in the adult industry to either send a surfer to a localized version of a paysite or, send them off to a dialer in an attempt to monetize them it works by checking either the surfers browser language, country ip address or, both of these.

    Putting Geo Targeting To Work.

    Despite what most webmasters may think about the complexities of geo targeting and the GEO::IP Module for Perl they are easy to use and implement and, furthermore, there are cost free and highly effective ways of using geo targeting on your sites to monetize your international traffic. Lets look at two easy free solutions for webmasters to geo target their traffic.

    JavaScript.

    Using JavaScript you could add the following code between your HTML <head> and </head> tags to redirect international traffic to a specific url on your server:

    <SCRIPT LANGUAGE=”JavaScript1.2″>
    <!– Begin
    if (navigator.appName == ‘Netscape’)
    var language = navigator.language;
    else
    var language = navigator.browserLanguage;

    if (language.indexOf(‘en’) > -1) document.location.href = ‘english.shtml';
    else if (language.indexOf(‘nl’) > -1) document.location.href = ‘dutch.shtml';
    else if (language.indexOf(‘fr’) > -1) document.location.href = ‘french.shtml';
    else if (language.indexOf(‘de’) > -1) document.location.href = ‘german.shtml';
    else if (language.indexOf(‘ja’) > -1) document.location.href = ‘japanese.shtml';
    else if (language.indexOf(‘it’) > -1) document.location.href = ‘italian.shtml';
    else if (language.indexOf(‘pt’) > -1) document.location.href = ‘portuguese.shtml';
    else if (language.indexOf(‘es’) > -1) document.location.href = ‘Spanish.shtml';
    else if (language.indexOf(‘sv’) > -1) document.location.href = ‘swedish.shtml';
    else if (language.indexOf(‘zh’) > -1) document.location.href = ‘chinese.shtml';
    else
    document.location.href = ‘english.shtml';
    // End –>
    </script>

    Php.

    Using .php you can add the following to your sites pages to redirect traffic based on their browser language.

    <?
    $user_lan = $HTTP_ACCEPT_LANGUAGE;

    if($user_lan==’de’) {
    ## German
    $redir_url = “http://www.germanlanguagepageurl.com”;

    } elseif($user_lan==’fr’) {
    ## French
    $redir_url = “http://www.frenchlanguagepageurl.com”;

    } elseif($user_lan==’it’) {
    ## Italian
    $redir_url = “http://www.italianlanguagepageurl.com”;

    } elseif($user_lan==’es’) {
    ## Spain
    $redir_url = “http://www.spanishlanguagepageurl.com”;

    ## US traffic or Rest of world not defined above
    } else {
    $redir_url = “http://www.yourmainpageurl.com”;

    }

    header(“Location: $redir_url”);
    exit;

    ?>

    As you can see from the above two examples, monetizing your international foreign traffic is a relatively easy step and, with the right sponsors to send this traffic to you will be able to reap the profits of countries that other webmasters are not targeting yet.

    Geo Targeting – Overview.

    When all is said and done geo targeting can be a cost effective way to increase your bottom line profits, whether using a free method as described above or, using one of the many filtering services available on the net you should start to monetize your global surfers. One thing you should also consider however is that far from sending your international traffic to a dialer you should search around and find a good mix of adult affiliate programs with a variety of localized paysites, this in itself will enable you to see just how much traffic you have been wasting by using dialers in the past and, it will also allow you to realize the full potential of your new found geo targeted adult traffic.

    Article written by Lee

  • Fighting Back Against Computer Viruses

    Date: 2011.02.22 | Category: General | Response: 0

    Only 10 years ago, the only way that a virus could have infected your computer was through you physically putting a floppy disk into your hard drive and running the program file that initiated the virus however, today’s viruses have become more complex in how they operate and, can spread like wild fire if they are not recognized as a virus doing damage not only to your computer but that of your associates and peers.

    So how do we minimize the risk of virus infection on our computers which, lets face it, for the Adult Webmaster are our livelihood and, need to be protected at all cost. This is what we will take a brief look at in this article, how to minimize the risk of getting a computer virus and, heaven forbid, how to deal with a virus if you actually have one on your machine already.

    Computer Viruses – A Little History.

    1981 – The first widely accepted computer virus is recognized. The ‘Elk Cloner’ virus was spread through floppy disks and, whilst by today’s standards, it was not a high risk virus it, non the less was still a virus, in fact, all this first virus did was to display a message on your computer screen.

    1991 – The ‘Tequila’ virus was the first recognized ‘polymorphic virus’ meaning, it was the first virus which actually altered its state once on your machine in order to remain undetected.

    2001 – 2001 saw the spawn of several highly potent ‘worm’ orientated viruses from the ‘Gnuman’ virus, the ‘LogoLogic Worm’ and the famed (is that the right word?) ‘Nimda worm’. These ‘worms’ were the first of their kind to actually spread themselves across the World Wide Web through infecting users email clients (specifically Outlook Express) and sending itself to every email address the infected machine had in the address book.

    Virus Infections – Prevention Better Than Cure.

    So we now have a little background on the main viruses over the last 20 years we will take a look at the ways in which we can prevent viral infections on our computers after all, as the subtitle for this paragraph says, a prevention is far better than having to find a cure.

    Here are a few simple ways you can help to prevent the spread and infection of computer viruses.

    1) Install a good anti-virus software package on ALL of your computers, Norton and, Mcafee, both have some good products to aid you in preventing viral infections of your PC. Also, you should ensure the software package you do install has updated virus definitions, a new virus is created almost every week so you need to ensure your software keeps up-to-date on the latest infections.

    2) Scan any and all email attachments you receive, irrespective of whether you know the sender or not, anything that you get sent in an email should be scanned, as mentioned above, there are complex viruses that can ‘send’ themselves from other peoples email address book, what happens if you are on the address book of an infected users computer? Can you afford not to be online?

    3) Ensure that, once you have an Anti virus program installed on your machine that you have it setup to start each and every time that your computer boots, the first thing that should be loading on your machine is an anti-virus program, you only have one chance to prevent infection.

    4) Avoid downloading .exe and .com files to your desktop unless, you know the person you are downloading the files from and, even then, before you launch or open the files ensure you scan them independently.

    Virus Prevention – Overview.

    Hopefully this article will have given you some insight into how you can protect both your computer and the computers of your associates and peers from computer viruses and, hopefully, by following the guidelines within this article your time spent online will be productive and virus free. One thing you should remember however is that no matter how hard we all try to remain virus free, someone will create a new program that ultimately, could infect our machines no matter how hard we try.

    Article written by Lee

  • Straight From The Horses Mouth – Get Googlized

    Date: 2011.02.24 | Category: Search Engine Optimization | Response: 0

    Many webmasters wonder how to ensure their sites will be included in Google’s index of web sites. Although Google crawls more than a billion pages, it’s inevitable some sites will be missed. When Google does miss a site, it’s frequently for one of these reasons:

    * The site is not well connected through multiple links to others on the web.
    * The site launched after Google’s last crawl was completed.
    * The design of the site makes it difficult for Google to effectively crawl its content (excessive frames, tables, etc).

    Google’s intent is to represent the content of the Internet fairly and accurately. To help make that goal a reality, we offer this guide to building a “crawler-friendly” site. There are no guarantees a site will be found by our crawler, but following these guidelines should increase the probability that your site will show up in Google search results.

    Do…
    Provide high-quality content on your page – especially your home page.
    If you follow only one tip from this page, this should be it. Our crawler indexes web pages by analyzing the content of the pages themselves. Google will index your site better if your pages contain useful information. Plus, your site has a better chance of becoming a favorite among web surfers and being linked to by others if the information it contains is relevant and useful.

    Submit your site to the appropriate category in a web directory.
    Listing your site in the Open Directory Project http://www.dmoz.org/ or Yahoo! http://www.yahoo.com/ increases the likelihood it will be seen by robot crawlers and web surfers.

    Pay attention to HTML conventions.

    Make sure that your <TITLE> and <ALT> tags are accurate and descriptive. Also, check your <A HREF> tags for errors since broken or improperly formatted links can prevent Google from indexing your page.

    Make use of the robots.txt file on your web server.
    This file tells crawlers which directories can or cannot be crawled. Make sure it’s current for your site so that you don’t accidentally block our crawler. Visit: http://www.robotstxt.org/wc/faq.html for a FAQ answering questions regarding robots and how to control them once they visit your site.

    Ensure that your site is accessible through HTML hyperlinks.
    Generally, your site is crawlable if the pages are connected to each other with ordinary HTML links. If certain areas are not linked, you may be excluding older browsers, differently-abled users, and Google. Google can crawl content from a database or other dynamically generated content as long as it can be found by following links. If you have many unlinked pages, you may want to create a jump page from which the crawler can find all of your pages.

    Build your site with a logical link structure.
    A hierarchical link structure is not only beneficial to you, but also to Google. More of your site can be crawled if it is laid out in with a clear architecture.

    Don’t…
    Fill your page with lists of keywords, attempt to “cloak” pages, or put up “crawler only” pages.
    If your site contains pages, links or text that you do not intend visitors to see, Google considers them deceptive and may ignore your site.

    Feel obligated to purchase a search optimization service.
    Some companies “guarantee” your site a place near the top of a results page. While legitimate consulting firms can improve your site’s flow and content, others employ deceptive tactics to try and fool search engines. Be careful – if your domain is affiliated with one of these services, it could be permanently banned from our index, we have found search engine optimization software like Web Position Gold works best but, again use it in moderation.

    Use images to display important names, content or links.
    Our crawler does not recognize text contained in graphics.
    Use ALT tags if the main content and key words on your page cannot be formatted in regular HTML.

    Provide multiple copies of a page under different URLs
    Many sites offer text-only or printer-friendly versions of pages that contain the same content as the graphic-enriched version of the page. While Google crawls these pages, duplicates are removed from our index. In order to ensure that we have the desired version of your page, place the other versions in separate directories and use the robots.txt file to block our crawler.

    Article written by a Google employee

  • Pay Per Click Or Search Engine Optimization

    Date: 2011.02.24 | Category: Paid Traffic, Search Engine Optimization | Response: 0

    What would you choose to run your business? Well each has their own benefits and drawbacks over one another.

    PPC or SEO The Breakdown.

    Pay Per Click or, PPC as it is most often referred to looks to the novice to be the better option for ‘immediate’ traffic results, you enter your desired keywords, place your minimum / maximum bid amount and you are set for top PPC engine listings for as long as you can maintain the balance in your engine account.

    Search Engine Optimization or, SEO on the other hand, is the more traditional way of attaining high ranking search engine pages. Either you or an SEO expert optimizes your websites pages and random other elements of your website and hopefully within a month or two, you achieve high rankings in the major search engines.

    So Which One? PPC or SEO?

    Generally speaking, SEO work is most commonly more cost effective to your business than utilizing PPC results to gain your traffic, you could pay an SEO expert anywhere from $500 plus to optimize your site and get high rankings indefinitely or, you could put that $500 into a PPC engine account and get high rankings until such time as your account balance runs dry.

    But lets look at this in terms of actual traffic…

    Say you get 1000 visitors to your SEO based website which you paid $500 for, each visitor has cost you $0.50c now lets say your site remains at the top of the engines for a few months perhaps even years each month you receive another 1000 visitors to your site, you have basically cut the cost of each surfer hitting your site down to less than a penny per hit (not taking into account bandwidth costs obviously).

    Now, on the other hand, you want to attract 1000 visitors from your chosen keywords via the PPC engines, most Pay Per Click search engines have a minimum bid amount of $0.05c per hit so right away in your first month, you could receive a potential 10k hits however, as most of you who have already tried your hand at the PPC engines will know, getting 10k hits for one or more keyword at a cost of $0.05 is hard to do, in fact, some would say almost impossible. Non the less let us keep going with this minimum bid amount for the time being.

    Immediately, you can see that you are already restricted to the actual amount of traffic you can receive from the PPC results to 10k hits however, this isn’t the case with the SEO traffic, you could potentially hit your top chosen keyword and stay there until another site out-optimizes you or, your site needs to be optimized again.

    Ultimately, the reasons you will choose over one or the other will be for either ease of traffic generation, PPC will allow you to gain almost instantaneous targeted traffic form the second you open your PPC account up until the point when your account funds empty whilst, SEO work will give you long term targeted traffic over time and, in most instances, this SEO traffic can last for years making the cost of the initial SEO work minimal.

    In Closing..

    Search Engine Optimization can last you years and years whilst Pay Per Click results can diminish in a relatively short amount of time depending on the amount of bid needed to achieve top listings.

    However lets look at a third option, using both PPC results and SEO results in conjunction with each other to minimize the traffic you lose from your SEO work and, to minimize the traffic you lose from your PPC results this will afford you the time to see what works with your Search Engine Optimized sites whilst being able to play with the targeting of keywords on your PPC traffic, once you have both types of search engine figured out, you can put them both together and use them to increase the traffic to your site for years to come.

    Article written by Lee

Premium Sponsors















Categories

Site Links