• Mirroring Adult Sites – Stage Four

    Date: 2011.02.23 | Category: Promotion, Tutorials | Response: 0

    We are now going to take our basic template and start to make some pages for the link lists for this you will need to open up your original warning page. Once you have this page open you need to visit a couple of the link lists that you use. Go to the submissions page of the link list and download their reciprocal buttons.

    In the same way that we did with the TGP pages we created you now need to edit the warning page you have open and insert the recip links on your warning page, ideally and, as most link lists will require this, the recip links should be placed above the enter and exit links.

    Once you have done this, you now need to save your newly edited page as linklist1.html in the /FreeSite/LinkList/ folder, do this for as many of the link lists that you wish to submit to, each time saving the new page in the /FreeSite/LinkList/ folder.

    Also at this stage, we should add a console to our link list page, ideally a blur console with links going to your surfer trap FPA’s. You can find the Java coding to create a blur console on the tutorial about building a surfer trap.

    Once we have done this, we now have some more sites created. However, this time, the page/s we have just created are designed to build the traffic we have to our site/s.

    Once you have these pages saved you now need to re-open your original warning page once more.

    This time, we are going to optimize the page for the search engines, again, in the same way we added our ALT tags on the surfer trap we are going to add them to the warning page however, we are also going to optimize our keywords, description and, site title.

    Once we have optimized our original warning page we should now save this as se1.html in the /FreeSite/Engine/ folder we created at the start of this tutorial. Once one SE page has been created, you need to create a selection of others making sure that you swap around the meta tags to ensure they are all different to each other, I would suggest making three variations of these SE pages each time saving them in the /FreeSite/Engine/ folder.

    The next thing we are going to do is to ‘dirty’ these SE pages up some more, pretty much in the same manner as we did with the surfer trap tutorial. What we need to do is add a small pop up console to the three SE pages along with a blur console, these consoles can be the same ones that we used for our surfer trap or, we can create new ones however, I think the smart thing to do would be if we used the ones we currently have online. So once we have added these consoles we need to save the pages.

    Also, to these SE pages we should add our Banner exchange code, as these will be going into the search engines there are no rules as to what we can and can not do on these SE pages so we should make sure that even if they just get one hit, we get additional traffic from them. By adding a banner exchange code as we did on the surfer tutorial we can drive additional traffic either to our sites or to our surfer trap, where you send this traffic is your choice. If you want to sign up for our new banner exchange you can find it at http://www.pornclient.com.

    Now we have the completed site template built. We need to submit the pages and sites we have just completed to the TGP’s, AVS’s, Link Lists and, the Search Engines.

    Once we have submitted these sites all that we need to do to create a new set of sites is to find another set of 50 images, rename the images as we did at the start of this tutorial, pic1.jpg, thumb1.jpg etc and simply upload the site to our server without the full size images and the thumbnails. Once the HTML pages are online, we now have to upload the full size images and the thumbnails and we have a set of sites using different content to our original one built in less than 5 minutes.

    The best thing with these sites is that if a any time we want to alter the layout, all we need to do is edit the HTML of the version we have saved on our HD and we have a new template ready to use.

    I hope you have found some use out of this tutorial and, I am sure you can think of other ways in which we can use templates of this style, this tutorial was just detailing some of the basics however, we can make the template as simple or as complicated as we like. Our only limits are what we make for ourselves.

    Article written by Lee

  • Using TITLE Tags effectively

    Date: 2011.02.24 | Category: Search Engine Optimization, WebDesign | Response: 0

    Microsoft Internet Explorer has supported the TITLE attribute for links since version 4.0. Netscape began supporting it in version 6.0

    The TITLE attribute allows you to enter descriptive text about a link that’s displayed when the mouse moves over the link. For an example, move your mouse over the link below:

    European Webmasters

    The HTML source code for this link is shown below:

    <a href="http://www.europeanwebmasters.com"
      title="Click Here To Learn Foreign Adult Marketing">European Webmasters</a>

    If your browser supports the TITLE attribute you should see a small popup box appear, similar to the way alternate text for images is displayed when you move your mouse over the image.

    This attribute is similar to the ALT attribute for the IMG tag, both in its appearance and its limitations. The value of the attribute must be inside quotation marks, and it can only contain plain text (no HTML tags). You can include carriage returns in the title, and these will be displayed in the text. The maximum length of the TITLE text for Internet Explorer is very large; we’ve tested it with over 500 characters. Try to keep your descriptions below 25 words, though; any more than this can be hard to read and, because the search engine spiders can read this text you may be considered as spamming the engines.

    The TITLE attribute is useful in places where your HTML design limits the length of your link text. That’s often the case for links in a navigation bar, especially if your page has a multi-column layout. Try using the TITLE attribute to give your visitors extra navigation information.

    In addition you can also use the TITLE tag on standard text for extra information where you just don’t have the space to be as affluent as you would like. An example of this is shown below:

    Use of the TITLE tag on plain text

    The HTML code for this is also shown below:

    <font size=”2″ face=”Verdana” color=”#000080″ title=”Use Of The TITLE Tag On Plain Text”>Use of the TITLE tag on plain text</font>

    This attribute is part of the HTML 4.0 standard. Internet Explorer, Netscape 6.x and Opera (Version 3.0 or higher) support it. However, the attribute degrades gracefully, visitors with non-supporting browsers (like Netscape 4.x) would see the link as if you hadn’t used the attribute.

    As you can see from the above examples by using the TITLE tags effectively you can double, if not treble your use of keywords however, caution should be paid when using these features of the tag so that you don’t inadvertently spam the search engines.

    Article Written By Lee

  • Everything You Need To Know About Meta Tags And More!

    Date: 2011.02.22 | Category: Search Engine Optimization, WebDesign | Response: 0

    Everything You Need To Know About Meta Tags And More!

    There’s a plethora of different META Tags that you can make use of on your site. Because there are so many, it’s impossible for me to cover all of them in this article. However, I will describe the most common ones. But let’s start from the beginning, shall we?

    What Are META Tags?

    META tags are similar to standard HTML tags. However, there is one big difference: You have to insert all META tags between the <head> </head> tags on your page(s). META tags are used primarily to include information about a document. The META tags will be invisible to your site’s visitors, but will be seen by browsers and search engines.

    For The Search Engines.

    Several of the major search engines make use of the META Keyword tag, and virtually all of them make use of the Description tag. These tags help the search engine spider determine the content of your web site so that it can be indexed properly.

    This is what they look like:

    <META name=”keywords” content=”Webmaster Resource Site”>
    <META name=”description” content=”Online Webmaster Resource Site”>

    You can also instruct the search engine robot/spider how to index your site using the robots META Tag. This is what it looks like:

    <META name=”robots” content=”noindex,nofollow”>

    What the tag above does is tell the spider not to index the page that it appears on, and not to follow the links on that page. Here’s a complete list of attributes for the robots tag:

    index – the default, the page is added to the search engine database

    noindex – the page is not added to the search engine database

    nofollow – the spider doesn’t follow the links on that page

    none – same as “noindex, no follow”

    To use any of these attributes simply replace the “noindex,nofollow” text in the example above with whatever you want to use. If you need to use more than one attribute, separate them with a comma.

    Client Pull.

    Have you ever seen a page that automatically refreshes to another URL after a few seconds? Did you ever wondered how it was done? I’ll tell you. The page made use of one of the http-equiv META Tags to automatically “pull” you to a different page after a few seconds. Here’s what the code looks like:

    <META http-equiv=”refresh” content=”1; url=newpage.htm”>

    The value of content denotes how many seconds will pass before the new page is called for. If you want it to happen as soon as a person hits that page, then set the value to “0”.

    Prevent Caching.

    If you want to prevent a page being saved in your visitor’s cache you can do so by inserting the following three tags:

    <META HTTP-EQUIV=”expires” CONTENT=”0″>
    <META HTTP-EQUIV=”Pragma” CONTENT=”no-cache”>
    <META HTTP-EQUIV=”Cache-Control” CONTENT=”no-cache”>

    Why would you want a page not to be cached? If your site is updated frequently and you always want your visitors to see the newest content/changes, or if you want to ensure that a new banner is loaded each time from your server when a page is accessed, you’ll want to use the tags above.

    A quick note here on banner caching: In addition to using the tags above, you’ll want to append random numbers at the end of the tag calling the image.

    Rating Your Content.

    By using the rating META tag on your site, you can specify the appropriateness of your web site for kids. The tag looks like this:

    <META name=”rating” content=”general”>
    In addition to the general rating, you can use three others. They are:

    mature

    restricted

    14 years

    Misc. Tags

    Below I’ve listed several other popular tags that you may have seen around the web.

    The generator META tag is used to specify what program was used to create your web site. Many HTML editing tools automatically insert this so that a company can gauge their market penetration. The tag looks like this:

    <META name=”generator” content=”program name and version”>

    The author tag is used to identify the author of a page. Simply replace “author’s name” with your name or email address.

    <META name=”author” content=”author’s name”>
    The copyright tag identifies the individual or company that holds the copyright to a particular page. This is what it looks like:

    <META name=”copyright” content=”This page and all its contents are copyright 2003 by Lee Windsor. All Rights Reserved.”>

    I hope this article gave you some insight into the usage of Meta Tags and how they can benefit your day to day business.

    Article written by Lee

  • Filtering Adult Traffic For More Sales

    Date: 2011.02.22 | Category: Traffic | Response: 0

    When webmasters build sites there seems to be the train of thought that just because they use one of the top converting sponsors site they will make sales regardless of what type of traffic they are sending.

    Unfortunately, this is often not true.

    One of the first steps that we should take as webmasters when it comes to creating a new site, whether it be a free, avs or a TGP site is to learn the art of targeting our sponsors tour with the content and, more importantly, the traffic we are going to use.

    taking it to the simplistic levels, a straight teen sponsors wont convert on a site that gets mostly mature traffic, the niches, although complementary, are so far apart its unrealistic.

    That said, lets take a look at some options we can all implement on our sites to better target and filter our traffic to ensure optimum conversions.

    One of the simplest ways to target and filter your traffic is to build yourself a hub or a surfer trap. Of course, we also have to take into consideration the type of traffic we are currently receiving to our site and, that we ‘could’ receive in the future.

    Just because we receive 10,000 hits to our teen TGP doesn’t mean that ALL of that traffic IS looking for a teen site, this is where our surfer trap or hub site can come into play.

    Even if your main focus is the teen niche, you should make sure that the first thing the surfer see’s when visiting any type of site you build is a multi-site FPA (Full Page Ad) where applicable.

    This then starts your filtering process. If they know that they can find other niches than teen on your TGP then chances are, they will go to them either because that is what their niche is or, they are curious about that niche.

    Another thing we can do as webmasters is to ensure we use the headers and footers of our pages in the proper fashion, offer our surfers a choice of the main niches, Asian, Gay, Ebony, Teen, Ect using only text links and, you will probably find that by lining these text links to your hub or surfer trap that you actually have a diverse range of traffic available to you.

    In addition, rather than sending our surfers directly to our sponsors, send them via a filter page whether this is an FPA, or a page full of text links you made yourself detailing the benefits of the sponsor you are going to send them to, again, you will let the surfer know where they are going and, give them a choice as to whether they want to continue to that site or, choose an alternative.

    Overall, webmasters who target their surfers by niche can, for the most part have some excellent sales with a variety of sponsors that otherwise they would not be able to see.

    Target, Filter, Micro-Niche, however you do it, make sure your traffic is not wasted and, more importantly, make sure your sponsor is able to offer your surfer what you are telling them is inside the site.

    Article written by Lee

  • Building A Surfer Trap – Stage 1

    Date: 2011.02.21 | Category: Traffic | Response: 0

    The last 2 weeks I have had a LOT of webmasters email and ICQ me regarding traffic and, not getting any sales.

    So, I thought it would be a good idea to make a step by step tutorial on building your own surfer trap.

    Now these are some of the arguments I have heard why people haven’t already built one over the last few weeks…

    * To complicated to follow.
    * I want surfers returning to my site.
    * Don’t have the time.

    The first two however seem to be the ones that I hear over and over again.

    First of all, a surfer trap IS NOT complicated, in fact, it is probably one of, if not THE easiest type of site to build.

    Secondly, Why get a surfer visiting your site over and over again if they are not paying for trial memberships? All you are doing is wasting your bandwidth.

    Thirdly, A surfer trap can be built over a period of time, my very first trap was built within 1 hour my next was built in a day and my next one was built in a week.

    So onto the start of building our surfer trap…

    The first thing that I want you to do is to look at the Multi-Site FPA’s that ARS in the marketing section and choose ONE of them.

    Download it, optimize the keywords on the FPA, make sure you add meta tags, titles descriptions and Alt tags on all images.

    Once you have done this, you now need to search through ALL of the single site FPA’s and download and optimize one for each of the ARS sites that you have on your Multi-Site FPA. (Even the gay sites!)

    This should take you 2-5 minutes for EACH FPA, any longer than that and you are doing TOO much work!

    Once you have all of the single FPA’s downloaded, optimized and uploaded onto your server that is the first stage completed.

    Article written by Lee

  • Protecting The Kids – Europe’s Role

    Date: 2011.02.24 | Category: 2257 | Response: 0

    As adult webmasters we have certain obligations we must fulfill daily whether these be, legal or moral.

    One issue that is constantly cropping up in our industry is that of minors (Children) accessing adult websites and, more importantly, the ease of which they can do.

    Recent figures released by one of Europe’s top internet analysis companies show that children in Europe and, specifically within the United Kingdom are more likely to visit a website and that, ‘showing that around one in five Internet kids (290,000) check out the sex sites’.

    These figures are certainly alarming to say the least, bearing in mind they are specifically talking about the United Kingdom, that leaves a vast majority of Europe, as yet, un-analyzed.

    We should all adapt to new methods in order to protect minors from accessing our site whether that includes placing your sites behind AVS systems which, in my opinion is one of the best ways and, most profitable of protecting minors or, utilizing ‘web safe’ software packages that are available to block access to adult sites completely we all need to adopt these practices.

    The fact of the matter remains that, in this day and age, no matter what precautions we take in order to prevent access there will, undoubtedly, be a majority of children who can and will, get access to adult sites.

    Does that mean we should compromise on the quality of our sites? No
    Does that mean we need to think harder about how we develop our sites? Yes

    It takes very little effort to subscribe to the services to block access and, you can get a single line of code to place within your html from ICRA which will at least, deter underage surfers from accessing your sites.

    In addition to the above mentioned service, you can also place a link on your warning page/s to the following services for parents to block access from children to our sites, SAFESURF, CYBERPATROL, NETNANNY and, SURFWATCH amongst others. Its not just about utilizing software and ratings on our sites though, we can go much further than that and, for instance, using nothing but softcore content on our sites will protect to some extent the materials available to children but, in addition you will also notice your conversion ratios will go up! I have and, I’m sure you have, heard it time and time again.. Make them ‘Pay for pink’ and, by using softcore images on your sites that’s exactly what you are doing.

    I for one certainly hope that you take heed at this article, protecting our children is paramount if, we are to succeed in the business, it is getting harder and harder to control who accesses our site but, we can at least take a step in the right direction.

    Article written by Lee

  • Why Cant I Get Indexed By The Search Engines?

    Date: 2011.02.24 | Category: Search Engine Optimization | Response: 0

    Unfortunately, this is an all too common question. If it makes you feel any better, you’re not the only one frustrated about the length of time it takes to be indexed, or the many pitfalls involved. It often takes anywhere from two days to as much as six months to be listed on a search engine. For example, last month Excite finally updated its index for the first time since last August! Luckily, Excite is the most extreme case lately, but waiting several weeks to a month can also be extremely frustrating especially when your livelihood depends partly on these search engines.

    The Web Position Submitter report will give you current time estimates for each engine so you’ll know what to expect. However, an engine at any time could choose to delay their indexing beyond the “norm” for maintenance or other reasons. On the flip side, you could get lucky and submit just a couple days before an engine does a complete refresh of their database. Therefore, submission times can never be an exact science since we’re all ultimately at the mercy of the engine.

    If you’ve submitted your site and have waited the estimated time to be indexed and there’s still no listing, what do you do now?

    Here are 16 tips that should help you solve this problem:

    1. First, be sure you’re not already indexed but just don’t know it. Unfortunately, none of the major engines are kind enough to e-mail or notify you as to if and when you’ve been indexed.

    The method to determine if a page or domain has been indexed varies from one engine to another, and in many cases, it’s difficult to tell for sure. Never assume that you’re not indexed just because you searched for a bunch of keywords and you never came up in the first few pages of results. You could be in there but buried near the bottom.

    In addition, it’s not very practical to check the status of a number of pages on each major engine each week. Fortunately, Web Position has a URL verification feature in the Reporter that makes this process much easier. Each time you run a mission, it will report which URLs exist and do not exist in each engine. If you’re using Web Position and are not finding your URLs after submitting, be sure to see this page for common pitfalls to watch out for:

    http://www.webposition.com/urlnotfoundhelp.htm

    2. Make sure you have uploaded the pages to your site before submitting them. This one seems obvious, but submitting a page that does not exist or submitting with a subtle typo in the URL is a goof we might all make at one time or another. If you’re using Web Position’s Submitter, there’s a checkbox on tab 2 that forces Web Position to verify that all your URLs are valid before submitting them.

    3. If you have information inside frames, that can cause problems with submissions. It’s best if you can create non-framed versions of your pages. You should then submit the non-frames versions of your pages which can of course point to your framed Web site. Alternatively, you can enter your relevant text within the NOFRAMES area of a framed page which most search engine spiders will read.

    4. Search engine spiders cannot index sites that require any kind of registration or password. A spider cannot fill out a form of any kind. The same rule applies regarding indexing of content from a searchable database, because the spider cannot fill out a form to query that database. The solution is to create static pages that the engines will be able to find.

    5. Dynamic pages often block spiders. In fact, any URL containing special symbols like a question mark (?) or an ampersand (&) will be ignored by many engines.

    6. Most engines cannot index text that is embedded in graphics. Text that appears in multimedia files (audio and video) cannot be indexed by most engines. Information that is generated by Java applets or in XML coding cannot be indexed by most engines.

    7. If your site has a slow connection or the pages are very complex and take a long time to load, it might time out before the spider can index all the text. For the benefit of your visitors and the search engines, limit your page size to less than 60K. In fact, most Webmasters recommend that your page size plus the size of all your graphics should not exceed 50K-70K. If it does, many people on dial up connections will leave before the page fully loads.

    8. If you submit just your home page, don’t expect a search engine to travel more than one or two links away from the home page or the page that you submitted. Over time they may venture deeper into your site, but don’t count on it. You’ll often need to submit pages individually that appear further down into your site or have no link from the home page.

    9. If your Web site fails to respond when the search engine spider pays a visit, you will not be indexed. Even worse, if you are indexed and they pay a visit when your site is down, you’ll often be removed from their database! Therefore, it pays to have a reliable hosting service that is up 99.5% of the time. However, at some point a spider is going to hit that other 0.5% and end up yanking your pages by mistake. Therefore, it pays to keep a close eye on your listings.

    10. If you have ever used any questionable techniques that might be considered an overt attempt at spamming (i.e., excessive repetition of keywords, same color text as background, or other things that the Web Position Page Critic warns you about), an engine may ignore or reject your submissions. If you’re having trouble getting indexed in the expected amount of time, make sure your site is spam-free.

    11. If your site contains redirects or meta refresh tags these things can sometimes cause the engines to have trouble indexing your site. Generally they will index the page that it is redirecting TO, but if it thinks you are trying to “trick” the engine by using “cloaking” or IP redirection technology, there’s a chance that it may not index the site at all.

    12. If you’re submitting to a directory site like Yahoo, Open Directory, NBCI.com, LookSmart, or others, then a human being will review your site. They must decide the site is of sufficient “quality” before they will list it. I recommend you read the submission guide on the directory tab of the WebPosition Submitter. It contains tips to improve your chances of obtaining a good listing on these directories.

    13. A number of engines no longer index pages residing on many common free web hosting services. The common complaint from the engines is that they get too many “junk” or low-quality submissions from free web site domains. Therefore, they often choose not to index anyone from those domains or they limit submissions from them. It’s always best to buy your own domain name (very important) and place it on a respected, paid hosting service to avoid being discriminated against.

    14. Some engines have been known to drop pages that cannot be traveled to from the home page. HotBot has been rumored to do this. You may want to consider submitting your home page that links either directly or indirectly to your doorway pages.

    15. Make sure you’re submitting within the recommended limits. Some engines do not like more than a certain number of submissions per day for the same domain. If you exceed the limit, you may find that all your submissions are ignored. Fortunately, WebPosition’s submitter will warn you regarding current limits and recommend you stay within them. Some submission consultants feel it is dangerous to submit more than ONE page a day to a engine for a given Web site. For those who wish to be ultra-conservative in their approach, the Web Position Submitter includes a checkbox to limit submissions to one URL per day per engine.

    16. Last but not least, sometimes the engines just lose submissions at random through technical errors and bugs. Therefore, some people like to resubmit once or twice a month for good merit in case they do lose a submission. Certainly if you’ve followed all the “rules” and are still not listed, re-submit! Sometimes a little persistence is all that’s needed.

    If any of the above scenarios apply to your submission, you should make the necessary adjustments and re-submit. If that still does not work, you should consider e-mailing or calling the search engine and asking them politely why you have not been indexed yet. Sometimes they will reply back with “Sorry, there was a problem with our system and I’ve now made sure you’ll be indexed within the next couple days.” Or, sometimes they’ll tell you why you were not indexed. In other cases, they will ignore your e-mail and you’ll have to keep e-mailing or calling them until they respond. Still, it’s definitely worth the effort to get your site listed with the major engines assuming you also take the time to optimize your pages so you’ll achieve top rankings.

    Article written by Lee

  • Marketing Sites With Slogans

    Date: 2011.02.22 | Category: Promotion | Response: 0

    With the constant changes happening in the adult industry webmaster need to look towards alternative ways of marketing not only their own sites but the sites of their sponsors. This is where something that we have all seen daily can come in handy, what is this marketing tool? Simple, Slogans.

    Slogans – What Are They?

    A slogan is a short, memorable advertising phrase for example, ‘Just Do it’ – Nike. By utilizing a slogan on your site you are not only giving your surfers something other by which they can remember your site by but, you are actively building up your brand which, in turn, can mean more repeat traffic and ultimately, more sales.

    How To Make A Slogan.

    So now you know what a slogan is how do you go about creating one? Simple, you should think of something short, usually three to four word slogans work the best again, as with the Nike example above ‘Just Do it’ its short, sweet and memorable, try to make your own slogan fit within this guideline and you are half way done. Another thing that you may like to consider is making your slogan practical or humorous, something witty can often stay in a persons mind a lot longer than something serious, play on peoples memory and ensure they remember your brand.

    Where To Use Your Slogan.

    Now that you have thought up a slogan which you feel will work on your traffic and sites you need to consider the places where you can use it for maximum impact, the most obvious place to use your slogan is on your site design, either at the top or the bottom of your pages, make it clear and more importantly, make it stand out. In addition, if you have any banners or buttons created, ensure you utilize the slogan on those creatives, the goal is to have your slogan seen by as many people, as quickly as possible.

    Article written by Lee

  • European Surfers – Porn High On The List!

    Date: 2011.02.22 | Category: General | Response: 0

    Adult web sites were visited by 3.8 million home Internet users in the UK in June 2001, according to Internet monitoring company Net Value. This accounts for more than a quarter of the entire UK home Internet population.

    UK visitors to adult websites spent more than 45 minutes on these sites each in June. Adult sites make up 41.2 per cent of all the sites visited in the UK with a significant audience. Students (23.2 per cent), manual workers (15 per cent), and professionals (12.8 per cent) are the largest groups of UK visitors to adult sites.

    However Germany has the largest audience for online porn in Europe with more than 5.3 million Germans visiting an adult site in June. They also spent more time on these type of sites than any other nationality – each German visitor spent almost an hour on adult sites in June. Net Value data also showed that more than half (52.3 per cent) the websites visited in June by a significant number of German users were of an adult orientated nature.

    Spain has the greatest proportion of visitors to pornographic sites – 40 per cent of Spanish home Internet users visiting a pornography site in June 2001.

    “Pornography has become a lucrative online business, with many companies now recognizing it as a valuable additional revenue stream,” Net Value’s Alki Manias said. “And this would seem to be a stable business area – the popularity of online pornography continues unchecked.”

    Adult sites are also popular amongst the earliest adopters of the Internet. People who first got online at home before 1997 account for 18.5 per cent of visitors to adult sites, though they only make up 14.1 per cent of the whole UK online population.

    In the UK, Lastminute.com has just introduced an auction for adult toys and aids, whilst in Germany, Freenet.de and T-Online are also moving into the adult business.

    The following charts detail various demographics relating to European surfers usage of adult sites:

    Visitors To Adult Sites In Europe

    Time Spent On Adult Sites By European Surfers

    Number of adult websites visited, June 2001

    Number of adult websites visited Total number of Websites visited* % of websites which have adult content
    Germany 4,140 7,909 52.3
    Norway 3,847 7,756 49.6
    Denmark 3,675 7,661 48.0
    Italy 3,131 6,579 47.6
    Sweden 2,419 5,231 46.2
    Spain 3,069 7,019 43.7
    UK 3,879 9,411 41.2
    France 4,348 11,845 36.7

    * Sites with a significant reach

    UK visitors to adult websites by occupation, June 2001
    Occupations Of Porn Surfers In The United Kingdom

    As you can now see, we, as adult webmasters, have a considerable chunk of the online adult industry sitting in our laps and, we should concentrate on tapping into this market as more and more American companies are realizing the importance of these figures we will inevitably reach a stage where the European market will be passed across to our US counterparts.

  • What Is The Golden Ratio?

    Date: 2011.02.24 | Category: Sponsors | Response: 0

    Its a little known marketing technique that say that, as long as the longest side on any oblong is 1.6 times the length of the shortest side of the same oblong people will be more likely to prefer that shape over any other shape.

    Why do they prefer this size? Because it occurs naturally and, subconsciously EVERY LIVING PERSON is attracted to this trait.

    An example of this is as follows:

    1) Stand Up
    2) Measure the distance from your head to your feet and write this measurement down
    3) Measure the distance from your Navel write this figure down.

    The length of your ENTIRE body is 1.6 times longer than from your Navel to your feet!!

    The same is also true for the length of your arm from shoulder to finger tip, then from elbow to finger tip.

    The same is true for credit cards, the longest side is 1.6 times longer than than the shorter side!

    Think about this for a second, banner ads, what shape are they? Oblong except, they don’t employ the ‘golden ratio’ what if, just what if, you used some banner ads that employed this ‘golden ratio’ who says banners have to be 468×60 in size?

    Why not try making some ‘basic’ banners in paint or a decent graphics program utilizing the 1.6 rule and see if your click thru’s increase, personally, I will bet my earnings last month that they will.

    On the same way of thinking towards marketing, table ads, money bars, etc, why not incorporate this idea into them also? It cant hurt and who knows it may just work.

    the math to work this ‘golden ratio’ out is as follows for Image a above:

    shortest side (60) x 1.6 = (90) Longest side

    You don’t even need to be a genius to make your own ads using this method as the people at Microsoft supplied you with a calculator.

    Article written by Lee

Premium Sponsors















Categories

Site Links