-
What’s In A Niche?
Many webmasters realize what the different ‘fetishes’ the surfers have are but, how many of us realize these are actually niches and, how many of realize that these niches can be broken down into smaller sub niches of the particular fetish the surfer is looking for?
Lets take an example of say, the ‘lesbian’ niche.
Of course we all know what the lesbian niche entails, two women getting naked with each other right? Wrong.
There is much more to it than that, for example, obviously lesbian sites appeal to, well, they appeal to guys who want to fantasize about getting it on with two women but, what about the female surfers? Sure, some women probably join lesbian paysites too.
That’s one ‘sub-niche’ of lesbian right there, this can be broken down even further though for examples, Ebony Lesbians, Teenage Lesbians, Mature Lesbians, the list goes on and on and on.
So how do we choose a niche to market in adult?
The first and foremost thing you should try when choosing ‘your’ niche is to try something that either excites you or, appeals to you in a non sexual manner, if you are into Lesbian Midgets getting it on, then start off with that for your niche.
Likewise, if guys getting down and dirty is the thing that floats your boat, choose that as your niche, you should already know enough about the niche you choose primarily to be able to sell something to the surfer. If it excites you, it will more than likely excite a surfer and result in a sale.
Of course, the above statement isn’t always true, I know plenty of straight webmasters who market the gay niche VERY successfully, they have found a method that works for them and developed their business from there it probably wasn’t easy at the start but, nothing easy will make you the ‘real’ money that there is in adult.
You need to experiment, constantly adapt and change the way you do business, if you try something and it doesn’t work, ask yourself why, what could you have changed that might have made a difference and, instead of changing it on the first site you built, change it on your next one, that way, even though it may not work at this moment in time, you will no doubt find it will work at a later date just as you will find that something that works well for you now may not work at all later on in your business career.
Once you have mastered a general niche you should then think about trying some of the sub niches that are available to you, there is more to teen than than just cute teenage girls lying naked on a bed and getting screwed rotten by some guy. just as you saw there was more to the lesbian niche above.
Take your sponsors tour for ideas, often you will notice things that spark some though and, once you have that spark, you should ensure it develops into a flame.
Even something as simple as a tattoo on a model or a piercing can be the tart of a whole new niche, look at the things that other webmasters may not be, just because there isn’t a proliferation of teenage pierced lesbian sites, doesn’t mean that someone doesn’t want to join one and, who knows, that weird sexual fetish you build a site for now may become popular in 6 months time.
Basically, what I am trying to say is that as adult webmasters we need to investigate all aspects of the niche we want to or, currently do market, just because a surfer comes to the front page of a lesbian site, doesn’t mean that they wont also be interested in an ebony site, its all down to how we market our sponsors paysites as webmasters.
Article written by Lee
-
Formatting HTML Text Using Tags
Formatting HTML Text Using Tags.
As we know there are many elements to a website from graphics to text. In this article we are going to have a closer look at text and, more importantly how we format the text to look like we want it to.The text of the HTML pages we create, as you already know goes inside the <body> tag of our HTML pages but, unless you are using a WYSIWYG editor you can not just click a button and make it appear bold or italic therefore we need to get to know some of the text property tags that can be used to enhance our HTML text. Lets take a look at some of these now.
<b> Any text inside these two tags will appear bold on our page</b>
<i> Any text inside these two tags will appear in italic on our page</i>
<u> Any text inside these two tags will appear to be underlined on our page</u>
<big> Any text inside these two tags will appear BIG on our page</big>
<small> Any text inside these two tags will appear small on our page</small>
<sub> Any text inside these two tags will appear subscript on our page</sub>
<sup> Any text inside these two tags will appear as superscript on our page</sup>There are also things called ‘heading tags’ these will work the same as the formatting tags mentioned above however there are only six of them and they look like this <hx> with the x being replaced by a number from 1-6 the lower this number is in the tag the LARGER our text will become so for example:
<h1> Will be the largest heading text</h1>
<h2> Will be the next smallest heading tag</h2>
<h3> Will be smaller again </h3>
<h4> Will be one size smaller again</h4>
<h5> Will be the second to smallest heading size</h5>
<h6> Is the smallest of the heading tags</h6>Also, you should remember that it *IS* possible to use more than one of the text tags in any single line of portion of text on our web page so for example, if i wanted to have bold underlined italic text my tags for the text would look like this:
<b><i><u>This text is bold, in italics, and underlined</u></i></b>
You will notice from the above example that the tags were opened and closed in the same order they were created this doesn’t have to be done like this but, in the long run, it is easier for you as a webmaster to code your pages this way.
Hopefully this article has given you a further understanding on how we can format out text and you will be bale to put this into practice on the next site you build.
Article written by Lee
-
Blogging For Traffic – An Alternate View
Just recently the adult industry seems to have gone ‘blog’ crazy in fact, it is almost a daily occurrence that webmasters mention a blog of some description on several message forums however, when it comes down to it, does blogging really matter at all? This is what we will take a look at in this article.
What Is A Weblog Or Blog?
A Weblog, or ‘Blog’ for short, is a kind of website or a part of a website. It is usually, but not always, run by a single person and they publish bits of writing on the Weblog. These bits of writing, called ‘entries’ or ‘posts,’ generally appear on the front page of the weblog in reverse chronological order, that is, with the newest entry at the top of the page, with older entries progressively further down.So How Can Blogging Help Me?
Quite honestly, I don’t think it can. Think about it for a moment, you spend an hour or so a day writing a page in your online journal only for it to be erased or archived at a later date, doesn’t that seem strange to you? Instead of blogging why don’t you actually do something constructive such as making a .html page out of the text you were going to use for your blog, keyword it up and upload that directly onto your server? After all this in itself will benefit you in the long term much more than a quick page that will eventually disappear from existence. Of course there will always be those people who do successfully use blogs as a part of their marketing strategy and, kudos to them however, looking at the bigger picture, it makes much more sense to have a fresh HTML page which will remain indefinitely on your site that you can keyword and SEO for the search engines rather than something that is going to become relatively short lived.Blogging – Overview.
To be perfectly honest with you, all of this talk of creating a blog to get additional traffic is, in itself just ‘hype’ unless you have a source of blog owners to link back to your blog or, an hour or so to kill each day wasting your time writing your blog your time really would be much better served coding HTML pages. After all, HTML pages have worked for the last 30 years, why all of a sudden have webmasters stopped using them?Article written by Lee.
-
Source Code – Eliminating HTML Margins
Since the inception of the very first HTML document, margins existed that surrounded the document, like in the case with this document.
The margins are there to act as a cushion between the edges of the browser and the contents of the page, so it’s easier to read the content.
As good as margins are, wouldn’t it be nice to be able to eliminate them in some cases, where the design of the page calls for their termination? While most webmasters think it is not possible to “flush out” the default margins of a HTML document, it actually is, although only possible in IE 3+, and NS 4+. In this tutorial, I’ll show you how.
Eliminating The Document Margin In IE 3+
To get rid of the document margin in IE 3 and above, so content presses right against the edges of the browser, use the IE exclusive “leftmargin” and “topmargin” attributes:<body leftmargin=”0″ topmargin=”0″>
It’s as simple as that!
Eliminating The Document Margin In NS 4+
To get rid of the document margin in NS 4 and above, use the NS 4 exclusive “marginwidth” and “marginheight” attributes instead:<body marginwidth=”0″ marginheight=”0″>
Putting It All Together.
By putting the two techniques together, we can create a document that flushes out its’ margins in both IE 3+ and NS 4+:<body leftmargin=”0″ topmargin=”0″ marginwidth=”0″ marginheight=”0″>
Try out the above coding on your page, and you’ll see what I mean.
Article written by Lee
-
Why Cant I Get Indexed By The Search Engines?
Unfortunately, this is an all too common question. If it makes you feel any better, you’re not the only one frustrated about the length of time it takes to be indexed, or the many pitfalls involved. It often takes anywhere from two days to as much as six months to be listed on a search engine. For example, last month Excite finally updated its index for the first time since last August! Luckily, Excite is the most extreme case lately, but waiting several weeks to a month can also be extremely frustrating especially when your livelihood depends partly on these search engines.
The Web Position Submitter report will give you current time estimates for each engine so you’ll know what to expect. However, an engine at any time could choose to delay their indexing beyond the “norm” for maintenance or other reasons. On the flip side, you could get lucky and submit just a couple days before an engine does a complete refresh of their database. Therefore, submission times can never be an exact science since we’re all ultimately at the mercy of the engine.
If you’ve submitted your site and have waited the estimated time to be indexed and there’s still no listing, what do you do now?
Here are 16 tips that should help you solve this problem:
1. First, be sure you’re not already indexed but just don’t know it. Unfortunately, none of the major engines are kind enough to e-mail or notify you as to if and when you’ve been indexed.
The method to determine if a page or domain has been indexed varies from one engine to another, and in many cases, it’s difficult to tell for sure. Never assume that you’re not indexed just because you searched for a bunch of keywords and you never came up in the first few pages of results. You could be in there but buried near the bottom.
In addition, it’s not very practical to check the status of a number of pages on each major engine each week. Fortunately, Web Position has a URL verification feature in the Reporter that makes this process much easier. Each time you run a mission, it will report which URLs exist and do not exist in each engine. If you’re using Web Position and are not finding your URLs after submitting, be sure to see this page for common pitfalls to watch out for:
http://www.webposition.com/urlnotfoundhelp.htm
2. Make sure you have uploaded the pages to your site before submitting them. This one seems obvious, but submitting a page that does not exist or submitting with a subtle typo in the URL is a goof we might all make at one time or another. If you’re using Web Position’s Submitter, there’s a checkbox on tab 2 that forces Web Position to verify that all your URLs are valid before submitting them.
3. If you have information inside frames, that can cause problems with submissions. It’s best if you can create non-framed versions of your pages. You should then submit the non-frames versions of your pages which can of course point to your framed Web site. Alternatively, you can enter your relevant text within the NOFRAMES area of a framed page which most search engine spiders will read.
4. Search engine spiders cannot index sites that require any kind of registration or password. A spider cannot fill out a form of any kind. The same rule applies regarding indexing of content from a searchable database, because the spider cannot fill out a form to query that database. The solution is to create static pages that the engines will be able to find.
5. Dynamic pages often block spiders. In fact, any URL containing special symbols like a question mark (?) or an ampersand (&) will be ignored by many engines.
6. Most engines cannot index text that is embedded in graphics. Text that appears in multimedia files (audio and video) cannot be indexed by most engines. Information that is generated by Java applets or in XML coding cannot be indexed by most engines.
7. If your site has a slow connection or the pages are very complex and take a long time to load, it might time out before the spider can index all the text. For the benefit of your visitors and the search engines, limit your page size to less than 60K. In fact, most Webmasters recommend that your page size plus the size of all your graphics should not exceed 50K-70K. If it does, many people on dial up connections will leave before the page fully loads.
8. If you submit just your home page, don’t expect a search engine to travel more than one or two links away from the home page or the page that you submitted. Over time they may venture deeper into your site, but don’t count on it. You’ll often need to submit pages individually that appear further down into your site or have no link from the home page.
9. If your Web site fails to respond when the search engine spider pays a visit, you will not be indexed. Even worse, if you are indexed and they pay a visit when your site is down, you’ll often be removed from their database! Therefore, it pays to have a reliable hosting service that is up 99.5% of the time. However, at some point a spider is going to hit that other 0.5% and end up yanking your pages by mistake. Therefore, it pays to keep a close eye on your listings.
10. If you have ever used any questionable techniques that might be considered an overt attempt at spamming (i.e., excessive repetition of keywords, same color text as background, or other things that the Web Position Page Critic warns you about), an engine may ignore or reject your submissions. If you’re having trouble getting indexed in the expected amount of time, make sure your site is spam-free.
11. If your site contains redirects or meta refresh tags these things can sometimes cause the engines to have trouble indexing your site. Generally they will index the page that it is redirecting TO, but if it thinks you are trying to “trick” the engine by using “cloaking” or IP redirection technology, there’s a chance that it may not index the site at all.
12. If you’re submitting to a directory site like Yahoo, Open Directory, NBCI.com, LookSmart, or others, then a human being will review your site. They must decide the site is of sufficient “quality” before they will list it. I recommend you read the submission guide on the directory tab of the WebPosition Submitter. It contains tips to improve your chances of obtaining a good listing on these directories.
13. A number of engines no longer index pages residing on many common free web hosting services. The common complaint from the engines is that they get too many “junk” or low-quality submissions from free web site domains. Therefore, they often choose not to index anyone from those domains or they limit submissions from them. It’s always best to buy your own domain name (very important) and place it on a respected, paid hosting service to avoid being discriminated against.
14. Some engines have been known to drop pages that cannot be traveled to from the home page. HotBot has been rumored to do this. You may want to consider submitting your home page that links either directly or indirectly to your doorway pages.
15. Make sure you’re submitting within the recommended limits. Some engines do not like more than a certain number of submissions per day for the same domain. If you exceed the limit, you may find that all your submissions are ignored. Fortunately, WebPosition’s submitter will warn you regarding current limits and recommend you stay within them. Some submission consultants feel it is dangerous to submit more than ONE page a day to a engine for a given Web site. For those who wish to be ultra-conservative in their approach, the Web Position Submitter includes a checkbox to limit submissions to one URL per day per engine.
16. Last but not least, sometimes the engines just lose submissions at random through technical errors and bugs. Therefore, some people like to resubmit once or twice a month for good merit in case they do lose a submission. Certainly if you’ve followed all the “rules” and are still not listed, re-submit! Sometimes a little persistence is all that’s needed.
If any of the above scenarios apply to your submission, you should make the necessary adjustments and re-submit. If that still does not work, you should consider e-mailing or calling the search engine and asking them politely why you have not been indexed yet. Sometimes they will reply back with “Sorry, there was a problem with our system and I’ve now made sure you’ll be indexed within the next couple days.” Or, sometimes they’ll tell you why you were not indexed. In other cases, they will ignore your e-mail and you’ll have to keep e-mailing or calling them until they respond. Still, it’s definitely worth the effort to get your site listed with the major engines assuming you also take the time to optimize your pages so you’ll achieve top rankings.
Article written by Lee
-
Saving Time Online – Shortcut Keys
The one thing that i have started to do just lately is use shortcut keys. These are actually very easy to get used to and, will save you a lot of effort over a period of time.
Below you will find a list of the more common shortcut keys in both Windows and Internet Explorer.
Even if you get used to using one or two of these shortcut keys you will notice the time you start save over a week, hell, i don’t even type the whole domain into my IE address bar anymore!
Here are the shortcuts for Windows along with the description of what they do:
Alt + F – File menu options in current program.
Alt + E – Edit options in current program
F1 – Universal Help.
Ctrl + A – Select all text.
Ctrl + X – Cut selected item.
Shift + Del – Cut selected item.
Ctrl + C – Copy selected item.
Ctrl + Ins – Copy selected item
Ctrl + V – Paste
Shift + Ins – Paste
Home – Goes to beginning of current line.
Ctrl + Home – Goes to beginning of document.
End – Goes to end of current line.
Ctrl + End – Goes to end of document.
Shift + Home – Highlights from current position to beginning of line.
Shift + End – Highlights from current position to end of line.
Ctrl + Left – Moves one word to the left at a time.
Ctrl + Right – Moves one word to the right at a time.
Ctrl + Backspace – Delete word to the left of cursor.
Ctrl + Del – Delete word to the right of cursor.
Alt + Tab – Switch between open applications.
Alt + Shift + Tab – Switch backwards between open applications.
Ctrl + Esc – Bring Up start button.
Alt + Esc – Switch Between open applications on taskbar.
F2 – Renames selected Icon
F3 – Starts find from desktop
F4 – Opens the drive selection when browsing.
F5 – Refresh Contents
Alt + F4 – Closes Current open program.
Ctrl + F4 – Closes Window in Program
Alt + Enter – Opens properties window of Selected icon or program.
Shift + F10 – Simulates right click on selected item.
Shift + Del – Delete programs/files without throwing into the recycle bin.
Holding Shift – Boot safe mode or by pass system files.
Holding Shift – When putting in an audio cd will prevent CD Player from playing.The following list are shortcut keys when using Internet Explorer:
Alt + Left – Back a page.
Alt + Right – Forward a page.
F5 – Refresh current page / frame.
Esc – Stop page or download from loading.
Ctrl + Enter – Quickly complete an address.
Ctrl + N – Open New browser window.
Ctrl + P – Print current page / frame.Hopefully this list of shortcut keys will get you saving some time even if its just a few minutes a day. After all, we all know the one thing webmasters need more of is time.
Article written by Lee
-
Straight From The Horses Mouth – Get Googlized
Many webmasters wonder how to ensure their sites will be included in Google’s index of web sites. Although Google crawls more than a billion pages, it’s inevitable some sites will be missed. When Google does miss a site, it’s frequently for one of these reasons:
* The site is not well connected through multiple links to others on the web.
* The site launched after Google’s last crawl was completed.
* The design of the site makes it difficult for Google to effectively crawl its content (excessive frames, tables, etc).Google’s intent is to represent the content of the Internet fairly and accurately. To help make that goal a reality, we offer this guide to building a “crawler-friendly” site. There are no guarantees a site will be found by our crawler, but following these guidelines should increase the probability that your site will show up in Google search results.
Do…
Provide high-quality content on your page – especially your home page.
If you follow only one tip from this page, this should be it. Our crawler indexes web pages by analyzing the content of the pages themselves. Google will index your site better if your pages contain useful information. Plus, your site has a better chance of becoming a favorite among web surfers and being linked to by others if the information it contains is relevant and useful.Submit your site to the appropriate category in a web directory.
Listing your site in the Open Directory Project http://www.dmoz.org/ or Yahoo! http://www.yahoo.com/ increases the likelihood it will be seen by robot crawlers and web surfers.Pay attention to HTML conventions.
Make sure that your <TITLE> and <ALT> tags are accurate and descriptive. Also, check your <A HREF> tags for errors since broken or improperly formatted links can prevent Google from indexing your page.
Make use of the robots.txt file on your web server.
This file tells crawlers which directories can or cannot be crawled. Make sure it’s current for your site so that you don’t accidentally block our crawler. Visit: http://www.robotstxt.org/wc/faq.html for a FAQ answering questions regarding robots and how to control them once they visit your site.Ensure that your site is accessible through HTML hyperlinks.
Generally, your site is crawlable if the pages are connected to each other with ordinary HTML links. If certain areas are not linked, you may be excluding older browsers, differently-abled users, and Google. Google can crawl content from a database or other dynamically generated content as long as it can be found by following links. If you have many unlinked pages, you may want to create a jump page from which the crawler can find all of your pages.Build your site with a logical link structure.
A hierarchical link structure is not only beneficial to you, but also to Google. More of your site can be crawled if it is laid out in with a clear architecture.Don’t…
Fill your page with lists of keywords, attempt to “cloak” pages, or put up “crawler only” pages.
If your site contains pages, links or text that you do not intend visitors to see, Google considers them deceptive and may ignore your site.Feel obligated to purchase a search optimization service.
Some companies “guarantee” your site a place near the top of a results page. While legitimate consulting firms can improve your site’s flow and content, others employ deceptive tactics to try and fool search engines. Be careful – if your domain is affiliated with one of these services, it could be permanently banned from our index, we have found search engine optimization software like Web Position Gold works best but, again use it in moderation.Use images to display important names, content or links.
Our crawler does not recognize text contained in graphics.
Use ALT tags if the main content and key words on your page cannot be formatted in regular HTML.Provide multiple copies of a page under different URLs
Many sites offer text-only or printer-friendly versions of pages that contain the same content as the graphic-enriched version of the page. While Google crawls these pages, duplicates are removed from our index. In order to ensure that we have the desired version of your page, place the other versions in separate directories and use the robots.txt file to block our crawler.Article written by a Google employee
-
Fighting Back Against Computer Viruses
Only 10 years ago, the only way that a virus could have infected your computer was through you physically putting a floppy disk into your hard drive and running the program file that initiated the virus however, today’s viruses have become more complex in how they operate and, can spread like wild fire if they are not recognized as a virus doing damage not only to your computer but that of your associates and peers.
So how do we minimize the risk of virus infection on our computers which, lets face it, for the Adult Webmaster are our livelihood and, need to be protected at all cost. This is what we will take a brief look at in this article, how to minimize the risk of getting a computer virus and, heaven forbid, how to deal with a virus if you actually have one on your machine already.
Computer Viruses – A Little History.
1981 – The first widely accepted computer virus is recognized. The ‘Elk Cloner’ virus was spread through floppy disks and, whilst by today’s standards, it was not a high risk virus it, non the less was still a virus, in fact, all this first virus did was to display a message on your computer screen.1991 – The ‘Tequila’ virus was the first recognized ‘polymorphic virus’ meaning, it was the first virus which actually altered its state once on your machine in order to remain undetected.
2001 – 2001 saw the spawn of several highly potent ‘worm’ orientated viruses from the ‘Gnuman’ virus, the ‘LogoLogic Worm’ and the famed (is that the right word?) ‘Nimda worm’. These ‘worms’ were the first of their kind to actually spread themselves across the World Wide Web through infecting users email clients (specifically Outlook Express) and sending itself to every email address the infected machine had in the address book.
Virus Infections – Prevention Better Than Cure.
So we now have a little background on the main viruses over the last 20 years we will take a look at the ways in which we can prevent viral infections on our computers after all, as the subtitle for this paragraph says, a prevention is far better than having to find a cure.Here are a few simple ways you can help to prevent the spread and infection of computer viruses.
1) Install a good anti-virus software package on ALL of your computers, Norton and, Mcafee, both have some good products to aid you in preventing viral infections of your PC. Also, you should ensure the software package you do install has updated virus definitions, a new virus is created almost every week so you need to ensure your software keeps up-to-date on the latest infections.
2) Scan any and all email attachments you receive, irrespective of whether you know the sender or not, anything that you get sent in an email should be scanned, as mentioned above, there are complex viruses that can ‘send’ themselves from other peoples email address book, what happens if you are on the address book of an infected users computer? Can you afford not to be online?
3) Ensure that, once you have an Anti virus program installed on your machine that you have it setup to start each and every time that your computer boots, the first thing that should be loading on your machine is an anti-virus program, you only have one chance to prevent infection.
4) Avoid downloading .exe and .com files to your desktop unless, you know the person you are downloading the files from and, even then, before you launch or open the files ensure you scan them independently.
Virus Prevention – Overview.
Hopefully this article will have given you some insight into how you can protect both your computer and the computers of your associates and peers from computer viruses and, hopefully, by following the guidelines within this article your time spent online will be productive and virus free. One thing you should remember however is that no matter how hard we all try to remain virus free, someone will create a new program that ultimately, could infect our machines no matter how hard we try.Article written by Lee
-
Mirroring Adult Sites – Stage Four
We are now going to take our basic template and start to make some pages for the link lists for this you will need to open up your original warning page. Once you have this page open you need to visit a couple of the link lists that you use. Go to the submissions page of the link list and download their reciprocal buttons.
In the same way that we did with the TGP pages we created you now need to edit the warning page you have open and insert the recip links on your warning page, ideally and, as most link lists will require this, the recip links should be placed above the enter and exit links.
Once you have done this, you now need to save your newly edited page as linklist1.html in the /FreeSite/LinkList/ folder, do this for as many of the link lists that you wish to submit to, each time saving the new page in the /FreeSite/LinkList/ folder.
Also at this stage, we should add a console to our link list page, ideally a blur console with links going to your surfer trap FPA’s. You can find the Java coding to create a blur console on the tutorial about building a surfer trap.
Once we have done this, we now have some more sites created. However, this time, the page/s we have just created are designed to build the traffic we have to our site/s.
Once you have these pages saved you now need to re-open your original warning page once more.
This time, we are going to optimize the page for the search engines, again, in the same way we added our ALT tags on the surfer trap we are going to add them to the warning page however, we are also going to optimize our keywords, description and, site title.
Once we have optimized our original warning page we should now save this as se1.html in the /FreeSite/Engine/ folder we created at the start of this tutorial. Once one SE page has been created, you need to create a selection of others making sure that you swap around the meta tags to ensure they are all different to each other, I would suggest making three variations of these SE pages each time saving them in the /FreeSite/Engine/ folder.
The next thing we are going to do is to ‘dirty’ these SE pages up some more, pretty much in the same manner as we did with the surfer trap tutorial. What we need to do is add a small pop up console to the three SE pages along with a blur console, these consoles can be the same ones that we used for our surfer trap or, we can create new ones however, I think the smart thing to do would be if we used the ones we currently have online. So once we have added these consoles we need to save the pages.
Also, to these SE pages we should add our Banner exchange code, as these will be going into the search engines there are no rules as to what we can and can not do on these SE pages so we should make sure that even if they just get one hit, we get additional traffic from them. By adding a banner exchange code as we did on the surfer tutorial we can drive additional traffic either to our sites or to our surfer trap, where you send this traffic is your choice. If you want to sign up for our new banner exchange you can find it at http://www.pornclient.com.
Now we have the completed site template built. We need to submit the pages and sites we have just completed to the TGP’s, AVS’s, Link Lists and, the Search Engines.
Once we have submitted these sites all that we need to do to create a new set of sites is to find another set of 50 images, rename the images as we did at the start of this tutorial, pic1.jpg, thumb1.jpg etc and simply upload the site to our server without the full size images and the thumbnails. Once the HTML pages are online, we now have to upload the full size images and the thumbnails and we have a set of sites using different content to our original one built in less than 5 minutes.
The best thing with these sites is that if a any time we want to alter the layout, all we need to do is edit the HTML of the version we have saved on our HD and we have a new template ready to use.
I hope you have found some use out of this tutorial and, I am sure you can think of other ways in which we can use templates of this style, this tutorial was just detailing some of the basics however, we can make the template as simple or as complicated as we like. Our only limits are what we make for ourselves.
Article written by Lee
-
The Death Of The AVS – The Birth Of The AEN (Adult Entertainment Network)
Well after a long and tedious few months of speculation, rumors and facts being presented to those of us in the adult industry working on AEN sites (Formerly known as AVS sites) it seems that Visa has put a strain on AEN companies to start re-working and, renaming their sites to fall under the ‘Adult Entertainment Network’ name. What does this mean to us as webmasters and, more importantly to the industry as a whole? This is what we shall address in this article.
A Little Visa History Lesson.
This whole mess started in October of 2002 when, jointly, CCBill, Epoch and, iBill issued a new set of operating guidelines that had been passed onto them by Visa. In effect, this meant that 3rd party credit card processors had to be re-classified as an ‘IPSP’ or ‘Internet Payment Service Provider’ this also in turn meant that webmasters needed to start paying ‘fees’ to Visa (via their IPSP) in order to continue charging their surfers for access to products and services by using Visa issued credit and / or debit cards.In addition to the initial $750 however, there is also a need to pay an ongoing fee of $375, two thirds of this fee goes directly to Visa and, the remaining balance goes to issuing banks and IPSP’s for administrative charges.
Shortly after this announcement was made the industry then got dealt another blow by the processing companies, this time, in the form of Paypal announcing they would no longer accept any adult orientated payment through their system. This went into effect on May 12th 2003 and, after initial speculations as to what could and could not be charged to Paypal accounts, it seemed to be the final nail in the coffin for some of the smaller webmasters.
However, the adult industry is strong and rallied together getting as much information as possible about alternates to both of the above situations and companies such as Stormpay and off-shore IPSP’s started to see an increase in revenues from adult payments.
As The Waves Settled – The Next Blow.
Given all of the financial troubles of the previous 10 months in respect of online adult payments the industry as a whole started to settle itself again, of course there was the usual posting on forums, conversion ratio and shaving discussions happening on the main forums as there always was but, in addition to these conversations more rumors started, this time, regarding the AVS business model.Almost as quickly as these rumors started however, they seemed to have fizzled out into nothing but a distant memory but, two months later after the first round of initial conversations on the adult industry message forums, the same AVS business model was dealt another devastating blow.
It seems Visa has its sights set on AVS (Age Verification Service / Systems) with a press release being issued first by Sex Key, then by Free Age Card, which was followed by similar announcements from Global Male Pass and Gay Passport informing adult webmasters that they had 2 days to make their current AVS sites conform to the new requirements set down by Visa.
These requirements were that.
1) Adult webmasters can no longer use the term ‘AVS’ on their sites.
2) Adult webmasters can no longer use the term ‘Adult Verification Service’ or ‘Age Verification Service’ on their sites.
3) Adult webmasters can no longer state that surfers joining up to their AVS will have ‘Access to thousands of other AVS protected sites’.As usual Adult Webmasters started to question the reasons behind this sudden and apparent change of AVS rules on almost every forum and, whilst most questions went unanswered the one thing that was answered (honestly) was that these new rules had been set down by current AVS IPSP’s who, had had these rules issued tot hem by Visa.
So, webmasters started to change out their tour page text to conform to these new rules and, for the most part, it seemed like that was all that was needed to be ‘safe’ working the AVS business model. Little did they know…
One Last Blow To The AVS Webmasters.
Several days after the initial press releases from some of the top AEN (Formerly AVS) companies was made it seems that in between all of the posting ont he boards yet another issues comes to light – That Visa is aiming to have AVS webmasters now pay a $750 fee for the benefit of being able to keep using this business model as stated by Craig Tant (Sex Key) ‘we believe is that every site will have to be registered and pay the $750.’This beleif, although somewhat vague tends to draw conclusions that, through no fault of their own, the AVS / AEN business model in the adult industry as we currently know it is changing and, changing for the worse. If webmasters are forced to pay Visa a $750 registeration fee per AEN system or per AEN site then a lot of todays industry leaders are going to come unstuck, especially those that focus solely on the AVS / AEN model for the primary source of income.
Adult Entertainment Networks – What’s Next?
This is the question that is currently being asked by many webmasters and, the truth of the matter is simply that nobody knows – Except for Visa, the next few days, weeks and months are going to be telling on the industry as a whole from what happens to current sites inside the AVS / AEN companies websites to what happens when surfers start to propagate back onto the adult internet looking for their fix of high quality porn, two things however are definitive throughout all of this, Visa is homing in on the adult industry and, every webmaster needs to be concerned about what will happen next.Article written by Lee
Premium Sponsors
Categories
- 2257
- Billing Solutions
- Blogging
- Branding
- Content
- Domain Names
- Employment
- Forms & Contracts
- General
- Hosting
- Link Lists
- Opt-in Mail
- Paid Traffic
- Pic Posts
- Promotion
- Scripts
- Search Engine Optimization
- Sponsors
- TGP
- Traffic
- Tutorials
- Viral Marketing
- WebDesign
- Writing