Online note taking app - Notes.ionotes

Fast | Easy | Short

Online Note Services -

Change Engineering Search Engine Ranking Algorithms
Back inside 1997 Used to do some research so that they can reverse-engineer algorithms used by lookup engines. In of which year, the huge ones included AltaVista, Webcralwer, Lycos, Infoseek, and a very few others.

I had been able to mostly declare my research a success. Within fact, it absolutely was so accurate that in one case I used to be able to create a program that produced the same look for results as one of the search engines like google. This article explains could did it, and how its still beneficial today.

Step 1: Decide Rankable Traits

The first thing to perform is make a list of what a person want to calculate. I came way up with about fifteen different possible methods to rank an online page. They integrated things like:

-- keywords in name

- keyword denseness

- keyword regularity

- keyword within header

- keyword in ALT labels

- keyword focus (bold, strong, italics)

- keyword inside body

- keyword in url

instructions keyword in domain or sub-domain

: criteria by location (density in name, header, body, or even tail) etc

Phase 2: Invent the New Keyword

The 2nd step is to be able to determine which keyword to try with. Typically the key is to decide on a word that will does not are present in any terminology on the planet. Otherwise, you will not be able to isolate your variables for this kind of study.

I used to work at an organization called Interactive Visuallization, and our internet site was Riddler. com along with the Commonwealth System. At the time, Riddler was the largest entertainment web site, in addition to CWN was among the top trafficked sites on the net (in the most notable 3). I looked to our co-worker Carol and mentioned I needed a new fake word. The lady gave me "oofness". I did a quick search and it also was not found on any search motor.

Note that an unique word can in addition be used to determine who has ripped content from your own web sites onto their own. Given that all my check pages are absent (for several years now), a search on the search engines shows some sites that did backup my pages.

Step 3: Create Test Internet pages

The next thing to do was to create test pages. I took the home page regarding my now defunct Amiga search engine "Amicrawler. com" and even made about seventy-five copies of that. Then i numbered each file 1. html code, installment payments on your html... seventy five. html.

For every single measurement criteria, I made at least 3 or more html files. Regarding example, to calculate keyword density within title, I customized the html titles of the first 3 files in order to look such as this:

one. html:

installment payments on your code:

3. html:

Typically the html files involving course contained the rest of my home page. I then logged in my notebook that files 1 instructions 3 were key word density in subject files.

I recurring this type involving html editing regarding about 75 or so files, right up until I had every criteria covered. The particular files where then uploaded to the web server and even placed in a similar directoty so that search engines like yahoo can locate them.

Step four: Wait for Search Engines to Index Check Webpages

Over the particular next few days, many of the webpages started appearing within search engines. Nevertheless a site want AltaVista might only show 2 or perhaps 3 pages. Infoseek / Ultraseek at that time was doing real time indexing so I have got to test everything instantly. In some situations, I had to hold back a few several weeks or months with regard to the pages to have indexed.

Simply typing the keyword "oofness" would bring up all pages found that had that keyword, in the order ranked by simply the search engine. Since only my pages contained that will word, I would not have competing pages to confound me.

Step five: Study Results

To my surprise, many search engines had very poor rank methodology. Webcrawler utilized an easy word thickness scoring system. Throughout fact, I got in a position to write the program that gave the very same search engine results as Webcrawler. That's right, merely give it a list of 12 urls, and this will rank these people in the exact same same order while Webcrawler. Employing this plan I would make any of my pages rank #1 basically wanted in order to. Problem is obviously that Webcrawler failed to generate any targeted visitors even if I actually was listed amount 1, so My partner and i would not bother along with it.

AltaVista replied best most abundant in number of keywords within the title of typically the html. It placed a few pages way at the end, but I actually don't recall which criteria performed worst. Along with the rest involving the pages graded somewhere in the middle. Overall, AltaVista only cared concerning keywords within the title. Everything else didn't seem to issue.

A couple of years later, I repeated this check with AltaVista plus found it was offering high preference to be able to domain names. Thus i added a wildcard to my DNS and web server, make keywords within the sub-domain. Voila! All of my personal pages had #1 ranking for any kind of keyword I select. This obviously directed to one trouble... Competiting web web sites don't like losing their top positions and will carry out anything to safeguard their very own rankings because it fees them traffic.

Some other Methods of Tests Search Engines

My partner and i is going to be able to quickly list many other things that may be done to be able to test search engines like yahoo codes. But these are all lengthy topics to discuss.

I tested a few search engines simply by uploading large copies of the dictionary, and redirecting any traffic to a secure web page. I also tried them by indexing massive quantities associated with documents (in the millions) under numerous domain names. My partner and i found in general of which there are quite few magic keywords found in the majority of documents. The fact still remains that will a few key word search times prefer "sex", "britney spears", etc introduced visitors but most usually do not. Hence, most webpages never saw any people traffic.


Unfortunately t here have been some drawbacks in order to getting listed #1 for a whole lot of keywords. I actually found that this ticked off some sort of lot of people who competing web sites. They can usually start by burning my winning strategy (like placing keywords in the sub-domain), and then repeat the process themselves, in addition to flood the search engines with a hundred times more web pages than the one page I acquired made. It manufactured it worthless to be able to compete for prime keywords.

And 2nd, certain data are unable to be measured. You can utilize tools like Alexa to determine traffic or Google's internet site: domain. com to be able to find out the amount of listings a domain has, but until you have a whole lot of this files to measure, you'll not get any useable readings. What excellent is it intended for you to try out and beat a new major web site for the major key word when they already need millions of visitors per day, a person don't, in fact it is portion of the search engine ranking?

Bandwidth and resources could become a problem. I have had website sites where 73% of my traffic was search powerplant spiders. And that they slammed my sites every second associated with every day for years. I would virtually get 30, 1000 hits from the Google spider every single day, in improvement to other bots. And unlike just what THEY believe, these people aren't as friendly as they claim.

Another drawback is that in case you are doing this for the corporate web web site, it might not really look so very good.

For instance , you might recall a few weeks ago any time Google was captured using shadow web pages, and of study course claimed they have been only "test" pages. Right. Does Search engines have no dev servers? No staging servers? Are they smart enough in order to make shadow internet pages hidden from normal users but is not good enough to cover dev or test webpages from normal customers? Have they certainly not figured out exactly how an URL or IP filter performs? Those pages have got to have served a new purpose, and these people didn't want many people to understand this. Maybe we were holding just weather balloon webpages?

I recall obtaining some pages that will were placed by a hot online as well as print tech magazine (that wired people into the electronic digital world) on research engines. They'd placed numerous blank landing pages using typeface colors matching the particular background, which covered large quantities associated with keywords for their biggest competitor. Perhaps that they wanted to pay out digital homage in order to CNET? Again, it was probably back found in 1998. In reality, they were operating articles at the time about how this is wrong in an attempt to trick search motors, yet they were doing it them selves.


While this specific methodology is great for learning a few things about search engines, generally speaking I actually would not recommend making this typically the basis to your internet site promotion. The amount of pages to remain competitive against, the quality of your website visitors, the shoot-first mentality involving search engines, and many other factors will provide evidence that there are far better methods to do net site promotion.

This particular methodology works extremely well intended for reverse engineering some other products. For example , any time I worked with Agency. com carrying out stats, we used a product manufactured by a serious small software company (you might be using one of their fine os products right now) to analyze web site server logs. The particular problem is that this took more as opposed to the way 24 hours to evaluate 1 days worthy of of logs, therefore it was never ever up to time. A little little of magic plus a little bit of perl had been able to produce the same reports in 45 minutes simply simply by feeding a similar wood logs into both methods until the effects came out the particular same and every situation was accounted for.
My Website:
what is is a web-based application for taking notes. You can take your notes and share with others people. If you like taking long notes, is designed for you. To date, over 8,000,000,000 notes created and continuing...


  • * You can take a note from anywhere and any device with internet connection.
  • * You can share the notes in social platforms (YouTube, Facebook, Twitter, instagram etc.).
  • * You can quickly share your contents without website, blog and e-mail.
  • * You don't need to create any Account to share a note. As you wish you can use quick, easy and best shortened notes with sms, websites, e-mail, or messaging services (WhatsApp, iMessage, Telegram, Signal).
  • * has fabulous infrastructure design for a short link and allows you to share the note as an easy and understandable link.

Fast: is built for speed and performance. You can take a notes quickly and browse your archive.

Easy: doesn’t require installation. Just write and share note!

Short:’s url just 8 character. You’ll get shorten link of your note when you want to share. (Ex: )

Free: works for 12 years and has been free since the day it was started.

You immediately create your first note and start sharing with the ones you wish. If you want to contact us, you can use the following communication channels;

Email: [email protected]




Regards; Team

Shortened Note Link
Looding Image
Long File

For written notes was greater than 18KB Unable to shorten.

To be smaller than 18KB, please organize your notes, or sign in.