
 URL Gen vs 3.0 (ssdl). Copyrighted FREEware. Update 1st January 2002.
 Copyright All Rights Reserved June 1998/2002.
 Designed, created and programmed by Charles Peter White,
 26, Oakdene, Stourport-On-Severn, Worcestershire, DY13 9NF, England. UK.

 This program is dedicated to my wife Pauline, with a special dedication to
 our best friend David B. Bransfield, who died suddenly on the 10th of
 February 2000. UK.

 Information.
 ============
 This version 3.0 is FREE to everyone! No Adware or any form of advertising!
 This program cannot be sold, hired or distributed in any financial form,
 that includes exchanges in Money, Barter or anything else!
 Reverse engineering is strictly forbidden!! This program is copyrighted and
 All Rights are Reserved by me. This version contains minor enhancements and
 ERROR corrections.
 Please read this document for further details, or see update information.

 Legal Stuff.
 ============
 Every effort has been made to ensure that 'URL Generator' is Bug free and
 capable of doing the tasks as described. However, the burden of responsibilty
 must remain yours if you choose to use this program in any form. No one may
 distribute this program as part of any commercial package/product, or part
 of anything to be sold or hired! The created 'Site Search' is for personal
 use ONLY! By reading this documentation and by running this program, you
 agree to accept the terms and assume all liabilities, irrespective.

 What does UrlGen do? READ THIS PLEASE !!
 --------------------====================
 Almost every selection is on the screen at ALL times. The design, colour and
 layout are all to ensure your concentration for the best results possible.
 It will extract various forms of URL (HYPER) LINKS from your ASCII HTML
 files. It will allow creation of LARGE HTML and Java 'Site Searches', under
 certain conditions. It can extract and SORT (in alphabetical order) to
 various file formats, including the ability to extract individual files and
 add to a single file. Your Master files must NOT have the extension of .HTL
 or A0-Z9.HTM, as this program uses those to extract to. DELETE: THIS CAN
 DELETE YOUR MASTERS. OPERATE WITH CARE!! You can re-sort extracted link files
 in either Java 'Site Search' format or through normal HTML extraction type.
 Site Search/Java to HTML format, even COMPRESS Java ARRAY Site Search
 Data (or de-compress) and compress the FULL Site Search system.  This
 program can create a Java 'Site Search' with all the necessary code to
 execute in your Web Site or other media types, and REMOVE Duplicate URLS.
 The code within the Site Search will refer to URL LINK addresses and to a
 pages within your Web Site or other media (depending on what you have in-
 your pages). This gives your viewers the ability to search and 'go to'
 almost anywhere ! Normal HTML extractions can be compressed as well,
 removing any Carriage Returns and Form Feeds, or de-compressed. You can use
 other compressors to complete the task (HTML- Shrinker is good). UrlGen
 allows Single or Multiple selections with various setup extraction types,
 all of which can be saved at any time. EXTRACTS JAVASCRIPT format.
 various HTML facilities. 'Cross Index' TABLE placement. URL - JAVASCRIPT
 'ERROR' LINKS now sent to a file for inspection, which now allows all file
 extractions to be generally clean ready for use. Lots more features.
                                                                  Page 1.


                                    INDEX.
                                    ======
          Page....................Description........................
            1......Introduction and Legal notice & What URL GEN does.
            3......System requirements.
            3......Program Keys, Mouse and other devices usage.
            4......Configuration files of CFG and LOG.
            5......Default settings. The Files and Directory Window.
            5......Viewing an ASCII File.
            6......What are SSDLAPP.HTL and URLGEN.HTM special files.
            6......DELETING Files. Extracting LINKS to single files.
            7......Tagging - Selecting. Extracting a series of files.
            7......Extracting files to APPEND to SSDLAPP file.
            7......Multiple Tag file APPEND to SSDLAPP file.
            7......Single file APPEND to SSDLAPP file.
            8......'A to Z' alphabetical extractions.
            8......'Cross Index' page linking/Table system.
            8......'Site Search': 'Normal' or 'Big File' feature.
            8......Creating a final URLGEN.HTM special file.
            9......What type of URL LINKS (Hyper links).
            9......C G I, Click type Traders and Java Links.
            9......Extract Mailto, personal files and Strip HTTP://.
            9......Installation.
            10.....How do I use it ?.
            10.....ESCape Key. F = Files. Q = Quit.
            10.....General Function Key descriptions.
            11.....Remove or Keep <b><font> type HTML commands.
            12.....EMPTY. Change error urls to 'Web Site address'.
            12.....OTHER. Change all other urls to 'Web Site address'.
            12.....BGPROPERTIES settings for 'Full Site Search'.
            12.....TABLE A - Z settings for 'FULL Site Search'.
            12.....My first extraction. (It is painless).
            13.....Create and Append to SSDLAPP file.
            13.....Best method of extracting. (use this always).
            13.....Re-sorting/Extracting from a created files.
            14.....Creating your final APPENDED HTML-Site Search.
            14.....ADDING to an already created Site Search file.
            14.....Site Search creation. The Best method.
            15.....Altering your created 'Site Search' Code.
            15.....Adding User SYNTAX checking for Site Search.
            16.....TARGET PAGE requirements.  FULL Site Search.
                                              HTML Data files.
            16.....REMOVE DUPLICATE URL (Web address).
            17.....RIP OUT unwanted characters from descriptions.
                   From FRONT and END.
            18.....MID RIP unwanted characters From within the text.
            19.....SYMBOLS on file/directory window.
            19.....Single file ERROR messages.
            19.....Javascript extractions and Links.
            20.....PROBLEM SOLVING.
            23.....URLERR.TXT bad links file.
            24.....EXTRACTING NORMAL HTTP:// with or without CGI,IMG etc.
            24.....Author's comments and Site details.
            UPDATE information, see update.txt file included.

                                                                  Page 2.
 PC System requirements:-
 System : 80 by 25 screen: MSDOS 3.3 onwards, Windows 3x,9x,NT,2000,ME.
 Memory required   : Anything above 640k. SYSTEM: Composit or Colour.
 Original File size: 62,400 bytes or very large for Site Searches.
 Created file size : Depends. Appended to SSDLAPP.HTL: 9 Mega Bytes +.
 Mode of entry     : Keyboard and or Microsoft emulation Mice / Tablets etc.
 Execution         : URLGEN.EXE to enter the system.
 Configuration file: LOG holds RIP settings, while CFG is Mid Rip/Program.
 NOTE. Depending on method, the configuration files will be written to the
 exited directory. Windows: Create a PIF file, it will write to site stated.

 Program Keys / Mouse Usage:-
 F1. Remove or Keep <B><FONT> type HTML commands within displayed links.
 F2. 'Cross Index' system's TABLE 'Left', 'Center' or 'Center with border'.
 F3. Select extraction method, HTML or 'JAVA' Format.
 F4. Java or HTML. HTML NORMAL, A to Z, Index 'AS IS' or 'Full' creations.
 F5. Extract Links to append to SSDLAPP.HTL file.
 F7. My copyright and your information.
 F8. Compress/Decompress HTML/Java data or 'Site Search' systems.
 F9. Selection system for various Link extraction methods.
 F10. Sort and/or Create URLGEN.HTM final 'FULL Site Search'.
      With 'Site Search' creation on BIG files.
 ENTER. Select current high-lighted or Tagged files and extract and sort
        to the same named file or files. Tagging prompts User for selection.
 CTRL/T. Select or De-select tagging Files.
 CTRL/S. Save settings to file named URLGEN.CFG. No display, just does it.
 TAB key. Switch 'Keyboard use' between 'MID RIP' window and Main program.
 CURSOR Left - Right: Set depth of Full Site Search A-Z 'base' selector bar.
 /. A-Z File creation output file type. ie; A0 - Z0 to A9 - Z9. 0-9 max.
 B. Switch ON - OFF 'BgProperties' command in Full Site Search creation.
 D. Delete items. WARNING..CAN ALSO DELETE MASTER FILES, TAKE CARE!!
 E. Extract Everything, excluding Java or Files, Mail To: etc. 3 selections.
 F. Select file type for directory display. Max files per directory: 3999.
 H. Remove Duplicate LINK addresses on creation of HTML DATA.
 J. Remove Duplicate LINK addresses on creation of JAVA DATA types.
 M. Select characters to Mid RIP, Starting & Ending with alpha[Link display].
 0. OTHER: Put Web address in displayed link in ALL non error URLs.
 P. TARGET Page assignments for both HTML Links and FULL 'Site Search'.
 R. RIP characters from 'Start' and 'End' of displayed links, moving inwards.
 T. or 'Space Bar'.  'Tag' or 'Untag' a file or a series of files.
 Y. EMPTY Error URLs: Place Web address in displayed selection.

 Page Up / Down, Home, End and Cursor keys = Movement through directory
 window and the ASCII File VIEW facility.
 Specific Keys for 'Site Search'....
 1. Increase counter by ten. (+) Key will increment by one hundred.
 2. Decrease the counter by ten. Minus (-) Key will decrease by one hundred.
 3. Increase by one. 4. Decrease by one.
 0. Set Array counter to one. Automatic if creating FULL Site Search.

 Mouse or Pointing device:-
 Defaults : Left = high-light a file, double click = Tag file.
 Right = ENTER (to select current file or select Change directory).
 To save Settings: Mouse selects the top right box. [] Ascii 254.
 SWAP Button: Select 'Mouse = Left', either button. Actions will be reversed.
 Page Up/Down, Home, End, In / De-creasing 'Site Search' numbers all require
 single press of the button for one action or hold down to allow continuous
 selections. If some devices have problems with second button, select [ENTER].
                                                                  Page 3.
 Configuration Files CFG and LOG.
 ================================
 THIS IS IMPORTANT TO YOU. PLEASE READ !
 ----------------------------------------
 URLGEN.CFG.
 -------------
 It is supplied already set up for general usage, setting all the functions
 that you see on the screen. If it is not available on the directory you
 want to operate on, built-in DEFAULT will be used and those are saved.
 As you alter things and continue to create other masterpieces, the CFG file
 will be updated with those settings.
 DO NOT attempt to use OLD versions of CFG. They will not operate !

 When you exit the program, depending on the option you choose, CFG
 will be written to that directory. You can Exit to either the current
 directory you are working on, or to the original from where this program
 was called from. (DOS users).
 The CFG file needs no alteration from you as it is done for you.

 URLGEN.LOG. (See RIP feature on page 17.) Maximum 100 characters.
 -------------
 This is a very special ASCII file which you can alter to suit your needs.
 This RIPS unwanted characters from descriptive URL Links. These characters
 are the ones that will be RIPPED out from the Beginning and End of the
 description lines. A DEFAULT is supplied within. If you have <<< word >>>,
 this will be ripped, as it is considered to be an HTML command. DO NOT use
 this form of syntax to highlight text. Some LINKS have Asterix and other
 characters trying to get the LINK noticed. Something like this...

 * * * * * * * Get a fantastic URL LINKS EXTRACTOR from me * * * * * *

 In this case you would put the Asterix character in the LOG file. When RIP
 is switched on, it will remove HTML commands like <B> etc and then look for
 the characters in the RIP LOG file. If those characters are found,it will
 RIP them out until it gets the first sensible Alphabetical character (A-Z or
 0 to 9). It does this from BOTH ENDS. The result will look like this..

 Get a fantastic URL LINKS EXTRACTOR from me

 and displayed just as you see. Not only does it look better, it saves a lot
 of room, assisting you to put even MORE links into your finished files.

  *** YOU SHOULD USE THIS FEATURE IF YOU WANT THINGS SORTED CORRECTLY. ***

 This program has one reserved character, a 'space' ascii 32, hex 20 If you
 add these to the LOG file, it will be ignored. Use ONLY what is needed.
 DO NOT ADD YOUR LANGUAGE ALPHABETICAL CHARACTERS LIKE A TO Z in either Upper
 and or Lower case. If you do that, you could end up without descriptions
 altogether. Now that's a novelty. OK, you can always remove them from the
 file and away you go again.

 NON ALPHABETICAL DESCRIPTIONS. (Dead links).
 ==============================
 If Descriptions only contain NON-Alphabetical characters or are empty, this
 program will place A SINGLE exclamation mark or leave the offending 'non'
 alpha character to indicate 'Empty' or 'Problem' description.
 These links are directed to URLERR.TXT file for inspection.
                                                                  Page 4.
 DEFAULT SETTINGS.
 =================
 All 'Selection' displays remain in the program colour until you change from
 the 'Default' settings. This gives you an instant view of what has changed.

 File and Directory Window. (Sorted in Ascending order).
 =======================================================
 The directories are marked with a block on the far left. The name next
 to that block is a directory. If you select a named directory you will
 be taken to that. There may be other directories listed after selecting.
 These are know as the 'Child'. Choosing one of them moves down the tree.
 A single DOT indicates the ROOT. Selecting this will take you there. A double
 DOT indicates going back one directory, or 'Back up the TREE'. Nothing too
 complicated, keep selecting until you get the idea.

 When selecting the actions required (far right) you press ENTER to complete
 the task. ENTER is the second button (default) on the mouse/pointing device.

 To create your final file (URLGEN.HTM), you must create the APPENDED file
 called SSDLAPP.HTL, from which URLGEN.HTM is created.
 Having selected the required settings on the far right, you press F10,
 whereupon the file will be created.
 No selections in the Directory / File window will apply when creating your
 masterpiece.
 There is no facility to change drives. The idea is to execute this program
 on the device required. It is therefore important to place it on any
 device that your system refers to, or from within the PATH command.
 If your media is completely empty, then URLGEN will just display and exit.

 To View a file (ASCII).
 =======================
 Highlight a file and select 'V' to View.
 Normal cursor key directions apply allowing you to move through the file.
 Viewing was created to allow you to examine the files created using the
 extraction method you selected. This allows alterations until you get the
 result needed.

 Pointing device highlights the file required and then selects 'View' from
 the bottom right window. Moving around the display, selects PageUp / Down,
 Home, End and the lines above and below the displayed text.

 Search Directory. ( A 'Word' Search and not the 'File' entry system.)
 =================
 This facility has been REMOVED.

 SSDLAPP.HTL and URLGEN.HTM (special files).
 =========================================
 These two files will be added to the bottom of the directory listing,
 providing there is more than one file available.
 They are created if you want to extract from your files and create
 a single large file with all, or some of those extractions.

 SSDLAPP.HTL.
 -----------
 This is the file that extractions are added (appended) to, allowing
 you to extract from, and create to your final URLGEN.HTM file, if required.
                                                                  Page 5.
 It is best to extract to single .HTL files and then TAG them to append to
 create this file. Why? Because you get First Characters in Uppercase and
 all the necessary extraction selections carried out. If your Master HTML
 URL descriptions contain leading NON 'Alpha' characters, switch RIP 'ON'.
 It can be created by selecting F5 key, switching the APPEND feature ON.
 All files will be appended to it until you switch the facility OFF.

 If the format is in the form of Java/'Site Search' data, you CAN extract
 and re-sort in HTML format as well as the other way round !
 If this file is too large to pass through URLGEN, then you can use the
 utility SSDLAPP3.EXE. You must have ssdlapp.htl in HTML format to use this
 utility. Before you use it, please read the document.

 Compress feature is allowed on SSDLAPP.HTL. You MUST de-compress it first if
 creating URLGEN.HTM. Compress can be used when creating single extractions
 and on the final URLGEN.HTM file.

 URLGEN.HTM.
 -----------
 This file is created after you have completed the task of extracting your
 links to SSDLAPP file. When ready, you select the type of creation method
 with F8 if compress is wanted and then press the F10 key to complete the
 task. If this file has previously been created, you will be informed and
 allowed to either keep it or re-write it with the new information.
 If you want to keep it, I suggest that you exit the program and re-name it
 with the .HTM or HTML extension.
 The FULL Site Search, URLGEN.HTM will contain all the operating codes and
 the links ready to RUN.

 Creating a normal HTML source file into URLGEN.HTM is not allowed, as you
 can convert SSDLAPP.HTL or any other major files just at a flick of F3.
 The resulting Java 'Site Search' can not be converted into normal HTML.
 To do this, you will have to edit out the Java code leaving the array
 lines.

 DELETE FILES. (IMPORTANT!!!).
 -------------
 You can delete single/Tagged files as and when the need arises.
 Make sure you are not on a high-lighted 'Void/Empty' file. This was done
 to make sure that you KNOW EXACTLY what proceedure you are asking for.
 Just select or 'Tag' the files and press the 'K' key, follow prompt. You
 are allowed to delete ANY file, so GREAT CARE is needed.
 In Windows, if you need to undelete, use a utility called 'BigBin', it's
 FREE and available from good Freeware resources.

 Extracting to Single Files.
 ===========================
 Highlight a file, select and set the settings on the far right window and
 then press ENTER. Make sure that APPEND (F5) is switched OFF.
 The file will be loaded, extracted and sorted and written to a file of the
 same name but with the extension of .HTL.
 It will be listed in the directory as soon as it is created.
 There will be letters (and maybe symbols) listed on the far right of the
 highlighted file. Just examine the ERROR Table in this manual to see what
 they mean. Better still, why not print the page for easy reference.

 If you view the created .HTL and it is extracted in the wrong format,
 or has the wrong numbering system for Site Search Data, you can do one
                                                                  Page 6.
 of two things.
 Re-extract from your master file with the correct settings or select the
 created .HTL File. If that file is in the Site Search data format, you can
 change the numbering system by selecting 1, 2, 3, 4, 0, + or - and press
 ENTER. You CAN NOW create an HTML file from that format !
 If the .HTL file is in HTML format, then you can extract to a Site Search
 format, or re-extract it to itself. Useful if you want to alter things on
 then fly. Each file can be re-sorted using the same format.

 Tagging / Selecting (Multiple to same named .HTL extractions).
 ==============================================================
 Tagged files are not maintained when changing directory or file spec.
 There was no need for this, as one hopes that tasks would be completed
 before using that facility. If you want to do a complete series of files,
 just select them with either the SPACE or 'T' key. MOUSE just double clicks
 on button 2. A block will appear on the right side of each file name.
 When all files have been selected, set the settings on the far right window
 with APPEND OFF, and press ENTER. As simple as that. You will be given the
 choice of either extract from the highlighted file or from all the Tagged
 ones. Select 'T'. All the files will be extracted to individually named files
 with the extension of .HTL. Tagging .HTL files Only, can be re-vamped.
 If you select HTL files ONLY, then changes can be made according to the
 method required.

 Multiple Tagging to SSDLAPP.HTL (append to this file).
 -------------------------------
 SSDLAPP.HTL is only created when you select the option to get all output
 directed to it. The 'APPEND' feature must be switched ON.
 Having Tagged / Selected the files required, set APPEND (F5)'ON' and the
 required extraction method with F3 key. Press ENTER and Select 'T' tagged.
 All files will be extracted and appended to SSDLAPP.HTL. Full sort does not
 take place at this time. You CAN extract HTML format from a Site Search
 format. To re-sort, highlight SSDLAPP.HTL and Press ENTER, select 'C'. Must
 be under 62400 bytes. It is best to switch OFF the APPEND feature first.

 Single files to SSDLAPP.HTL.
 ============================
 Providing the APPEND (F5) feature is ON, any file you select will be
 extracted and attached to the end of the file. Remember, you have the 
 facility to re-extract and re-sort ssdlapp at any time, providing it is
 below 62400 bytes. If more than one file is Tagged, select current 'C'.

 A0 - Z0 HTML files. 4 features. Select / to change. 0 to 9 maximum!
 ================================-----------------------------------
 In HTML mode, F4 has four selections: Normal, A-Z: Extract and append to
 selected files of 'Alpha' content to the correct 'alpha' file. Any file
 with links starting with 'A' or below, will be extracted and appended to the
 correct created alpha file A0.HTL and so on. These created A0 to Z0 files,
 MUST sorted afterwards! Providing under 62,400 bytes, all should be well.
 If A-Z or Index creation, then 'append' will be switched 'OFF'. You can
 override this by pressing F5. Index As-Is, Creates a FULL HTML 'Cross Index'
 system, using only what's available. Finally, Index FULL, HTML 'Cross Index'
 system with all of the A to Z links intact. If you opt to use the F10
 function, these features will be switched OFF.
 Select '/' to change A0 to A1 to 9, then back to 0. This allows you to
 extract to, and create another system, if you desire. Normal: creates HTL
 files. A-Z, creates and or appends to existing alpha files, according to
 what is available. Index As-Is: Creates both HTL (just showing extracted
                                                                  Page 7.
 links) and HTM, the final created FULL 'cross Index' pages, which you can
 use straight away. Index FULL: Creates Both HTM and HTL files, with FULL
 A to Z selections. The HTL become 'Masters', allowing further manipulation.
 Note! If an already extracted 'Index' system exists, any existing files will
 be appended too. This was felt to be an advantage if additional insertions
 were needed. Always complete one task, move the created files and rename
 in lower-case. Complete the HTML project. Then start the next project.
 F2. TABLE of A-Z Cross Index systems: Left, Centered or Center with border.
 F2, F4 and [/] selections are not preserved on exiting, so setup is required.
 Once A-Z are sorted, you can then do what you like with them. If you want to
 create a Site Search, tag all of them and then re-sort them as Site Search
 data. Then sellect F5 (append to ssdlapp.htl) and append them to ssdlapp.htl
 file. Viewing the created file, you will see they are in Java data format.
 Make sure that F4 shows 'Normal', F3 and that F6 append is switched to 'NO'.
 Then press F10. Selecting F3 is not necessary as long as SSDLAPP.HTL has
 code in it.

 Do NOT try to extract 'A - Z Index' or 'Index As - is' HTL files to
 themselves. The result will be an HTL file being extracted and being
 appended to itself, thus making it TWICE as long and containing repeated
 URLS. The size of the file may end up to large for a 'Sort' and or removal
 of Duplicate addresses.

 Site Search. F10. 'BIG FILE over 65535 bytes).
 ==============================================
 Select Yes to create or NO to exit. Nothing else takes place. This only 
 adds the necessary bits of code to the created URLGEN.HTM files. It is wise
 to ensure that ssdlapp.htl is FULLY sorted or appended to in the correct
 sequence for any Site Search Data creation (create and sort a0 to z0 first,
 then append) be it for Java DATA or the FULL site search. Large HTML will 
 not be acted upon. If ssdlapp.htl is in the wrong format, just tag the
 created A0-Z0 files and set F3 to 'Site Search'. Press ENTER, then 'T' for
 all Tagged items. Once converted, delete ssdlapp.htl with F6. Select F5 to
 append, then ENTER, then 'T'. SSDLAPP.HTL will now be in Java format and in
 correct assending order. Then Press F10 to complete the task.

 Creating your Final URLGEN.HTM file.
 ====================================
 SSDLAPP.HTL must be present first. Make sure that it is not over 62400 bytes
 in size. URLGEN.HTM allows you to get all the information which was added to 
 ssdlapp and create your final FULL but simple Site Search for your Web site
 or other media. Extracting to single files was fine, but what if you wanted
 all the LINKS to create an INDEX or something of that nature. Make sure that
 ssdlapp.htl is in the format you want. When all is correct, press F10.
 Ssdlapp.htl will be loaded, links will be extracted and sorted and urlgen.htm
 will be created. Previous version can be deleted.

 What types of URL LINKS.
 ========================
 These are also called Hyper Links. Those are the selections you can make
 on any given Web Site page which, once selected, will carry you to another
 page or give a required action.
 'Site Search and Data Links' gives you various forms of extraction, trying
 to cover as many as possible.
 The SORT works on any description that is given (or displayed) within the
 URL hyper LINK code. The code itself consists of the following...
                                                                  Page 8.
 <A HREF="HTTP://www.m.com/index.htm/ " target="main"><b>description<b> </a>.
 or as used by vendors that embed C G I / Java script.
 These take various forms with pictures and customer code which I will not
 list here. Suffice to say, this program will try to get your required setup.
 Things like <A HREF="JAVASCRIPT  can also be extracted.
 There is also a facility for you to extract your MAILTO:,direct files and
 Javascript to your Web site and create a separate file if needs must.
 As you know, there is no need to keep having the full URL within your
 HTML source code, so files and pictures etc load and display.
 If you have a GIF in your root directory, you just name the file.GIF.
 If the file Paul.htm is in a directory off the root, say C G I, you just
 say <A HREF="../C G I/paul.htm">description if needed</a>.

 Selecting F9 key will roll through the selections. I suggest that you test
 things out until the required extraction method is obtained.

 C G I and Java.
 ---------------
 These are used by various sources. People allowing pictorial advertising or
 'Click Trade' as they are called.  MID RIP and RIP must be 'NO' and Keep
 <b> must be YES using the F1 key. Do not bother extracting these to create
 a Site Search. The information given is not really relevant and descriptions
 are more often than not, pictures. You can see for yourself.
 However, if you have accounts and want to keep displays in Alphabetical
 order, then this may be of use to you. DO NOT remove HTML commands within.

 Strip HTTP://.
 -------------
 You can select to strip all references of HTTP:// within the Hyper Link.
 If the link looks like this...

 <A HREF="HTTP://www.fuddyduddy.connectfree.co.uk/index.htm">my site</A>

 The result will look like this...

 <A HREF="www.fuddyduddy.connectfree.co.uk/index.htm">my site</A>

 This is automatically used to create a 'Site Search' data file. The Default
 display method code is set with the HTTP:// embedded in the code. You can
 alter this to suit your needs after you have created the FULL Site Search.
 Extracted and forgotten to SORT? Select your created .HTL files. Set F9 to
 'EVERYTHING POSSIBLE' and press ENTER.

 There are 5 options altogether. Each has its good points. I suggest that
 you test things out, one at a time until the correct method is found for
 your specific needs.

 Installation.
 =============
 Copy everything into an empty directory. Get some of your HTML files and
 put them into that directory. If the file extensions are HTML or bigger than
 three characters, re-name them to .HTM. Files with eight characters or more
 will be displayed truncated, containing the tile '~' and a number. Nothing
 to worry about. URLGEN's created HTL files can always be re-named when
 finished. DO NOT USE the extension of .HTL. This program will extract and
 create the same file names but with the .HTL extension.
 Make sure that DOS and or Windows knows where you put URLGEN. This allows
 you to call it from anywhere. In Autoexec.bat just ensure that the PATH
                                                                  Page 9.
 command contains your directory and urlgen.exe. Set windows to Maximize
 the screen and to exit when finished. Operates in Composit or Colour.

 How Do I use it ?
 =================
 Execute URLGEN and the directory window on the left side will show you the
 files. At the top of that window will be the root and child directories.
 The block with one DOT is the ROOT directory while the block with two dots
 indicate any Parent directories available.
 Shown on the top left of your screen (or in windows) will be displayed
 the current directory you are in. First get used to moving around with both
 the Mouse and the keyboard. Once you know how to change things then you are
 ready to start. You may rename URLGEN.EXE but nothing else. URLGEN.CFG
 and URLGEN.LOG must remain the same.

 ESCAPE key and its other selections. F = File, ESCape = Return and Q = Quit.
 ============================================================================
 There are three selections when you select to quit from the program.
 The first lets you return to the program by pressing ESCape again.
 The second allows you to select the type of files you want displayed within
 the files window. Selecting 'F' will present you with a window and a small
 suggestion window.
 If you press ESCAPE and nothing was entered, URLGEN will ensure that all
 files are displayed. The prompt window shows what can be typed on the outset,
 or you can have your own selection. Only a single request is allowed. Do not
 bother to type more than one request. If nothing was found, the files window
 will inform you, and you will be returned to the entry window.
 The final selection to Quit presents you with another window, allowing you
 to select the exit method or return to the program.

 General Selection Descriptions.
 ===============================
 URL Descriptions.
 -----------------
 These must be contained within the links and not after the </a> expression.
 These are what you select to 'go somewhere'. Anything after is ignored.

 HTML commands like <B> and <FONT> will be automatically stripped within the
 displayed selectable link. An exclamation mark will be inserted in to empty
 links and URLERR.TXT created. This allows you to see what needs editing.

 Simple Example.
 ---------------
 <A HREF="http://www.me.com/m.htm"><B>See Me</B> and <b>I</b> can see you.</a>

 F1 to Keep or Remove <B> type commands within the desplayed link.
 =================================================================
 When selecting to 'Remove', some results may be far from perfect. Depending
 if you use RIP and MID Rip features and what type of additional characters
 are used to highlight the link feature, the results differ. You can always
 create JAVA data files first. Then, when the result is perfect,
 convert to HTML(Normal).

 Scripts.
 --------
 In general, will not be removed unless enclosed in <>.
                                                                 Page 10.
 Site Search DATA. (Java).
 -------------------------
 All of which is extracted in a special format via normal HTML code.

 F10. FULL JAVA Site Search.
 ===========================
 This contains ALL the necessary code and the extracted converted data from
 your files. A very nice utility for your Web Site and other media.
 The directive http:// is only contained in the additional code when the link
 or links are displayed after a search or a request for help.
 This code can be removed, after URLGEN.HTM has been created, allowing your
 links to access your other media. You could manually ADD http:// in some or
 all of the DATA lines, therefore giving access to both InterNet Web sites
 and your own media or server.

 F8. Compression of CODE (HTML and JAVA).
 ======================================
 This compresses BOTH methods. HTML removes Carriage Returns and Form Feeds.
 This assists in stopping others from taking your hard work and claiming it
 their own. It also makes the file very much smaller and speeds loading.
 You can De-compress the file by switching the feature OFF and re-selecting
 the compressed .HTL file. Takes but a moment on small files.
 Java ARRAY files you switch to 'JAVA'. If you want to de-compact, then
 convert it to HTML (F3) and back to 'JAVA'. Make sure that you have
 'Compress' switched OFF when converting into HTML format.
 On the final results, other compressors can be used to compact even more.
 It will not operate on any Highly compressed files other than on its own
 HTML and HTL extracted files and normal HTML / ASCII source files.

 F9. Extraction Types.
 =====================
 Five in all. Each has its good points and should be examined until the best
 result is found to suit your needs.
 ------
 Extract Everything[E] (3 selections here):-
 ---------------------
 1). Pulls out everything possible.
 2). Everything but NOT JAVA. As it states. Java references will not be
     extracted.
 3). Everything but NOT Files, Mail To:. 'Root' web files and user 'MailTo:'
     links are not extracted. 
 ------
 'No C G I Trade', and 'Also CGI Trade', IMG etc, are now combined in to one
 selection. These are set according to the way you switch 'Mid Rip', 'Rip' &
 F1 ( Keep <b><font> type commands).
 Settings:-
 1) 'Mid Rip' and 'Rip' to 'YES' and F1 to 'NO'.
    Strips and removes unwanted characters including <b><font> type commands.
 2)a. 'Mid Rip' and 'Rip' facilities 'off' and F1 to keep HTML commands
      within displayed Url selection, will result in extractions as normal.
   b. A possible 3rd setting with can be achieved with F1 set to 'NO'.
      This depends on the layout of you original HTML document content.

 <A HREF="JAVASCRIPT
 -------------------
 Extracts only links with the above command. Sort, Target etc work normally.
                                                                 Page 11.
 Mailto, Files:-
 --------------
 Extracts only direct Web Site URL addresses. Those that do not contain
 http://, www and other directives. Any MAILTO (E-Mail links) will be
 extracted. JAVASCRIPT=A HREF will NOT be listed.

 Strip http://:-
 --------------
 This removes ALL http:// directives. Very useful in creating HTML files for
 inclusion in other files which have those directive.
 You now have the ability to have links refering to...
 a). Http:// Internet Sites (outside your Web Site or other media).
 b). Directly to your Home Site Pages (your own server, cd and diskettes etc).
 c). Or BOTH. By combining different extraction methods, you can achieve
     this easily.

 NEW FACILITIES.
 ===============
 1). EMPTY. Key (Y). [ERROR URLS without displayed links] 2 settings.
 -------------------
 Allowing you to make them NORMAL by inserting the 'Web Address' into the
 empty link instead of the default exclamation mark.
 These links will then be added normally to your listings, but NOT to the
 URLERR file. A very useful addition to this utility.

 2). OTHER. Key (O). [Applies other links but NOT Error one]. 2 settings.
 -------------------
 Make pre-defined displayed links show the 'Web Address' instead of what they
 originally contained.
 
 Using in combinations, you have upto 4 different settings.
 These two facilities settings are NOT saved in the configuration file CFG.
 You will have to select them as and when needed.

 3). BgProperties. Key (B). For 'FULL Site Search' creation.
 --------------------------
 You can now switch 'ON' or 'OFF' the facility to include the HTML command to
 'lock' the background with the displayed page or not.
 If 'Locked', the background will scroll with the normal displayed page as
 you scroll up and down.
 'Un-Locked' will result in the background being stationery whil the rest of
 the page scrolls. A nice effect. Internet Exploror allows this while some
 others may not! So test your reslts with other browsers when possible.

 4). TABLE. Keys (Cursor Left and Right). For 'Full Site Search creation.
 ----------------------------------------
 The 'A to Z' selection bar displayed at the bottom of each page can NOW be
 SET to a different depth. The Default is 18 lines. Maximum is 99.
 ------
 ------
 My first extraction.
 ====================
 There are so many ways to extract. I think it's best for you to highlight
 one of your Master files containing some links and try this:
 Set APPEND to NO (F5), Set to CREATE HTML (F3). Set COMPRESS to NO (F8),
 Select F9 extraction method to HTTP://NO C G I etc. Mid Rip & Rip to YES.
 Press 'H' to remove Duplicate URL Link addresses.
 Now press ENTER on your highlighted Master File (or right mouse click).
                                                                 Page 12.
 Your Master file was loaded, extracted and sorted with the resulting file
 displayed as the same name but with the extension of .HTL.
 The Master File (if all was well) should have a smiling face displayed on
 the far right side of its name. The created .HTL file should have a symbol
 of 'h'. This indicates the Links extracted were in HTML data format ONLY.

 Highlight the created file and press 'V' to view. A window will open at
 the bottom of the screen.
 This allows you to see if the correct format resulted. You may have links
 without descriptions listed at the top with an exclamation mark or a single
 no alpha character. This is for you to alter. Also see the URLERR.TXT file.
 The rest of the descriptions are listed in Ascending order. The idea of
 allowing the creation of single files helps people to add new documentation
 to already created data or Site Search files. To exit from the viewing
 window, press ESCape and you will be returned to main window. You have now
 created your extraction file and viewed it. Could not be simpler.
 Highlight the same Master File and set F3 to Create 'JAVA' data.
 Press ENTER. You will see an 's' symbol on the far right of the created
 .HTL file. This is a Site Search/Java Data File. View it.

 Press zero key on top of the keyboard. Select Master and ENTER. View it.
 You have now created a Data File for the Site Search. Simple isn't it.
 If your master File did not have any LINKS to extract according to the method
 you selected, you will see an 'e' or 'n' symbol. The .HTL file of the same
 name will not be changed. I say this as you could already have created other
 ata extractions. Only a 'smiling Face' indicates a good extraction and that
 the .HTL or SSDLAPP file has been appended to. Multi tagging and selecting
 ENTER will result in individual files being created.

 Best method of extracting.
 ==========================
 ALWAYS create individual .HTL files. Then TAG those and select to APPEND
 to SSDLAPP.HTL. This way, you will have everything as setup.
 When creating and APPENDING to the special file SSDLAPP.HTL, I suggest that
 you use the HTML extraction method at all times. With this, you have the
 ability to re-extract in other formats. Then, if a Site Search data is
 required, just highlight this file, switch APPEND to NO, switch the numbering
 system to 1. Press ENTER and there you have it.

 Re-Sorting and Extracting from an already created .HTL or Site Search file.
 ===========================================================================
 Highlight it and press ENTER. If it's an HTML formatted file, you can
 change the extraction method and re-create a new file. Using this method has
 all the facilities available to you and SORTS the result. Very Useful.

 Create and APPEND to SSDLAPP.HTL file.
 ======================================
 Could not be simpler. Highlight a file and select to APPEND (F5 to YES).
 Press ENTER. Providing there are LINKS to extract according to your
 requested method, SSDLAPP.HTL will have them inserted. The symbol 's' will
 be displayed and a 'Smiling Face' on your Master File, indicating all went
 well. Now HighLight another file. With the same selections, press ENTER.
 The extractions will be added (APPENDED) to SSDLAPP.HTL. View it and see.
 Having both HTML and Site Search data within SSDLAPP.HTL file is not
 allowed. URLGEN has a 'Protect' system with a warning.
 You can ONLY extract in the format that SSDLAPP.HTL was originally created.
 It is better this way, keeping things neat and tidy.
                                                                 Page 13.
 Creating your final APPENDED FULL Site Search.
 ==============================================
 Just Tag or individually select the files with APPEND to YES. Select ENTER.
 If you have Multi-tagged files, you will be prompted to either select
 the Current highlighted file or the Tagged files. Select 'T' and away it
 goes, extracting and appending to SSDLAPP.HTL. Sorting is not done at this
 stage of the creation process.
 Individual files selected without Tagging will be dealt with immediately.
 View the final SSDLAPP.HTL file and check for the correct extraction.
 You can Highlight it and re-extract. You can do everything again in either
 format. When all is correct, press F5 to switch APPEND to 'NO' and then
 press the F10 key. If SSDLAPP has not been created you will be informed.
 F10 key requires it. If the file is present, it will be loaded, sorted and
 the final file called URLGEN.HTM will be created. A symbol 'f' for a FULLY
 created Site Search, which includes all the necessary code to execute on
 your Web site or other media. A symbol of 'h' indicates HTML sorted DATA.
 ALL The Links will be Sorted correctly in that file by description.
 LARGE SSDLAPP.HTL will need my routine SSDLAPP?.EXE. Down load it.

 ADDING to an already created Site Search.
 =========================================
 This really is simple and only takes a few steps to complete the task.
 Depending on how many files you need to add and the speed of your computer
 system, it should take but a few minutes at most.

 First, take the Created Site Search file and load it into your HTML editor
 or your ASCII editor. Then delete the first and second parts of the
 actual Site Search Code. Just leave the urlsArray[] and NamesArray[] data
 lines in that file. Everything above and below goes.

 Then save that file as SSDLAPP.HTL. Run URLGEN, switch APPEND to YES (F5),
 select the extraction method and other refinements with the creation method
 to Site Search or HTML if you are using HTML formated file. Then TAG all the
 HTML Master files you need to add to the Site Search. When ready press
 ENTER. The extractions from your master files will be appended to 
 ssdlapp.htl.
 When ready, set APPEND to NO (F5), If SSDLAPP is in HTML format, select
 extraction method as HTML (F3). Finally, press F10 to create your new and
 FULL Site Search with all the new data.
 If SSDLAPP.HTL is over 62400 bytes, the code will just be added. If this is
 the case, create A-Z, sort them, append them to SSDLAPP.HTL and then Press
 F10. You can get my other utility SSDLAPP3 for very large files. Available
 as FREEWARE from good Shareware/Freeware suppliers.

 Site Search Creation. The Best method. (Alway keep numbers low) !!
 ======================================
 I do suggest that you always make SSDLAPP.HTL in HTML format and create
 your FULL Site Search or Data to the final file URLGEN.HTM.
 Rename SSDLAPP.HTL as your master source. Then, at any time you can append
 to it, re-sort it and then extract and create the Site Search methods in
 URLGEN.HTM. When ready, just press F10. The numbering system only goes to a
 maximum of 65500.
 That's it in a 'Nut Shell'. You can extract individual files and sort them.
 You can extract selected files and create a file with those Links within it.
 You can re-extract and sort single or the large SSDLAPP appended file.
 You can extract one format from a multi-format file and create exactly what
 you want out of that file.
 You can create a FULL Site Search for your Web site or other media, and
                                                                 Page 14.
 finally, extract single or multiple files to create a single sorted HTML
 or Site Search data files for other uses.

 IF RE-SORTING, REMEMBER TO ALWAYS SET THE COUNTER TO THE SAME START NUMBER.
 OTHERWISE THE NUMBERING SYSTEM CONTINUES FROM WHERE IT LEFT OFF.
 This is 'VITAL' if you want to TEST this extraction program! (Magazines etc).

 Altering the Site Search Code.
 ==============================
 It is easy to alter the code even when compressed.
 When the URLGEN.HTM final file is created, the second part of the code
 will have two lines in it looking like this..

 document.write("a href='http://" + links[q].URL + "'>" and so on..

 If you want the links to access the 'root' of your directory, then remove
 the 'http:// and the ' in "'>", like this...

 document.write("a href=" + links.URL + ">"  and so on.

 If you want to create a CD and use this Search, all you need is to extract
 links from your own site files. Then, when each link is selected, it jumps
 to your pages. You can also ADD other LINK extraction types to give you
 an 'alround' ability to search not only your site pages, but also to
 search and 'go to' outside your site. Now that's very useful.
 To do this, you simply extract using 'Mailto: files', which give your own
 files and append to SSDLAPP. Then extract other files with the method
 of HTTP:// + NO C G I etc and APPEND again to SSDLAPP. View the result
 and make sure all is correct. You can manually correct anything using your
 HTML or ASCII editor. When all is complete, create the FULL Site Search.
 The file will have two types of links. One containing direct hyper links
 to sites and the other linking things to your own pages.

 ADDING USER ENTRY SYNTAX CHECKING FOR 'SITE SEARCH'.
 ---------------------------------------------------
 If you need it, just remove this code (after data, second part):

 function validate_form() {
 txt = document.form.search.value;
 jsi = new makeLinks(linksize);
 searchLinks(jsi, txt);
 document.write("<hr>");
 }

 and REPLACE it with THIS.....

 function validate_form() {
 txt = document.form.search.value;
 IF (txt.indexOf(".") != -1){
 alert("Search is invalid! Contains a '.'"); return;}
 else {
 jsi = new makeLinks(linksize);
 searchLinks(jsi, txt);
 document.write("<hr>");
 }

 NOTE. You can add more 'IF' statements to check for unwanted keys.
 If Compression was used, this part will be compressed. Just examine the
                                                                 Page 15.
 lines and alter/add to suit. Keep the code as small a possible. This helps
 in cutting down unnecessary big file creations which take longer to laod.

 TARGET PAGE requirements. DO NOT USE under_score, IE4 below causes faults.
 ==========================================================================
 Only 12 characters maximum. These can be in Upper or Lower case, or mixed.
 If your file does not have TARGET pages throughout it, and you want to ADD
 your own, then please try to keep the original file no bigger than 50k.
 URLGEN Checks as it inserts the new target name. Remember that TARGET=" is
 eight characters, plus the length of your page name plus ". If you have
 the name as eight characters then there will be 17 extra characters per Link.
 300 links times 17 = 5100 bytes. Please keep you master files short.
 There is no point having anything over 40k irrespective of the connection.
 The bigger the file, the more storage space needed on your Web Site and the
 less information you can supply, the slower things will load.

 Site Search, Java. (DATA (Array) files do not require this!)
 ------------
 The DEFAULT does not have a TARGET name within the special code that is used
 to create your final FULL program. The ordinary data does not have targets.
 Select method of extraction as 'Site Search' (F3). Then press 'P'. Type in
 the TARGET name and press ENTER. F10 creates the FULL Site Search.
 To remove the name, just delete it and press ENTER. This will revert to the
 default. If you selected the option by mistake, press ESCape and things will
 remain as before.
 
 Normal HTML format. The creation of .HTL files.
 -------------------
 Select the method of extraction as HTML (F3). Then press 'P'. Type in the
 page name you require. When finished, press ENTER, or again on file to do it.
 To leave the name you typed before without accepting any new entry, or if
 you selected this option by mistake, press ESCape. To remove all TARGETS
 already assigned to the HTML CODE, type REMOVE (in any case). This will seek
 and destroy all TARGET PAGE assignments, but not any of those used within
 C G I directives. To remove them would render the code useless.

 Mouse, other device users. First button= Upper case and display selections.
 Second button= Lower case and ENTER on the name typed in.
 Bottom of the screen will show you what is assigned, and to what.

 DUPLICATE URL address removal. One of the BEST features of this program!
 ==============================
 You can elect to remove all duplicate lines containing the same Web addresses
 within the URL (Not descriptions).
 J=Java or H=HTML. Switch to 'Yes' on the system you want this feature.
 MAKE SURE THAT YOU EXTRACT TO SINGLE .HTL FILES FIRST! This will help you!
 When extracting and APPENDING files to SSDLAPP.HTL, it will NOT remove
 duplicate lines in that file. To do this, you wait until you have extracted
 everything. Providing the final creation is not greater than 62400 bytes,
 you can highlight it, set APPEND to 'NO' and select ENTER. It will react to
 the settings made and remove ALL duplicates. If bigger, use SSDLAPP3.EXE.
 Providing the original extraction method allowed more types to be extracted,
 you can select another method and re-extract. That means, re-creating
 itself with the new system. If you have the TAGGING system active and you
 want to re-extract ssdlapp.htl file, you highlight SSDLAPP.HTL and select it
 with the 'C' option.  Removing duplicates from single files TO single files,
 press ENTER once. The numbering system must be set according to your needs.
 If descriptions are the same, the URL addresses are more than likely NOT.
                                                                 Page 16.
 You are advised to Check them! You can use A-Z, then sort normally and
 finally append to ssdlapp.htl.

 RIP Unwanted Characters from LINK Descriptions. Unique to URLGEN.
 ===============================================
 This KILLS as many unwanted characters from BOTH the Front and End.
 To remove commands like <b><FONT> press (F1) or again to switch OFF.
 DO NOT USE THIS FACILITY ON EXTRACTING C G I TRADER TYPE LINKS !!!!!
 ALL YOUR DESCRIPTION IMAGES ETC WILL BE EMPTY. IF THE ASCII CFG FILE IS NOT
 PRESENT, THE DEFAULTS WILL BE AVAILABLE.

 URGENT !!
 ------
 Please check the resulting files. Descriptions with apostroph or an inverted
 commer as starting words should be dealt with correctly. Some problems may
 occour as not every combination has been checked for. One tries.
 Commands like <B> within descriptions' to will be automatically removed and
 the insertion of an exclamation or a non alpha character in empty lines will
 be done. WHY this feature ?
 Some descriptions may have things like Asterixs to get people to notice the
 LINK. If you do not want these types of characters, this will attempt to
 remove them. It will also remove Leading Space Characters and any finishing
 ones. This feature has been updated and operated much better than before.
 There is a Default character 'rejection' list within this program, which will
 be activated if URLGEN.LOG file is not present on the directory requested.
 You can save both (CFG and LOG) files at any time by pressing CTRL/S key.
 You can edit LOG ASCII file with your ASCII type editor.
 Examine the descriptions for unwanted characters and ADD them to the file.
 DO NOT USE anything below the 'SPACE' key (32 decimal / 20 hex) as these are
 used for computer control. If you are using the Universal Character set on
 your system, DO NOT TYPE in your native language character set.
 ie; If English, then A to Z in BOTH upper and lower case. DO NOT DO IT!

 URLGEN.LOG will be automatically saved with URLGEN.CFG everytime you
 extract or request to 'Exit'. With 'the 'Exit' request, if you decide to go
 back to the program, your LOG and CFG file will still be saved on the
 directory you are currently using. When 'Exiting', depending which method
 you choose, they are also saved, but onto the exit request directory.

 Things to WATCH OUT For. (Problem Characters and positions in Description).
 ------------------------
 If there are HTML commands between the description, they will be removed.
 Any description that does not have ALPHA characters will be destroyed.
 In its place will be a SINGLE 'Exclamation Mark' to indicate this. These
 errors are saved into a file named URLERR.TXT for examination later.
 Run your HTML creations in your Browser. Look for any problem links.
 The Site Search DATA may not display. You need to create the FULL routine
 to view the results in your browser. You can 'view' or use ASCII editor.

 Characters set by Default. (See Page 4. URLGEN.LOG).
 --------------------------
 space "!<>'#$%&)*+,-\:;=?@\[]^_`{ char124 }~ and some universal BLOCK
 drawing characters. The 'Space' character is embedded in this program.
 DO NOT use <<<  >>> to highlight text. It will be ripped out as commands.
 ALWAYS CHECK THE RESULTING FILES. It is in your interest to make sure that
 ALL results are as you want them. Missing extraction commands are normally
 caused by the insertions here. 100 characters maximum.
                                                                 Page 17.
 MID RIP. Remove selected Characters from within the Description. Unique!
 ========================================================================
 This function allows you to select certain characters that you do not want
 appearing within the descriptions. It has a built-in checking system for
 things like 'Forward Slash' between words, or a series of them.
 It will also check and remove any extra spaces. Any characters selected
 will be removed. 'Alpha' and 'Numeric' characters are not selectable.
 HTTP:// in description link required? Remove / and semi colon from list.

 Operations. Independant or the 'End Rip' feature !
 -----------
 Select the 'Mid Rip' function with the 'M' key. A window will pop up showing
 the top half with the characters you can select. The bottom area will display
 any characters you have chosen to remove from descriptions.

 If you already have URL links that show the 'Web address' as the selector,
 Then '/' forward slash should be removed from 'mid rip's' selection if you
 want them to be extracted correctly. If you do not do this, all that will
 be displayed is 'HTTP:/...'. This indecates that a 'Web address' was
 present.

 Keyboard.
 ---------
 Use the 'TAB' key to switch from the Main screen and the 'Mid Rip' feature.
 When you have access, there will be a flashing message (not in Windows) at
 the bottom right. Use the Cursor Keys to move around BOTH parts of 'Mid Rips'
 window. Press 'Space' or 'ENTER' Keys to Select or De-select any characters
 in either half. Selecting 'Empty' will CLEAR everything! To exit, press
 'ESCape' or the 'TAB' key to return you to the main operating window.

 Mouse or other pointing Devices.
 --------------------------------
 The same options as the keyboard. Selecting either area will allow removal
 of any characters listed in the lower part of the window, while selecting
 any character in the top window, will also insert characters not listed in
 the lower section. You can ALSO select any of the 'Main Programms' features.

 SAVING the selected deletion characters.
 ----------------------------------------
 These will be saved into URLGEN.CFG file whenever you select CTRL/S, or
 the ESCape window. You do not have to exit the program for them to be saved.
 When you run URLGEN.EXE again, all your previous settings and displays will
 be ready and waiting for you. CFG is automatically saved when extracting.

 Why this feature ?
 ------------------
 Believe it or not, this is VERY useful indeed! Just like the RIP feature,
 there may be multiple useage of 'Asterix' and or other symbols to give
 various words 'Visual Impact'. However, you may want to keep your exractions
 looking 'Neat' and 'Tidy', and even try to compact the files. This would
 allow additional LINKS to be added without creating 'loading' problems on
 your Web site. It also make things look more professional.

 Site Search/Java Data File.
 ---------------------------
 'RIP' and 'Mid Rip' will not operate on these .HTL file. You can always
 convert to HTML, do all the necessary extraction methods and then convert
 back. Just select the F3 Key to swap between methods.
                                                                 Page 18.
 Press 'ENTER' on the .HTL file for each method. As simple as that!
 If creating a Final FULL Search, then ssdlapp.htl will operate correctly.

 ERROR MESSAGES.
 ===============
 Symbols shown within File Name Directory.
 =========================================
 Smiley face (ASCII 2). This file is perfect and was extracted from.
          a.  File is Alright and APPENDED to SSDLAPP.HTL file.
          e.  Cannot Extract, Mixed, No MAILTO: or links errors.
          h.  HTML normal code created. Extracts links only.
          l.  Too LARGE to load. Over 62400 bytes. Split it.
          n.  No URL Links Detected for extraction. Please check.
 Ascii 254.  This is a TAGGED file for multiple file actions.
          s.  Site Search format using Array[numbers].
          f.  FULL Site Search created routine, ready to run.
          v.  Void - Empty, awaiting your final input.
          t.  TARGET page assignment makes file too large. Aborted!

      Files that are too large, just split them and run them through.
  The 'Symbols' are not maintained when changing directories or file spec.
 To get the description current file and examine the above table.
 I suggest you print the 'spec' table for easy access.
 It is a good idea to always switch on the 'Remove Duplicate URL' feature.
 This could make the files a lot smaller when extracting.

 JAVASCRIPT extractions and links.
 =================================
 The Web address (URL) may have spaces at the beginning, URLGEN will remove
 them. JAVA links, <a href="javascript: etc>more</a>, can be extracted with
 either 'EVERYTHING', 'Files Only' and NOW using JAVASCRIPTS own extraction
 method, by selecting (F9). Some editing may be necessary if JAVASCRIPTs own
 system is not selected.
 The extractions will be sorted in to accending order as normal and various
 requirements will be acted on as before. Remember to extract removing all
 duplicates, as it is the best methoid to use.

 Converting JAVASCRIPT to normal HTML:-
 =====================================
 The extractions of links containing <A HREF="JAVASCRIPT: can be converted
 quite easily with the required TARGET page assigned. Just select HTML method
 of extraction and press ENTER.

 Messages when selecting Single formats:-
 =======================================
 URL Generator will try to trap most problems.
 If you have chosen to extract in one format and the file is for another,
 then you will be informed. This is for appending to SSDLAPP.HTL.
 If files have been reserved for certain actions, like Read ONLY, the Viewing
 facility will inform you to change the Attribute before you can view it.
 This must be done through other means, outside this program.

 Every effort has been made to trap errors. However you should not alter
 URLGEN's own created code and pass them through again. Unless you are
 altering the code 'as is' or adding the same type of line content. There may
 be things that you have added that are not covered here. Once you have
 completed all of your tasks, add your other code, pictures etc.
 Later on, if you need to create additional files, or ADD to already created 
                                                                 Page 19.
 data, or even extract, you can use this program.

 PROBLEM SOLVING.
 ================
 This section will assist you in various ways, please read !

 A). Duplicate URL address removal.  THEY DO NOT REMOVE !

 When this feature is switched ON, you need only press ENTER once.
 Make sure that you have switched this feature to 'YES'. Press 'H' key for
 HTML and or 'D' key for Site Search. Some options only operate in HTML.

 SSDLAPP.HTL. If you have TAGGING feature ON, then Highlight it and press
 ENTER and select the 'C' option ! This is a MUST ! If you do not select 'C',
 all the files TAGGED will be appended again to ssdlapp.htl.

 B). SOME URL LINKS ARE THE SAME WITH EXCEPTIONS OF A FORWARD SLASH AT END.
     CAN I REMOVE THEM ?

 No. I decided to leave this as standard. It's the correct syntax in html.

 C). HOW DO I REMOVE LINKS THAT HAVE THE SAME ADDRESS BUT DIFFERENT USERS
     AND OR PAGES ?

 This will have to be done manually with your ASCII or HTML Editor.
 Sometimes, a company or person may have different users and or pages,
 although through the main site. I personally would remove these and keep just
 the major name. ie /Peter/index.html and  /Peter/software.htm would result in
 me keeping /Peter/index.html.  Why clutter up your system with hundreds of
 links pointing to one person's different pages, when you could just have one
 LINK to his or her main site.  In some cases it would be advantageous for you
 to have these other LINKs. It really depends on the type of Web Site or CD
 media you are creating.  This option was deliberately left for you to choose.

 Target Page Removal. (A suggestion).
 ====================================
 It is best to REMOVE any HTML target pages assigned at the outset. You can
 always type the new name just before you finish up. This will keep things
 neat and tidy and get you into a good routine. DO NOT USE UNDERSCORE!

 SSDLAPP.HTL.
 ===========
 APPENDING TO SSDLAPP.HTL. THE COMMANDS IN THE DESCRIPTION ARE NOT REMOVED. 
 Make sure that set up keys are 'ON'. When completed, you select ssdlapp.htl
 with ENTER.  If the format is HTML, set up other options and press ENTER.
 If the format is 'Site Search', convert it to HTML, then set up options and
 press ENTER. To get back to Site Search', Switch to it (F3) and press ENTER.
 It is now back as it was with all commands (except SCRIPTS) removed.
 If items are TAGGED, select 'C' (current file) option.
 
 'Site Search/Java' Numbering ARRAYS. (Line count).
 ==================================================
 All files will be sequential unless you elect to change the numbers for each
 individual file. If you TAG a number of files, then the resulting number -1
 is the LINKS total for all those files. To get a TRUE figure, you can use
 DUPLICATE Remove feature. Sequence will be broken if other files are random,
 or if you do the same file more than once without setting the start number.
                                                                 Page 20.
 CLICK TRADE TYPE LINKS. C G I and others.
 =========================================
 D). WHY DO THEY NOT SHOW WHEN I REMOVE <B> AND OTHER COMMANDS FROM WITHIN
     THE DESCRIPTION USING THE 'F1' FEATURE.

 You MUST NOT remove any commands from these types of links. The description
 usually contains images (IMG) and other code needed by the vendor.
 If you must have C G I or Trade type Links extracted, then it is wise to
 extract everything, then load the file in to your ASCII or HTML editor and
 delete ALL other Links. On the other hand, you could delete the contents of
 the descriptions and add your own, You MUST have permission to do this!

 E). IF URLGEN STRIPS OFF END SEQUENCES FROM THE DESCRIPTIONS, THEN WHY,
     WHEN TARGET PAGES ARE ASSIGNED, DOES IT SAY FILE TOO BIG ?

 Adding or removing TARGET Pages always comes first. It does not take into 
 account things to come. Anything before has already been done and therefor
 the calculation works on that. Up to, and including that point, knowing the
 file size is a 'must', and that is what governs what passes or not.
 Try a smaller 'name' for a start, it just may pass.

 F). CAN NOT RE-CREATE FROM THE SAME .HTL FILES.

 Usually it's to do with the method of extraction you are trying to use.
 After creating your .HTL file, press F9 to switch the extraction method to
 'EVERYTHING' and then re-extract it. If that fails, select another option.
 There may be a time when it refuses to extract. Simply select the MASTER
 HTML file and create a fresh .HTL file. Takes but a moment and is always
 worth the effort.

 G). AFTER TAKING OUT ALL THE HTML COMMANDS FROM WITHIN THE DESCRIPTION,
     HOW DO I MAKE THE RESULTING FILES DISPLAY IN 'BOLD' OR HAVE OTHER
     COMMANDS ?

 Depends on the type of HTML Editor you use. If you can edit while seeing
 the FULL display (displays like a  browser) then you can simply highlight
 the area you want and select the colour, size and anything else.
 If your Editor only operates in ASCII text mode, then you will have to
 manually insert open commands like <B> (for BOLD) just before the first
 character of the description. When finished, you will have to put the close
 commands after the last character of the description. Remember that the
 last 'open' command MUST be the FIRST to be switched OFF and so on.
 ie; URL goes here"><B><FONT="#FFFFFF">This is a test</FONT></B>.
 Providing that you do not have 'Paragraphs <P>','Tables <T' somewhere in
 the code, you can put the commands at the very beginning and close at the
 very end. With all Paragraphs <P>, Tables etc, you put the code directly
 after it and CLOSE them just before the close of the paragraph </P>.
 However, it must be done according to normal HTML standards. Check manuals
 and see. A Table may have Paragraphs within it. You need to know where your
 LINK code is before you insert code. Always have a HTML reference book.

 .HTL.  Convert Site Search/Java To HTML format.
 -----------------------------------------------
 H). WHEN I CONVERT, SOME OF THE SETTINGS ARE NOT ACTED UPON, WHY ?

 The routine does it as fast as possible. After converting, just select the
 file again with ENTER and all the settings operate normally.
 Remember, if you need TARGET for HTML, make sure you have set it up.
                                                                 Page 21.
 I). SCRIPT COMMANDS WITHIN THE DESCRIPTION WILL NOT BE REMOVED !
 This program will not do this at present. I am looking into it.

 J). I HAVE MADE AN SSDLAPP.HTL FILE WHICH IS BIG, WHAT NOW ?
 VIEW your creation first and check it properly. Always Check creations!
 Make sure you have created it with all the options, TARGET (if required) and
 filling empty descriptions with an exclamation mark (always best) and so on.
 You can create a Big Site Search via the F10 key or use SSDLAPP3.EXE to
 finish the .HTL file or create other stuff. MAke A-Z first. That may help.

 K). SOME OF THE KEYS DO NOT WORK !
 It's NOT the program. If parts of your Disk are becomming weak, generally
 through Reading and Writing a lot, or if there is a problem with memory
 allocation, you could have a problem with the operating system or the file
 on your drive has become corrupt. Re-install URLGEN on a different directory
 (do not delete the old one). Test the new installation. If it works, fine.
 Check the URLGEN.EXE's file size. Anything bigger or smaller usually means
 you have a 'virus' on the system. If that is not the problem, has there been
 any problems over the last couple of weeks or so. Disks report errors,
 programs crash, strange displays, unknown messages and various other things.
 Do not get confused over bad connections on your modem. That problem is more
 often than not your line rental company. Make sure that you are NOT connected
 to a DACS unit. This is known not to be suitable for digital use. It's used
 mainly in the (UK) by Bt's PSDN network (voice lines, information correct up
 to March 9th 2000).

 L). THE DESCRIPTIONS ARE CORRUPTED !
 ===================================
 This can only happen when either URLGEN.CFG or the main program EXE file
 becomes corrupted on your storage media. First delete the CFG file and
 try again. If that is not the solution, re-install the entire package.
 If it persists, then you could have disk, hardware or 'Virus' problems.

 M). A0 to Z0.HTM files.
 =======================
 Some Files are to big! Just examine them and remove any known unwanted links
 to reduce the size. Or, you could just split them and re-do them, or use
 ssdlapp3.exe to complete the task. Get it from my sites or other sources.

 N). Creating a BIG Site Search.
 ===============================
 How do I do it? Select HTML and A-Z (F4) and tag all the files you need.
 Then press ENTER and then the 'T' key. Once finished, untag all, then tag
 A0.HTL to Z0.HTM files. Select 'Site Search' and REMOVE Site Search
 duplicates. Press ENTER and then the 'T' key. A0 to Z0 will be re-done,
 sorted and duplicates removed. Delete ssdlapp.htl then switch 'append' ON
 for ssdlapp.htl and press F10, then select 'Yes'.

 O). <B> type commands are incomplete.
 =====================================
 Just set for a Java. Switch Keep <b> type commands to 'YES' and
 remove <> from MID Rips list. Create the file. Then highlight it, select
 HTML format/Normal and do it. Examine the result. It should be OK.
 Some Hyper links displays have silly 'stand out' type characters.
 It is a good idea to edit these where needs must.
                                                                 Page 22.
 P). Using the F = Files command.
 ================================
 This is to assist those of you that may have forgotten.
 Items wanted ONLY.
 ------------------
 A0.HTM to Z0.HTM                           Type ?0*.HTM
 A0.HTM to Z0.HTM and A0.HTL to Z0.HTL      Type ?0*.*
 All .HTL files                             Type *.HTL
 ALL .HTM files                             Type *.HTM

 and so on...

 Q). Deleting Files.
 ===================
 If you have only .HTL files listed, and you delete the lot. The directory
 listing appears EMPTY. If you have other files with different extensions
 within that directory, they need calling with the FILES facility. 
 Press ESCape, press 'F', type *.* and press ENTER. They will be listed.
 If, on the other hand, there are no remaining files, one of three things
 will happen.
 A). If there are other directories present, you will see both 'blocks'
     allowing selection to move 'back up' the tree.
 B). If it is the root, then the program will exit.
 C). Sometimes, a 'blank' directory listing will be displayed. All you do,
     is select the 'FILES' facility and type *.* pressing ENTER.
     If others files or directories are present, they will be displayed.
     If NOT, the program will exit.

 UrlGen does not have a 'select Drive' facility. If the drive is completely
 EMPTY, it is designed to exit.

 R). Creating FULL 'Cross INDEXING' pages with 'As-Is' and 'FULL'.
 =================================================================
 If you select master files directly and use either of these facilities,
 the final creations will NOT HAVE SORTED links. Some people prefer this.
 To get a TRUE SORT, you must first select A-Z and create the .HTL files.
 Then Untag everything, Tag the new A0.HTL files and set extraction type
 to HTML ONLY and extract/sort. Once completed, you select one of the INDEX
 methods and create the final A0-Z0.HTM files.

 S). HTTP only displays as url link.
 ===================================
 Remove both the 'forward slash' and 'semi colon' from 'Mid Rips' List.
 
 URLERR.TXT Error links extraction.
 ==================================
 Most empty links will be redirected and appended (added) to this file ready
 for you to examine and alter if necessary. Here you can load it up in your
 HTML editor, correct the necessary lines by inserting the descriptions. Then
 rename the file and run it through this program in the usual way.
 CGI/IMG: If they do not have displayed >TEXT</A> in the LINK, they too will
 be directed to this file.

 Master Files that have links marked with a single exclamation mark, will be
 extracted to the URLERR file. However, those links will have an extra
 exclamation mark added to indicate this. If on the other-hand the master
 file links contain TWO exclamation marks, they will be extracted and placed
 at the begining of the HTL file. They are considered as 'TRUE' descriptions.
                                                                 Page 23.
 IMG, CGI etc:-
 Links containing a picture as the selection ONLY (without a text selection),
 will also be considered as an ERROR. Although not strictly correct.
 The too will be placed into the URLERR file. Not a bad thing. If you need
 a listing of those files, just load the URLERR file into your editor,
 cut all other links out of it, and you now have a normal IMG, GCI reference
 file - html file. Save it under another name with the extension of HTM or
 HTML etc.

 Each session will be listed with the offending file names and its links.
 This applies to both true HTML and JAVA. Duplicates are NOT removed.
 You may find some error links get through placing themselves at the top
 of individually created files, or in either or both the A0.HTL and Z0.HTL
 file depending on their ascii value. Just check them as normal.

 URLERR:- Do NOT try to extract links listed within this file. It will not
 be allowed! This will contain links that are 'void', extracting will only
 produce the same ERROR file. That's why it's not an selectable option.

 You can delete this file using this program, 'DOS' or 'Windows' using
 whatever is required.

 EXTRACTING NORMAL HTTP:// with or without CGI, IMG etc.
 =======================================================
 Extract CGI,IMG etc.
 ====================
 Set F9 key selection to 'HTTP://NO CGI,IMG etc'.
 If MID RIP, RIP are set to 'YES', or F1 'Keep <B><FONT>' is set to 'NO',
 Url links that have a displayed text link ( >text</a> ) will be included
 in the extraction. Those without, will be directed to URLERR.TXT file.
 You can always edit that file and rename it for storage, if required.
 
 DO NOT Extract CGI,IMG etc.
 ===========================
 Set F9 key selection to 'HTTP:// & CGI,IMG etc'.
 To get this, you must make sure that 'MID RIP' and 'RIP' are set to 'NO',
 and F1 'Keep <B><FONT>' is switched to display 'YES'.
 Those without a displayed text link ( >text</A> ), will be directed to
 URLERR.TXT file. You can always edit that file and rename it for storage.

 NOTE!!
 If F1 is set to 'NO' at any time, CGI/IMG etc; will NOT, depending on link
 syntax, be extracted. Need testing to get results needed.

 My comments and Web Sites.
 ==========================
 For those of you who want to know, I have not been taught programming and
 do not do it for any form of living.
 This Program, a combination of my utilities, come from 'a need', where nothing
 seemed to be available else where, at the time.
 Every bit of code was created by me. No 3rd party modules were used.
 This program contains as many features as I could think of.
 Without these, it would have taken me almost forever to achieve the results
 needed. In a matter of minutes (depending on speed of your system), you too
 can extract, sort and create some very nice files.
 An additional utility SSDLAPP3, can help to create very large systems and
 a useful FULL Alphabetical 'Cross Indexed' paging system. Remember, very
 large 'FULL Site Search Systems' can also be created for most media.
                                                                 Page 24.
 There are now a few other similar utilities around. Although nice, some lack
 the features needed, while others are for specific uses and written for
 Windows environment only. SSDL/UrlGen offers many more features and in
 general, produces better results (my personal opinion). Still, I am sure that
 there will be many more 'new commers' to rival my little routine.
 Let us hope so!
 Remember to check ALL resulting files! It is in your interest to do so.

 I hope you will enjoy the results as much as I have.
 Have a look at one of my sites. All my software is available with some
 information pages and useful 'hints' and 'tips' on various subjects.

     TELL PEOPLE, MAGAZINES, WEB SITES AND OTHERS ABOUT my UrlGen utility.
     ---------------------------------------------------------------------
 Please ADD my CPWHITE.GIF and a LINK (to the first site listed below) on your
 web site. If you know of other FREE Url Extractors like this, let me know
 and at what URL address they are located. They may get added to my sites.
 Best wishes and regards.
 Charles Peter White (author).
 E-Mail:
 webmaster@alltodaysnews.com   {- (these are the main two sites.)
 fuddyduddy@connectfree.co.uk  {- (------''-------)
 fuddydyddy@telinco.co.uk
 fuddyduddy@freenet.co.uk

 My Personal Sites:
 ------------------
 http://www.alltodaysnews.com   All sorts of great topics. (use for links..)
 http://www.fuddyduddy.connectfree.co.uk     ............  (OR this one.)
 http://www.cpwhite.free-online.co.uk
 http://fuddyduddy.future.easyspace.com
 http://www.freenetpages.co.uk/fuddyduddy
 http://www.fuddyduddy.telinco.co.uk
 http://members.tripod.co.uk/fuddyduddy
 http://www.fuddyduddy.f2s.com
 http://website.lineone.net/~cp_white

 More being added all the time.
 All program updates will be placed on as many sites as possible, as and
 when the need arises.
                                                                  Pge 25.
