Patent application title: DOCUMENT SEARCH APPARATUS
Inventors:
Hitoshi Takeya (Tokyo, JP)
Oki Data Corporation
Assignees:
OKI DATA CORPORATION
IPC8 Class: AG06F1730FI
USPC Class:
707728
Class name: Post processing of search results ranking search results relevance of document based on features in query
Publication date: 2013-08-29
Patent application number: 20130226917
Abstract:
A document search apparatus for searching stored document data based on a
specified search word is provided with a search word input unit for
inputting the search word, a search processing unit for counting the
number of times the search word appears in the document data, and a
display control unit for changing, based on the number of times the
search word appears in the document data counted by the search processing
unit, an amount of the document data displayed. The document search
apparatus is capable of lightening the burden placed on the user in
finding the document of interest and in search refinement by displaying
the search result in an easy-to-see graphical layout.Claims:
1. A document search apparatus for searching stored document data based
on a specified search word, comprising: a search word input unit for
inputting the search word; a search processing unit for counting the
number of times the search word appears in the document data; and a
display control unit for changing, based on the number of times the
search word appears in the document data counted by the search processing
unit, an amount of the document data displayed.
2. The document search apparatus according to claim 1, wherein the display control unit increases the amount of the document data displayed as the number of times the search word appears in the document data increases.
3. The document search apparatus according to claim 1, further comprising a search result memory unit, wherein the search processing unit stores a filename corresponding to the document data and the number of times the search word appears in the document data.
4. The document search apparatus according to claim 1, further comprising a display unit, wherein the display control unit sorts displaying sequences of the searched document data based on the number of times the search word appears in the document data and renders the display unit display the search result.
5. The document search apparatus according to claim 4, wherein the display control unit makes the display sequences of the searched document data in a descending order according to the number of times the search word appears in the document data.
6. The document search apparatus according to claim 5, wherein the display control unit increases the amount of the document data displayed as the number of times the search word appears in the document data increases.
7. The document search apparatus according to claim 6, wherein the display control unit displays the filename corresponding to the document data as a search result.
8. The document search apparatus according to claim 7, wherein the display control unit further displays a portion of the document data as a search result.
9. The document search apparatus according to claim 8, wherein the display unit displays the amount of the document data in a manner depending on the font size of the filename.
10. The document search apparatus according to claim 9, wherein the display unit displays the amount of the document data, which sets a display size of the document data.
11. The document search apparatus according to claim 10, wherein the display amount of the document data depends on the number of the lines of the document data.
12. A document search method for searching stored document data based on a specified search word, comprising the steps of: inputting the search word; counting the number of times the search word appears in the document data; and changing an amount of the document data to be displayed, based on the number of times the search word appears in the document data.
Description:
CROSS-SECTION TO RELATED APPLICATIONS
[0001] This application is a continuation of copending U.S. application Ser. No. 12/171,412 filed Jul. 11, 2008, the entire disclosure of which is incorporated herein by reference.
BACKGROUND OF THE INVENTION
[0002] 1. Field of the Invention
[0003] This invention relates to a document search apparatus.
[0004] 2. Description of Related Art
[0005] In order to improve efficiency in creating a new document with word processing software on a computer, it is important to find and reuse a similar document which was made in the past. However, finding a document from among many files stored in a computer hard disk or a network storage is a time consuming job, and sometimes, a user fails to find the document that the user is looking for even though the user believes that the document should be stored somewhere in the hard disk.
[0006] Various document search technologies have been developed to search the document stored in a storage apparatus. For example, Japanese Patent Application Publication No. H7-175825 (hereinafter referred to as JP '825) discloses a document search and display method that aims to improve convenience in document searching and displaying. A document search device of JP '825 is targeted for searching structured documents, and takes advantage of structural information of the documents to display a part of the document in an attempt to improve the convenience in displaying the search results.
[0007] The document search device of JP '825 improves the convenience in displaying the search result, for example, by allowing the user to see a layout display of the search result. But the document search device of JP '825 still places a great burden on the user in finding a document of interest from among many items listed in the search results and in finally determining that the document of interest does not exist.
SUMMARY OF THE INVENTION
[0008] This invention is made in consideration of such problems as described above, and provides a document search apparatus capable of lightening the burden placed on the user in finding the document of interest and in search refinement by displaying the search result in an easy-to-see graphical layout.
[0009] In order to solve problems as described above, a document search apparatus for searching stored document data based on a specified search word is provided with a search word input unit for inputting the search word, a search processing unit for counting the number of times the search word appears in the document data, and a display control unit for changing, based on the number of times the search word appears in the document data counted by the search processing unit, an amount of the document data displayed.
[0010] In accordance with a preferred embodiment, when a search word is inputted to the search word input unit, the search processing unit searches the document data for counting the number of times the search word appears in the document data. The display control unit renders a display unit display the search result and changes an amount of the document data to be displayed based on the number of times the search word appears in the document data counted by the search processing unit.
DETAILED DESCRIPTION OF THE DRAWINGS
[0011] This invention may take physical form in certain parts and arrangements of parts, a preferred embodiment and method of which will be described in detail in this specification and illustrated in the accompanying drawings which form a part hereof, and wherein:
[0012] FIG. 1 is a block diagram describing a structure of an image forming apparatus according to the first embodiment;
[0013] FIG. 2 is an example of document data used in the image forming apparatus according to the first embodiment;
[0014] FIG. 3 is an example of the document data used in the image forming apparatus according to the first embodiment;
[0015] FIG. 4 is a flow chart describing a series of steps from a step for inputting the document data to the image forming apparatus via a data input unit to a step for outputting via a data output unit;
[0016] FIG. 5 is a diagram describing an example of field name-value pairs according to the first embodiment of this invention;
[0017] FIG. 6 is an example describing a document data file and field data stored in a data storage unit;
[0018] FIG. 7 is a flowchart describing a search processing and a display processing of the search result according to the first embodiment of this invention;
[0019] FIG. 8 is a figure describing an example of a search result according to the first embodiment of this invention;
[0020] FIG. 9 is a figure describing an example of the search result according to the first embodiment of this invention;
[0021] FIG. 10 is a block diagram describing a structure of the image forming apparatus according to the second embodiment;
[0022] FIG. 11 is a flow chart describing a series of steps from a step for inputting non-document data to the image forming apparatus via the data input unit to a step for outputting via the data output unit;
[0023] FIG. 12 is a flowchart describing steps executed by a corresponding template search unit according to the second embodiment of this invention;
[0024] FIG. 13 is a figure describing a preprocessing prior to calculating a degree of similarity according to the second embodiment;
[0025] FIG. 14 is a block diagram illustrating functions of a computer serving as a document search apparatus of the third embodiment;
[0026] FIG. 15 is a flowchart describing a search processing operation performed by a computer; and
[0027] FIG. 16 is a figure describing an example of a screen for showing search results on a display unit.
PREFERRED EMBODIMENTS
First Embodiment
[0028] FIG. 1 is a block diagram describing a structure of an image forming apparatus 100 having a search device according to the first embodiment. The image forming apparatus 100 such as an electrophotographic printer has a data input unit 101, a field extraction unit 103, a data storage unit 104, a data output unit 105, a data processing unit 106, a search word input unit 107, a search result output unit 108, a panel image processing unit 109, a search processing unit 110, an importance evaluation unit 111, a printing unit 120, and a field data template memory 121.
[0029] The data input unit 101 is a means to input a document file as document data into the image forming apparatus 100. The data input unit 101 communicates with a host computer as an information processing apparatus, a scanner, a data storage such as a USB (Universal Serial Bus) memory, a facsimile machine, and the like.
[0030] The field extraction unit 103 extracts field names and field values form the document data inputted from the data input unit 101 to generate field data.
[0031] The field data template memory 121 stores later-described field data templates.
[0032] The data storage unit 104 is a large capacity data storage such as a hard disk. The data storage unit 104 includes a field data storage unit 1041 and a document data storage unit 1042. The field data generated by the field extraction unit 103 is stored in the field data storage unit 1041, and the document data inputted to the data input unit 101 is stored in the document data storage unit 1042. The field data is associated with the corresponding document data, from which the field data has been extracted. The document data is also associated with the corresponding field data.
[0033] The data processing unit 106 performs a series of processings relating to image formation and the like from data input to data output.
[0034] The data output unit 105 outputs data and the like processed by the data processing unit 106. The data output unit 105 communicates with the host computer, the scanner, the data storage such as the USB (Universal Serial Bus) memory, the facsimile machine, and the like.
[0035] The printing unit 120 prints data and the like processed by the data processing unit 106.
[0036] The search word input unit 107 is a character input interface for inputting a search word. The search processing unit 110 searches the field data storage unit 1041 for the inputted search word. The search result output unit 108 displays a search result in response to a user's search request. In the present embodiment, the search word input unit 107 and the search result output unit 108 are composed of a touch panel, and for example, an LCD (Liquid Crystal Display) and the like.
[0037] The panel image processing unit 109 performs processings related to a search word input operation or a search result output operation on the touch panel.
[0038] The search processing unit 110 interprets the search request inputted by the user, and accesses the data storage unit 104 to retrieve from the data storage unit 104 the document file and the field data corresponding to the user's search request.
[0039] The importance evaluation unit 111 calculates an importance rank of each document file stored in the data storage unit 104. The importance rank is calculated from information such as time and date of printing, the number of printed copies, and the number of times the document file is selected from the search result. The more frequently the document file is printed or selected, the higher the importance rank thereof becomes.
[0040] A series of processings of the present embodiment is categorized into three stages, namely, a storing processing of the document data, a search processing, and a display processing of the search result. First, the storing processing of the document data will be described. The storing processing of the document data is performed every time the document data is inputted to the image forming apparatus 100. For example, a text file and a binary file generated by document generation software such as word processor software, spreadsheet software, and the like are the document data to be inputted to the image forming apparatus 100 via the data input unit 101 and stored in the data storage unit 104 of the image forming apparatus 100.
[0041] An example of a field data template is shown in FIG. 2. The field data templates are stored in the field data template memory 121. The document search device in the image forming apparatus 100 of this invention receives the document file, for example, a trip report as shown in FIG. 3 from the host computer (information processing apparatus) via the data input unit 101. The trip report in FIG. 3 includes field data pairs consisting of field names such as "destination" and field values such as "Shibaura" arranged in a table. The document file of the trip report in FIG. 3 is not a database file but a mere document file of word processor software. Thus, such field names and field values in the trip report are indicated by tables, spaces, tabs, or underlines, which can be easily recognized by human but are difficult to be accurately recognized by computer software. That is, it is difficult for the document search device to indentify which text strings are field names and which text strings are field values. To solve this problem, the document search device of this invention uses the field data template as shown in FIG. 2 to identify the positions of such field names and field values in the document file.
[0042] The trip report in the example of FIG. 2 has fields "destination", "purpose", "route", "expense", "the number of people on the trip", and "detail". In addition, the trip report has administrative field data "creation time", "modification time", and "creator", which serve as administrative information of the document. The user generates the document file such as the trip report as shown in FIG. 3 by filling out the field data template for the trip report as shown in FIG. 2.
[0043] FIG. 4 is a flowchart showing a series of steps from a step of input of the document data via the data input unit 101 of the image forming apparatus 100 to a step of output via the data output unit 105 or printing with the printing unit 120. First, when the user inputs the document data to the image forming apparatus 100, the field extraction unit 103 makes determination at S301 as to whether the document data includes the field data. Where the field extraction unit 103 determines that the document data includes the field names and field values (S301 YES), the field extraction unit 103 extracts the field names and field values by comparing the document file with the corresponding field data template memorized in the field data template memory 121, and generates field name-value pairs as shown in FIG. 5. Subsequently, the field extraction unit 103 generates the field data (FIG. 6) by combining a filename of the document data file, a template name, and the extracted field name-value pairs (S302).
[0044] The field data generated by the field extraction unit 103 is associated with the document file from which the field data are extracted, and the field data associated with the document file is stored to the field data storage unit 1041 (S304).
[0045] Subsequently, at S305, the importance evaluation unit 111 calculates the importance rank of each document file stored in the data storage unit 104. The importance rank is calculated from information such as time and date of printing, the number of printed copies, and the number of times the document file is selected from the search result.
[0046] Subsequently, the data processing unit 106 forms an image based on the inputted document data at S306, and outputs the formed image to the data output unit 105 or prints the formed image with the printing unit 120 (S307). Further, where the field extraction unit 103 determines that the document data does not include the field names or the field values (S301 NO), steps S302 to S305 are not executed, and instead, a data processing for direct output is performed.
[0047] The search processing and the display processing of the search result according to the present embodiment are hereinafter described with reference to a flowchart in FIG. 7. First, the user inputs the search word with the search word input unit 107 on the touch panel, or alternatively, with a keyboard of the host computer where the host computer is connected to the image forming apparatus 100 (S401). Operation with the search word input unit 107 on the touch panel is processed by the panel image processing unit 109. The search processing unit 110 receiving a search instruction from the panel image processing unit 109 searches the field data storage unit 1041 for the search word inputted to the search word input unit 107, and extracts the document data file associated with the search word.
[0048] In a case where the field value corresponding to the search word inputted by the user exists in the field data storage unit 1041 (S402 YES), the search processing unit 110 provides the panel image processing unit 109 with an instruction to display the search result on the touch panel. As shown in FIG. 8, the search result output unit 108 receiving the instruction to display the search result from the panel image processing unit 109 displays a search result screen showing the search result on the touch panel. At this moment, an axis selection box in the search result screen of FIG. 8 displays the selectable field names held in the field data storage unit 1041 (S403). The axis selection box can also display the selected field names. Then, the user selects one of the field names displayed in the axis selection box (S404). The user can simultaneously choose as much as two field names, that is, one field name for a vertical axis, and the other field name for a horizontal axis as shown in FIG. 8. If there still exists the selectable field name and if the number of the selected field names is less than two, the user is prompted to select the field name again (S405 YES).
[0049] When the fields are selected for the vertical axis and the horizontal axis, the storage data files are shown represented by icons in a coordinate system composed of the selected vertical axis and the selected horizontal axis as shown in FIG. 8. FIG. 8 shows the search result in a case where the entered search words, namely "trip report" and "Shibaura", are searched in the data storage unit 104. In FIG. 8, "destination", "purpose", "route," expense", "the number of people on the trip", "creation time", "modification time", and "creator" are displayed as the selectable field names for the axes, and the "creation time" and "modification time" are selected for the axes that are used for displaying the search result.
[0050] As shown in FIG. 8, each of the icons representing the document files is displayed in a different size. This is because the sizes of the icons are determined according to the importance rank calculated by the importance evaluation unit 111. The size of the icon of the document file becomes larger, as the importance rank thereof becomes higher. When the evaluation calculation is performed, the time and date of printing, the number of printed copies, and the number of times the document file is selected from the search result are taken into consideration as described above. In addition, the number of times the document file is listed up in the search result and the number of times the document file is selected for enlarged display are also considered. By clicking any one of the icons representing the document files, the document file is enlarged and displayed as shown in FIG. 9.
[0051] If there is not the selectable field name anymore or if the number of the selected field data is two, the search result is displayed in the coordinate system composed of the selected axes. Then, the user selects the document file to retrieve or print (S406), so that the selected document is outputted via the data output unit 105 or printed with the printing unit 120, and this series of steps is terminated (S407). Alternatively, in a case where the search word entered by the user at S402 does not exist in the field data in the field data storage unit, this series of steps is terminated without executing the steps S403 to S406 (S402 NO). However, in a case where a termination is not selected at step S407, this series of steps can be started from the step for the input of the search word (S407 NO).
[0052] As hereinabove described, the image forming apparatus according to the first embodiment of this invention enables the user to easily find the document data that the user is looking for because the user can easily find the document data from the search result displayed in an easy-to-see visual layout in the coordinate system composed of the vertical and horizontal axes. Any field can be chosen for the vertical or horizontal axis, and the icons representing the document files in the search result are two-dimensionally sorted based on the fields selected for the axes. In addition, the size of each icon representing the document file is enlarged or reduced according to the previously calculated importance rank. Therefore, the user can instinctively choose the document file that the user is looking for.
[0053] For example, when the user is looking for the trip report created in around July and vaguely remembers that the expense is rather expensive but does not remember other precise criteria for the search, the user can choose the field name "creation time" and "expense" as the vertical and horizontal axes in the coordinate system, and the user can enter "July" as the search word, so that the document files created in around July are displayed two-dimensionally sorted according to "creation time" and "expense". The user can compare the files with each other with respect to the creation time and the expense on the same search result screen. Thus, the user can find the document file that he is looking for, even where the user does not remember the exact criteria of the search.
Second Embodiment
[0054] In the second embodiment, image data such as in a TIFF (Tagged Image File Format) image is stored as non-document data, which the first embodiment does not target for storing. A structure of the image forming apparatus 200 according to the second embodiment is substantially the same as a structure of the image forming apparatus 100 according to the first embodiment. In the below description, units and operation similar to the first embodiment are omitted, and only different units or operation will be described below. FIG. 10 is a block diagram describing the structure of the image forming apparatus 200 having a search device according to the second embodiment. In addition to the image forming apparatus 100 of the first embodiment, the image forming apparatus 200 according to the second embodiment has a data judgment unit 112 and a corresponding template search unit 113.
[0055] The data judgment unit 112 makes a judgment as to whether data inputted by the user via the data input unit is the document data or the non-document data.
[0056] The corresponding template search unit 113 searches the field data template memory 121 of the image forming apparatus 200 for the field data template corresponding to the inputted non-document data.
[0057] FIG. 11 is a flowchart describing a series of steps from the step of the input of the non-document data via the data input unit 101 of the image forming apparatus 100 to the step of the output via the data output unit 105 or printing with the print unit 120. First, when the user inputs the data to the image forming apparatus 200, the data judgment unit 112 judges as to whether the inputted data is the document data or the non-document data. Where the document data is inputted to the image forming apparatus 200, the image forming apparatus 200 operates in the same way as the image forming apparatus 100 of the first embodiment, that is, the image forming apparatus 200 executes the storing of the document data, the searching of the document data, and the displaying of the search result in the same way as the first embodiment.
[0058] Where the inputted data is judged to be the non-document data, the corresponding template search unit 113 searches the field data template memory 121 for the corresponding field data template (S501). FIG. 12 is a flowchart describing the steps executed by the corresponding template search unit 113. When the non-document data is inputted, the corresponding template search unit 113 supplies the data to an OCR (Optical Character Reader) to extract character string data from the image data (S601). Subsequently, the corresponding template search unit 113 determines a degree of matching between the extracted character string data and the field data template by searching the extracted character string data for each of the field names in the field data template. The degree of matching is determined for each of the field data templates in the field data template memory 121 (S602). The degree of matching is a value calculated by dividing the number of the matched field names in the field data template by the number of total field names in the field data template. A predetermined number of the field date templates having highest and subsequent degrees of matching are chosen as the template candidates in the order of the degree of matching (S603).
[0059] The degree of matching is calculated as described above, and the several template candidates are chosen. However, the template candidate having a high value in the degree of matching may not be necessarily be the field data template matching to the inputted non-document data, because the degree of matching is calculated based on only the number of the matching field names, that is, the degree of matching does not reflect the position where the matching field name appears in the template candidate. Accordingly, the degree of matching may become high in a case where the non-document data having many words that happen to be the same as the field names in the template candidate even though the non-document data has nothing to do with the template candidate. Thus, in addition to the degree of matching, a degree of similarity based on bit map comparison is calculated as described below.
[0060] To calculate the degree of similarity based on the bit map comparison, the template candidates are converted into the same image format as the inputted non-document data. The degree of similarity is calculated between the inputted non-document data and each of the template candidates. As a preprocessing prior to a calculation of the degree of similarity, a processing is executed to exclude the field value area from the target of calculation of the degree of similarity as shown in FIG. 13. As shown in FIG. 13, an area for the field values are excluded from the bit map comparison because the field values vary from document to document and this field value area is not needed to be compared for the determination of the corresponding field date template. For example, the degree of similarity is calculated as follows. The corresponding template search unit 113 generates binary bitmap images of the non-document data and the template candidate, and compares each bit therebetween as to whether the bit matches or not. The degree of similarity is calculated by dividing the number of matched bits by the total number of the bits (S604).
[0061] A criteria for evaluating an absolute level of similarity is required in the above-described calculation of the degree of similarity, but such absolute criteria is difficult to be determined because the inputted non-document data may be an scanned image having smears and deformed characters thereon or a facsimile-received image. Accordingly, instead of determining an absolute threshold value, a document randomly selected is used for comparison. The degree of similarity is calculated between the randomly selected document and each of the template candidates, and the degree of similarity therebetween having the largest value is determined as a threshold value. The template candidate having the degree of similarity less than the determined threshold value is excluded from the template candidates. In this way, the template candidate having matching words therein but is different in image is excluded from the template candidates.
[0062] Then, for each of the template candidates, a total degree is calculated by adding the degree of matching thereof and the degree of similarity thereof. The template candidate having the largest total degree is determined as the field data template corresponding to the non-document data (S502 YES). Then, the field extraction unit 103 extracts the field name-value pairs as text string data from the non-document data upon converting an image of the non-document data at a positions corresponding to positions of the field name-value pairs in the determined field data template. In a case where the text strings do not exist at the position corresponding to the field name-value pairs, a null character is extracted.
[0063] The extracted field name-value pairs are stored to the field data storage unit 1041 in the data storage unit 104, and the inputted non-document data is stored in the document data storage unit 1042 in the data storage unit 104 (S505). The field name-value pairs are associated with the corresponding non-document data, from which the field name-value pairs have been extracted. The non-document data is also associated with the corresponding field name-value pairs.
[0064] Subsequently, the importance evaluation unit 111 calculates the importance rank of the non-document stored in the data storage unit 104 at step S506. The importance rank is calculated from information such as time and date of printing, the number of printed copies, and the number of times the document file is selected from the search result.
[0065] At step S507, the data processing unit 106 performs image formation and the like according to an output format of the inputted non-document data, and outputs the non-document data via the data output unit 105 or print the non-document data with the printing unit 120 (S508). In a case where the corresponding field data template cannot be found (S502 NO), steps S505 to S507 are not executed, and instead, the data processing for direct output is performed.
[0066] The search processing and the display processing of the present embodiment are the same as those of the first embodiment, but the search result can be classified into either the document file or the non-document file. When the search results shows both of the document files and the non-document files in a same screen, the size of the icon of each of the non-document files is determined according to the importance rank of the non-document data that is calculated by multiplying the importance rank by the degree of similarity as weighing. The importance rank is calculated as described in the first embodiment. The degree of similarity is described above in the present embodiment.
[0067] As described above, the image forming apparatus according to the second embodiment can store the non-document data in addition to the document data. In the present embodiment, the corresponding template search unit 113 can extract the text string data from the non-document data with the OCR, and searches the stored field data templates for the template candidates corresponding to the inputted non-document data based on the field name-value pairs. Furthermore, the corresponding template search unit 113 excludes the template candidates having the degree of similarity less than the threshold value calculated as described above, so as to make sure that the non-document data not corresponding to any of the field data templates is not stored to the file storage unit. Furthermore, in the search result of the image forming apparatus according to the present embodiment, the document data and the non-document data can be displayed on a same screen in a classified manner. Alternatively, only the document data or only the non-document data can be displayed.
Third Embodiment
[0068] FIG. 14 is a block diagram illustrating functions of a computer 400 serving as a document search apparatus used in the present embodiment.
[0069] A document data input unit 201 inputs document data to be searched. Specifically, the document data input unit 201 is an input device such as a keyboard and a mouse or an interface device for inputting the document data as digital data from the outside of the computer 400.
[0070] A document data memory unit 202 is a memory unit storing the document data inputted from the document data input unit 201. Specifically, the document data memory unit 202 is a hard disk device and the like. In the present embodiment, the document data is stored therein with a unique filename given to each document data.
[0071] A search criteria input unit 203 inputs search criteria such as a search word and a search range for searching. Specifically, the search criteria input unit 203 is a keyboard, a touch panel, and the like.
[0072] A search processing unit 204 compares the document data memorized in the document data memory unit 202 with the search word inputted from the search criteria input unit 203 to count the number of the search words included in each of the searched document data. Specifically, the search processing unit 204 is, e.g., a program memorized in a ROM and a CPU executing the program.
[0073] A search result memory unit 205 is a memory unit storing a result of a search performed by the search processing unit 204. Specifically, the search result memory unit 205 is a semiconductor memory device such as a hard disk, a RAM, and the like. In the present embodiment, the unique filename of each document file is associated with the number of the search words included in the document file, and the filenames associated therewith and the number of the search words are memorized in the search result memory unit 205.
[0074] A display unit 207 is a unit for displaying the search result. For example, a display device such as an LCD, a CRT, and the like can be used as the display unit 207.
[0075] A display control unit 206 is a unit that uses the search result memorized in the search result memory unit 205 to process a display format according which the document data memorized in the document data memory unit 202 is displayed on the display unit 207.
[0076] FIG. 15 is a flowchart describing a search processing operation performed by the computer 400. It is assumed here that one or multiple document data to be searched is previously memorized in the document data memory unit 202 via the document data input unit 201.
[0077] The search processing operation of the computer 400 will be hereinafter described.
[0078] (Step S700)
[0079] First, a user, who does a search, uses the search criteria input unit 203 to input the search word serving as a key to perform the search.
[0080] (Step S701)
[0081] Subsequently, with the search criteria input unit 203, the user specifies document data to be searched from among the document data memorized in the document data memory unit 202.
[0082] (Step S702)
[0083] When a search start instruction is inputted from the search criteria input unit 203, the search processing unit 204 begins the search.
[0084] (Step S703)
[0085] The search processing unit 204 reads one document data from among the to-be-searched document data memorized in the document data memory unit, and counts the number of the search words included in the document data.
[0086] (Step S704)
[0087] The search processing unit 204 associates the filename of the document data, in which the search words have been counted, with the number of search words included in the document file, and memorizes the filename associated therewith and the number of search words in the search result memory unit 205.
[0088] (Step S705)
[0089] The search processing unit 204 makes a determination as to whether all of the to-be-searched document files are finished being searched. In a case where all of them are finished being searched, step S706 is proceeded to. In a case where all of them are not yet finished being searched, step S703 is returned to, and another document file is read.
[0090] (Step S706)
[0091] The display control unit 206 sorts the filenames memorized in the search result memory unit 205 in descending order by the number of the search words included therein.
[0092] (Step S707)
[0093] The display control unit 206 reads the document data corresponding to each of the sorted filenames from the document data memory unit 202, and displays portions of the document data together with the filenames on the display unit 207. The amount of the portion of the document data displayed on the display unit 207 is changed and outputted according to the number of the search words included in the document data and the ranking in the sorted order.
[0094] FIG. 16 is a figure describing an example of a screen for showing the search result on the display unit 220.
[0095] A field 301 displays the filename of the found document file and a ranking number in the sorted order.
[0096] A field 302 displays a portion of a content of the document data corresponding to the filename displayed in the field 301.
[0097] The field 301 and the field 302 are displayed as a pair for each found document file.
[0098] In FIG. 16, the document file including the largest number of the search words is "FILE0001". In the field 302 corresponding to "FILE0001", three lines of the document data are displayed.
[0099] The document file including the second largest number of the search words is "FILE0005". In the field 302 corresponding to "FILE0005", two lines of the document data are displayed.
[0100] In the present embodiment, a font size of the filename displayed in the field 301 serving as a title may be changed according the amount of the document data displayed in the field 302 as shown in FIG. 16.
[0101] Although it is assumed in the above description that the document data to be searched is in the document data memory unit 202 in the computer 400, the document data to be searched may be intensively stored in a single file server or distributed and stored in multiple file servers connected with the computer 400 via a network.
[0102] As hereinabove described, the amount of the displayed document data can be changed according to the number of the search words included in the found document data, and thus, a user can view the content of the useful document data in detail from among the search result.
[0103] The embodiments described above are preferred embodiments of this invention, but this invention is not limited thereto, and each structure can be changed as necessary as long as not deviating from the scope of the gist of this invention.
[0104] The foregoing description of preferred embodiments of the invention has been presented for purposes of illustration and description, and is not intended to be exhaustive or to limit the invention to the precise form disclosed. The description was selected to best explain the principles of the invention and their practical application to enable others skilled in the art to best utilize the invention in various embodiments and various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention should not be limited by the specification, but be defined by the claims set forth below.
User Contributions:
Comment about this patent or add new information about this topic:
People who visited this patent also read: | |
Patent application number | Title |
---|---|
20210395964 | SOUND WALL AND METHOD OF CONSTRUCTING THE SAME |
20210395963 | MICROTRENCHING SYSTEM HAVING A VACUUM HOSE SUPPORT AND METHOD OF MICROTRENCHING |
20210395962 | ROTOR DEPTH VISUAL INDICATION ZONES |
20210395961 | METHOD AND APPARATUS FOR DETECTING A DATA SET |
20210395960 | SLIPFORM CONCRETE PAVING MACHINE HAVING DOWEL BAR INSERTER MECHANISM WITH CONTINUOUS FLOOR |