Vishv Malhotra | IIT Guwahati (original) (raw)
Papers by Vishv Malhotra
Web Information Systems and Technologies, 2006
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the user's information needs is synthesized from these text documents. The methodology provides a way for creating web search query that matches the user's information need even when the user may have difficulty in doing so directly due to lack of experience in the query design or lack of familiarity of the search domain. A user survey has shown that the algorithmically formed query has recall coverage and precision characteristics better than those achieved by the experienced human web searchers.
Web Information Systems and Technologies, 2006
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the
This paper describes a distributed system which aggregates the unused and usually wasted processi... more This paper describes a distributed system which aggregates the unused and usually wasted processing capacity of idle workstations. The aggregation is achieved through the use of now ubiquitous internet infrastructure and web technology. And, it delivers a powerful yet inexpensive execution environment for computationally intensive applications. The prototype system described here makes use of Sun Microsystems Jini technology, particularly JavaSpaces, along with Java Web Start, to produce a dynamic, flexible and reliable system. Two example applications used to evaluate the system are described: (a) the n-Queens problem and (b) a parallel sorting (shearsort) application. The results of the evaluation clearly show that, for certain classes of applications, the system is capable of delivering significant performance.
Lecture Notes in Computer Science, 1986
ABSTRACT A new technique for automatic retargetable code generation is presented. This technique ... more ABSTRACT A new technique for automatic retargetable code generation is presented. This technique is based on the observation that a reduced instruction set subset can always be found for a target machine instruction set; the rest of the instructions are seen to be equivalent to sequence of the reduced instruction set instructions. Given a target machine description, a method is presented which partitions it into two: one partition contains the reduced instruction set; the other contains the rest of the instructions, along with corresponding simulation sequences — these simulation sequences are unique with respect to the partition. In view of the above, the code generation problem, more precisely the instruction selection problem, is seen as a sequence of two phases: The first phase generates reduced instruction set code; the second phase finds opportunities for optimization through compaction of sub sequences in the generated code, to complex instructions. The full capability of the target machine can be potentially utilized due to the latter phase. Retargetability is achieved by automatically generating the above phases from the information in the partition. The phases are based on tree-pattern-matching, implemented through LR parsing. The technique has been tested for the MC 68000; simple comparisons show the generated code to be satisfactory.
Lecture Notes in Business Information Processing, 2007
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the user's information needs is synthesized from these text documents. The methodology provides a way for creating web search query that matches the user's information need even when the user may have difficulty in doing so directly due to lack of experience in the query design or lack of familiarity of the search domain. A user survey has shown that the algorithmically formed query has recall coverage and precision characteristics better than those achieved by the experienced human web searchers.
Proceedings of the American Society for Information Science and Technology, 2009
Interactivity as a way to help users understand and/or explore complex information issues has bee... more Interactivity as a way to help users understand and/or explore complex information issues has been utilized successfully in information science areas such as information retrieval. In this paper we report experiences with utilizing interactivity in a different albeit related context: informing users about spam filtering processes. Spam filters that prevent spam from showing up in inboxes by filtering incoming messages according to their suspected level of "spamminess" are readily available. However, the complexity of modern spam filters aggregating evidence from various sources means few non-expert users of spam filters actually understand what their spam filters are doing. It also remains unclear how different spam filter settings would impact on what they see in their email inboxes. In this paper, we describe work into using interactive exploration as a way to increase users' understanding of how spam filters work. Lab-based evaluation of a prototype suggests a considerable increase in spam filtering understanding occurred even among subjects that stated they had prior spam filtering expertise.
Proceedings of IEEE TENCON '98. IEEE Region 10 International Conference on Global Connectivity in Energy, Computer, Communication and Control (Cat. No.98CH36229), 1998
Because of the mobility of the machines, it is difficult to provide a satisfactory solution to th... more Because of the mobility of the machines, it is difficult to provide a satisfactory solution to the problem of resource and load sharing. Mobile computers can change their position from time to time. In addition, their platforms are not only heterogeneous but also dynamic. This makes the resource and load sharing problem more difficult to tackle in such environments than
Revenue collection services are deemed essential for attracting quality services on the networks.... more Revenue collection services are deemed essential for attracting quality services on the networks. Generally these collections will be of small amounts but will be frequent. A water-tight authorisation check is not costeffective in such an environment. This paper introduces an inexpensive, robust and fast scheme for authorising the credit usage. The scheme is especially suited to small credit brokers as it does not involve any special hardware.
An algorithm to synthesise a web search query from example documents is described. A user searchi... more An algorithm to synthesise a web search query from example documents is described. A user searching for information on the Web can use a rudimentary query to locate a set of potentially relevant documents. The user classifies the retrieved documents as being relevant or irrelevant to his or her needs. A query can be synthesised from these categorised documents to perform a definitive search with good recall and precision characteristics.
Popular web search engines use Boolean queries as their main interface for users to search their ... more Popular web search engines use Boolean queries as their main interface for users to search their information needs. The paper presents results a user survey employing volunteer web searchers to determine the effectiveness of the Boolean queries in meeting the information needs. A metric for measuring the quality of a web search query is presented. This enables us to relate attributes of the search session and the Boolean query with its success. Certain easily identified characteristics of a good web search query are identified.
Lecture Notes in Computer Science, 1987
ABSTRACT Without Abstract
Information Processing Letters, 1988
This paper describes a distributed system which aggregates the unused and usually wasted processi... more This paper describes a distributed system which aggregates the unused and usually wasted processing capacity of idle workstations. The aggregation is achieved through the use of now ubiquitous internet infrastructure and web technology. And, it delivers a powerful yet inexpensive execution environment for computationally intensive applications. The prototype system described here makes use of Sun Microsystems Jini technology, particularly JavaSpaces, along with Java Web Start, to produce a dynamic, flexible and reliable system. Two example applications used to evaluate the system are described: (a) the n-Queens problem and (b) a parallel sorting (shearsort) application. The results of the evaluation clearly show that, for certain classes of applications, the system is capable of delivering significant performance.
Proceedings of the 1981 conference on …, 1981
This paper describes an algorithm whereby an initial, naïve user query to a search engine can be ... more This paper describes an algorithm whereby an initial, naïve user query to a search engine can be subsequently refined to improve both its recall and precision. This is achieved by manually classifying the documents retrieved by the original query into relevant and irrelevant categories, and then finding additional Boolean terms which successfully discriminate between these categories. Latent semantic analysis is used to weight the choice of these extra search terms to make the resulting queries more intuitive to users.
... O020-019O/88/S3.5O © 1988, Elsevier Science Publishers BV (North-Holland) 249 ... PROCESSING ... more ... O020-019O/88/S3.5O © 1988, Elsevier Science Publishers BV (North-Holland) 249 ... PROCESSING LETTERS 28 April 1988 Acknowledgment We gratefully acknowledge the suggestions made by David Gries, Christian Codognet, Philippe Codognet and Vipin Kumar to improve ...
Utilizing interactive visualizations of data sets as a way of helping users explore and understan... more Utilizing interactive visualizations of data sets as a way of helping users explore and understand complex relationships in data has a long tradition in Human Computer Interaction, dating back at least to dynamic queries and starfield displays developed in the early Nineties. In this paper we discuss how interactive visualizations can be used to help non-expert users understand how different spam filter settings impact on what they will see in their email inboxes. The key idea is that users can visually explore the positive or ...
Web Information Systems and Technologies, 2006
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the user's information needs is synthesized from these text documents. The methodology provides a way for creating web search query that matches the user's information need even when the user may have difficulty in doing so directly due to lack of experience in the query design or lack of familiarity of the search domain. A user survey has shown that the algorithmically formed query has recall coverage and precision characteristics better than those achieved by the experienced human web searchers.
Web Information Systems and Technologies, 2006
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the
This paper describes a distributed system which aggregates the unused and usually wasted processi... more This paper describes a distributed system which aggregates the unused and usually wasted processing capacity of idle workstations. The aggregation is achieved through the use of now ubiquitous internet infrastructure and web technology. And, it delivers a powerful yet inexpensive execution environment for computationally intensive applications. The prototype system described here makes use of Sun Microsystems Jini technology, particularly JavaSpaces, along with Java Web Start, to produce a dynamic, flexible and reliable system. Two example applications used to evaluate the system are described: (a) the n-Queens problem and (b) a parallel sorting (shearsort) application. The results of the evaluation clearly show that, for certain classes of applications, the system is capable of delivering significant performance.
Lecture Notes in Computer Science, 1986
ABSTRACT A new technique for automatic retargetable code generation is presented. This technique ... more ABSTRACT A new technique for automatic retargetable code generation is presented. This technique is based on the observation that a reduced instruction set subset can always be found for a target machine instruction set; the rest of the instructions are seen to be equivalent to sequence of the reduced instruction set instructions. Given a target machine description, a method is presented which partitions it into two: one partition contains the reduced instruction set; the other contains the rest of the instructions, along with corresponding simulation sequences — these simulation sequences are unique with respect to the partition. In view of the above, the code generation problem, more precisely the instruction selection problem, is seen as a sequence of two phases: The first phase generates reduced instruction set code; the second phase finds opportunities for optimization through compaction of sub sequences in the generated code, to complex instructions. The full capability of the target machine can be potentially utilized due to the latter phase. Retargetability is achieved by automatically generating the above phases from the information in the partition. The phases are based on tree-pattern-matching, implemented through LR parsing. The technique has been tested for the MC 68000; simple comparisons show the generated code to be satisfactory.
Lecture Notes in Business Information Processing, 2007
This paper presents an algorithm to improve a web search query based on the feedback on the viewe... more This paper presents an algorithm to improve a web search query based on the feedback on the viewed documents. A user who is searching for information on the Web marks the retrieved (viewed) documents as relevant or irrelevant to further expose the information needs expressed in the original query. A new web search query matching this improved understanding of the user's information needs is synthesized from these text documents. The methodology provides a way for creating web search query that matches the user's information need even when the user may have difficulty in doing so directly due to lack of experience in the query design or lack of familiarity of the search domain. A user survey has shown that the algorithmically formed query has recall coverage and precision characteristics better than those achieved by the experienced human web searchers.
Proceedings of the American Society for Information Science and Technology, 2009
Interactivity as a way to help users understand and/or explore complex information issues has bee... more Interactivity as a way to help users understand and/or explore complex information issues has been utilized successfully in information science areas such as information retrieval. In this paper we report experiences with utilizing interactivity in a different albeit related context: informing users about spam filtering processes. Spam filters that prevent spam from showing up in inboxes by filtering incoming messages according to their suspected level of "spamminess" are readily available. However, the complexity of modern spam filters aggregating evidence from various sources means few non-expert users of spam filters actually understand what their spam filters are doing. It also remains unclear how different spam filter settings would impact on what they see in their email inboxes. In this paper, we describe work into using interactive exploration as a way to increase users' understanding of how spam filters work. Lab-based evaluation of a prototype suggests a considerable increase in spam filtering understanding occurred even among subjects that stated they had prior spam filtering expertise.
Proceedings of IEEE TENCON '98. IEEE Region 10 International Conference on Global Connectivity in Energy, Computer, Communication and Control (Cat. No.98CH36229), 1998
Because of the mobility of the machines, it is difficult to provide a satisfactory solution to th... more Because of the mobility of the machines, it is difficult to provide a satisfactory solution to the problem of resource and load sharing. Mobile computers can change their position from time to time. In addition, their platforms are not only heterogeneous but also dynamic. This makes the resource and load sharing problem more difficult to tackle in such environments than
Revenue collection services are deemed essential for attracting quality services on the networks.... more Revenue collection services are deemed essential for attracting quality services on the networks. Generally these collections will be of small amounts but will be frequent. A water-tight authorisation check is not costeffective in such an environment. This paper introduces an inexpensive, robust and fast scheme for authorising the credit usage. The scheme is especially suited to small credit brokers as it does not involve any special hardware.
An algorithm to synthesise a web search query from example documents is described. A user searchi... more An algorithm to synthesise a web search query from example documents is described. A user searching for information on the Web can use a rudimentary query to locate a set of potentially relevant documents. The user classifies the retrieved documents as being relevant or irrelevant to his or her needs. A query can be synthesised from these categorised documents to perform a definitive search with good recall and precision characteristics.
Popular web search engines use Boolean queries as their main interface for users to search their ... more Popular web search engines use Boolean queries as their main interface for users to search their information needs. The paper presents results a user survey employing volunteer web searchers to determine the effectiveness of the Boolean queries in meeting the information needs. A metric for measuring the quality of a web search query is presented. This enables us to relate attributes of the search session and the Boolean query with its success. Certain easily identified characteristics of a good web search query are identified.
Lecture Notes in Computer Science, 1987
ABSTRACT Without Abstract
Information Processing Letters, 1988
This paper describes a distributed system which aggregates the unused and usually wasted processi... more This paper describes a distributed system which aggregates the unused and usually wasted processing capacity of idle workstations. The aggregation is achieved through the use of now ubiquitous internet infrastructure and web technology. And, it delivers a powerful yet inexpensive execution environment for computationally intensive applications. The prototype system described here makes use of Sun Microsystems Jini technology, particularly JavaSpaces, along with Java Web Start, to produce a dynamic, flexible and reliable system. Two example applications used to evaluate the system are described: (a) the n-Queens problem and (b) a parallel sorting (shearsort) application. The results of the evaluation clearly show that, for certain classes of applications, the system is capable of delivering significant performance.
Proceedings of the 1981 conference on …, 1981
This paper describes an algorithm whereby an initial, naïve user query to a search engine can be ... more This paper describes an algorithm whereby an initial, naïve user query to a search engine can be subsequently refined to improve both its recall and precision. This is achieved by manually classifying the documents retrieved by the original query into relevant and irrelevant categories, and then finding additional Boolean terms which successfully discriminate between these categories. Latent semantic analysis is used to weight the choice of these extra search terms to make the resulting queries more intuitive to users.
... O020-019O/88/S3.5O © 1988, Elsevier Science Publishers BV (North-Holland) 249 ... PROCESSING ... more ... O020-019O/88/S3.5O © 1988, Elsevier Science Publishers BV (North-Holland) 249 ... PROCESSING LETTERS 28 April 1988 Acknowledgment We gratefully acknowledge the suggestions made by David Gries, Christian Codognet, Philippe Codognet and Vipin Kumar to improve ...
Utilizing interactive visualizations of data sets as a way of helping users explore and understan... more Utilizing interactive visualizations of data sets as a way of helping users explore and understand complex relationships in data has a long tradition in Human Computer Interaction, dating back at least to dynamic queries and starfield displays developed in the early Nineties. In this paper we discuss how interactive visualizations can be used to help non-expert users understand how different spam filter settings impact on what they will see in their email inboxes. The key idea is that users can visually explore the positive or ...