Web Caching Research Papers - Academia.edu (original) (raw)

Web caching has been proposed as an effective solution to the problems of network traffic and congestion, Web objects access and Web load balancing. This paper presents a model for optimizing Web cache content by applying either a genetic... more

Web caching has been proposed as an effective solution to the problems of network traffic and
congestion, Web objects access and Web load balancing. This paper presents a model for optimizing Web cache
content by applying either a genetic algorithm or an evolutionary programming scheme for Web cache content
replacement. Three policies are proposed for each of the genetic algorithm and the evolutionary programming
techniques, in relation to objects staleness factors and retrieval rates. A simulation model is developed and long term
trace-driven simulation is used to experiment on the proposed techniques. The results indicate that all evolutionary
techniques are beneficial to the cache replacement, compared to the conventional replacement applied in most
Web cache server. Under an appropriate objective function the genetic algorithm has been proven to be the best
of all approaches with respect to cache hit and byte hit ratios

Internet is the boon in modern era as every organization uses it for dissemination of information and e-commerce related applications. Sometimes people of organization feel delay while accessing internet in spite of proper bandwidth.... more

Internet is the boon in modern era as every organization uses it for dissemination of information and e-commerce related applications. Sometimes people of organization feel delay while accessing internet in spite of proper bandwidth. Prediction model of web caching and prefetching is an ideal solution of this delay problem. Prediction model analysing history of internet user from server raw log files and determine future sequence of web objects and placed all web objects to nearer to the user so access latency could be reduced to some extent and problem of delay is to be solved. To determine sequence of future web objects, it is necessary to determine proximity of one web object with other by identifying proper distance metric technique related to web caching and prefetching. This paper studies different distance metric techniques and concludes that bio informatics based distance metric techniques are ideal in context to Web Caching and Web Prefetching.

This study deals with the problem of maintaining large hierarchy folder replicated in a distributed environment. It was found that this problem afflicted a number of important applications, such as synchronization of Hierarchy Folders... more

This study deals with the problem of maintaining large hierarchy folder replicated in a distributed environment. It was found that this problem afflicted a number of important applications, such as synchronization of Hierarchy Folders between Peer-to-peer environments, synchronization of data between accounts or devices, content distribution and web caching networks, web site mirroring, storage networks, and large scale web search and mining. At the core of the problem lay the File in Hierarchy Folder synchronization challenge. This challenge posed the question; “Given two versions of files inside Folders on different machines, call outdated hierarchy and a current one, how can we update the outdated version with a minimum communication cost, by exploiting the significant similarity between the versions?” Although a popular open source tool for this problem called RSYNC is being used in hundreds of thousands of servers around the world, only very few attempts have been made to impro...

Designing E-Business applications in an efficient way has become a competitive necessity rather than a competitive advantage. One of the most important goals for many organi- zations is to satisfy their clients’... more

Designing E-Business applications in an efficient
way has become a competitive necessity rather than a competitive
advantage. One of the most important goals for many organi-
zations is to satisfy their clients’ service level agreements with
respect to the response time and throughput. Adopting Service
Oriented Architecture (SOA) during design and implementation
promotes communication with the external and internal business
entities. Web services are one of the popular technologies to
achieve SOA solutions. Lookup web services are broadly used
by many service consumers to fetch data which are used by their applications. In this paper we focus on how to efficiently build lookup web services using design patterns. Our goal is to improve the response time (latency) and throughput of lookup web services

Permission to make digital or hard copies of portions of this work for personal or classroom use is granted provided that the copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the... more

Permission to make digital or hard copies of portions of this work for personal or classroom use is granted provided that the copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise requires prior specific permission by the publisher mentioned above.

Web caching has been considered as a powerful solution to deal with the growth of Web traffic. Several studies have shown that caching documents throughout the Internet can save network bandwidth and reduce document access latency.... more

Web caching has been considered as a powerful solution to deal with the growth of Web traffic. Several studies have shown that caching documents throughout the Internet can save network bandwidth and reduce document access latency. However, this technique has introduced new problems such as maintaining the document coherency and selecting the next document to be removed. With the continuous increase in demand for documents, the Web cache servers are becoming the new bottleneck. A need for better resource management is becoming urgent in order to reduce the overhead sustained by Web cache servers. In this paper, a number of Web replacement policies are discussed and compared on the basis of trace-driven simulations. The impact of the Web cache server configuration is pointed out through a set of experiments that use the cache size as a tuning parameter

Web caching is a well-known strategy for improving performance of Web-based system by keeping web objects that are likely to be used in the near future close to the client. Most of the current Web browsers still employ traditional caching... more

Web caching is a well-known strategy for improving performance of Web-based system by keeping web objects that are likely to be used in the near future close
to the client. Most of the current Web browsers still employ traditional caching policies that are not efficient in web caching. This research proposes a splitting
browser cache to two caches, instant cache and durable cache. Initially, a web object is stored in instant cache, and the web objects that are visited more than
the pre-specified threshold value will be moved to durable cache. Other objects are removed by Least Recently Used (LRU) algorithm as instant cache is full. More
significantly, when the durable cache saturates, a fuzzy system is employed in classifying each object stored in durable cache into either cacheable or
uncacheable object. The old uncacheable objects are candidate for removing from the durable cache. By implementing this mechanism, the cache pollution can
be mitigated and the cache space can be utilized effectively. Experimental results have revealed that the proposed approach can improve the performance up to
14.8% and 17.9% in terms of hit ratio (HR) compared to LRU and Least Frequently Used (LFU). In terms of byte hit ratio (BHR), the performance is improved up to
2.57%, compared to LRU and LFU.

Amazon Web Services(AWS) provides various solutions for small, medium and enterprise level organizations. These solutions cover a wide range of problems ranging from hosting a website, computing a complex AI algorithm, storing terabytes... more

Amazon Web Services(AWS) provides various solutions for small, medium and enterprise level organizations. These solutions cover a wide range of problems ranging from hosting a website, computing a complex AI algorithm, storing terabytes of data. Depending on the volume of data and rate at which it is requested, different caching solutions may be needed. AWS provides a wide range of caching solutions at different stages of architecture, this article will focus on what is caching and different caching strategies. We will conclude the article by providing some of industry standard best practices for using caching in your architecture.

Abstract. This work presents an extensive evaluation of the request filtering in hierarchy of proxy caches. Using the recently proposed ADF (Aggregation, Disaggregation and Filtering) model as well as entropy as metric for Web traffic... more

Abstract. This work presents an extensive evaluation of the request filtering in hierarchy of proxy caches. Using the recently proposed ADF (Aggregation, Disaggregation and Filtering) model as well as entropy as metric for Web traffic characterization, we evaluate how locality of reference changes as the streams of requests pass through a hierarchy of caches. Moreover, we propose the use of average entropy for comparing the locality of reference of different streams and present how a proxy server can dynamically calculate the entropy of ...

Web caching has received considerable attention in the past few years due to its effectiveness in reducing client response time and network traffic. Conventional memory caching algorithms have been proved to be insufficient in the context... more

Web caching has received considerable attention in the past few years due to its effectiveness in reducing client response time and network traffic. Conventional memory caching algorithms have been proved to be insufficient in the context of web caching because of variable ...

The World Wide Web has come to be a great part of our daily life, yet user observed latency is still a problem that needs a proper means of handling. Even though earlier attempts focused on caching as the chief solution to tackling this... more

The World Wide Web has come to be a great part of our daily life, yet user observed latency is still a problem that needs a proper means of handling. Even though earlier attempts focused on caching as the chief solution to tackling this issue, its success was extremely limited. Prefetching has come to be the primary technique in supplementing caching towards soothing the latency problem associated with the contemporary Internet.
However, existing approaches in prefetching are extremely limited in their ability to employ application level web document relationship which is often visible only to the content developer. This is because most approaches are access history based schemes that make future users’ access prediction only based on past user access. Attempts to incorporate prefetching schemes that utilize semantic information with those that use users past access history are extremely limited in their extensibility. In this work we present a novel framework that enables integration of schemes from both worlds of prefetching (i.e., history based and semantic schemes) without the need for a major modification to the algorithms. When there is a need/possibility to capture new application level context, a new algorithm could be developed to do so and then it can be integrated into the framework.
Since each participating scheme is merely viewed as an algorithm that produces a list of candidate objects that are likely to be accessed in the near future, the framework can entertain any one of the existing prefetching schemes. With its adaptive weight management technique the framework adjusts the effect of each algorithm in the overall prediction to parallel with its observed performance so far.
We have found this formwork to be less aggressive than its contemporary counterparts which is extremely important for resource constrained mobile devices that have come to be the major means of access by users of the current web.

The benefits of Web cache prefetching are well understood, and so prefetching has been implemented in a number of commercial products. This paper argues that the current support for prefetching in HTTP/1.1 is insufficient because... more

The benefits of Web cache prefetching are well understood, and so prefetching has been implemented in a number of commercial products. This paper argues that the current support for prefetching in HTTP/1.1 is insufficient because prefetching with GET is not good.Existing prefetching implementations can cause problems with undesirable side-effects and server abuse, and the potential for these problems may thwart additional prefetching development and deployment. We make some initial suggestions of extensions to HTTP that would allow for safe prefetching, reduced server abuse, and differentiated Web server quality of service. It is our hope that this paper will restart a dialog on these issues that will move in time into a standards development process.

In this paper, we introduce group participation support into decentralized peer-to-peer (P2P) Web caching systems. The support allows P2P Web caching systems to have more participators, and as a result, realizes highly scalable caching... more

In this paper, we introduce group participation support into decentralized peer-to-peer (P2P) Web caching systems. The support allows P2P Web caching systems to have more participators, and as a result, realizes highly scalable caching systems. An implementation design of systems is described in this paper. Some experimental results show that every node in our system can get better performance by

This article introduces a novel family of decentralised caching policies for wireless networks, referred to as spatial multi-LRU. Based on these, cache inventories are updated in a way that provides content diversity to users that are... more

This article introduces a novel family of decentralised caching policies for wireless networks, referred to as spatial multi-LRU. Based on these, cache inventories are updated in a way that provides content diversity to users that are covered by, and thus have access to, more than one station. Two variations are proposed, the multi-LRU-One and-All, which differ in the number of replicas inserted in the involved edge caches. Che-like approximations are proposed to accurately predict their hit probability under the Independent Reference Model (IRM). For IRM traffic multi-LRU-One outper-forms multi-LRU-All, whereas when the traffic exhibits temporal locality the-All variation can perform better.

The large popularity of web services and applications makes their performance very critical. Reducing the latency of retrieve web pages has become a real challenge. Caching is widely used for this purpose. Web caching is a well-known... more

The large popularity of web services and applications makes their performance very critical. Reducing the latency of retrieve web pages has become a real challenge. Caching is widely used for this purpose. Web caching is a well-known strategy for improving performance of Web-based system by keeping web objects that are likely to be used in the near future close to the client. Most of the current Web browsers still employ traditional caching policies that are not efficient in web caching. This research proposes a splitting browser cache to two caches, instant cache and durable cache. Initially, a web object is stored in instant cache, and the web objects that are visited more than the pre-specified threshold value will be moved to durable cache. Other objects are removed by Least Recently Used (LRU) algorithm as instant cache is full. More significantly, when the durable cache saturates, a fuzzy system is employed in classifying each object stored in durable cache into either cacheable or uncacheable object. The old uncacheable objects are candidate for removing from the durable cache. By implementing this mechanism, the cache pollution can be mitigated and the cache space can be utilized effectively. Experimental results have revealed that the proposed approach can improve the performance in terms of hit ratio (HR) and Latency Saving Ratio (LSR) when compared to LRU and Least Frequently Used (LFU).

—Information dissemination as the sole functionality driving the current Internet trend has been of keen interest for its manageability. Information Centric Network (ICN) proposed as a new paradigm shift to mitigate the predicted traffic... more

—Information dissemination as the sole functionality driving the current Internet trend has been of keen interest for its manageability. Information Centric Network (ICN) proposed as a new paradigm shift to mitigate the predicted traffic of the current Internet. However, caching as an advantageous building block of ICN is faced with the challenges of content placement, content replacement and eviction. The current practice of ICN caching has given birth to the problems of content redundancy, path redundancy and excessive wastage of bandwidth. This study analyze the intelligence in cache content management to palliate the gross expenses incurred in the ICN practice. The use of the current factors in previous studies in recency and frequency in content usage plays delicate roles in our study. Replacement strategies are agreed to influence the entire cache-hit, stretch and network diversity.

Handoff of Application Sessions Across Time and Space ... pa-per we address the issue of application session transfer for unin-terrupted data access across this diverse ... From this effort we have established baseline performance metrics... more

Handoff of Application Sessions Across Time and Space ... pa-per we address the issue of application session transfer for unin-terrupted data access across this diverse ... From this effort we have established baseline performance metrics and have found that the overhead is within ...

In recent research, the problem of document replacement in Web caches has received much interest. Web caches are different from system/processor caches because web caches have several additional criteria (frequency and recentness of... more

In recent research, the problem of document replacement in Web caches has received much interest. Web caches are different from system/processor caches because web caches have several additional criteria (frequency and recentness of pages, size of a document, cost of fetching a document etc.). It has been shown that, the classical LRU replacement policy performs poorly in Web caches because

Abstract. This work presents an extensive evaluation of the request filtering in hierarchy of proxy caches. Using the recently proposed ADF (Aggregation, Disaggregation and Filtering) model as well as entropy as metric for Web traffic... more

Abstract. This work presents an extensive evaluation of the request filtering in hierarchy of proxy caches. Using the recently proposed ADF (Aggregation, Disaggregation and Filtering) model as well as entropy as metric for Web traffic characterization, we evaluate how locality of reference changes as the streams of requests pass through a hierarchy of caches. Moreover, we propose the use of average entropy for comparing the locality of reference of different streams and present how a proxy server can dynamically calculate the entropy of ...

Internet is the boon in modern era as every organization uses it for dissemination of information and e-commerce related applications. Sometimes people of organization feel delay while accessing internet in spite of proper bandwidth.... more

Internet is the boon in modern era as every organization uses it for dissemination of information and e-commerce related applications. Sometimes people of organization feel delay while accessing internet in spite of proper bandwidth. Prediction model of web caching and prefetching is an ideal solution of this delay problem. Prediction model analysing history of internet user from server raw log files and determine future sequence of web objects and placed all web objects to nearer to the user so access latency could be reduced to some extent and problem of delay is to be solved. To determine sequence of future web objects, it is necessary to determine proximity of one web object with other by identifying proper distance metric technique related to web caching and prefetching. This paper studies different distance metric techniques and concludes that bio informatics based distance metric techniques are ideal in context to Web Caching and Web Prefetching.

Dynamic Web applications have gained a great deal of popularity. Improving the performance of these applications has recently attracted the attention of many researchers. One of the most important techniques proposed for this purpose is... more

Dynamic Web applications have gained a great deal of popularity. Improving the performance of these applications has recently attracted the attention of many researchers. One of the most important techniques proposed for this purpose is caching, which can be done at different locations and within different stages of the process of generating a dynamic Web page. Most of the caching schemes proposed in literature are lenient about the issue of consistency; they assume that users can tolerate receiving stale data. However, an important class of dynamic Web applications are those in which users always expect to get the freshest data available. Any caching scheme has to incur a significant overhead to be able to provide this level of consistency (i.e., strong consistency); the overhead may be so much that it neutralizes the benefits of caching. In this paper, three alternative architectures are investigated for dynamic Web applications that require strong consistency. A proxy caching scheme is designed and implemented, which performs caching at the level of database queries. This caching system is used in one of the alternative architectures. The performance experiments show that, despite the high overhead of providing strong consistency in database caching, this technique can improve the performance of dynamic Web applications, especially when there is a long network latency between clients and the (origin) server.

Information dissemination has been the practice of today’s Internet with media data overtaking the ordinary textual information. A need to subscribe to Information-centric network has been seen as an option to provide better manageability... more

Information dissemination has been the practice of today’s Internet with media data overtaking the ordinary textual information. A need to subscribe to Information-centric network has been seen as an option to provide
better manageability of the future Internet. Caching is therefore assumed to be the most beneficial practice
for fast delivery of information, saving cost and efficient bandwidth utilization. However, in the popular caching strategy, content and path redundancy has necessitated an extension for mitigating their effects. This paper
provides an approach of mitigating high content and path redundancy in information-centric network using
hypergraph properties of arcs and clustering. A comparison of some popular cache deployment strategies was
analyzed based on simulation results. The paper proposes a cache deployment strategy using hypergraph with less content and path redundancy to improve content delivery in ICN.

Abstract—Remote visualization is an enabling technology aiming to resolve the barrier of physical distance. Although many researchers have developed innovative algorithms for remote visualization, previous work has focused little on... more

Abstract—Remote visualization is an enabling technology aiming to resolve the barrier of physical distance. Although many researchers have developed innovative algorithms for remote visualization, previous work has focused little on systematically investigating optimal ...

The benefits of the ubiquitous caching in ICN are profound, such features make ICN promising for content distribution, but it also introduces a challenge to content protection against the unauthorized access. The protection of a content... more

The benefits of the ubiquitous caching in ICN are profound, such features make ICN promising for content distribution, but it also introduces a challenge to content protection against the unauthorized access. The protection of a content against unauthorized access requires consumer authentication and involves the conventional end-to-end encryption. However, in information-centric networking (ICN), such end-to-end encryption makes the content caching ineffective since encrypted contents stored in a cache are useless for any consumers except those who know the encryption key. For effective caching of encrypted contents in ICN, we propose a secure distribution of protected content (SDPC) scheme, which ensures that only authenticated consumers can access the content. SDPC is lightweight and allows consumers to verify the originality of the published content by using a symmetric key encryption. Moreover, SDPC naming scheme provides protection against privacy leakage. The security of SDPC was proved with the BAN logic and Scyther tool verification, and simulation results show that SDPC can reduce the content download delay.