Here's an analogy:
You're in your room working on a paper for class. You get to a point where you need to quote a periodical, but its in the car. You walk down to the car get the periodical, go back to your room and continue working on your paper. Repeat this 10 times for 10 different things.
Imagine instead you go down to the car once, take all the periodicals and put them on the desk next to you. Now when you want to quote a line all your materials are next to you. That is WAY faster. Same principle. Fetching data from disk (your hard drive) is amazingly slow. When we have to fetch data from disk its better to take a big block, push it into a faster kind of memory (some kind of RAM or register) and fetch data from those blocks instead.
Cache resides in between the CPU and the main memory. Whenever CPU requires some data, it first looks it in the cache. If the data is available in the cache[called hit], then the data is transferred to CPU.But in case , if data is not available in cache[called miss], then it has to be fetched into cache from the main memory. So using cache we can reduce the access time to fetch the data from memory by making it available in the cache.
Many CPUs have what is known as a CPU cache. The function of this CPU cache is to speed up access to data.
Cache is a high speed memory which is basically used for the following reason: As the speed of the main memory is not as much as the speed of the CPU.so just to compensate the speed mistmatch between the CPU and main memory the cache is used in between the two.so whenever the CPU asks for any data its being checked with the cache memory and if present then "cache hit" occurs or else "cache miss" occurs wher the CPU takes the data form the main memory and that data's cpoy is being send to the cache for any further operation where the CPU can request for the same data. Anand bhat(mca@kiit-870024)
Cache memory is the high speed memories which are repeatedly requested by the Cache client (CPU). Whenever the requested data from the cpu is present in the cache, it directly supply the data and is known as cache hit(fast) and when the data is not accessible in cache then cache access the block of the main memory and feed to the CPU and it is termed as cache miss (slow).
The only memory on the CPU is cache memory and it is only dependent on the CPU type and generation you use.
It takes a long time, in computer time, for the CPU to retrieve stuff out of RAM while running a program. So to speed things up cache was built. Cache holds a small part of the program you are working with and it is easy and fast for the CPU to retrieve it from cache. Cache is memory and it holds data for the CPU.
Answer:- The Cache memory is placed between the CPU and the main memory. It is a fast speed memory and is expensive and faster than the main memory. Cache memory is used to store the frequently accessed data of main memory. The instructions that are frequently used by the CPU are stored in the cache memory. It is used to reduce the average access time for address, instructions or data, which are normally stored in the main memory. Cache memory increases the operating speed of the system. But is much costlier than main memory. From economic considerations, the capacity of the cache memory is much less as compared to main memory..
There are different type of cache memory: processor cache memory, cache memory ram,1 cache memory l2, cache memory, CPU cache memory, disk cache memory, hard disk cache, cache memory motherboard.
Cache memory is a place for the CPU to store important data but still have quick access to it. The cache acts as a middleman, providing data to the CPU at full speed while it is obtaining data from the SDRAM.
It depends entirely.
Constantly needed data is stored in RAM, and depending on the system and the CPU used, it may be stored in cache memory. Cache memory is a special, high-speed type of RAM. It is more expensive and uses more current than DRAM, but it is used where speed is critical. Cache memory is generally inside of the CPU these days.
Cache memory is used to hold a copy of the data from the most frequently accessed memory locations to minimise calculation time. For example, a CPU's cache will hold this data so that it doesn't have to call data from the RAM for every calculation it makes, and merely draws from its local cache.
Amount of cache memory, speed of processor, speed of the front side bus