US20060002555A1 - Dynamic content-aware memory compression and encryption architecture - Google Patents
Dynamic content-aware memory compression and encryption architecture Download PDFInfo
- Publication number
- US20060002555A1 US20060002555A1 US10/869,984 US86998404A US2006002555A1 US 20060002555 A1 US20060002555 A1 US 20060002555A1 US 86998404 A US86998404 A US 86998404A US 2006002555 A1 US2006002555 A1 US 2006002555A1
- Authority
- US
- United States
- Prior art keywords
- application
- areas
- memory
- area
- memory hierarchy
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
- G06F12/08—Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
- G06F12/0802—Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
- G06F12/08—Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
- G06F12/12—Replacement control
- G06F12/121—Replacement control using replacement algorithms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2212/00—Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
- G06F2212/40—Specific encoding of data in memory or cache
- G06F2212/401—Compressed data
Definitions
- the present invention is related to memory architectures and, more particularly, to architectures for compression and encryption of memory.
- the present invention is directed to a methodology for compression/encryption that is content-aware.
- compression/decompression and encryption/decryption are done in multiple locations within the various levels of a memory hierarchy.
- Applications are segmented into different areas that can be compressed and/or encrypted using different algorithms. Critical segments that must be executed fast can be left uncompressed while other segments which are less critical but which are highly compressible can be compressed. Sensitive data areas can be encrypted while other less sensitive areas can be left unencrypted for speed of execution. Different compression and/or encryption schemes can be used for the various segments to maximize compressibility and performance.
- the present invention provides the flexibility to treat different regions of the application using different strategies and provides the ability to carry out such decisions dynamically during execution.
- FIG. 1 is an example architecture illustrating different aspects of the invention.
- FIG. 2 is a flowchart of processing performed prior to execution of an application, in accordance with an embodiment of the invention.
- FIG. 3 is a flowchart of processing performed during application runtime by memory management middleware or hardware, in accordance with an embodiment of the invention.
- FIG. 1 is a diagram of an example architecture that illustrates dynamic application partitioning and the use of compression/decompression at various levels of a memory hierarchy.
- a processor (CPU) 110 is depicted in FIG. 1 which includes an instruction cache 120 and a data cache 130 .
- the processor 110 has a bus which connects it to main memory 140 .
- the processor 110 notably, can be any conventional microprocessor and advantageously does not need to be modified in order to be utilized in the context of the present invention.
- FIG. 1 is only illustrative. The techniques disclosed herein can be readily utilized in other memory hierarchies with different (or no) caching architectures.
- compression is integrated into the architecture shown in FIG. 1 at two levels of the memory hierarchy—while advantageously segmenting both code and data into various areas.
- Most code areas are stored in compressed format in both the instruction cache 120 and the main memory 140 .
- a decompression engine 150 is placed between the instruction cache 120 and the processor 110 . If code is not modified during runtime, a compression engine is not required to be integrated with the decompression engine 150 .
- most data areas are stored in compressed format in main memory 140 while in decompressed format in the data cache 130 . Therefore, a compression/decompression engine 160 is required which is placed between main memory 140 and the CPU/data cache. Note that since data needs to be written during runtime, a compression engine is also necessary.
- FIG. 2 sets forth a flowchart of processing performed on an application prior to execution on the architecture shown in FIG. 1 .
- the executable application is segmented into a plurality of areas at step 201 .
- a typical executable application for example, can contain code and data segments.
- One simple technique of application segmentation can be achieved by separating code from data regions within the application. Handling these areas differently is useful as they have different statistical properties, and, more importantly, they are treated in a different manner during execution.
- Instruction code for example, typically is not modified during execution (with the exception of self-modifying code, which is not often used) while data is often modified.
- an appropriate compression scheme is selected for each segment of the application. Different compression schemes can be used for the various segments to maximize compression results and performance. Different code areas, for example, can contain a different statistical mix of instructions making a particular choice of algorithm more suitable. Large data areas, on the other hand, are often initially set to zeroes, which implies that an algorithm such as run length encoding would be a very effective compression approach.
- the present invention advantageously provides the flexibility to apply different strategies to different segments of the application.
- the mechanisms described above are readily extendable to encryption.
- the application can be segmented into areas that must be protected since they have either sensitive data or code—and other areas that can be left unencrypted for speed of execution.
- a suitable encryption scheme is selected, and the segment is encrypted using the selected encryption scheme.
- memory management middleware 145 is provided which handles the movement and processing of the different application segments. It is preferable that the memory management middleware 145 act in a dynamic manner, as illustrated by FIG. 3 .
- FIG. 3 is a flowchart of processing performed by the memory management middleware, in accordance with a preferred embodiment of another aspect of the invention.
- statistics are gathered during the application runtime.
- the particular form of the statistics gathered does not affect the nature of the invention, although it is preferable to gather information which reflects how often particular segments are being requested for processing as well as other metrics which may be used to aid the performance of the particular compression/encryption algorithms selected.
- a variety of dynamic operations can be carried out by the memory management middleware based on the dynamic statistical information that is gathered during the application's runtime, as reflected in steps 302 - 305 in FIG. 3 .
- the compression of code/data that is requested too often can be moved “closer” to main memory.
- code that is executed very often can be designated to fall into a category of areas that are decompressed between main memory and the cache, rather than between the cache and the CPU. Fast execution of such code could warrant the increase in code size.
- such segments of the code/data can be designed to utilize the compression/decompression engine 160 (or even be designated to avoid compression and use the bypass route 170 only) rather than the decompression engine 150 between the instruction cache 120 and the CPU 110 .
- existing code or data areas can be recompressed according to the new statistical information.
- the statistical properties of data that changes can warrant a change in compression algorithm that more effectively deals with the nature of the data as it currently exists in memory.
- run length encoding may effectively deal with initially empty data areas, as those same data areas fill up with live data during runtime, a shift in compression strategy could be advantageous to maximizing performance of the architecture.
- step 304 data that is never or infrequently requested can be recompressed using stronger compression algorithms to save memory space.
- the selection of compression algorithm often reflects a tradeoff between speed of execution and the amount of space that is saved through the use of compression. Where the segment of the application is infrequently accessed, the increase in memory savings may readily justify the decrease in execution speed from using a slower but more effective compression algorithm.
- the areas can be rearranged according to the frequency of access.
- the memory management system is content-aware and is able to adapt to the particular application being executed and learn during its execution so that both compression/encryption and performance are optimized as much as possible.
- the different segments of the application can be rearranged and reallocated during runtime based on the appropriate performance metrics collected.
Abstract
Description
- The present application is related to co-pending commonly-assigned United States utility patent applications “MEMORY COMPRESSION ARCHITECTURE FOR EMBEDDED SYSTEMS,” Attorney Docket No. 03041, Serial No. to be assigned, and “MEMORY ENCRYPTION ARCHITECTURE,” Attorney Docket No. 03042, Serial No. to be assigned, both filed contemporaneously with the present application and both of which are incorporated by reference herein.
- The present invention is related to memory architectures and, more particularly, to architectures for compression and encryption of memory.
- Compression and encryption techniques are well-known. A recent development has been to use techniques such as compression to reduce the size of main memory in a computer architecture. See, e.g., M. Kjelso et al., “Main Memory Hardware Data Compression,” 22nd Euromicro Conference, pages 423-30, IEEE Computer Society Press (September 1996). For example, researchers at IBM have developed the “MXT” architecture for servers which performs compression and decompression during runtime of an application when transferring data from the L3 cache to main memory and vice versa. See Tremaine et al., “IBM Memory Expansion Technolog (MXT),” IBM J. Res. & Dev., Vol. 45, No. 2 (March 2001). See also U.S. Pat. Nos. 5,761,536, 5,812,817, and 6,240,419, which are incorporated by reference herein. Similarly, encryption has been utilized in the prior art to protect sensitive code or data stored in memory.
- Despite the advances in compression and encryption, prior art use of application compression and encryption techniques typically rely on the following constraints. First, compression/decompression and encryption/decryption is typically applied at a specific level of the memory hierarchy. Second, once that level of the memory hierarchy is pre-specified, a specific compression and/or encryption algorithm is selected, namely an algorithm that is suitable for that level of the memory hierarchy. Thus, solutions currently available will provide a compression or encryption scheme that may be optimal with regards to a portion of an application's code or data but that may be suboptimal with regards to much of the rest of the application code or data. This is particularly of concern in embedded systems, where space constraints and security issues typically exist.
- Accordingly, there is a need for an architecture that can handle compression and encryption in a more flexible and efficient manner.
- The present invention is directed to a methodology for compression/encryption that is content-aware. In accordance with an embodiment of the invention, compression/decompression and encryption/decryption are done in multiple locations within the various levels of a memory hierarchy. Applications are segmented into different areas that can be compressed and/or encrypted using different algorithms. Critical segments that must be executed fast can be left uncompressed while other segments which are less critical but which are highly compressible can be compressed. Sensitive data areas can be encrypted while other less sensitive areas can be left unencrypted for speed of execution. Different compression and/or encryption schemes can be used for the various segments to maximize compressibility and performance.
- The present invention provides the flexibility to treat different regions of the application using different strategies and provides the ability to carry out such decisions dynamically during execution. These and other advantages of the invention will be apparent to those of ordinary skill in the art by reference to the following detailed description and the accompanying drawings.
-
FIG. 1 is an example architecture illustrating different aspects of the invention. -
FIG. 2 is a flowchart of processing performed prior to execution of an application, in accordance with an embodiment of the invention. -
FIG. 3 is a flowchart of processing performed during application runtime by memory management middleware or hardware, in accordance with an embodiment of the invention. -
FIG. 1 is a diagram of an example architecture that illustrates dynamic application partitioning and the use of compression/decompression at various levels of a memory hierarchy. A processor (CPU) 110 is depicted inFIG. 1 which includes aninstruction cache 120 and adata cache 130. Theprocessor 110 has a bus which connects it tomain memory 140. Theprocessor 110, notably, can be any conventional microprocessor and advantageously does not need to be modified in order to be utilized in the context of the present invention. - It should be noted that the memory hierarchy depicted in
FIG. 1 is only illustrative. The techniques disclosed herein can be readily utilized in other memory hierarchies with different (or no) caching architectures. - In accordance with an embodiment of an aspect of the invention, compression is integrated into the architecture shown in
FIG. 1 at two levels of the memory hierarchy—while advantageously segmenting both code and data into various areas. Most code areas are stored in compressed format in both theinstruction cache 120 and themain memory 140. Adecompression engine 150 is placed between theinstruction cache 120 and theprocessor 110. If code is not modified during runtime, a compression engine is not required to be integrated with thedecompression engine 150. In contrast, most data areas are stored in compressed format inmain memory 140 while in decompressed format in thedata cache 130. Therefore, a compression/decompression engine 160 is required which is placed betweenmain memory 140 and the CPU/data cache. Note that since data needs to be written during runtime, a compression engine is also necessary. -
FIG. 2 sets forth a flowchart of processing performed on an application prior to execution on the architecture shown inFIG. 1 . - In accordance with a preferred embodiment of the present invention, the executable application is segmented into a plurality of areas at
step 201. A typical executable application, for example, can contain code and data segments. One simple technique of application segmentation can be achieved by separating code from data regions within the application. Handling these areas differently is useful as they have different statistical properties, and, more importantly, they are treated in a different manner during execution. Instruction code, for example, typically is not modified during execution (with the exception of self-modifying code, which is not often used) while data is often modified. - Thus, it is preferable to do a statistical analysis of the application prior to execution and to segment the application based on the statistical properties of the different areas of the application. This can result in code being separated from data or, as depicted in
FIG. 1 , different areas of code and data being segmented as well. At steps 202-205, an appropriate compression scheme is selected for each segment of the application. Different compression schemes can be used for the various segments to maximize compression results and performance. Different code areas, for example, can contain a different statistical mix of instructions making a particular choice of algorithm more suitable. Large data areas, on the other hand, are often initially set to zeroes, which implies that an algorithm such as run length encoding would be a very effective compression approach. The present invention advantageously provides the flexibility to apply different strategies to different segments of the application. Once an appropriate compression algorithm has been selected for a particular segment atstep 203, the segment is compressed and the index tables updated atstep 204. Notably, for certain critical segments of the application, the appropriate choice may be to leave the segment uncompressed. Finally, atstep 206, the application may be executed. - The mechanisms described above are readily extendable to encryption. The application can be segmented into areas that must be protected since they have either sensitive data or code—and other areas that can be left unencrypted for speed of execution. For those segments that are to be encrypted, a suitable encryption scheme is selected, and the segment is encrypted using the selected encryption scheme.
- As depicted in
FIG. 1 ,memory management middleware 145 is provided which handles the movement and processing of the different application segments. It is preferable that thememory management middleware 145 act in a dynamic manner, as illustrated byFIG. 3 . -
FIG. 3 is a flowchart of processing performed by the memory management middleware, in accordance with a preferred embodiment of another aspect of the invention. Atstep 301, statistics are gathered during the application runtime. The particular form of the statistics gathered does not affect the nature of the invention, although it is preferable to gather information which reflects how often particular segments are being requested for processing as well as other metrics which may be used to aid the performance of the particular compression/encryption algorithms selected. A variety of dynamic operations can be carried out by the memory management middleware based on the dynamic statistical information that is gathered during the application's runtime, as reflected in steps 302-305 inFIG. 3 . - At
step 302, the compression of code/data that is requested too often can be moved “closer” to main memory. Given the typical overhead in using compression/encryption, it is generally beneficial to compress and decompress at levels of the memory hierarchy that are far away from the CPU. Thus, code that is executed very often can be designated to fall into a category of areas that are decompressed between main memory and the cache, rather than between the cache and the CPU. Fast execution of such code could warrant the increase in code size. Thus, with reference toFIG. 1 , such segments of the code/data can be designed to utilize the compression/decompression engine 160 (or even be designated to avoid compression and use thebypass route 170 only) rather than thedecompression engine 150 between theinstruction cache 120 and theCPU 110. - At
step 303, existing code or data areas can be recompressed according to the new statistical information. For example, the statistical properties of data that changes can warrant a change in compression algorithm that more effectively deals with the nature of the data as it currently exists in memory. Thus, although run length encoding may effectively deal with initially empty data areas, as those same data areas fill up with live data during runtime, a shift in compression strategy could be advantageous to maximizing performance of the architecture. - At
step 304, data that is never or infrequently requested can be recompressed using stronger compression algorithms to save memory space. The selection of compression algorithm often reflects a tradeoff between speed of execution and the amount of space that is saved through the use of compression. Where the segment of the application is infrequently accessed, the increase in memory savings may readily justify the decrease in execution speed from using a slower but more effective compression algorithm. - At
step 305, the areas can be rearranged according to the frequency of access. Thus, the memory management system is content-aware and is able to adapt to the particular application being executed and learn during its execution so that both compression/encryption and performance are optimized as much as possible. The allocation of different portions of the application to particular compression approaches and to particular locations in the memory hierarchy, thus, need not be static. The different segments of the application can be rearranged and reallocated during runtime based on the appropriate performance metrics collected. - The above implementation is merely illustrative of the invention. It will be understood by those of ordinary skill in the art that various changes may be made that are within the scope of the invention, which is to be limited only by the appended claims.
Claims (23)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/869,984 US7474750B2 (en) | 2004-06-16 | 2004-06-16 | Dynamic content-aware memory compression and encryption architecture |
PCT/US2005/020016 WO2006009617A2 (en) | 2004-06-16 | 2005-06-07 | Dynamic content-aware memory compression and encryption architecture |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/869,984 US7474750B2 (en) | 2004-06-16 | 2004-06-16 | Dynamic content-aware memory compression and encryption architecture |
Publications (2)
Publication Number | Publication Date |
---|---|
US20060002555A1 true US20060002555A1 (en) | 2006-01-05 |
US7474750B2 US7474750B2 (en) | 2009-01-06 |
Family
ID=35513954
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/869,984 Expired - Fee Related US7474750B2 (en) | 2004-06-16 | 2004-06-16 | Dynamic content-aware memory compression and encryption architecture |
Country Status (2)
Country | Link |
---|---|
US (1) | US7474750B2 (en) |
WO (1) | WO2006009617A2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060174137A1 (en) * | 2005-02-03 | 2006-08-03 | International Business Machines Corporation | Memory controller with performance-modulated security |
US20080059728A1 (en) * | 2006-09-06 | 2008-03-06 | David Michael Daly | Systems and methods for masking latency of memory reorganization work in a compressed memory system |
US20080172525A1 (en) * | 2007-01-15 | 2008-07-17 | Hitachi, Ltd. | Storage system and method of controlling a storage system |
CN104008064A (en) * | 2013-02-25 | 2014-08-27 | 国际商业机器公司 | Method and system for multi-level memory compression |
US20160188212A1 (en) * | 2014-12-30 | 2016-06-30 | International Business Machines Corporation | Data storage system selectively employing multiple data compression techniques |
WO2016160129A1 (en) * | 2015-03-27 | 2016-10-06 | Intel Corporation | Cache-less split tracker architecture for replay protection trees |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4615337B2 (en) * | 2005-03-16 | 2011-01-19 | 株式会社日立製作所 | Storage system |
US8108692B1 (en) | 2006-06-27 | 2012-01-31 | Siliconsystems, Inc. | Solid-state storage subsystem security solution |
US8356184B1 (en) | 2009-06-25 | 2013-01-15 | Western Digital Technologies, Inc. | Data storage device comprising a secure processor for maintaining plaintext access to an LBA table |
US8583608B2 (en) * | 2010-12-08 | 2013-11-12 | International Business Machines Corporation | Maximum allowable runtime query governor |
US9305142B1 (en) | 2011-12-19 | 2016-04-05 | Western Digital Technologies, Inc. | Buffer memory protection unit |
US9851916B2 (en) | 2016-02-17 | 2017-12-26 | International Business Machines Corporation | Dynamic compression for runtime services |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5057837A (en) * | 1987-04-20 | 1991-10-15 | Digital Equipment Corporation | Instruction storage method with a compressed format using a mask word |
US5386469A (en) * | 1993-08-05 | 1995-01-31 | Zilog, Inc. | Firmware encryption for microprocessor/microcomputer |
US5455576A (en) * | 1992-12-23 | 1995-10-03 | Hewlett Packard Corporation | Apparatus and methods for Lempel Ziv data compression with improved management of multiple dictionaries in content addressable memory |
US5654703A (en) * | 1996-06-17 | 1997-08-05 | Hewlett-Packard Company | Parallel data compression and decompression |
US5761536A (en) * | 1996-08-21 | 1998-06-02 | International Business Machines Corporation | System and method for reducing memory fragmentation by assigning remainders to share memory blocks on a best fit basis |
US5764994A (en) * | 1996-09-16 | 1998-06-09 | International Business Machines Corporation | Method and system for compressing compiled microcode to be executed within a data processing system |
US5812817A (en) * | 1994-10-17 | 1998-09-22 | International Business Machines Corporation | Compression architecture for system memory application |
US5825878A (en) * | 1996-09-20 | 1998-10-20 | Vlsi Technology, Inc. | Secure memory management unit for microprocessor |
US5920723A (en) * | 1997-02-05 | 1999-07-06 | Hewlett-Packard Company | Compiler with inter-modular procedure optimization |
US5999189A (en) * | 1995-08-04 | 1999-12-07 | Microsoft Corporation | Image compression to reduce pixel and texture memory requirements in a real-time image generator |
US6175896B1 (en) * | 1997-10-06 | 2001-01-16 | Intel Corporation | Microprocessor system and method for increasing memory Bandwidth for data transfers between a cache and main memory utilizing data compression |
US6195743B1 (en) * | 1999-01-29 | 2001-02-27 | International Business Machines Corporation | Method and system for compressing reduced instruction set computer (RISC) executable code through instruction set expansion |
US6202152B1 (en) * | 1998-01-27 | 2001-03-13 | Philips Semiconductors, Inc. | System and method for accessing information decrypted in multiple-byte blocks |
US6233674B1 (en) * | 1999-01-29 | 2001-05-15 | International Business Machines Corporation | Method and system for scope-based compression of register and literal encoding in a reduced instruction set computer (RISC) |
US6240419B1 (en) * | 1996-02-20 | 2001-05-29 | International Business Machines Corporation | Compression store addressing |
US6330675B1 (en) * | 1998-02-13 | 2001-12-11 | Liquid Audio, Inc. | System and method for secure transfer of digital data to a local recordable storage medium |
US6442680B1 (en) * | 1999-01-29 | 2002-08-27 | International Business Machines Corporation | Method and system for compressing reduced instruction set computer (RISC) executable code |
US6581131B2 (en) * | 2001-01-09 | 2003-06-17 | Hewlett-Packard Development Company, L.P. | Method and apparatus for efficient cache mapping of compressed VLIW instructions |
US6819271B2 (en) * | 1999-01-29 | 2004-11-16 | Quickshift, Inc. | Parallel compression and decompression system and method having multiple parallel compression and decompression engines |
US6822589B1 (en) * | 1999-01-29 | 2004-11-23 | Quickshift, Inc. | System and method for performing scalable embedded parallel data decompression |
US6865664B2 (en) * | 2000-12-13 | 2005-03-08 | Conexant Systems, Inc. | Methods, systems, and computer program products for compressing a computer program based on a compression criterion and executing the compressed program |
-
2004
- 2004-06-16 US US10/869,984 patent/US7474750B2/en not_active Expired - Fee Related
-
2005
- 2005-06-07 WO PCT/US2005/020016 patent/WO2006009617A2/en active Application Filing
Patent Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5057837A (en) * | 1987-04-20 | 1991-10-15 | Digital Equipment Corporation | Instruction storage method with a compressed format using a mask word |
US5455576A (en) * | 1992-12-23 | 1995-10-03 | Hewlett Packard Corporation | Apparatus and methods for Lempel Ziv data compression with improved management of multiple dictionaries in content addressable memory |
US5386469A (en) * | 1993-08-05 | 1995-01-31 | Zilog, Inc. | Firmware encryption for microprocessor/microcomputer |
US5812817A (en) * | 1994-10-17 | 1998-09-22 | International Business Machines Corporation | Compression architecture for system memory application |
US5999189A (en) * | 1995-08-04 | 1999-12-07 | Microsoft Corporation | Image compression to reduce pixel and texture memory requirements in a real-time image generator |
US6240419B1 (en) * | 1996-02-20 | 2001-05-29 | International Business Machines Corporation | Compression store addressing |
US5654703A (en) * | 1996-06-17 | 1997-08-05 | Hewlett-Packard Company | Parallel data compression and decompression |
US5761536A (en) * | 1996-08-21 | 1998-06-02 | International Business Machines Corporation | System and method for reducing memory fragmentation by assigning remainders to share memory blocks on a best fit basis |
US5764994A (en) * | 1996-09-16 | 1998-06-09 | International Business Machines Corporation | Method and system for compressing compiled microcode to be executed within a data processing system |
US5825878A (en) * | 1996-09-20 | 1998-10-20 | Vlsi Technology, Inc. | Secure memory management unit for microprocessor |
US5920723A (en) * | 1997-02-05 | 1999-07-06 | Hewlett-Packard Company | Compiler with inter-modular procedure optimization |
US6175896B1 (en) * | 1997-10-06 | 2001-01-16 | Intel Corporation | Microprocessor system and method for increasing memory Bandwidth for data transfers between a cache and main memory utilizing data compression |
US6202152B1 (en) * | 1998-01-27 | 2001-03-13 | Philips Semiconductors, Inc. | System and method for accessing information decrypted in multiple-byte blocks |
US6330675B1 (en) * | 1998-02-13 | 2001-12-11 | Liquid Audio, Inc. | System and method for secure transfer of digital data to a local recordable storage medium |
US6233674B1 (en) * | 1999-01-29 | 2001-05-15 | International Business Machines Corporation | Method and system for scope-based compression of register and literal encoding in a reduced instruction set computer (RISC) |
US6195743B1 (en) * | 1999-01-29 | 2001-02-27 | International Business Machines Corporation | Method and system for compressing reduced instruction set computer (RISC) executable code through instruction set expansion |
US6442680B1 (en) * | 1999-01-29 | 2002-08-27 | International Business Machines Corporation | Method and system for compressing reduced instruction set computer (RISC) executable code |
US6819271B2 (en) * | 1999-01-29 | 2004-11-16 | Quickshift, Inc. | Parallel compression and decompression system and method having multiple parallel compression and decompression engines |
US6822589B1 (en) * | 1999-01-29 | 2004-11-23 | Quickshift, Inc. | System and method for performing scalable embedded parallel data decompression |
US6865664B2 (en) * | 2000-12-13 | 2005-03-08 | Conexant Systems, Inc. | Methods, systems, and computer program products for compressing a computer program based on a compression criterion and executing the compressed program |
US6581131B2 (en) * | 2001-01-09 | 2003-06-17 | Hewlett-Packard Development Company, L.P. | Method and apparatus for efficient cache mapping of compressed VLIW instructions |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060174137A1 (en) * | 2005-02-03 | 2006-08-03 | International Business Machines Corporation | Memory controller with performance-modulated security |
US7694152B2 (en) * | 2005-02-03 | 2010-04-06 | International Business Machines Corporation | Memory controller with performance-modulated security |
US20080059728A1 (en) * | 2006-09-06 | 2008-03-06 | David Michael Daly | Systems and methods for masking latency of memory reorganization work in a compressed memory system |
WO2008030672A2 (en) * | 2006-09-06 | 2008-03-13 | International Business Machines Corporation | Systems and methods for masking latency of memory reorganization work in a compressed memory system |
WO2008030672A3 (en) * | 2006-09-06 | 2008-05-08 | Ibm | Systems and methods for masking latency of memory reorganization work in a compressed memory system |
US8122216B2 (en) | 2006-09-06 | 2012-02-21 | International Business Machines Corporation | Systems and methods for masking latency of memory reorganization work in a compressed memory system |
US20080172525A1 (en) * | 2007-01-15 | 2008-07-17 | Hitachi, Ltd. | Storage system and method of controlling a storage system |
US7953939B2 (en) * | 2007-01-15 | 2011-05-31 | Hitachi, Ltd. | Storage system and method of controlling a storage system |
CN104008064A (en) * | 2013-02-25 | 2014-08-27 | 国际商业机器公司 | Method and system for multi-level memory compression |
US20140244603A1 (en) * | 2013-02-25 | 2014-08-28 | International Business Machines Corporation | Multi-Level Memory Compression |
US20140244962A1 (en) * | 2013-02-25 | 2014-08-28 | International Business Machines Corporation | Multi-Level Memory Compression |
US9852056B2 (en) * | 2013-02-25 | 2017-12-26 | International Business Machines Corporation | Multi-level memory compression |
US9852055B2 (en) * | 2013-02-25 | 2017-12-26 | International Business Machines Corporation | Multi-level memory compression |
US20160188212A1 (en) * | 2014-12-30 | 2016-06-30 | International Business Machines Corporation | Data storage system selectively employing multiple data compression techniques |
US10101938B2 (en) * | 2014-12-30 | 2018-10-16 | International Business Machines Corporation | Data storage system selectively employing multiple data compression techniques |
WO2016160129A1 (en) * | 2015-03-27 | 2016-10-06 | Intel Corporation | Cache-less split tracker architecture for replay protection trees |
US9678894B2 (en) | 2015-03-27 | 2017-06-13 | Intel Corporation | Cache-less split tracker architecture for replay protection trees |
Also Published As
Publication number | Publication date |
---|---|
WO2006009617A3 (en) | 2007-03-15 |
US7474750B2 (en) | 2009-01-06 |
WO2006009617A2 (en) | 2006-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2006009617A2 (en) | Dynamic content-aware memory compression and encryption architecture | |
US6658549B2 (en) | Method and system allowing a single entity to manage memory comprising compressed and uncompressed data | |
US7302543B2 (en) | Compressed memory architecture for embedded systems | |
US20070005625A1 (en) | Storage architecture for embedded systems | |
US7849241B2 (en) | Memory compression method and apparatus for heterogeneous processor architectures in an information handling system | |
US7831796B2 (en) | Apparatus, system, and method for dynamically allocating main memory among a plurality of applications | |
US8131927B2 (en) | Fast accessible compressed thin provisioning volume | |
US5812817A (en) | Compression architecture for system memory application | |
US7181457B2 (en) | System and method for utilizing compression in database caches to facilitate access to database information | |
Debray et al. | Profile-guided code compression | |
CA2511752C (en) | Method and apparatus for morphing memory compressed machines | |
US20070005911A1 (en) | Operating System-Based Memory Compression for Embedded Systems | |
US7243191B2 (en) | Compressing data in a cache memory | |
US6349375B1 (en) | Compression of data in read only storage and embedded systems | |
US20060026200A1 (en) | Method and system for shared object data member zones | |
US20060233367A1 (en) | System and methods for an overlay disk and cache using portable flash memory | |
Franaszek et al. | Algorithms and data structures for compressed-memory machines | |
US6175896B1 (en) | Microprocessor system and method for increasing memory Bandwidth for data transfers between a cache and main memory utilizing data compression | |
CA2647259A1 (en) | High-speed data compression based on set associative cache mapping techniques | |
US20220012184A1 (en) | Accessing compressed computer memory | |
Carvalho et al. | Understanding cache compression | |
US20220011941A1 (en) | Managing free space in a compressed memory system | |
EP2469446B1 (en) | Anti-virus method for scrambled digital content. | |
CN101390043A (en) | Storage architecture for embedded systems | |
Lekatsas et al. | A unified architecture for adaptive compression of data and code on embedded systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NEC LABORATORIES AMERICA, INC., NEW JERSEY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEKATSAS, HARIS;HENKEL, JOERG;CHAKRADHAR, SRIMAT T;AND OTHERS;REEL/FRAME:014973/0040;SIGNING DATES FROM 20040630 TO 20040802 |
|
AS | Assignment |
Owner name: NEC CORPORATION,JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NEC LABORATORIES AMERICA, INC.;REEL/FRAME:023957/0816 Effective date: 20100216 Owner name: NEC CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NEC LABORATORIES AMERICA, INC.;REEL/FRAME:023957/0816 Effective date: 20100216 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20170106 |