Harriet Coverston

Last updated

Harriet Coverston is an American computer scientist and software developer focused on large-scale secondary storage environments, who has previously participated to various kernel developments in HPC systems. [1] Harriet is an expert in large scale archiving systems, having participated to several large projects and product developments.

Contents

Career

She started her career at Lawrence Livermore National Laboratory in 1967 working on CDC 7600 Livermore Timesharing System [2] [3] within a team of 5 people. She stayed until 1974 and then moved to Control Data Corporation for 12 years to collaborate on the Cyber 205 Operating System and CDCNET.

In January 1986, she co-founded LSC (Large Scale Configurations) where she was vice-president of technology more than 15 years. LSC was a ISV developing SAM-QFS and QFS, two software dedicated to manage cold and archived data. QFS is a high performance file system and SAM is advanced storage management.

In 2001, Sun Microsystems acquired LSC and Harriet Coverston became distinguished engineer at Sun, a position she occupied until 2010. Oracle closed the acquisition of Sun early 2010. [4]

She co-founded Versity in March 2011 with Bruce Gilpin and acts as its CTO since its inception. [5] Versity is a storage ISV developer of archiving software. [6] [7] [8]

Education

Harriet Coverston received a BA in Mathematics from Florida State University.[ citation needed ]

Patents

Harriet participates and co-owns 15 patents: [9]

Related Research Articles

A disk operating system (DOS) is a computer operating system that resides on and can use a disk storage device, such as a floppy disk, hard disk drive, or optical disc. A disk operating system provides a file system for organizing, reading, and writing files on the storage disk, and a means for loading and running programs stored on that disk. Strictly, this definition does not include any other functionality, so it does not apply to more complex OSes, such as Microsoft Windows, and is more appropriately used only for older generations of operating systems.

In computing, a file server is a computer attached to a network that provides a location for shared disk access, i.e. storage of computer files that can be accessed by workstations within a computer network. The term server highlights the role of the machine in the traditional client–server scheme, where the clients are the workstations using the storage. A file server does not normally perform computational tasks or run programs on behalf of its client workstations.

<span class="mw-page-title-main">CDC 6600</span> Mainframe computer by Control Data

The CDC 6600 was the flagship of the 6000 series of mainframe computer systems manufactured by Control Data Corporation. Generally considered to be the first successful supercomputer, it outperformed the industry's prior recordholder, the IBM 7030 Stretch, by a factor of three. With performance of up to three megaFLOPS, the CDC 6600 was the world's fastest computer from 1964 to 1969, when it relinquished that status to its successor, the CDC 7600.

<span class="mw-page-title-main">Inter-process communication</span> How computer operating systems enable data sharing

In computer science, inter-process communication (IPC), also spelled interprocess communication, are the mechanisms provided by an operating system for processes to manage shared data. Typically, applications can use IPC, categorized as clients and servers, where the client requests data and the server responds to client requests. Many applications are both clients and servers, as commonly seen in distributed computing.

Virtual Storage Access Method (VSAM) is an IBM direct-access storage device (DASD) file storage access method, first used in the OS/VS1, OS/VS2 Release 1 (SVS) and Release 2 (MVS) operating systems, later used throughout the Multiple Virtual Storage (MVS) architecture and now in z/OS. Originally a record-oriented filesystem, VSAM comprises four data set organizations: key-sequenced (KSDS), relative record (RRDS), entry-sequenced (ESDS) and linear (LDS). The KSDS, RRDS and ESDS organizations contain records, while the LDS organization simply contains a sequence of pages with no intrinsic record structure, for use as a memory-mapped file.

Checkpointing is a technique that provides fault tolerance for computing systems. It basically consists of saving a snapshot of the application's state, so that applications can restart from that point in case of failure. This is particularly important for long running applications that are executed in failure-prone computing systems.

<span class="mw-page-title-main">CDC 7600</span> 1967 supercomputer

The CDC 7600 was designed by Seymour Cray to be the successor to the CDC 6600, extending Control Data's dominance of the supercomputer field into the 1970s. The 7600 ran at 36.4 MHz and had a 65 Kword primary memory using magnetic core and variable-size secondary memory. It was generally about ten times as fast as the CDC 6600 and could deliver about 10 MFLOPS on hand-compiled code, with a peak of 36 MFLOPS. In addition, in benchmark tests in early 1970 it was shown to be slightly faster than its IBM rival, the IBM System/360, Model 195. When the system was released in 1967, it sold for around $5 million in base configurations, and considerably more as options and features were added.

In computing, the Global File System 2 or GFS2 is a shared-disk file system for Linux computer clusters. GFS2 allows all members of a cluster to have direct concurrent access to the same shared block storage, in contrast to distributed file systems which distribute data throughout the cluster. GFS2 can also be used as a local file system on a single computer.

<span class="mw-page-title-main">CDC STAR-100</span>

The CDC STAR-100 is a vector supercomputer that was designed, manufactured, and marketed by Control Data Corporation (CDC). It was one of the first machines to use a vector processor to improve performance on appropriate scientific applications. It was also the first supercomputer to use integrated circuits and the first to be equipped with one million words of computer memory.

Lustre is a type of parallel distributed file system, generally used for large-scale cluster computing. The name Lustre is a portmanteau word derived from Linux and cluster. Lustre file system software is available under the GNU General Public License and provides high performance file systems for computer clusters ranging in size from small workgroup clusters to large-scale, multi-site systems. Since June 2005, Lustre has consistently been used by at least half of the top ten, and more than 60 of the top 100 fastest supercomputers in the world, including the world's No. 1 ranked TOP500 supercomputer in November 2022, Frontier, as well as previous top supercomputers such as Fugaku, Titan and Sequoia.

Hierarchical storage management (HSM), also known as Tiered storage, is a data storage and Data management technique that automatically moves data between high-cost and low-cost storage media. HSM systems exist because high-speed storage devices, such as solid state drive arrays, are more expensive than slower devices, such as hard disk drives, optical discs and magnetic tape drives. While it would be ideal to have all data available on high-speed devices all the time, this is prohibitively expensive for many organizations. Instead, HSM systems store the bulk of the enterprise's data on slower devices, and then copy data to faster disk drives when needed. The HSM system monitors the way data is used and makes best guesses as to which data can safely be moved to slower devices and which data should stay on the fast devices.

In computer science, execute in place (XIP) is a method of executing programs directly from long-term storage rather than copying it into RAM. It is an extension of using shared memory to reduce the total amount of memory required.

The Network Livermore Timesharing System is an operating system that was actively developed at Lawrence Livermore Laboratory from 1979 until about 1988, though it continued to run production applications until 1995. An earlier system, the Livermore Time Sharing System had been developed over a decade earlier.

QFS is a filesystem from Oracle. It is tightly integrated with SAM, the Storage and Archive Manager, and hence is often referred to as SAM-QFS. SAM provides the functionality of a hierarchical storage manager.

A memory-mapped file is a segment of virtual memory that has been assigned a direct byte-for-byte correlation with some portion of a file or file-like resource. This resource is typically a file that is physically present on disk, but can also be a device, shared memory object, or other resource that an operating system can reference through a file descriptor. Once present, this correlation between the file and the memory space permits applications to treat the mapped portion as if it were primary memory.

A clustered file system (CFS) is a file system which is shared by being simultaneously mounted on multiple servers. There are several approaches to clustering, most of which do not employ a clustered file system. Clustered file systems can provide features like location-independent addressing and redundancy which improve reliability or reduce the complexity of the other parts of the cluster. Parallel file systems are a type of clustered file system that spread data across multiple storage nodes, usually for redundancy or performance.

The Sort/Merge utility is a mainframe program to sort records in a file into a specified order, merge pre-sorted files into a sorted file, or copy selected records. Internally, these utilities use one or more of the standard sorting algorithms, often with proprietary fine-tuned code.

Panasas is a data storage company that creates network-attached storage for technical computing environments.

BeeGFS is a parallel file system, developed and optimized for high-performance computing. BeeGFS includes a distributed metadata architecture for scalability and flexibility reasons. Its most used and widely known aspect is data throughput.

Quantcast File System (QFS) is an open-source distributed file system software package for large-scale MapReduce or other batch-processing workloads. It was designed as an alternative to the Apache Hadoop Distributed File System (HDFS), intended to deliver better performance and cost-efficiency for large-scale processing clusters.

References

  1. "Short bio on Sandia Labs Laboratories".
  2. "Livermore time-sharing system for the CDC 7600 (DBLP)".
  3. Sutherland, George G.; Coverston, Harriet G.; Dubois, Pierre J.; Emery, Donald R.; Kent, Douglas A.; Lund, Paul E.; Powles, George H.; Storch, F. David; Tennant, Lee O.; von Buskirk, Donald L. (1971). "Livermore time-sharing system for the CDC 7600 (AMC)". ACM Sigops Operating Systems Review. 5 (2–3): 6–26. doi:10.1145/1232916.1232917.
  4. "The end of Oracle HSM".
  5. StorageNewsletter (May 9, 2011). "Harriet Coverston, co-founder and CTO of Versity".
  6. Chris Mellor (June 21, 2016). "New storage upstart Versity offers S3 object storage interface".
  7. Chris Mellor (April 9, 2014). "Cray flogs 1PB of TASty storage to German übercomputing boffins, leaves them in tiers".
  8. Chris Mellor (November 6, 2013). "Supercomputer-maker Cray offers cold storage to hungry Big Data boys".
  9. "Patents by Inventor Harriet G. Coverston".
  10. "File archiving system and method". April 21, 2015.
  11. "Delegation in a file system with distributed components". September 6, 2011.
  12. "File system with distributed components". April 24, 2011.
  13. "Method and system for collective file ac cess using an mmap (memory-mapped file)". January 12, 2010.
  14. "Method and system for collective file ac cess using an mmap (memory-mapped file)". May 22, 2008.
  15. "Dynamic routing of I/O requests in a multi-tier storage environment". May 26, 2009.
  16. "Dynamic data migration in a multi-tier storage system". May 26, 2009.
  17. "File archiving system and method". July 3, 2008.
  18. "Delegation in a file system with distributed components". March 13, 2008.
  19. "File system with distributed components". February 28, 2008.
  20. "Method, system, and program for providing data to an application program from a file in a file system". January 2, 2003.
  21. "Method, system, and program for managing files in a file system". January 2, 2003.
  22. "Archiving file system for data servers in a distributed network environment". June 9, 1998.
  23. "Method and apparatus for insuring recovery of file control information for secondary storage systems". April 2, 1996.
  24. "Method and apparatus for file storage allocation for secondary storage using large and small file blocks". September 26, 1995.