Programming Language Design and Implementation (conference)

Last updated

Programming Language Design and Implementation (PLDI) is an academic conference in computer science, in particular, in the study of programming languages and compilers. PLDI is organized by the Association for Computing Machinery under the SIGPLAN interest group.

Contents

History

The precursor of PLDI was the Symposium on Compiler Optimization, held July 2728, 1970 at the University of Illinois at Urbana-Champaign and chaired by Robert S. Northcote. That conference included papers by Frances E. Allen, John Cocke, Alfred V. Aho, Ravi Sethi, and Jeffrey D. Ullman. The first conference in the current PLDI series took place in 1979 under the name SIGPLAN Symposium on Compiler Construction in Denver, Colorado. The next compiler construction conference took place in 1982 in Boston, Massachusetts. The compiler construction conferences then alternated with SIGPLAN Conferences on Language Issues until 1988, when the conference was renamed to PLDI. From 1982 until 2001, the conference acronym was SIGPLAN 'xx. Starting in 2002, the initialism became PLDI 'xx, and in 2006 it became PLDI xxxx.

Conference locations and organizers

Related Research Articles

SIGPLAN is the Association for Computing Machinery's Special Interest Group on programming languages.

In compiler theory, partial redundancy elimination (PRE) is a compiler optimization that eliminates expressions that are redundant on some but not necessarily all paths through a program. PRE is a form of common subexpression elimination.

<span class="mw-page-title-main">Robert Harper (computer scientist)</span> Computer scientist

Robert William "Bob" Harper, Jr. is a computer science professor at Carnegie Mellon University who works in programming language research. Prior to his position at Carnegie Mellon, Harper was a research fellow at the University of Edinburgh.

The International Parallel and Distributed Processing Symposium is an annual conference for engineers and scientists to present recent findings in the fields of parallel processing and distributed computing. In addition to technical sessions of submitted paper presentations, the meeting offers workshops, tutorials, and commercial presentations & exhibits. IPDPS is sponsored by the IEEE Computer Society's Technical Committee on Parallel Processing.

Brent Hailpern is a computer scientist retired from IBM Research. His research work focused on programming languages, software engineering, and concurrency.

The International Conference on Functional Programming (ICFP) is an annual academic conference in the field of computer science sponsored by the ACM SIGPLAN, in association with IFIP Working Group 2.8. The conference focuses on functional programming and related areas of programming languages, logic, compilers and software development.

The annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages (POPL) is an academic conference in the field of computer science, with focus on fundamental principles in the design, definition, analysis, and implementation of programming languages, programming systems, and programming interfaces. The venue is jointly sponsored by two Special Interest Groups of the Association for Computing Machinery: SIGPLAN and SIGACT.

In computer science, pointer analysis, or points-to analysis, is a static code analysis technique that establishes which pointers, or heap references, can point to which variables, or storage locations. It is often a component of more complex analyses such as escape analysis. A closely related technique is shape analysis.

Thread Level Speculation (TLS), also known as Speculative Multi-threading, or Speculative Parallelization, is a technique to speculatively execute a section of computer code that is anticipated to be executed later in parallel with the normal execution on a separate independent thread. Such a speculative thread may need to make assumptions about the values of input variables. If these prove to be invalid, then the portions of the speculative thread that rely on these input variables will need to be discarded and squashed. If the assumptions are correct the program can complete in a shorter time provided the thread was able to be scheduled efficiently.

Matthew Flatt is an American computer scientist and professor at the University of Utah School of Computing in Salt Lake City. He is also the leader of the core development team for the Racket programming language.

In computing, compiler correctness is the branch of computer science that deals with trying to show that a compiler behaves according to its language specification. Techniques include developing the compiler using formal methods and using rigorous testing on an existing compiler.

In computer science, region-based memory management is a type of memory management in which each allocated object is assigned to a region. A region, also called a zone, arena, area, or memory context, is a collection of allocated objects that can be efficiently reallocated or deallocated all at once. Memory allocators using region-based managements are often called area allocators, and when they work by only "bumping" a single pointer, as bump allocators.

Mary Lou Ehnot Soffa is an American computer scientist noted for her research on compilers, program optimization, system software and system engineering.

Barbara G. Ryder is an American Computer Scientist noted for her research on programming languages and more specifically, the theoretical foundations and empirical investigation of interprocedural compile-time analysis.

Susan Beth Horwitz was an American computer scientist noted for her research on programming languages and software engineering, and in particular on program slicing and dataflow-analysis. She had several best paper and an impact paper award mentioned below under awards.

<span class="mw-page-title-main">Kathleen Fisher</span> American computer scientist

Kathleen Shanahan Fisher is an American computer scientist who specializes in programming languages and their implementation.

Differential testing, also known as differential fuzzing, is a popular software testing technique that attempts to detect bugs, by providing the same input to a series of similar applications, and observing differences in their execution. Differential testing complements traditional software testing, because it is well-suited to find semantic or logic bugs that do not exhibit explicit erroneous behaviors like crashes or assertion failures. Differential testing is sometimes called back-to-back testing.

References

  1. "PLDI 2023". pldi23.sigplan.org. Retrieved 2023-09-20.
  2. "PLDI 2022". pldi22.sigplan.org. Retrieved 2022-06-14.
  3. "PLDI 2021". pldi21.sigplan.org. Retrieved 2022-06-14.
  4. "PLDI 2020". pldi20.sigplan.org. Retrieved 2022-06-14.