© Oxford University
This series is host to episodes created by the Department of Computer Science, University of Oxford, one of the longestestablished Computer Science departments in the country.
The series reflects this department's worldclass research and teaching by providing talks that encompass topics such as computational biology, quantum computing, computational linguistics, information systems, software verification, and software engineering.
en
Thu, 16 May 2024 10:35:46 +0100
http://podcasts.ox.ac.uk
Computer Science
Oxford University
Oxford University
podcasts@it.ox.ac.uk
no
http://mediapub.it.ox.ac.uk/sites/fred/files/images/albumcovers/computerscience.jpg
Computer Science
http://podcasts.ox.ac.uk

1
computer
protocols
open access
The talk will emphasize the diversity of mathematical tools necessary for understanding blockchain protocols and their applications The talk will emphasize the diversity of mathematical tools necessary for understanding blockchain protocols and their applications (e.g., distributed computing, game theory, mechanism design, and continuoustime stochastic processes) and the immediate practical impact that mathematical work on this topic has had (e.g., Ethereum's EIP1559 and LVR for automated market makers).
http://rss.oucs.ox.ac.uk/tag:20240516:103546:000:file:325015:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20240430comscistrachey.mp3
The talk will emphasize the diversity of mathematical tools necessary for understanding blockchain protocols and their applications
The talk will emphasize the diversity of mathematical tools necessary for understanding blockchain protocols and their applications The talk will emphasize the diversity of mathematical tools necessary for understanding blockchain protocols and their applications (e.g., distributed computing, game theory, mechanism design, and continuoustime stochastic processes) and the immediate practical impact that mathematical work on this topic has had (e.g., Ethereum's EIP1559 and LVR for automated market makers).
computer,protocols,open access,20240430
Tim Roughgarden
3729
Thu, 16 May 2024 10:31:10 +0100

2
computer science
stochastic
strachey
classical ssp
Professor Christel Baier delivers the Hillary Term 2024 Strachey Lecture Abstract: The classical stochastic shortest path (SSP) problems asks to find a policy for traversing a weighted stochastic graph until reaching a distinguished goal state that minimizes the expected accumulated weight. SSP problems have numerous applications in, e.g., operations research, artificial intelligence, robotics and verification of probabilistic systems. The underlying graph model is a finitestate Markov decision process (MDP) with integer weights for its stateaction pairs. Prominent results are the existence of optimal memoryless deterministic policies together with linear programming techniques and value and policy iteration to compute such policies and their values. These results rely on the assumption that the minimum under all proper policies that reach the goal state almost surely exists. Early work on the SSP problems goes back to the 1960s1990s and makes additional assumptions. Complete algorithms that only require the existence of proper policies combine these techniques with a preanalysis of end components, an elegant graphtheoretic concept for MDPs that has been introduced by de Alfaro in the late 1990s. The talk will start with a summary of these results. The second part of the talk presents more recent results for variants of the classical SSP. The conditional and partial SSP drop the assumption that the goal state must be reached almost surely and ask to minimize the expected accumulated weight under the condition that the goal will be reached (conditional SSP) resp. assign value 0 to all paths that do not reach the goal state (partial SSP). Other variants take into account aspects of riskawareness, e.g., by studying the conditional valueatrisk or the variancepenalized expected accumulated weight. While the classical SSP problem is solvable in polynomial time, such nonclassical SSP problems are computationally much harder. For the general case, the decidability status of such nonclassical SSP problems is unknown, but they have been shown to be at least as hard as the Skolem problem (and even as the positivity problem). However, for nonpositive weights, the conditional, partial and variancepenalized SSP problem are solvable in exponential time with a PSPACE lower bounds for acyclic MDPs
Speaker bio:
Christel Baier is a full professor and head of the chair for Algebraic and Logic Foundations of Computer Science at the Faculty of Computer Science of the Technische Universität Dresden since 2006. Since 2022 she holds an honorary doctorate (Dr. rer. nat. h.c.) from RWTH Aachen. From the University of Mannheim she received her Diploma in Mathematics in 1990, her Ph.D. in Computer Science in 1994 and her Habilitation in 1999. She was an associate professor for Theoretical Computer Science at the University of Bonn from 1999 to 2006. She was member of the DFG (German Research Foundation) review board for computer science from 2012 to 2019, and is a member of the DFG senate commission on Research Training Groups since 2020. Since 2011 she is a member of Academia Europaea. Her research focuses on modeling, specification and verification of reactive systems, quantitative analysis of stochastic systems, probabilistic model checking, temporal and modal logics and automata theory.
http://rss.oucs.ox.ac.uk/tag:20240416:100936:000:file:324350:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20240206strachey.mp3
Professor Christel Baier delivers the Hillary Term 2024 Strachey Lecture
Professor Christel Baier delivers the Hillary Term 2024 Strachey Lecture Abstract: The classical stochastic shortest path (SSP) problems asks to find a policy for traversing a weighted stochastic graph until reaching a distinguished goal state that minimizes the expected accumulated weight. SSP problems have numerous applications in, e.g., operations research, artificial intelligence, robotics and verification of probabilistic systems. The underlying graph model is a finitestate Markov decision process (MDP) with integer weights for its stateaction pairs. Prominent results are the existence of optimal memoryless deterministic policies together with linear programming techniques and value and policy iteration to compute such policies and their values. These results rely on the assumption that the minimum under all proper policies that reach the goal state almost surely exists. Early work on the SSP problems goes back to the 1960s1990s and makes additional assumptions. Complete algorithms that only require the existence of proper policies combine these techniques with a preanalysis of end components, an elegant graphtheoretic concept for MDPs that has been introduced by de Alfaro in the late 1990s. The talk will start with a summary of these results. The second part of the talk presents more recent results for variants of the classical SSP. The conditional and partial SSP drop the assumption that the goal state must be reached almost surely and ask to minimize the expected accumulated weight under the condition that the goal will be reached (conditional SSP) resp. assign value 0 to all paths that do not reach the goal state (partial SSP). Other variants take into account aspects of riskawareness, e.g., by studying the conditional valueatrisk or the variancepenalized expected accumulated weight. While the classical SSP problem is solvable in polynomial time, such nonclassical SSP problems are computationally much harder. For the general case, the decidability status of such nonclassical SSP problems is unknown, but they have been shown to be at least as hard as the Skolem problem (and even as the positivity problem). However, for nonpositive weights, the conditional, partial and variancepenalized SSP problem are solvable in exponential time with a PSPACE lower bounds for acyclic MDPs
Speaker bio:
Christel Baier is a full professor and head of the chair for Algebraic and Logic Foundations of Computer Science at the Faculty of Computer Science of the Technische Universität Dresden since 2006. Since 2022 she holds an honorary doctorate (Dr. rer. nat. h.c.) from RWTH Aachen. From the University of Mannheim she received her Diploma in Mathematics in 1990, her Ph.D. in Computer Science in 1994 and her Habilitation in 1999. She was an associate professor for Theoretical Computer Science at the University of Bonn from 1999 to 2006. She was member of the DFG (German Research Foundation) review board for computer science from 2012 to 2019, and is a member of the DFG senate commission on Research Training Groups since 2020. Since 2011 she is a member of Academia Europaea. Her research focuses on modeling, specification and verification of reactive systems, quantitative analysis of stochastic systems, probabilistic model checking, temporal and modal logics and automata theory.
computer science,stochastic,strachey,classical ssp,20240206
Christel Baier
3429
Tue, 06 Feb 2024 17:28:27 +0000

3
Data Protection
strachey
personal data
algorithm
privacy
computer science
In this terms Strachey lecture, Professor Monika Henzinger gives an introduction to differential privacy with an emphasis on differential private algorithms that can handle changing input data. Decisions are increasingly automated using rules that were learnt from personal data. Thus, it is important to guarantee that the privacy of the data is protected during the learning process. To formalize the notion of an algorithm that protects the privacy of its data, differential privacy was introduced. It is a rigorous mathematical definition to analyze the privacy properties of an algorithm – or the lack thereof. In this talk I will give an introduction to differential privacy with an emphasis on differential private algorithms that can handle changing input data.
Monika Henzinger is a professor of Computer Science at the Institute of Science and Technology Austria (ISTA). She holds a PhD in computer science from Princeton University (New Jersey, USA), and has been the head of research at Google and a professor of computer science at EPFL and the University of Vienna.
Monika Henzinger is an ACM and EATCS Fellow and a member of the Austrian Academy of Sciences and the German National Academy of Sciences Leopoldina. She has received several awards, including an honorary doctorate from TU Dortmund University, Two ERC Advanced Grant, the Leopoldina Carus Medal, and the Wittgensteinpreis, the highest science award of Austria.
The Strachey Lectures are generously supported by OxFORD Asset Management
http://rss.oucs.ox.ac.uk/tag:20231113:232647:000:file:323617:audio
https://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20231107comscistracheyhezinger.mp3
In this terms Strachey lecture, Professor Monika Henzinger gives an introduction to differential privacy with an emphasis on differential private algorithms that can handle changing input data.
In this terms Strachey lecture, Professor Monika Henzinger gives an introduction to differential privacy with an emphasis on differential private algorithms that can handle changing input data. Decisions are increasingly automated using rules that were learnt from personal data. Thus, it is important to guarantee that the privacy of the data is protected during the learning process. To formalize the notion of an algorithm that protects the privacy of its data, differential privacy was introduced. It is a rigorous mathematical definition to analyze the privacy properties of an algorithm – or the lack thereof. In this talk I will give an introduction to differential privacy with an emphasis on differential private algorithms that can handle changing input data.
Monika Henzinger is a professor of Computer Science at the Institute of Science and Technology Austria (ISTA). She holds a PhD in computer science from Princeton University (New Jersey, USA), and has been the head of research at Google and a professor of computer science at EPFL and the University of Vienna.
Monika Henzinger is an ACM and EATCS Fellow and a member of the Austrian Academy of Sciences and the German National Academy of Sciences Leopoldina. She has received several awards, including an honorary doctorate from TU Dortmund University, Two ERC Advanced Grant, the Leopoldina Carus Medal, and the Wittgensteinpreis, the highest science award of Austria.
The Strachey Lectures are generously supported by OxFORD Asset Management
Data Protection,strachey,personal data,algorithm,privacy,computer science,20231107
Monika Henzinger
3288
Mon, 13 Nov 2023 23:06:13 +0000

4
strachey
ai
It’s said that Henry Ford’s customers wanted a “a faster horse”. If Henry Ford was selling us artificial intelligence today, what would the customer call for, “a smarter human”? That’s certainly the picture of machine intelligence we find in science fiction narratives, but the reality of what we’ve developed is far more mundane.
Car engines produce prodigious power from petrol. Machine intelligences deliver decisions derived from data. In both cases the scale of consumption enables a speed of operation that is far beyond the capabilities of their natural counterparts. Unfettered energy consumption has consequences in the form of climate change. Does unbridled data consumption also have consequences for us?
If we devolve decision making to machines, we depend on those machines to accommodate our needs. If we don’t understand how those machines operate, we lose control over our destiny. Much of the debate around AI makes the mistake of seeing machine intelligence as a reflection of our intelligence. In this talk we argue that to control the machine we need to understand the machine, but to understand the machine we first need to understand ourselves.
Neil Lawrence is the inaugural DeepMind Professor of Machine Learning at the University of Cambridge where he leads the University’s flagship mission on AI, AI@Cam. He has been working on machine learning models for over 20 years. He recently returned to academia after three years as Director of Machine Learning at Amazon. His main interest is the interaction of machine learning with the physical world. This interest was triggered by deploying machine learning in the African context, where ‘endtoend’ solutions are normally required. This has inspired new research directions at the interface of machine learning and systems research, this work is funded by a Senior AI Fellowship from the Alan Turing Institute. He is interim chair of the advisory board of the UK’s Centre for Data Ethics and Innovation and a member of the UK’s AI Council. Neil is also visiting Professor at the University of Sheffield and the cohost of Talking Machines.
THE STRACHEY LECTURES ARE GENEROUSLY SUPPORTED BY OxFORD ASSET MANAGEMENT
http://rss.oucs.ox.ac.uk/tag:20230918:154647:000:file:323161:audio
https://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20230502comsciscrachey.lawrence.mp3
It’s said that Henry Ford’s customers wanted a “a faster horse”. If Henry Ford was selling us artificial intelligence today, what would the customer call for, “a smarter human”?
It’s said that Henry Ford’s customers wanted a “a faster horse”. If Henry Ford was selling us artificial intelligence today, what would the customer call for, “a smarter human”? That’s certainly the picture of machine intelligence we find in science fiction narratives, but the reality of what we’ve developed is far more mundane.
Car engines produce prodigious power from petrol. Machine intelligences deliver decisions derived from data. In both cases the scale of consumption enables a speed of operation that is far beyond the capabilities of their natural counterparts. Unfettered energy consumption has consequences in the form of climate change. Does unbridled data consumption also have consequences for us?
If we devolve decision making to machines, we depend on those machines to accommodate our needs. If we don’t understand how those machines operate, we lose control over our destiny. Much of the debate around AI makes the mistake of seeing machine intelligence as a reflection of our intelligence. In this talk we argue that to control the machine we need to understand the machine, but to understand the machine we first need to understand ourselves.
Neil Lawrence is the inaugural DeepMind Professor of Machine Learning at the University of Cambridge where he leads the University’s flagship mission on AI, AI@Cam. He has been working on machine learning models for over 20 years. He recently returned to academia after three years as Director of Machine Learning at Amazon. His main interest is the interaction of machine learning with the physical world. This interest was triggered by deploying machine learning in the African context, where ‘endtoend’ solutions are normally required. This has inspired new research directions at the interface of machine learning and systems research, this work is funded by a Senior AI Fellowship from the Alan Turing Institute. He is interim chair of the advisory board of the UK’s Centre for Data Ethics and Innovation and a member of the UK’s AI Council. Neil is also visiting Professor at the University of Sheffield and the cohost of Talking Machines.
THE STRACHEY LECTURES ARE GENEROUSLY SUPPORTED BY OxFORD ASSET MANAGEMENT
strachey,ai,20230502
Neil Lawrence
3026
Mon, 04 Sep 2023 12:36:43 +0100

5
symmetry
mathematics
engineering
An introduction to algorithmic aspects of symmetry and similarity, ranging from the fundamental complexity theoretic "Graph Isomorphism Problem" to applications in optimisation and machine learning Symmetry is a fundamental concept in mathematics, science and engineering, and beyond. Understanding symmetries is often crucial for understanding structures. In computer science, we are mainly interested in the symmetries of combinatorial structures. Computing the symmetries of such a structure is essentially the same as deciding whether two structures are the same ("isomorphic"). Algorithmically, this is a difficult task that has received a lot of attention since the early days of computing. It is a major open problem in theoretical computer science to determine the precise computational complexity of this "Graph Isomorphism Problem".
http://rss.oucs.ox.ac.uk/tag:20230216:092234:000:file:321292:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20230207strachey.mp3
An introduction to algorithmic aspects of symmetry and similarity, ranging from the fundamental complexity theoretic "Graph Isomorphism Problem" to applications in optimisation and machine learning
An introduction to algorithmic aspects of symmetry and similarity, ranging from the fundamental complexity theoretic "Graph Isomorphism Problem" to applications in optimisation and machine learning Symmetry is a fundamental concept in mathematics, science and engineering, and beyond. Understanding symmetries is often crucial for understanding structures. In computer science, we are mainly interested in the symmetries of combinatorial structures. Computing the symmetries of such a structure is essentially the same as deciding whether two structures are the same ("isomorphic"). Algorithmically, this is a difficult task that has received a lot of attention since the early days of computing. It is a major open problem in theoretical computer science to determine the precise computational complexity of this "Graph Isomorphism Problem".
symmetry,mathematics,engineering,20230207
Martin Grohe
3634
Thu, 16 Feb 2023 09:05:46 +0000

6
logic
reasoning
neural
symbolic
data
society
computational social science
An overview of work on probabilistic soft logic (PSL), an SRL framework for largescale collective, probabilistic reasoning in relational domains and a description of recent work which integrates neural and symbolic (NeSy) reasoning. Our ability to collect, manipulate, analyze, and act on vast amounts of data is having a profound impact on all aspects of society. Much of this data is heterogeneous in nature and interlinked in a myriad of complex ways. From information integration to scientific discovery to computational social science, we need machine learning methods that are able to exploit both the inherent uncertainty and the innate structure in a domain. Statistical relational learning (SRL) is a subfield that builds on principles from probability theory and statistics to address uncertainty while incorporating tools from knowledge representation and logic to represent structure. In this talk, I’ll overview our work on probabilistic soft logic (PSL), an SRL framework for largescale collective, probabilistic reasoning in relational domains. I’ll also describe recent work which integrates neural and symbolic (NeSy) reasoning. I’ll close by highlighting emerging opportunities (and challenges!) in realizing the effectiveness of data and structure for knowledge discovery.
Bio:
Lise Getoor is a Distinguished Professor in the Computer Science & Engineering Department at UC Santa Cruz, where she holds the Jack Baskin Endowed Chair in Computer Engineering. She is founding Director of the UC Santa Cruz Data Science Research Center and is a Fellow of ACM, AAAI, and IEEE. Her research areas include machine learning and reasoning under uncertainty and she has extensive experience with machine learning and probabilistic modeling methods for graph and network data. She has over 250 publications including 13 best paper awards. She has served as an elected board member of the International Machine Learning Society, on the Computing Research Association (CRA) Board, as Machine Learning Journal Action Editor, Associate Editor for the ACM Transactions of Knowledge Discovery from Data, JAIR Associate Editor, and on the AAAI Executive Council.. She is a Distinguished Alumna of the UC Santa Barbara Computer Science Department and received the UC Santa Cruz Women in Science & Engineering (WISE) award. She received her PhD from Stanford University in 2001, her MS from UC Berkeley, and her BS from UC Santa Barbara, and was a professor at the University of Maryland, College Park from 20012013.
THE STRACHEY LECTURES ARE GENEROUSLY SUPPORTED BY OxFORD ASSET MANAGEMENT
http://rss.oucs.ox.ac.uk/tag:20221028:163849:000:file:319714:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20221027comlabcomscistracheygetoor.mp3
An overview of work on probabilistic soft logic (PSL), an SRL framework for largescale collective, probabilistic reasoning in relational domains and a description of recent work which integrates neural and symbolic (NeSy) reasoning.
An overview of work on probabilistic soft logic (PSL), an SRL framework for largescale collective, probabilistic reasoning in relational domains and a description of recent work which integrates neural and symbolic (NeSy) reasoning. Our ability to collect, manipulate, analyze, and act on vast amounts of data is having a profound impact on all aspects of society. Much of this data is heterogeneous in nature and interlinked in a myriad of complex ways. From information integration to scientific discovery to computational social science, we need machine learning methods that are able to exploit both the inherent uncertainty and the innate structure in a domain. Statistical relational learning (SRL) is a subfield that builds on principles from probability theory and statistics to address uncertainty while incorporating tools from knowledge representation and logic to represent structure. In this talk, I’ll overview our work on probabilistic soft logic (PSL), an SRL framework for largescale collective, probabilistic reasoning in relational domains. I’ll also describe recent work which integrates neural and symbolic (NeSy) reasoning. I’ll close by highlighting emerging opportunities (and challenges!) in realizing the effectiveness of data and structure for knowledge discovery.
Bio:
Lise Getoor is a Distinguished Professor in the Computer Science & Engineering Department at UC Santa Cruz, where she holds the Jack Baskin Endowed Chair in Computer Engineering. She is founding Director of the UC Santa Cruz Data Science Research Center and is a Fellow of ACM, AAAI, and IEEE. Her research areas include machine learning and reasoning under uncertainty and she has extensive experience with machine learning and probabilistic modeling methods for graph and network data. She has over 250 publications including 13 best paper awards. She has served as an elected board member of the International Machine Learning Society, on the Computing Research Association (CRA) Board, as Machine Learning Journal Action Editor, Associate Editor for the ACM Transactions of Knowledge Discovery from Data, JAIR Associate Editor, and on the AAAI Executive Council.. She is a Distinguished Alumna of the UC Santa Barbara Computer Science Department and received the UC Santa Cruz Women in Science & Engineering (WISE) award. She received her PhD from Stanford University in 2001, her MS from UC Berkeley, and her BS from UC Santa Barbara, and was a professor at the University of Maryland, College Park from 20012013.
THE STRACHEY LECTURES ARE GENEROUSLY SUPPORTED BY OxFORD ASSET MANAGEMENT
logic,reasoning,neural,symbolic,data,society,computational social science
Lise Getoor
3819
Thu, 27 Oct 2022 13:29:04 +0100

7
strachey
computing
c++
Stroustrup discusses the development and evolution of the C++, one of the most widely used programming languages ever. The development of C++ started in 1979. Since then, it has grown to be one of the most widely used programming languages ever, with an emphasis on demanding industrial uses. It was released commercially in 1985 and evolved through one informal standard (“the ARM”) and several ISO standards: C++98, C++11, C++14, and C++17. How could an underfinanced language without a corporate owner succeed like that? What are the key ideas and design principles? How did the original ideas survive almost 40 years of development and 30 years of attention from a 100+ member standards committee? What is the current state of C++ and what is likely to happen over the next few years? What are the problems we are trying to address through language evolution?
http://rss.oucs.ox.ac.uk/tag:20230118:090611:000:file:301923:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/strachey/20171121_stracheystroustrup.mp3
Stroustrup discusses the development and evolution of the C++, one of the most widely used programming languages ever.
Stroustrup discusses the development and evolution of the C++, one of the most widely used programming languages ever. The development of C++ started in 1979. Since then, it has grown to be one of the most widely used programming languages ever, with an emphasis on demanding industrial uses. It was released commercially in 1985 and evolved through one informal standard (“the ARM”) and several ISO standards: C++98, C++11, C++14, and C++17. How could an underfinanced language without a corporate owner succeed like that? What are the key ideas and design principles? How did the original ideas survive almost 40 years of development and 30 years of attention from a 100+ member standards committee? What is the current state of C++ and what is likely to happen over the next few years? What are the problems we are trying to address through language evolution?
strachey,computing,c++,20171121
Bjarne Stroustrup
3532
Tue, 12 Dec 2017 12:20:39 +0000

8
computer science
cryptography
universal machine
artificial intelligence
Professor Andrew Hodges author of 'Alan Turing: The Enigma' talks about Turing's work and ideas from the definition of computability, the universal machine to the prospect of Artificial Intelligence. In 1951, Christopher Strachey began his career in computing. He did so as a colleague of Alan Turing, who had inspired him with a 'Utopian' prospectus for programming. By that time, Turing had already made farreaching and futuristic innovations, from the definition of computability and the universal machine to the prospect of Artificial Intelligence. This talk will describe the origins and impacts of these ideas, and how wartime codebreaking allowed theory to turn into practice. After 1951, Turing was no less innovative, applying computational techniques to mathematical biology. His sudden death in 1954 meant the loss of most of this work, and its rediscovery in modern times has only added to Turing's iconic status as a scientific visionary seeing far beyond his short life.
Andrew Hodges is the author of Alan Turing: The Enigma (1983), which inspired the 2014 film The Imitation Game.
The Strachey Lectures are generously supported by OxFORD Asset Management.
http://rss.oucs.ox.ac.uk/tag:20230118:090851:000:file:297755:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/20161031_comsci_hodges.mp3
Professor Andrew Hodges author of 'Alan Turing: The Enigma' talks about Turing's work and ideas from the definition of computability, the universal machine to the prospect of Artificial Intelligence.
Professor Andrew Hodges author of 'Alan Turing: The Enigma' talks about Turing's work and ideas from the definition of computability, the universal machine to the prospect of Artificial Intelligence. In 1951, Christopher Strachey began his career in computing. He did so as a colleague of Alan Turing, who had inspired him with a 'Utopian' prospectus for programming. By that time, Turing had already made farreaching and futuristic innovations, from the definition of computability and the universal machine to the prospect of Artificial Intelligence. This talk will describe the origins and impacts of these ideas, and how wartime codebreaking allowed theory to turn into practice. After 1951, Turing was no less innovative, applying computational techniques to mathematical biology. His sudden death in 1954 meant the loss of most of this work, and its rediscovery in modern times has only added to Turing's iconic status as a scientific visionary seeing far beyond his short life.
Andrew Hodges is the author of Alan Turing: The Enigma (1983), which inspired the 2014 film The Imitation Game.
The Strachey Lectures are generously supported by OxFORD Asset Management.
computer science,cryptography,universal machine,artificial intelligence,20161031
Andrew Hodges, Mike Wooldridge
4042
Wed, 02 Nov 2016 12:49:49 +0000

9
quantum physics
quantum computing
science
computing
technology
Dr Scott Aaronson (MIT, UT Austin) gives the 2016 Strachey lecture. In the near future, it will likely become possible to perform specialpurpose quantum computations that, while not immediately useful for anything, are plausibly hard to simulate using a classical computer. These "quantum supremacy experiments" would be a scientific milestonedecisively answering quantum computing skeptics, while casting doubt on one of the foundational tenets of computer science, the Extended ChurchTuring Thesis. At the same time, these experiments also raise fascinating questions for computational complexity theorists: for example, on what grounds should we believe that a given quantum system really is hard to simulate classically?
Does classical simulation become easier as a quantum system becomes noisier? and how do we verify the results of such an experiment? In this lecture, I'll discuss recent results and open problems about these questions, using three proposed "quantum supremacy experiments" as examples: BosonSampling, IQP / commuting Hamiltonians, and random quantum circuits.
Based partly on joint work with Alex Arkhipov and with Lijie Chen.
The Strachey Lectures are generously supported by OxFORD Asset Management.
http://rss.oucs.ox.ac.uk/tag:20230118:090919:000:file:296066:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/20160524comsciaaronson.mp3
Dr Scott Aaronson (MIT, UT Austin) gives the 2016 Strachey lecture.
http://creativecommons.org/licenses/byncsa/2.0/uk/
Dr Scott Aaronson (MIT, UT Austin) gives the 2016 Strachey lecture. In the near future, it will likely become possible to perform specialpurpose quantum computations that, while not immediately useful for anything, are plausibly hard to simulate using a classical computer. These "quantum supremacy experiments" would be a scientific milestonedecisively answering quantum computing skeptics, while casting doubt on one of the foundational tenets of computer science, the Extended ChurchTuring Thesis. At the same time, these experiments also raise fascinating questions for computational complexity theorists: for example, on what grounds should we believe that a given quantum system really is hard to simulate classically?
Does classical simulation become easier as a quantum system becomes noisier? and how do we verify the results of such an experiment? In this lecture, I'll discuss recent results and open problems about these questions, using three proposed "quantum supremacy experiments" as examples: BosonSampling, IQP / commuting Hamiltonians, and random quantum circuits.
Based partly on joint work with Alex Arkhipov and with Lijie Chen.
The Strachey Lectures are generously supported by OxFORD Asset Management.
Creative Commons AttributionNonCommercialShare Alike 2.0 UK: England & Wales; http://creativecommons.org/licenses/byncsa/2.0/uk/
quantum physics,quantum computing,science,computing,technology
Scott Aaronson
4321
Tue, 14 Jun 2016 11:14:45 +0100

10
artificial intelligence
ai
Google
Deep Mind
computer science
In this talk Demis Hassabis discuss's what is happening at the cutting edge of AI research, its future impact on fields such as science and healthcare, and how developing AI may help us better understand the human mind. Strachey Lecture 2016, generously supported by Oxford Asset Management. Dr. Demis Hassabis is the CoFounder and CEO of DeepMind, the world’s leading General Artificial Intelligence (AI) company, which was acquired by Google in 2014 in their largest ever European acquisition. Demis will draw on his eclectic experiences as an AI researcher, neuroscientist and videogames designer to discuss what is happening at the cutting edge of AI research, its future impact on fields such as science and healthcare, and how developing AI may help us better understand the human mind.
http://rss.oucs.ox.ac.uk/tag:20230118:090951:000:file:294814:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/20160224stracheylecturenoq+a.mp3
In this talk Demis Hassabis discuss's what is happening at the cutting edge of AI research, its future impact on fields such as science and healthcare, and how developing AI may help us better understand the human mind.
In this talk Demis Hassabis discuss's what is happening at the cutting edge of AI research, its future impact on fields such as science and healthcare, and how developing AI may help us better understand the human mind. Strachey Lecture 2016, generously supported by Oxford Asset Management. Dr. Demis Hassabis is the CoFounder and CEO of DeepMind, the world’s leading General Artificial Intelligence (AI) company, which was acquired by Google in 2014 in their largest ever European acquisition. Demis will draw on his eclectic experiences as an AI researcher, neuroscientist and videogames designer to discuss what is happening at the cutting edge of AI research, its future impact on fields such as science and healthcare, and how developing AI may help us better understand the human mind.
artificial intelligence,ai,Google,Deep Mind,computer science,20160224
Demis Hassabis
3308
Fri, 26 Feb 2016 14:37:13 +0000

11
computer science
BX
viewupdate
lens
monad
modeldriven development
A reconstruction (slides and voiceover) of a talk given at the Summit on Advances in Programming Languages (snapl.org/2015) in May 2015. Bidirectional transformations inherently involve state effects. Modelling them that way allows the incorporation of other effects too, such as I/O, nondeterminism, and exceptions. We briefly outline the construction.
http://rss.oucs.ox.ac.uk/tag:20230118:091025:000:file:291752:audio
http://media.podcasts.ox.ac.uk/comlab/comsci/20151028_comscigibbons.mp3
A reconstruction (slides and voiceover) of a talk given at the Summit on Advances in Programming Languages (snapl.org/2015) in May 2015.
http://creativecommons.org/licenses/byncsa/2.0/uk/
A reconstruction (slides and voiceover) of a talk given at the Summit on Advances in Programming Languages (snapl.org/2015) in May 2015. Bidirectional transformations inherently involve state effects. Modelling them that way allows the incorporation of other effects too, such as I/O, nondeterminism, and exceptions. We briefly outline the construction. Creative Commons AttributionNonCommercialShare Alike 2.0 UK: England & Wales; http://creativecommons.org/licenses/byncsa/2.0/uk/
computer science,BX,viewupdate,lens,monad,modeldriven development,20151028
Jeremy Gibbons
316
Tue, 17 Nov 2015 14:29:46 +0000