Past events
Logic Colloquium
December 06, 2024, 4:10 PM
732 Evans Hall (Note the location change)
Isaac Goldbring
UC Irvine
On the undecidability of the QWEP for C*-algebras
In his landmark 1993 paper, Kirchberg introduced a property of C*-algebras called the QWEP, which stands for “quotient of the weak expectation property.” As the name suggests, the property is defined by the fact that the algebra is a quotient of a C*-algebra with the weak expectation property, which was a property introduced by Lance years earlier in connection with the theory of tensor products of C*-algebras. While at first glance this seems to be a strange property, Kirchberg showed that whether or not every C*-algebra has the QWEP is equivalent to the famous Connes Embedding Problem (CEP) from von Neumann algebra theory. The CEP remained open for nearly 50 years until its recent refutation in 2020 via a result in quantum complexity theory (as well as the equivalence with Kirchberg’s QWEP conjecture).
Several years ago, I showed that the QWEP is an axiomatizable property of C*-algebras. In this talk, I will present joint work with Aruseelan and Hart where we show that the QWEP does not have an effective axiomatization and, in fact, there can be no effectively axiomatizable satisfiable theory of C*-algebras all of whose models have the QWEP (modulo some nontrivially conditions). The proofs use the connection with the quantum complexity results mentioned above as well as other techniques from C*-algebra theory.
Logic Colloquium
November 15, 2024, 4:10 PM
60 Evans Hall
Eyal Kaplan
UC Berkeley
The potentialist principle
A Σ2 statement ϕ is called possible if, for every ordinal alpha, ϕ holds in a generic extension which preserves Vα (set of sets of rank less than α). Woodin’s Σ2 potentialist principle is the statement that every possible Σ2 sentence is true. In a joint work with Omer Ben-Neria and Gabriel Goldberg, we proved that a supercompact cardinal suffices to force the potentialist principle. Also, the principle implies the consistency of a Woodin cardinal. In the talk, we will present the principle and sketch the proof that it can be forced from a supercompact cardinal.
Logic Colloquium
November 01, 2024, 4:10 PM
60 Evans Hall
Felix Weilacher
UC Berkeley
Shannon’s Theorem and the Unbalanced Matching Problem in the Measurable Context
A theorem of Shannon states that any multigraph of maximum degree d admits a proper edge coloring using at most 3d/2 colors. We investigate the status of this theorem in the setting of “descriptive” combinatorics. That is, we are interested in finding Borel, measurable, etc. edge colorings of Borel graphs on standard Borel spaces.
We focus on the measurable setting as this is where most of the work needed to be done. There, we obtain a full generalization of Shannon’s result: Any Borel multigraph of maximum degree d on a standard probability space admits a Borel edge coloring using at most 3d/2 colors almost everywhere. Similarly in the Baire category setting.
We also prove that losing a null/meager set is not necessary for graphs of subexponential growth rate.
The proof uses a coloring procedure first used in distributed computing by Ghaffari, Kuhn, Maus, and Uitto. In the measurable setting, the most difficult step turns out to be the following result, interesting in its own right: Let G be a Borel bipartite multigraph on a standard probability space where all vertices on one side have larger degree than all vertices on the other. Then there is a Borel matching of G covering almost all the large degree vertices. Though for measure preserving graphs this follows immediately from the well known Lyons-Nazarov theorem, the general result requires new ideas inspired by Grebik’s recent proof of the measurable Vizing’s theorem.
This is joint work with Anton Bernshteyn and Matt Bowen.
Logic Colloquium
October 18, 2024, 4:10 PM
60 Evans Hall
Andrew Bacon
USC
How to take a mathematical model seriously
In modern logic and in philosophy it is common to take interpreted languages to be the primary vehicle for representing the world, with models often playing a supporting role for establishing metatheoretic results. In the sciences, by contrast, it is much more common to represent the world directly using interpreted mathematical models. Part of the reason for this difference has to do with the absence of intended models for key mathematical and philosophical languages. For instance, the languages of set theory, higher-order logic, quantified modal logic, and languages with empty names all lack intended models. Nonetheless, for these languages we can construct interpreted models that partially capture their subject matter, treating some aspects of the model (such as its cardinality) as an “artefact”. In this talk I will present a framework for theorizing about subject matters, partially interpreted models, and artefacts using some categorical ideas. I will then introduce the concept of notionally sound and complete model, a model that not only gets all the truths correct but is intended in a stronger sense that can be made precise.
Logic Colloquium
October 04, 2024, 4:10 PM
60 Evans Hall
Elliot Glazer
Epoch AI
Versatility of Random Reals
This talk is to show off the versatility of random real forcing for obtaining relative consistency results (and some positive results) over a wide variety of theories. When one wishes to find the nicest way the reals can behave regarding some problem, it is common to first analyze the question in the Solovay model or even determinacy models for their strong regularity properties, at the expense of large cardinal strength. But these constructions can often be carried out in random real extensions, achieving relative consistency over ZF, Z, and nth order arithmetic, while also maintaining fragments of choice far beyond DC.
I’ll go over some of the key features of random real forcing and identify some commonalities with and advantages over Cohen forcing. Then I will discuss some recent results that rely on random reals, pertaining to projective models of arithmetic, planar combinatorics, and prediction puzzles.
Logic Colloquium
September 20, 2024, 4:10 PM
60 Evans Hall
Ted Slaman
UC Berkeley
Extending Borel’s Conjecture from Measure to Dimension
Borel (1919) defined a set of real numbers A to have strong measure zero if for every sequence of positive numbers (epsilon_i: i in omega) there is an open cover of A, (U_i: i in omega), such that for each i, the diameter of U_i is less than epsilon_i. Besicovitch (1956) showed that A has strong measure zero if and only if A has strong dimension zero, which means that for every gauge function f, A is null for its associated measure H^f. We say that a subset of A of R^n has strong dimension f if and only if H^f(A)>0 and for every gauge function g of higher order H^g(A)=0. Here, g has higher order than f when lim_{t to 0^+} g(t)/f(t)=0.
Borel conjectured that a set of strong measure zero must be countable. This conjecture naturally extends to the assertion that a set has strong dimension f if and only if it is sigma-finite for H^f. Sierpinski (1928) used the continuum hypothesis to give a counterexample to Borel’s conjecture and Besicovitch (1963) did the same for its generalization. Laver (1976) showed that Borel’s conjecture is relatively consistent with ZFC, the conventional axioms of set theory including the axiom of choice. We will discuss the proof that its generalization to strong dimension is also relatively consistent with ZFC.
Logic Colloquium
September 13, 2024, 4:10 PM
60 Evans Hall
Forte Shinko
UC Berkeley
Hyperfiniteness for graphs of slow intermediate growth
A definable graph on a standard Borel space is hyperfinite if it is the increasing union of component-finite definable subgraphs. Hyperfiniteness is a strong form of amenability, and it is a long-standing open problem to determine whether every amenable graph is in fact hyperfinite. We are quite far from resolving the problem, which is most notably still open for Schreier graphs of solvable groups, although there is a positive answer for nilpotent and polycyclic groups. Another natural class where the problem is open is graphs of subexponential growth, that is, graphs for which there is a subexponential function f(n) such that every n-ball has at most f(n) vertices. Recently, it was shown by Bernshteyn and Yu that every graph of polynomial growth is hyperfinite. We extend this to show that there is a constant 0 < c < 1 such that every graph of growth exp(n^c) is hyperfinite. This is joint with Jan Grebík, Andrew Marks and Václav Rozhoň.
Logic Colloquium
April 26, 2024, 4:10 PM
60 Evans Hall
Sergei Artemov
CUNY Graduate
Center
The Provability of Consistency
Gödel’s second incompleteness theorem does not apply to Hilbert’s approach to proving consistency in the way it was previously thought. This is because Hilbert’s understanding of consistency as a series of claims “D is consistent” for each PA-derivation D is not provably in PA equivalent to the Gödelian consistency formula Con(PA). Kripke showed in 2022 that Hilbert’s epsilon-substitution method for PA, HES, fails for non-Gödelian reasons. We modify HES to prove the Hilbertian consistency of PA within PA. As expected, our proof of consistency does not yield a PA-derivation of Con(PA). This undermines the unprovability of consistency thesis that “there exists no consistency proof of a system that can be formalized in the system itself” (Encyclopedia Britannica).
Full text: “Serial Properties, Selector Proofs, and the Provability of Consistency,” arXiv preprint arXiv:2403.12272, 2024.
Logic Colloquium
April 19, 2024, 4:10 PM
60 Evans Hall
Giovanni Sambin
Professor of
Mathematical Logic, Università di Padova
Positive Topology: A New Practice in Constructive Mathematics
Since my first serious encounter with axiomatic set theory ZFC (at Berkeley as a PhD student in 1971-72), I have never been able to make its strong idealizations my own. This has led me, after more than fifty years of small steps, to a conception of mathematics that is radically different from the common one.
The main idea is to really accept that mathematics is a human construction—the outcome of a dynamic process—along with all its consequences. Then it can be seen as a conquered and local truth, which means certain and reliable information, and not as a given and universal truth. It becomes crucial to avoid principles such as the Law of the Excluded Middle, the Power Set Axiom, and the Axiom of Choice, that destroy the difference in information, respectively, between ∃ and ¬∀¬, inductively generated domains and not, operations and functions. The resulting dynamic-minimalist foundation provides a solution to foundational problems. In particular, the corresponding formal system MF is provably consistent.
This talk (with the same title as a book to appear next summer with Oxford UP) gives an overview of what mathematics, in particular topology, has been developed using this foundation. Most importantly, it illustrates the vast new regions of mathematical thought that emerge, in particular the co-presence and connection between real-effective and ideal-infinitary aspects, in particular between discrete and continuum, made possible via pointfree topology.
The challenge now is to show that we could arrive at a new Kuhnian paradigm in mathematics. This will inevitably require a considerable collective effort.
Logic Colloquium
April 12, 2024, 4:10 PM
60 Evans Hall
Martin Zeman
UC Irvine
Games with Filters, Strong Ideals, and Iterated Club Shooting
In a joint work with Foreman and Magidor we establish connections between winning strategies for the “good” player in games with filters introduced by Welch, ideals with high degree of distributivity, and distributivity of Easton support iterations of club shooting forcing. Along the way we obtain a condition that guarantees sufficient distributivity of such forcing iterations, which is of independent interest.
Logic Colloquium
March 15, 2024, 4:10 PM
60 Evans Hall
Tom Benhamou
Rutgers University
Commutativity of Cofinal Types of Ultrafilters
The Tukey order finds its origins in the concept of Moore-Smith convergence in topology, and is especially important when restricted to ultrafilters with reverse inclusion. The Tukey order on ultrafilters over ω was studied intensively by many, but still contains several fundamental unresolved problems. After providing the topological motivation for the Tukey order, I will present recent developments in the theory of the Tukey order restricted to ultrafilters on measurable cardinals, and explain how different the situation is when compared to ultrafilters on ω.
In the second part of the talk, we will demonstrate how ideas and intuition from ultrafilters over measurable cardinals led to new results at the level of ω.
Logic Colloquium
March 01, 2024, 4:10 PM
60 Evans Hall
Mariana Vicaría
UCLA
Model Theory of Valued Fields
Model theory is a branch of mathematical logic that studies structures (that is sets equipped with relations, functions and constants) and their definable sets, that is the subsets of various cartesian powers that can be defined in terms of these distinguished constants, relations and functions via the logical connectives and quantifiers. There is a more general class of subsets that one could study, called the interpretable sets, obtained by taking the quotient of a definable set by a definable equivalence relation. A natural question is: given a structure can one classify the interpretable sets in that structure?
A valued field is a field K equipped with a distinguished subset 𝒪, a valuation ring.* Examples of valued fields are the p-adic field ℚp or the Laurent series over the complex numbers ℂ((t)). Given 𝒪 a valuation ring of a field and ℳ its maximal ideal, we commonly refer to the additive quotient 𝒪/ℳ as the residue field, while the multiplicative quotient K×/𝒪× is an ordered abelian group and it is called the value group.
One of the most striking results in the model theory of valued fields is the Ax-Kochen/Ershov theorem which roughly states that the first order theory of an un-ramified henselian valued field is completely determined by the first order theory of its residue field and its value group. A principle follows from this theorem: the model theory of a valued field is controlled by its residue field and its value group.
In this talk I will make a brief description of valued fields and their model theory. I’ll present how the problem of classifying interpretable sets in henselian valued fields can be approached in an Ax-Kochen style: What obstructions come from the residue field? and from the value group? I will conclude by presenting the classification of the interpretable sets in valued fields obtained in joint work with Rideau-Kikuchi, building on [1] and [2].
[1] M. Hils and S. Rideau-Kikuchi, Un principe D’Ax-Kochen-Ershov imaginaire, preprint, arXiv:2109.12189.
[2] M. Vicaría, Elimination of Imaginaries in C((t)), Journal of the London Mathematical Society, Vol. 108, 2, (2023), 482-544
* Let K be a field, a subring A ⊆ K is said to be a valuation ring of K if for any element x ∈ K \ {0}, either x ∈ A or x−1 ∈ A.
Logic Colloquium
February 16, 2024, 4:10 PM
60 Evans Hall
Jing Zhang
University of
Toronto
The Strengths and Weaknesses of the Second Uncountable Cardinal
We intend to use specific combinatorial problems concerning ultrafilters to illustrate the title. Indecomposable ultrafilters were introduced by Keisler and Prikry as a weakening of measures on measurable cardinals, loosely speaking, by not insisting on countable completeness. Silver asked whether an inaccessible cardinal carrying an indecomposable ultrafilter necessarily has to be measurable. Sheard answered this negatively. However, recently Goldberg showed Silver’s question has a positive answer above a strongly compact cardinal. We will show that strong forcing axioms, for example PFA, imply a positive answer to Silver’s question, giving evidence to the heuristic that strong forcing axioms assert ω2 is “supercompact”. Previous evidence includes failure of square principles and the singular cardinal hypothesis. Then we will define a family of weak indexed square principles and use them to demonstrate that the positive result we obtain is indeed optimal, thus demonstrating an extent of the heuristic and showing that ω2 is “mortal” after all. Joint work with Chris Lambie-Hanson and Assaf Rinot.
Logic Colloquium
February 02, 2024, 4:10 PM
60 Evans Hall
Paolo Mancosu and Guillaume
Massas
UC Berkeley
Totality, Regularity and Cardinality in Probability Theory
Recent developments in generalized probability theory have renewed a debate about whether regularity (i.e., the constraint that only logical contradictions get assigned probability 0) should be a necessary feature of both chances and credences. Crucial to this debate, however, are some mathematical facts regarding the interplay between the existence of regular generalized probability measures and various cardinality assumptions. Using an old theorem due to Zermelo, we improve on several known results in the literature regarding the existence of regular generalized probability measures. These results, together with recent results in Non-Archimedean Probability Theory (NAP), give, under the Generalized Continuum Hypothesis, necessary and sufficient conditions for the existence of regular generalized probability measures defined on the whole powerset of any sample space.
Logic Colloquium
January 19, 2024, 4:10 PM
60 Evans Hall
Lenore Blum
UC Berkeley
A Theoretical Computer Science Perspective on Consciousness and Artificial General Intelligence
The quest to understand consciousness, once the purview of philosophers and theologians, is now actively pursued by scientists of many stripes. We have defined the Conscious Turing Machine (CTM) for the purpose of investigating a Theoretical Computer Science (TCS) approach to consciousness. For this, we have hewn to the TCS demand for simplicity and understandability. The CTM is consequently and intentionally a simple machine. It is not a model of the brain, though its design has greatly benefited - and continues to benefit - from cognitive neuroscience, in particular the global (neuronal) workspace theory. Although it is developed to understand consciousness, the CTM offers a thoughtful and novel guide to the creation of an Artificial General Intelligence (AGI). For example, the CTM has an enormous number of powerful processors, some with specialized expertise, others unspecialized but poised to develop an expertise. For whatever problem must be dealt with, the CTM has an excellent way to utilize those processors that have the required knowledge, ability, and time to work on the problem, even if it, the CTM, is not aware of which of the processors these may be.
This is joint work with Manuel Blum and Avrim Blum.
Logic Colloquium
December 08, 2023, 4:10 PM
60 Evans Hall
John Wright
UC Berkeley
Nonlocal games, MIP* = RE, and the Connes Embedding Problem
A nonlocal game involves two players named Alice and Bob who want to work together to win as often as possible, but they are not allowed to communicate while the game is being played. Nonlocal games were introduced by John Bell in the 1960s in his classic resolution of the EPR paradox, where he showed that Alice and Bob can use quantum particles to “cheat” and outperform any classical strategy, and they have since become an important object of study in the field of quantum mechanics. Parallel to this, in the 1980s, nonlocal games were essentially rediscovered in a completely different context—theoretical computer science—where they play an important role in the study of interactive proof systems. The connection between these two areas was observed in a 2004 work of Cleve, Hoyer, Toner, and Watrous, who introduced a new complexity class named MIP*, corresponding to interactive proofs involving quantum provers. Since then, the study of MIP* has led to a fruitful exchange between the two fields, culminating in the 2020 proof that MIP* = RE by Ji, Natarajan, Vidick, Wright, and Yuen, a corollary of which was the resolution in the negative of both Connes’ embedding problem from operator algebras and Tsirelson’s problem from entanglement theory.
In this talk, I will survey the topics of nonlocal games and interactive proofs, and give a bird’s-eye view of the proof of MIP* = RE. No background in quantum mechanics, theoretical computer science, or operator algebras will be assumed.
Logic Colloquium
November 17, 2023, 4:10 PM
60 Evans Hall
Alexander Kechris
Caltech
The compact action realization problem
In this talk I will discuss realizations of countable Borel equivalence relations by continuous actions of countable groups, focusing in particular on the problem of realization by continuous actions on compact spaces and more specifically subshifts. This also leads to considering a natural universal space for actions and equivalence relations via subshifts and the study of the descriptive and topological properties in this universal space of various classes of countable Borel equivalence relations, especially the hyperfinite ones.
Logic Colloquium
November 03, 2023, 4:10 PM
60 Evans Hall
Meng-Che “Turbo” Ho
California
State University Northridge
Word problems of groups as ceers
Classically, the word problem of a group is the set of words equal to the identity of the group, and we analyze them using Turing reductions. In this talk, we consider the word problem of a group as a computably enumerable equivalence relation (ceer), namely, two words are equivalent if and only if they are equal in the group. We compare ceers using the computable reduction: E is reducible to F if there is a computable function f so that iEj if and only if f(i)Ff(j).
We will discuss some recent results and see that the landscape of word problems as ceers is very different from the classical theory. For instance, in the classical setting, any Turing degree can be realized as a word problem by first constructing a countable group and then embedding it into a finitely presented group via the Higman embedding theorem. However, we prove that in the ceer setting, there is a group G whose word problem is not universal, but for any nontrivial H, the free product G * H has a universal word problem.
This is a joint work with Uri Andrews and Luca San Mauro.
Logic Colloquium
October 13, 2023, 4:10 PM
60 Evans Hall
Valeria de Paiva
Topos
Institute
War Time Proofs and Futuristic Programs
Historians remind us that we cannot predict the future if we don’t understand the past. And the past sometimes has new explanations, if we look for them. I want to show you a particular instance of the Curry-Howard correspondence, based on the work of Kurt Gödel in the 40s that I claim is an example of this phenomenon of new explanations from the past. Gödel wanted to prove the consistency of arithmetic by restricted means. In the 40s he came up with the ingenious idea of interpreting any formula of arithmetic as a formula of the shape ∃u : U∀x : X.AD(u, x) where AD is quantifier-free. This was his famous Dialectica interpretation, published in 1958. Some 30 years after that, following Hyland’s suggestion, I showed how Gödel’s idea corresponds to a kind of linear logic type theory. This was interesting then because Linear Logic was just beginning and people didn’t know how good for Computer Science it was going to be. Now, 35 years later, Linear Logic is used in many advanced programming languages and type systems, and more applications are expected. But we’re still explaining and extending the tricks of the old master, as I hope to show you.
Logic Colloquium
September 29, 2023, 4:10 PM
60 Evans Hall
Antonina Kolokolova
Memorial
University of Newfoundland
Reasoning about complexity
What is the bound on the length of possible proofs of propositional formulas? Is P = NP? These questions, while deceptively simple to state, have eluded over half a century of concerted effort to resolve them. Could it be that the complexity of the meta-problem, of proving such computational complexity statements, is itself high – in a formal logical sense? In this talk, I will survey what is known about the complexity of reasoning about computational complexity, in particular in the setting of bounded arithmetic, including exciting new results that appeared in the last several years.
Logic Colloquium
September 15, 2023, 4:10 PM
60 Evans Hall
Scott Mutchnik
University of
Illinois Chicago
Perspectives on NSOP3
Since at least the first decade of the 21st century, two questions have troubled model theorists: how can we extend (neo)stability-theoretic methods beyond simplicity (and now, beyond NSOP2), and is NSOP2 equal to NSOP3? Recent progress on the equality of NSOP1 and NSOP2 has offered hope that these two questions are related. However, while new stability-theoretic relations such as Conant-independence and n-ð-independence have proven promising in making sense of the higher NSOPn hierarchy, positive global consequences of NSOPn, for n > 2, have continued to elude us. We discuss our recent finding of the first such results, on NSOP3. While it is still open whether NSOP3 coincides with NSOP2 (i.e. with NSOP1), these results are concrete in that they do not, in general, hold in NSOP4 theories. Previously, NSOP3 has been thought of as very different from NTP2, in the sense that there is no known NSOP3 NTP2 theory which is not also simple. It is therefore surprising that our results on NSOP3 theories illustrate similar behavior to NTP2 theories.
Logic Colloquium
September 01, 2023, 4:10 PM
60 Evans Hall
Snow Zhang
UC Berkeley, Department
of Philosophy
Imaging and Conditionalization Revisited
Many think that there are two kinds of supposition, indicative and subjunctive, and they call for different hypothetical revisions of one’s degrees of belief (credences). Rational credal revision under indicative supposition goes by Bayesian conditionalization, whereas rational credal revision under subjunctive supposition goes by imaging. But why is imaging the rational way to revise one’s credences under subjunctive supposition (assuming that it is)? The primary goal of this talk is to sketch an answer to this question by proving a Dutch book theorem and an accuracy-dominance theorem for imaging, parallel to de Finetti’s Dutch book and accuracy-dominance theorems for conditionalization. If time permits, we’ll discuss how these results shed light on the recent debates about probabilities of conditionals and partial supposition.
Logic Colloquium
May 05, 2023, 4:10 PM
60 Evans Hall
Joan Bagaria
Catalan Institute for
Research and Advanced Studies
Self-similarity in the Higher Infinite
I shall give an overview of recent results showing that many large-cardinal notions may be seen as particular instances of a general reflection principle: Structural Reflection.
Alfred Tarski Lectures
April 26, 2023, 4:00 PM
60 Evans
Richard Shore
Cornell
University
Reverse Mathematics: A Global View
We view the structure of reverse mathematics as a degree structure similar to that of the Turing degrees, 𝒟T with the ordering of Turing reducibility, ≤T. We define an ordering ≤P on sets of sentences of second order arithmetic S ≤P T ⇔ RCA0 ∪ T ⊢ φ for every φ ∈ S. As usual we consider the induced ordering ≤P on the equivalence classes s and t and the resulting structure 𝒟P. Important substructures are ℱP and ℛP consisting of the degrees of finitely and recursively axiomatizable sets of sentences. We prove a large number of global results about 𝒟P that differ remarkably from those for the analogous questions about 𝒟T and other degree structures. A few sample results are the following.
Theorem: 𝒟P is a complete algebraic lattice (with 0 and 1) and so pseudocomplemented. ℛP is an incomplete algebraic lattice. For each of them the compact elements are those in ℱP and the pseudocomplement of T in 𝒟P is ∨{φ | φ ∧ T = 0}. ℱP is the atomless Boolean algebra.
Theorem: The (first order) theories of ℱP and 𝒟P with ≤P (and so with ∨, ∧, 0, and 1) are decidable by applying major results of Tarski and Rabin.
Theorem: 𝒟P and ℱP have exactly 2ω many automorphisms
Theorem: There are only four finite sets which are definable in 𝒟P: ∅, {0}, {1} and {0,1}. There are only four countably infinite definable subsets of 𝒟P: ℱP, ℱP - {0}, ℱP - {1} and ℱP - {0,1}. In each case, no other such sets are fixed under all automorphisms of 𝒟P.
Theorem: Up to isomorphism, there are only four cones 𝒟Ps= {t | s ≤P t} in 𝒟P: {1}, {s, 1}, {s, s∨ t0, s∨ t1,1} and 𝒟P. We can characterize the few s that fall into each of the first three classes in terms of notions familiar in the general study of theories.
This analysis was prompted by my thinking about what I should talk about in these lectures. Much to my surprise, after I had worked out most of these results I discovered that Tarski had proven many of them some ninety years ago and so long before the rise of reverse mathematics.
Alfred Tarski Lectures
April 24, 2023, 4:00 PM
3 Physics
Richard Shore
Cornell
University
Reverse Mathematics: A Multiverse
The basic question that reverse mathematics attempts to address is how hard it is to prove particular theorems or carry out particular constructions of ordinary classical mathematics. Hard not in the sense of how many hours, days or years or cups of coffee it took a mathematician or many mathematicians to prove the theorem but hard in some more mathematical sense. This talk will be an introduction to the subject with hints of a new view. We primarily consider two related standard criteria of hardness and the relations between. One is based on the strength of formal systems of (second order) arithmetic in which theorems can be proved. The second is recursion theoretic and based on the computational complexity of the objects the theorems assert exist. We will also briefly discuss relations to common philosophical approaches to the foundations of mathematics as well as a number of newer approaches that vary the logic and formal systems allowed in the first case and variations on computation complexity notions in the second. Finally, we present a glimpse of a new setting for the universe of reverse mathematics. It is an analysis of the sort that has become common for recursion theoretic degree structures but yields drastically different results for proof theoretic strength.
Logic Colloquium
April 21, 2023, 4:10 PM
60 Evans Hall
Noah Schweber
Proof School
Ceers (on omega and above)
A ceer is a computably enumerable equivalence relation on the natural numbers. Ceers are preordered in a natural way, by saying E is reducible to F iff there is a computable function f such that for all x and y we have xEy ↔︎ f(x)Ff(y). Lots of specific properties of the preorder of ceers have been established; in another direction, however, we could look at the analogues of ceers in appropriate “generalized computability theories” and ask what properties of the ceers are (relatively) setting-independent.
In this talk we will see that the theory of the ceers is “robustly” as complicated as it could possibly be: there is a natural notion of “κ-ceer” for any infinite regular cardinal κ, and the theory of the κ-ceers is always maximally complicated in terms of κ-recursion theory. Interestingly, this seems to require a genuinely different argument once we look at uncountable κ. I will also present a separate argument which applies to a somewhat different class of generalized computability theories.
The first part of this work is joint with Uri Andrews and Andrea Sorbi, and the second part is joint with Uri Andrews, Steffen Lempp, and Manat Mustafa.
Logic Colloquium
April 07, 2023, 4:10 PM
60 Evans Hall
Zeynep Soysal
University of
Rochester
The Metalinguistic Construal of Mathematical Propositions
On metalinguistic construal, famously advanced by Robert Stalnaker, mathematical propositions are possible-worlds propositions about the relation between mathematical sentences and what these sentences express. In this talk, I’ll defend the metalinguistic construal of mathematical propositions against the commonly raised criticism that it yields a highly implausible account of mathematical content. My defense will draw upon on a conventionalist metasemantics for mathematics and build on my prior work addressing a different “internal” challenge faced by the metalinguistic construal, namely, that it ascribes too much mathematical knowledge on the standard possible-worlds account of belief and knowledge on which these are closed under entailment.
Logic Colloquium
March 31, 2023, 4:10 PM
60 Evans Hall
Ralf Schindler
WWU Münster
Martin’s Maximum, Woodin’s Pmax axiom (*), and Cantor’s Continuum Problem
In 2019, D. Asperó and I showed that Martin’s Maximum implies Woodin’s Pmax axiom (*), thereby amalgamating two prominent maximality principles which had sometimes been considered as competitors. Since then, variants of the “(*)-forcing” have been studied and applied to other problems. For instance, my student A. Lietz obtained a forcing extension in which NSω1 is ω1-dense. Many interesting questions remain.
Logic Colloquium
March 10, 2023, 4:10 PM
60 Evans Hall
Elaine Landry
UC Davis
As-If Mathematics were True
The aim of this talk is to show that when we shift our focus from solving philosophical problems to solving mathematical ones, we see that an as-if interpretation of mathematics can be used to provide an account of both the practice and the applicability of mathematics. I begin first with Plato to show that much philosophical milk has been spilt owing to our conflating the method of mathematics with the method of philosophy. I then use my reading of Plato to develop what I call as-ifism, the view that, in mathematics, we treat our hypotheses as if they were true first principles and we do this with the purpose of solving mathematical problems not philosophical ones. I next extend as-ifism to modern mathematics wherein the method of mathematics becomes the axiomatic method, noting that this engenders a shift from as-if hypotheses to as-if axioms. I next distinguish as-ifism from if-thenism, and use this to develop my structural as-ifist position. I end by showing that taking a methodological as-ifist route, by placing our focus on what is needed for the practice and applicability of mathematics, we are neither committed to the unconditional consistency of our mathematical axioms nor the unconditional truth of our background meta-mathematical theory. Simply, it is methodological considerations, and not metaphysical ones, that “condition” our as if assumptions of both the consistency of our mathematical axioms and the truth of our background meta-mathematical theory.
Logic Colloquium
February 10, 2023, 4:10 PM
60 Evans Hall
Hunter Spink
Stanford
University
Additive combinatorics at infinity in the presence of a lattice
(Joint with Spencer Dembner) If X is an algebraic or o-minimally definable set in C^n, we give a nice description of a set Y which together with exp(X) unions to the closure of exp(X), answering a question of Gallinaro. The key is a careful extension of the work of Peterzil-Starchenko using infinitesimal stabilizers to answer an analogous question about cocomplete lattice quotient maps of C^n.
Logic Colloquium
January 27, 2023, 4:10 PM
60 Evans Hall
Patrick Lutz
UCLA
The Solecki dichotomy and the Posner-Robinson theorem
The Solecki dichotomy in descriptive set theory, roughly stated, says that every Borel function on the real numbers is either a countable union of partial continuous functions or at least as complicated as the Turing jump. The Posner-Robinson theorem in computability theory, again roughly stated, says that every non-computable real looks like 0’ relative to some oracle. Superficially, these theorems look similar: both roughly say that some object is either simple or as complicated as the jump. However, it is not immediately apparent whether this similarity is more than superficial. If nothing else, the Solecki dichotomy is about third order objects—functions on the real numbers—while the Posner-Robinson theorem is about second order objects—individual real numbers. We will show that there is a genuine mathematical connection between the two theorems by showing that the Posner-Robinson theorem plus determinacy can be used to give a short proof of a slightly weakened version of the Solecki dichotomy (and vice-versa). The proof can also be immediately extended to prove a slightly weakened version of a recent theorem of Marks and Montalban which generalizes the Solecki dichotomy to higher levels of the Baire hierarchy.
Logic Colloquium
December 02, 2022, 4:10 PM
60 Evans Hall
Paolo Mancosu
UC Berkeley
How many points are in a line segment? From Grosseteste to numerosities.
In his commentary on Aristotle’s Physics, Robert Grosseteste (ca. 1175-1253), Oxford theologian and Chancellor of the University, wrote: “Moreover, [God] created everything by number, weight, and measure, and He is the first and most accurate Measurer. By infinite numbers which are finite to Him, he measured the lines which He created. By some infinite number which is fixed and finite to Him, He measured and numbered the one-cubit line; and by an infinite number twice that size, He measured the two-cubit line; and by an infinite number half that size, He measured the half-cubit line.” In Grosseteste’s account the numerosity of the points in a finite line segment covaries with the length of the line segment. This position gave rise to an interesting number of debates in the XIIIth century especially as a consequence of a challenge raised by the Oxford theologian Richard Fishacre (1205-1248) who set up a one to one correspondence between the points in line segments of different lengths. I will reconstruct some aspects of this medieval debate, connect it to later intuitions (Bolzano and Cantor), and then discuss recent results from the theory of numerosities to the effect that the counting of points in a line segment preserving the part-whole principle is compatible with Lebesgue measure. I conclude that Grosseteste’s intuitions can find a suitable mathematical implementation.
Logic Colloquium
November 04, 2022, 4:10 PM
60 Evans Hall
Thomas Scanlon
UC Berkeley
The perfectoid correspondence as a bi-interpretation
The Ax-Kochen-Ershov theorems express a sense in which the rings of p-adic numbers and of formal power series over the field of p elements are essentially the same: their first-order theories have the same limits along any non-principal ultrafilter on the primes. In celebrated work on “perfectoids”, Scholze established for a fixed prime number p an equivalences of categories between certain objects over perfectoid field of mixed characteristic and of the corresponding categories over the “tilt” of that field, a field of characteristic p, and then used these equivalences to transfer certain theorems from characteristic p to characteristic zero.
It would seem that first-order logic cannot explain such transfer theorems because the theories involved are explicitly different: in characteristic p, 1 + … + 1 (p-times) = 0, while in characteristic zero 1 + … + 1 ≠ 0. In joint work with Silvain Rideau-Kikuchi and Pierre Simon, we instead explain Scholze’s correspondence using bi-interpretations in logics slightly beyond first-order logic (either multi-sorted first-order logic or a very mild form of continuous logic).
Taking into account that this lecture is intended for the Logic Colloquium, I will focus on the logical issues (e.g. consequences of the existence of bi-interpretations) and will try to take a light touch with the algebra and geometry.
Logic Colloquium
October 21, 2022, 4:10 PM
60 Evans Hall
Wesley Holliday
UC Berkeley
Subsystems of classical logic and their semantics based on graphs
The starting points of my talk are two of the most important subsystems of classical logic arising in the foundations of mathematics and foundations of physics: intuitionistic logic (Heyting 1930) and orthologic (Birkhoff and von Neumann 1936). In a Fitch-style formulation of natural deduction, intuitionistic logic and orthologic can be obtained from a more basic logic, defined using only the introduction and elimination rules for the logical connectives, by the addition of rules of Reiteration and Reductio ad Absurdum, respectively. In this talk (based on https://arxiv.org/abs/2207.06993), I will characterize this “fundamental” logic both proof-theoretically and semantically. The semantic characterization is based on representation theorems for complete lattices with additional operations (e.g., negation, implication) using directed graphs.
Logic Colloquium
October 07, 2022, 4:10 PM
60 Evans Hall
Dino Rossegger
UC Berkeley and TU
Wien
The degrees of categoricity above 0″
The degree of categoricity of a computable structure is the least Turing degree that computes an isomorphism between any two computable isomorphic copies of the structure. It provides a robust measure of complexity for computable structures – if A has degree of categoricity d, then modulo d all its isomorphic copies have the same computational properties. Two of the main goals in this area are to obtain a characterization of the Turing degrees that are degrees of categoricity of computable structure and to see whether all degrees of categoricity are strong, i.e., arise from the isomorphisms between two copies of a structure.
In this talk I will introduce the methods and main results in the area and discuss a recent result of Csima and I that presents a breakthrough in our understanding of degrees of categoricity. We obtained a characterization of the degrees of categoricity above 0″ as what we call the treeable degrees above 0″ and showed that every degree of categoricity above 0″ is strong.
Logic Colloquium
September 23, 2022, 4:10 PM
60 Evans Hall
Branden Fitelson
Northeastern
University
Probabilities of Conditionals and Conditional Probabilities – Revisited
Lewis’s (1976) triviality argument against The Equation (a.k.a, Adams’s Thesis) rests on an implausibly strong presupposition about the nature of (epistemic) rational requirements. Interestingly, Lewis (1980) later rejected this presupposition. In his discussion of the Principal Principle, Lewis assumes something weaker, and more reasonable, about the nature of rational requirements. In this paper, I explain how to apply the insights of Lewis (1980) to repair Lewis’s earlier (1976) discussion. This leads to a more reasonable rendition of The Equation — one which is (a) immune from triviality, and (b) a better candidate for a (bona fide) rational requirement.
Logic Colloquium
September 09, 2022, 4:10 PM
60 Evans Hall
Andrew Marks
UCLA
Borel asymptotic dimension and hyperfiniteness of polycyclic groups
The simplest nontrivial class of Borel equivalence relations under Borel reducibility are the hyperfinite equivalence relations. This is made precise by the Glimm-Effros dichotomy of Harrington, Kechris, and Louveau. In 1984 Weiss posed the question of what countable groups have the property that all of their Borel actions are hyperfinite. In particular, he asked if this is exactly the class of amenable groups: a natural class of groups with “tame dynamics”. Informally, Weiss’s question asks if this notion of group theoretic tameness corresponds exactly to descriptive set theoretic simplicity. We will discuss a recent theorem that Weiss’s question has a positive answer for polycyclic groups, and we prove this using an adaptation of Gromov’s theory of asymptotic dimension to the Borel setting. This is joint work with Conley, Jackson, Seward, and Tucker-Drob.
Logic Colloquium
December 03, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Thomas Icard
Associate Professor of
Philosophy, Stanford University
Interleaving Logic and Counting
Reasoning with quantifiers in natural language combines logical and arithmetical features, transcending divides between qualitative and quantitative. This practice blends with inference patterns in “grassroots mathematics” such as pigeon-hole principles. Our topic is this cooperation of logic and counting, studied with small systems and gradually moving upward. We start with monadic first-order logic with counting. We provide normal forms that allow for axiomatization, determine which arithmetical notions are definable, and conversely, discuss which logical notions can be defined out of arithmetical ones, and what sort of (non-)classical logics are induced. Next we study a series of strengthenings in the same style, including second-order versions, systems with multiple counting, and a new modal logic with counting. As a complement to our fragment approach, we also discuss another way of controlling complexity: changing the semantics of counting to reason about “mass” or other aggregating notions than cardinalities. Finally, we return to natural language, confronting the architecture of our formal systems with linguistic quantifier vocabulary, modules such as monotonicity reasoning, and procedural semantics via semantic automata. We conclude with some thoughts on further entanglements of logic and counting in formal systems, on rethinking the qualitative/quantitative divide, and on empirical aspects of our findings. Joint work with Johan van Benthem.
Logic Colloquium
November 19, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Joel (Ronnie) Nagloo
Associate
Professor of Mathematics, University of Illinois Chicago
TBA
TBA
Logic Colloquium
November 05, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Snow Zhang
Bersoff Faculty Fellow
in Philosophy, New York University
Conglomerability, Countable Additivity and the Continuum Hypothesis
According to standard Bayesianism, rational agents should have degrees of belief (i.e. credences) that satisfy the Kolmogorov axioms of probability. In particular, rational credences should be countably additive. One of the strongest arguments for countable additivity is the conglomerability argument. Roughly, the idea is that rational agents should have credences that are countably additive because otherwise they would be disposed to change their credences in a predictable way. One objection against the conglomerability argument is that it begs the question: given the coherent theory of conditional probability, countable additivity is sufficient for conglomerability in countable partitions, but does not guarantee conglomerability in partitions of higher cardinality. But why should rational credences be conglomerable only in countable partitions, and not in uncountable partitions?
This talk explores whether the conglomerability argument succeeds if one adopts Kolmogorov’s theory of conditional probability instead. I argue that the answer is not obvious. In particular, under plausible assumptions, a natural generalization of the conglomerability argument fails because it entails both the continuum hypothesis and its negation.
Logic Colloquium
October 22, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Adam Bjorndahl
Associate Professor
of Philosophy, Carnegie Mellon University
The Epistemology of Nondeterminism
Propositional dynamic logic (PDL) is a framework for reasoning about nondeterministic program executions (or, more generally, nondeterministic actions). In this setting, nondeterminism is taken as a primitive: a program is nondeterministic iff it has multiple possible outcomes. But what does “possible” mean, here? This talk explores an epistemic interpretation: working in an enriched logical setting, we represent nondeterminism as a relationship between a program and an agent deriving from the agent’s (in)ability to adequately measure the dynamics of the program execution. More precisely, using topology and the framework of dynamic topological logic, we show that dynamic topological models can be used to interpret the language of PDL in a manner that captures the intuition above, and moreover that continuous functions in this setting correspond exactly to deterministic processes. We prove that certain axiomatizations of PDL remain sound and complete with respect to corresponding classes of dynamic topological models. We also extend the framework to incorporate knowledge using the machinery of subset space logic, and show that the topological interpretation of public announcements coincides exactly with a natural interpretation of test programs. Finally, we sketch a generalization of the topological paradigm in which the distinction between action and measurement (i.e, between functions and opens) is erased, highlighting some preliminary results in this direction.
Logic Colloquium
September 24, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Zoltán Vidnyánszky
Taussky-Todd
Instructor in Mathematics, Caltech
Borel combinatorics, the LOCAL model, and determinacy
In the first half of the talk, I will give a high-level overview of Borel combinatorics and Linial’s LOCAL model of distributed computing, and the surprising connection between the two fields. The second half of the talk will be centered around a particular instance of this connection. One of the most striking results of Borel combinatorics is Marks’ determinacy method: Marks proved the existence of d-regular acyclic Borel graphs with Borel chromatic number d + 1 using the Borel determinacy theorem. I will discuss how the adaptation of this method to the LOCAL world inspires a more general version, which leads to a very rich class of examples of 3-regular acyclic Borel graphs. This yields new proofs of several known results. As a new application, I will show that it is hard to decide whether the Borel chromatic number of a Borel graph is ≤ 3, even for acyclic 3-regular graphs (that is, such graphs form a Σ21-complete set).
Logic Colloquium
September 10, 2021, 4:10 PM
https://berkeley.zoom.us/j/94041701380?pwd=dVBobHNqSGNjT2ttdVpEMWxmcG93dz09
Registration with Zoom is required for access.
Andrew Marks
Professor of
Mathematics, UCLA
A dichotomy characterizing piecewise Baire class α functions
In the 1920s, Lusin asked whether every Borel function on 2ω is a union of countably many partial continuous functions (i.e. whether every Borel function is piecewise continuous). This question has a negative answer; an example of a non-piecewise continuous Borel function is the Turing jump. This is the only counterexample in one sense. Solecki and Zapletal have shown that every Borel function f is either piecewise continuous, or the Turing jump continuously reduces to f.
We generalize the Solecki-Zapletal dichotomy throughout the Borel hierarchy. Recall that a Borel function is Baire class α if and only if it is Σα + 10 measurable. We show that that every Borel function f is either piecewise Baire class α, or the complete Baire class α + 1 function (an appropriate iterate of the Turing jump) continuously reduces to f. Our proof uses an adaptation of Montalbán’s game metatheorem for priority arguments to boldface descriptive set theory. We also discuss some connections of these dichotomy theorems to Martin’s conjecture.
This is joint work with Antonio Montalbán.
Logic Colloquium
April 23, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Henry Towsner
Associate Professor
of Mathematics, University of Pennsylvania
Tameness and Hypergraph Regularity
Szemeredi’s regularity lemma says that any graph can be approximated by a finite grid of quasirandom graphs. A series of results show that various model theoretic tameness notions - NIP, stable, and distal - guarantee various additional properties of the approximating graphs. Once we generalize to hypergraphs - that is, to k-ary relations with k>2 - the situation becomes more complicated. Without assuming prior knowledge of graph or hypergraph regularity, we will discuss the generalizations to tame versions of weak hypergraph regularity, and our recent result with Chernikov giving a generalization of the NIP case to full hypergraph regularity using the notion of NIP_k hypergraphs.
Logic Colloquium
April 09, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Marcus Rossberg
Associate Professor
of Philosophy, University of Connecticut
An Inferentialist Redundancy Theory of Truth
I present a “fully schematic”, proof-theoretic account of higher-order logic. The framework allows for the explicit definition of truth predicates for arbitrarily strong theories formulated in the framework. Accordingly, for any reasonable language, a truth predicate is explicitly definable with purely logical, proof-theoretic means. Enough “truth” should thus be available to the inferentialist to do the theoretical work that opponents claim is wanting. Furthermore, deflationists claim that the truth-predicate does not express a substantive property, but is only required to express generalizations (and such like). On the account presented here, truth-predicates are purely logical, and indeed eliminable since they are logically definable. Deflationism appears to collapse into a redundancy theory of truth.
Logic Colloquium
March 12, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Nicholas Ramsey
Hedrick Visiting
Assistant Adjunct Professor, UCLA
Measures in model theory
A Keisler measure on a structure, in the sense of model theory, is a finitely additive probability measure on sets definable in that structure. By restricting attention to the smaller, and often more structured, algebra of definable sets, the study of these measures gives important information about the algebraic, dynamical, and combinatorial properties of the structures themselves. Keisler measures arise naturally in the study of groups via the notion of definable amenability and in asymptotic combinatorics via nonstandard counting measures. We will survey the constructions and results that established Keisler measures as core objects of study in model theory, and then describe a cluster of new results that give a new understanding of these measures among structures that exhibit different forms of combinatorial complexity.
Logic Colloquium
February 26, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Minh Chieu Tran
Kenna Visiting
Assistant Instructor, University of Notre Dame
Minimal and nearly minimal measure expansions in locally compact group
I will present a number of results from recent joint works with Yifan Jing and Ruixiang Zhang and discuss various connections to model theory. These include a classification of connected unimodular groups and compact subsets allowing minimal and nearly minimal measure expansions; this answers a question by Kemperman in 1964 and confirms the connected special cases of conjectures by Griesmer and Tao. Another result is a generalization of the Brunn-Minkowski inequality to all locally compact groups, and showing that this is sharp for the class of helix-free locally compact groups (containing all real linear algebraic groups, and more generally, o-minimally definable Lie groups); this resolves a problem proposed by Henstock and Macbeath in 1953 for these groups and answers questions by Hrushovski, McCrudden, and Tao.
Logic Colloquium
February 12, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Dima Sinapova
Associate Professor,
University of Illinois at Chicago
Iteration, reflection, and singular cardinals
Two classical results of Magidor are:
from large cardinals it is consistent to have reflection at ℵω + 1, and
from large cardinals it is consistent to have the failure of SCH at ℵω.
These principles are at odds with each other. The former is a compactness type principle. (Compactness is the phenomenon where if a certain property holds for every smaller substructure of an object, then it holds for the entire object.) In contrast, failure of SCH is an instance of incompactness. The natural question is whether we can have both of these simultaneously. We show the answer is yes.
We describe a Prikry style iteration, and use it to force stationary reflection in the presence of not SCH. Then we obtain this situation at ℵω. This is joint work with Alejandro Poveda and Assaf Rinot.
Logic Colloquium
January 29, 2021, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
James Walsh
Klarman Postdoctoral
Fellow in Philosophy, Cornell University
Reducing omega-model reflection to iterated syntactic reflection
Two types of principles are commonly called “reflection principles” in reverse mathematics. According to syntactic reflection principles for T, every theorem of T (from some complexity class) is true. According to semantic reflection principles, every set belongs to some (sufficiently correct) model of T. We will present a connection between syntactic reflection and semantic reflection in second-order arithmetic: For any Pi^1_2 axiomatized theory T, every set is contained in an omega-model of T if and only if every iteration of Pi^1_1 reflection for T along a well-ordering is Pi^1_1 sound. There is a thorough proof-theoretic understanding of the latter in terms of ordinal analysis. Accordingly, this reduction yields proof-theoretic analyses of omega-model reflection principles. This is joint work with Fedor Pakhomov.
Logic Colloquium
November 13, 2020, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Rehana Patel
AIMS (Senegal) and
MSRI
Combining logic and probability in the presence of symmetry
Among the many approaches to combining logic and probability, an important one has been to assign probabilities to formulas of a classical logic, instantiated from some fixed domain, in a manner that respects logical structure. A natural additional condition is to require that the distribution satisfy the symmetry property known as exchangeability. In this talk I will trace some of the history of this line of investigation, viewing exchangeability from a logical perspective, and report on the current status of a joint program of Ackerman, Freer and myself on countable exchangeable structures.
Logic Colloquium
October 30, 2020, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Anush Tserunyan
Assistant Professor
of Mathematics, McGill University
A backward ergodic theorem and its forward implications
In the classical pointwise ergodic theorem for a probability measure preserving (pmp) transformation T, one takes averages of a given integrable function over the intervals {x, T(x), T2(x), …, Tn(x)} in front of the point x. We prove a “backward” ergodic theorem for a countable-to-one pmp T, where the averages are taken over subtrees of the graph of T that are rooted at x and lie behind x (in the direction of T−1). Surprisingly, this theorem yields forward ergodic theorems for countable groups, in particular, one for pmp actions of finitely generated groups, where the averages are taken along set-theoretic (but right-rooted) trees on the generating set. This strengthens Bufetov’s theorem from 2000, which was the most general result in this vein. This is joint work with Jenna Zomback.
Logic Colloquium
October 16, 2020, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Agustin Rayo
Professor of
Philosophy, MIT and Professorial Fellow, University of Oslo
Transcendence and Triviality
I argue that the notion of a logical truth can be naturally extended to the notion of a transcendent truth. (Roughly, a sentence is transcendentally true if its truth at a world can be established by one’s metatheory without relying on information about that world.) Whether or not the transcendental truths go beyond the logical truths depends on subtle questions concerning the relationship between our language and the world it represents. I develop a picture on which arithmetical truths count as transcendentally true and use it to defuse a stubborn problem in the philosophy of mathematics.
Logic Colloquium
October 02, 2020, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Matthew
Harrison-Trainor
Post-doctoral Fellow, Victoria University of
Wellington and Massey University, New Zealand
Scott complexity of countable structures
Dana Scott proved that every countable structure has a sentence of the infinitary logic Lω1ω which characterizes that structure up to isomorphism among countable structures. Such a sentence is called a Scott sentence, and can be thought of as a description of the structure. The least complexity of a Scott sentence for a structure can be thought of as a measurement of the complexity of describing the structure. I will give an introduction to the area, and then focus on three subtopics: connections with computability, Scott complexity of particular structures, and complexity in classes of structures.
Logic Colloquium
September 18, 2020, 4:10 PM
https://berkeley.zoom.us/j/99271828753
Registration with Zoom is required for access.
Greg Restall
Professor of
Philosophy, University of Melbourne
Speech Acts and the Problem of Classical Natural Deduction
It is tempting to take the logical connectives, such as conjunction, disjunction, negation and the material conditional to be defined by the basic inference rules in which they feature. Systems of ‘natural deduction’ provide the basic framework for studying these inference rules. In natural deduction proof systems, well-behaved rules for the connectives give rise to intuitionistic logic, rather than to classical logic. Some, like Michael Dummett, take this to show that intuitionistic logic is on a sounder theoretical footing than classical logic. Defenders of classical logic have argued that other proof-theoretical frameworks, such as Gentzen’s sequent calculus, or a bilateralist system of natural deduction, can provide a proof-theoretic justification of classical logic. Such bilateralist defences of classical logic have significant shortcomings, in that the systems of proof offered are much less natural than existing systems of natural deduction. Neither sequent derivations nor bilateralist natural deduction proofs are good matches for representing the inferential structure of everyday proofs.
In this talk I clarify the shortcomings of existing bilateralist defences of classical proof, and, making use of recent results in the proof theory for classical logic from theoretical computer science (specifically, the λμ calculus of Michel Parigot), I show that the bilateralist can give an account of natural deduction proof that models our everyday practice of proof as well as intuitionist natural deduction, if not better.
Logic Colloquium
May 01, 2020, 4:10 PM
https://berkeley.zoom.us/j/98233524597 (this meeting is password-protected: email pierre.simon (at) berkeley.edu to ask for the password.)
Emily Riehl
Associate Professor of
Mathematics, John Hopkins University
∞-category theory for undergraduates
At its current state of the art, ∞-category theory is challenging to explain even to specialists in closely related mathematical areas. Nevertheless, historical experience suggests that in, say, a century’s time, we will routinely teach this material to undergraduates. This talk describes one dream about how this might come about — under the assumption that 22nd century undergraduates have absorbed the background intuitions of homotopy type theory/univalent foundations.
Logic Colloquium
April 17, 2020, 4:10 PM
https://berkeley.zoom.us/j/715184959
Gabriel Conant
Research Associate,
Cambridge
NIP approximate groups and arithmetic regularity
I will present recent work with Anand Pillay on the structure of finite approximate groups satisfying a local NIP assumption. Our results can be seen as a unification of work of Breuillard, Green, and Tao on the structure theory of approximate groups, and the model-theoretic study of “tame” arithmetic regularity.
Logic Colloquium
April 03, 2020, 4:10 PM
https://berkeley.zoom.us/j/985996921
Joshua Wiscons
Assistant Professor
of Mathematics, California State University
The geometry of involutions in groups of finite Morley rank
We discuss a three-dimensional “geometry of involutions’’ arising in certain groups of finite Morley rank (fMr) that possess a subgroup C whose conjugates (1) generically cover the group and (2) intersect trivially. The resulting geometry is that of a generically defined projective 3-space, and our focus is on whether it is genuine or not: a distinction that separates SO3(ℝ) (which is not of fMr) from PGL2(ℂ) (which is). Our main result identifies a condition on the subgroup C that, if violated, implies that the geometry is genuine and the group does not exist (as it cannot have fMr). As an application of this result, we unify and generalize several known results about the elimination of involutions from certain non-algebraic (but possibly existent and possibly ``definably linear’’) groups of fMr.
This is joint work with Adrien Deloro.
Logic Colloquium
March 20, 2020, 4:10 PM
Online talk through Zoom
Thomas Scanlon
Professor of
Mathematics, UC Berkeley
Model theory meets modeling: Canonical parameters, canonical bases, and parameter identifiability
With a common form of mathematical modeling, one considers a system of ordinary differential equations ẋ = f(x, μ, y), y = g(x, μ, u), and h(x, mu, u) = 0 where x is a vector of state variables, y is a vector of output variables, u is a vector input variables, and μ is a vector of constants called the parameters. For us, the functions f, g, and h will be rational functions with rational coefficients. A key step in using such a model is to identify the parameters and one says that the parameters are identifiable if they may be computed from a generic output of an experiment.
There may be algebraic reasons why the parameters cannot be identified. We relate this to the model theoretic (in the sense of logic) notion of canonical parameters. There may be subtler reasons why the parameters cannot be identified. We relate this to the model theoretic notion of canonical bases of types.
However, from the theory of canonical bases, we show that provided the parameters are canonical, they may be identified from a finite number of outputs, and we compute bounds on this finite number.
(This a report on joint work with Alexey Ovchinnikov, Anand Pillay, and Gleb Pogudin.)
Logic Colloquium
March 06, 2020, 4:10 PM
60 Evans Hall
Nam Trang
Assistant Professor of
Mathematics, University of North Texas
The Strength of Hom∞ Sealing
A set of reals is universally Baire if all of its continuous preimages in topological spaces have the Baire property. Hom∞ Sealing (or just Sealing) is a type of generic absoluteness condition introduced by Woodin that asserts in strong terms that the theory of the universally Baire sets cannot be changed by forcing. In some sense, the largest class of sets of reals for which this Shoenfield-type generic absoluteness can hold is the collection of the universally Baire sets. The Largest Suslin Axiom (LSA) is a strong determinacy axiom that asserts the largest Suslin cardinal is inaccessible for ordinal definable bijections. Let LSA − over − uB be the statement that in all (set) generic extensions there is a model of LSA whose Suslin, co-Suslin sets are the universally Baire sets. We prove that modulo a mild large cardinal hypothesis, Sealing is equiconsistent with LSA-over-UB. A consequence of this is that Sealing is consistent relative to PFA or the existence of a Woodin cardinal which is a limit of Woodin cardinals. We hope to outline some ideas involved in the proof as well as explain why the consistency strength of Hom∞ sealing being low is a hurdle for inner model theory. This is joint work with Grigor Sargsyan.
Logic Colloquium
February 21, 2020, 4:10 PM
60 Evans Hall
Rohan French
Assistant Professor of
Philosophy, UC Davis
First Steps in a Theory of Metainferences
According to the nonreflexive approach to the semantic paradoxes presented in French (2016), no inference is, strictly speaking, valid. Nonetheless, there are a number of inference, such as that from a pair of sentences to their conjunction, which strike many of us as being valid. In order to account for this, it is argued in French (2016) that what we are really judging to be valid are metainferences—relations between sets of inferences and inferences. In this talk I’ll introduce the notion of a metainference, providing a partial taxonomy of the kinds of formal objects we might take them to be, and how we might reason about them semantically, before going on to provide a proof-system which treats metainferences as first-class citizens, showing it to be sound and complete with respect to a particular notion of validity for metainferences.
Logic Colloquium
February 07, 2020, 4:10 PM
60 Evans Hall
Erik Walsberg
Visiting Assistant
Professor, UC Irvine
First order expansions of (R,<,+)
I will discuss recent work on first order expansions of (R,<,+). I will describe the general dichotomy between tame and wild expansions and then describe a strong recent result on NIP expansions.
December 06, 2019, 4:10 PM
60 Evans Hall
Dana Bartosova
Assistant Professor
of Mathematics, University of Florida
Questions about phase spaces of minimal Boolean flows
By a flow, we mean a continuous action of a topological groups G on a compact Hausdorff space X. We refer to X as the phase space of the flow. We are primarily interested in minimal flows, that is flows that have no non-trivial proper closed invariant subset. Among minimal flows, there exists a maximal one called the universal minimal flow, M(G), which admits a continuous homomorphism onto every minimal flow. When G is non-Archimedean, that is, admits a neighbourhood basis of the identity of open subgroups, then M(G) is 0-dimensional. These are exactly groups of automorphisms of first-order structures with the topology of pointwise convergence. If M(G) is 0-dimensional, we can think dually in terms of its algebra of clopen subsets. We summarize which algebras are known to appear as phase spaces of universal minimal flows and we pose questions about the unknown.
Logic Colloquium
November 15, 2019, 4:10 PM
60 Evans Hall
Kenny Easwaran
Associate Professor
of Philosophy, Texas A&M University
Realism in Mathematics: The Case of the Hyperreals (joint work with Henry Towsner)
Although there was controversy about the role of the Axiom of Choice in mathematics in the early 20th century, it has become an accepted part of mathematical practice. One interesting consequence of its acceptance has been the development of non-standard analysis, which studies systems that behave like the real numbers, with the addition of infinitely large and infinitely small elements. These have been used to develop various models of physical, biological, and philosophical phenomena. We argue that these models can be used to develop understanding, and to make predictions and perform some calculations. However, we claim that these models can’t correspond to reality in ways that some other mathematical models can. We claim that no physical meaning can be given to the claim that a specific non-standard number represents some quantity. Some uses of the standard real numbers have this feature (like their use in differential equations representing predator-prey systems), but there is no in-principle barrier to their correspondence with physical reality, the way we claim there is for non-standard reals. This feature derives from their dependence on the Axiom of Choice, rather than from the presence of infinitely large and infinitely small elements.
Logic Colloquium
November 15, 2019, 4:10 PM
60 Evans Hall
Kenny Easwaran
Associate Professor
of Philosophy, Texas A&M University
Realism in Mathematics: The Case of the Hyperreals
Although there was controversy about the role of the Axiom of Choice in mathematics in the early 20th century, it has become an accepted part of mathematical practice. One interesting consequence of its acceptance has been the development of non-standard analysis, which studies systems that behave like the real numbers, with the addition of infinitely large and infinitely small elements. These have been used to develop various models of physical, biological, and philosophical phenomena. We argue that these models can be used to develop understanding, and to make predictions and perform some calculations. However, we claim that these models can’t correspond to reality in ways that some other mathematical models can. We claim that no physical meaning can be given to the claim that a specific non-standard number represents some quantity. Some uses of the standard real numbers have this feature (like their use in differential equations representing predator-prey systems), but there is no in-principle barrier to their correspondence with physical reality, the way we claim there is for non-standard reals. This feature derives from their dependence on the Axiom of Choice, rather than from the presence of infinitely large and infinitely small elements.
(joint work with Henry Towsner)
Logic Colloquium
November 01, 2019, 4:10 PM
60 Evans Hall
Isaac Goldbring
Associate Professor
of Mathematics, UC Irvine
The Connes Embedding Problem and Model Theory
The Connes Embedding Problem is the following problem in von Neumann
algebras:
does every tracial von Neumann algebra embed into an ultrapower of a
particular von Neumann algebra, the so-called hyperfinite II1 factor R. After the work of many
mathematicians, this problem has been shown to have equivalent
reformulations in C*-algebras, operator systems,
quantum information theory, free probability, and geometric group
theory, to name a few.
In this talk, I will discuss the many model-theoretic reformulations of this problem, including connections with existentially closed von Neumann algebras and computability of theories. Some of the work presented in this talk is joint with Bradd Hart and Thomas Sinclair. All relevant terminology from operator algebras will be defined.
Logic Colloquium
October 18, 2019, 4:10 PM
60 Evans Hall
Sebastian Eterovic
Visiting
Assistant Professor of Mathematics, University of California,
Berkeley
Categoricity of Shimura Varieties
The question of categoricity of the universal cover of certain algebraic varieties begun with Zilber’s study of the complex exponential function, and his methods have allowed people to study the question on other arithmetic varieties. In this talk we will consider the case of Shimura varieties, and we will see how some ideas from model theory and the classification of abstract elementary classes, can have explicit reformulations in terms of arithmetic geometry.
Logic Colloquium
October 04, 2019, 4:10 PM
60 Evans Hall
Gabriel Goldberg
NSF Postdoctoral
Fellow, University of California, Berkeley
Comparison principles and very large cardinals
In 1947, Gödel proposed a program to solve the Continuum Problem, as well as many other unsolvable problems of set theory, by supplementing the traditional ZFC axioms with large cardinal axioms. Although this program has been remarkably successful, serious limitations have since been discovered: for example, it turns out that large cardinal axioms do not help resolve the Continuum Problem itself. This talk is about new set theoretic principles that serve to amplify traditional large cardinals by endowing them with a powerful structure theory analogous to both the Comparison Lemma of inner model theory and the Wadge order of descriptive set theory. The first part of this talk deals with one of these principles, the Ultrapower Axiom (UA) and its relationship with the long-standing problem of building inner models with supercompact cardinals. The second part of the talk explores more recent work on generalizations of UA, especially the refutation of the Extender Power Axiom. The talk concludes with some speculation about the prospect of solving the Continuum Problem using UA.
Logic Colloquium
September 20, 2019, 4:10 PM
60 Evans Hall
Rod Downey
Professor of
Mathematics, Victoria University of Wellington
Foundations of Online Structure Theory
You are in a situation were you are given information in bundles one at a time, and you must make some kind of decision based on what you have seen so far. For example, you are packing objects into bins. You are in an online situation. I will look at some work trying to give a model theory to such situations.
Logic Colloquium
September 06, 2019, 4:10 PM
60 Evans Hall
Antonio Montalban
Professor of
Mathematics, University of California, Berkeley
Transfinite Ramsey Theorem
We consider a version of the Ramsey Theorem for coloring of tuples within a finite set where the exponent is transfinite. That is, the tuples we color are gamma-large for some ordinal gamma. As in Ramsey theorem we ask: How large should a finite set of numbers be to ensure that, for all colorings of the gamma-large tuples with a certain finite number of colors, there exists a homogeneous set that is alpha-large? Again, alpha being an ordinal. The answer should be an ordinal, and its existence follows from the Galvin-Prikry Theorem. What we ask is about the precise value of the bound, given as a function of the ordinals alpha and gamma. We also consider computability theoretic and reverse mathematics issues related to this.
Alfred Tarski Lectures
April 26, 2019, 4:00 PM
60 Evans Hall
Thomas Hales
Andrew Mellon
Professor of Mathematics, University of Pittsburgh
Integrating with Logic
In 1995, Kontsevich introduced a new form of integration, call motivic integration. From the start, the development of motivic integration has been guided by model theory, especially quantifier elimination. One particularly useful result has been a far-reaching generalization of the Ax-Kochen-Ersov transfer principle in logic to integration. This talk will give a gentle introduction to motivic integration and will highlight some applications to the Langlands program.
Alfred Tarski Lectures
April 24, 2019, 4:00 PM
4 LeConte Hall
Thomas Hales
Andrew Mellon
Professor of Mathematics, University of Pittsburgh
Formalizing mathematics
This talk will describe a broad long-term research program to make formal proofs in mathematics a practical reality. A formal proof is a mathematical proof that has been checked exhaustively by computer, on the basis of the fundamental axioms of mathematics and the primitive inference rules of logic. The field has progressed to the point that it is now possible to give formal proofs of major theorems in mathematics.
Alfred Tarski Lectures
April 22, 2019, 4:00 PM
4 LeConte Hall
Thomas Hales
Andrew Mellon
Professor of Mathematics, University of Pittsburgh
A formal proof of the Kepler conjecture
The Kepler Conjecture asserts that no packing of congruent balls in three-dimensional Euclidean space can have density greater than that of the face-centered cubic packing. This talk will describe the history and proof of the conjecture, including early attempts to reduce the problem to a finite calculation, controversy surrounding claimed proofs, the announcement of a proof by Sam Ferguson and me more than 20 years ago, and finally a formal proof of the Kepler conjecture in the HOL Light proof assistant in 2014.
Logic Colloquium
April 19, 2019, 4:10 PM
60 Evans Hall
Martin Davis
Professor Emeritus,
Courant Institute, NYU
Encounters with infinity
I will survey some examples of mathematical developments that have led to reassessments of philosophical attitudes with respect to the infinite. I will also discuss the impact of Gödel undecidability on mathematical practice.
Logic Colloquium
April 05, 2019, 4:10 PM
60 Evans Hall
François Loeser
Professor of
Mathematics, Sorbonne Université
Uniform bounds in Number Theory via Model Theory
I will present some recent applications of Model Theory to uniform bounds in questions arising from Number Theory.
Logic Colloquium
March 22, 2019, 4:10 PM
60 Evans Hall
Sam van Gool
Utrecht University,
the Netherlands
Pro-aperiodic monoids and model theory
This talk is about joint work with Benjamin Steinberg, in which we apply model theory to study algebraic structures occurring in the theory of regular languages.
Regular languages can both be understood as monadic-second-order-definable classes of finite colored linear orders, and as subsets of the free monoid which induce a certain finite-index congruence, called the syntactic congruence. Definability in fragments of monadic second order logic then corresponds to the finite monoid quotient by the syntactic congruence having special properties. As the first and most famous instance of this correspondence theory, essentially due to Schützenberger, the class of languages definable in first-order logic coincides with the class of languages recognizable by an aperiodic (“subgroup-free”) monoid.
In this setting, questions about definability in logic thus lead to challenges that can often be resolved by studying finite monoids. Within this “finite algebra” approach to regular languages, free profinite objects are a useful tool, as they provide the appropriate notion of “equation”. In particular, the free pro-aperiodic monoid is a crucial algebraic object for studying the class of first-order definable languages.
The starting point of our work is the observation that Stone duality and Schützenberger’s Theorem together imply that elements of the free pro-aperiodic monoid may be viewed as elementary equivalence classes of pseudofinite words. Concretely, this means that one may ‘compute’ with elements of the free pro-aperiodic monoid as if they were finite words, in a way reminiscent of the methods of non-standard analysis. In particular, model theory provides us with saturated words in each class of pseudofinite words, i.e., words in which all possible factorizations are realized. We prove that such saturated words are stable under algebraic operations, and we give several applications of our new model-theoretic approach to pro-aperiodic monoids, including a solution to the word problem for omega-terms that avoids using McCammond’s normal forms, as well as new proofs and extensions of other structural results concerning pro-aperiodic monoids.
Reference: S. J. v. Gool and B. Steinberg, Pro-aperiodic monoids via saturated models, Israel J. Math., accepted (2019). Preprint: https://arxiv.org/abs/1609.07736
Logic Colloquium
March 08, 2019, 4:10 PM
60 Evans Hall
Dana S. Scott, FBA,
FNAS
Distinguished Research Associate, UC Berkeley Department
of Philosophy
Free Logic: its formalization and some applications
Prof. J. Karel Lambert (Emeritus, UC Irvine) gave the subject its name and its profile as a well defined field of research over 60 years ago. “Free logic” is logic free of existential presuppositions in general and with respect to singular terms in particular. This talk will discuss its formalization and models for both classical and intuitionistic logic. Applications to understanding properties of descriptions, virtual classes, and structures with partial functions will be presented.
A bibliography on free logic may be found here.
Logic Colloquium
February 22, 2019, 4:10 PM
60 Evans Hall
Rizos Sklinos
Lecturer, Stevens
Institute of Technology
Elementary theories of hyperbolic groups
The discovery of non euclidean geometry in the early nineteenth century had shaken the beliefs and conjectures of more than two thousand years and changed the picture we had for mathematics, physics and even philosophy. Lobachevsky and Bolyai independently around 1830 discovered hyperbolic geometry. A notable distinguish feature of hyperbolic geometry is its negative curvature in a way that the sum of angles of a triangle is less than π. Gromov much later in 1987 introduced hyperbolic groups which are groups acting “nicely” on hyperbolic spaces, or equivalently finitely generated groups whose Cayley graphs are “negatively curved”. Main examples are free groups and almost all surface groups. The fascinating subject of hyperbolic groups touches on many mathematical disciplines such as geometric group theory, low dimensional topology and combinatorial group theory. It is connected to model theory through a question of Tarski.
Tarski asked around 1946 whether non abelian free groups have the same common first order theory. This question proved extremely hard to answer and only after more than fifty years in 2001 Sela and Kharlampovich-Myasnikov answered it positively. Both works are voluminous and have not been absorbed yet. The techniques almost exclusively come from the disciplines mentioned above, hence it is no wonder that the question had to wait for their development. The great novelty of the methods and the depth of the needed results have made it hard to streamline any of the proofs. Despite the difficulties there is some considerable progress in the understanding of the first order theory of “the free group” and consequently first order theories of hyperbolic groups from the scopes of basic model theory, Shelah’s classification theory and geometric stability. In this talk I will survey what is known about these theories and what are the main open questions.
Logic Colloquium
February 08, 2019, 4:10 PM
60 Evans Hall
Szymon Toruńczyk
Assistant
Professor, University of Warsaw
Computation with Atoms
The framework of Sets with Atoms has its roots in the works of Fraenkel and Mostowski, and has been recently rediscovered in the Computer Science community. I will show how this framework can be used in order to represent certain infinite structures in a finite way which can be processed by algorithms. This turns out to be equivalent to model-theoretic interpretations. Other notions from model theory, notably countable-categoricity, play a role in the complexity analysis of such algorithms. This has applications to verification of infinite-state systems.
Logic Colloquium
January 25, 2019, 4:10 PM
60 Evans Hall
Theodore A. Slaman
Professor of
Mathematics, UC Berkeley
Algorithmic and Metric Aspects of Randomness and Approximation
Randomness and approximation have been studied from the separate points of view of Diophantine Approximation and of Recursion Theory. We will discuss the similarities, divergences and connections between the two.
Logic Colloquium
November 30, 2018, 4:10 PM
60 Evans Hall
Ryan Williams
Associate Professor
of EECS, Massachusetts Institute of Technology and Visiting Professor of
EECS, UC Berkeley (Fall 2018)
Thinking Algorithmically About Impossibility
Computational complexity lower bounds like P != NP assert impossibility results for all possible programs of some restricted form. As there are presently enormous gaps in our knowledge of lower bounds, a central question on the minds of today’s complexity theorists is: how will we find better ways to reason about all efficient programs? I argue that some progress can be made by (very deliberately) thinking algorithmically about the lower bound problem. Slightly more precisely, to prove a lower bound against some class C of programs, we can start by treating C as a set of inputs to another (larger) process, which is intended to perform some basic analysis of programs in C. By carefully studying the algorithmic “meta-analysis” of programs in C, we can learn more about the limitations of the programs being analyzed. This way of thinking has led to several recent advances in complexity lower bounds where no progress had been made for many years. Indeed, in several interesting cases, the only way we presently know how to prove lower bounds against some classes C is to directly design algorithms for analyzing C.
Logic Colloquium
November 09, 2018, 4:10 PM
60 Evans Hall
Mark Colyvan
Professor of
Philosophy and Philosophy of Mathematics, University of Sydney; Visiting
Professor and Humboldt Fellow, Ludwig-Maximilians University
If 13 were not prime: Mathematics and counterpossibles
Standard approaches to counterfactuals in the philosophy of explanation are geared toward causal explanation. I suggest how to extend the counterfactual theory of explanation to non-causal, mathematical explanation. The core idea here is to model impossible perturbations to the relevant mathematics while tracking the resulting differences to the explanandum (either physical or mathematical, depending on whether we are dealing with extra-mathematical or intra-mathematical explanation). This approach has the potential to provide a unified account of explanation across science, mathematics, and logic.
Logic Colloquium
October 26, 2018, 4:10 PM
60 Evans Hall
Jouko Väänänen
Professor of
Mathematics, University of Helsinki
On an extension of a theorem of Zermelo
Zermelo (1930) proved the following categoricity result for set theory: Suppose M is a set and E, E’ are two binary relations on M. If both (M, E) and (M, E’) satisfy the second order Zermelo–Fraenkel axioms, then (M,E) and (M, E’) are isomorphic. Of course, the same is not true for first order ZFC. However, we show that if first order ZFC is formulated in the extended vocabulary {E,E’}, then Zermelo’s result holds even in the first order case. Similarly, Dedekind’s categoricity result (1888) for second order Peano arithmetic has an extension to a result about first order Peano.
Logic Colloquium
October 12, 2018, 4:10 PM
60 Evans Hall
Pavel Pudlak
Professor, Czech
Academy of Sciences
Combinatorial principles from proof complexity
One of the goals of proof theory is to find combinatorial characterization of sentences provable in particular theories, i.e., to present these sentences as mathematical principles, rather than mere syntactical statements. While for strong theories these sentences tent to be incomprehensible, for weak theories we expect to find something familiar, or at least similar to well-known principles. In this talk I will report on the project to characterize provably disjoint NP pairs of sets in systems called Bounded depth Frege systems. The resulting characterization is expressed in terms of positional strategies in some combinatorial games that generalize the standard concept of a finite game.
Logic Colloquium
September 28, 2018, 4:10 PM
60 Evans Hall
Pierre Simon
Assistant Professor of
Mathematics, UC Berkeley
On geometric homogeneous structures
A first order structure is homogeneous if any partial automorphism defined on a finite set extends to an automorphism of the full structure. I will present the first steps towards a classification of homogeneous structures which have few finite substructures. Peter Cameron and Dugald Macpherson conjectured some 30 years ago that such structures are tree-like or order-like. Model-theoretic results on NIP theories can be used to classify the order-like case. Applications include the classification of homogeneous structures in a language consisting of n linear orders and their reducts.
Logic Colloquium
September 14, 2018, 4:10 PM
60 Evans Hall
Thomas Scanlon
Professor of
Mathematics, UC Berkeley
Elimination and consistency checking for difference equations (even though the theory is undecidable!)
In practice, an algebraic difference equation (of N variables) is given by a set Σ of polynomials in the variables {xi, j : 1 ≤ i ≤ N, j ∈ ℕ} and one looks for sequences of N-tuples of numbers ((a1, j)j = 0∞, …, (aN, j)j = 0∞) as solutions in the sense that for each P ∈ Σ, the equations P(a1, j, …, an, j; a1, j + 1, …, aN, j + 1; …; a1, d + j, …, aN, d + j) = 0 hold for all j ∈ ℕ. Read more algebraically, difference equations may be understood as atomic formulae in the language of difference rings, the language of rings augmented by a function symbol for the difference operator, and we seek the solutions in the ring of sequences treated as a difference ring by interpreting the difference operator as a shift.
The theory of difference fields, that is, of fields equipped with a distinguished endomorphism, has a model companion and the theory of this model companion is known to be decidable and to admit quantifier elimination in a reasonable expansion of the language of difference rings. We set out to extend the model theory of difference fields to produce algorithms to test for the solvability of difference equations in sequence rings and to solve the elimination problem for such difference equations.
Remarkably, the theory of such difference rings is undecidable, already in low quantifier complexity. However, we were able to adapt the methods behind the axiomatization of the theory of difference closed fields to produce efficient algorithms to solve both the consistency checking and the elimination problems for difference equations in sequence rings. Even more remarkably, ultraproducts of Frobenius automorphisms play a crucial role in verifying the correctness of our algorithms.
(This is a report on joint work with Alexey Ovchinnikov and Gleb Pogudin [available at arXiv:1712.01412] and an on-going project with them joined by Wei Li.)
Logic Colloquium
August 31, 2018, 4:10 PM
60 Evans Hall
Paolo Mancosu
Professor of
Philosophy, UC Berkeley
Neologicist Foundations: Inconsistent Abstraction Principles and Part-Whole
Neologicism emerges in the contemporary debate in philosophy of mathematics with Wright’s book Frege’s Conception of Numbers as Objects (1983). Wright’s project was to show the viability of a philosophy of mathematics that could preserve the key tenets of Frege’s approach, namely the idea that arithmetical knowledge is analytic. The key result was the detailed reconstruction of how to derive, within second order logic, the basic axioms of second order arithmetic from Hume’s Principle (HP) ∀C, D (#(C) = #(D) ↔︎ C ≅ D) (and definitions). This has led to a detailed scrutiny of so-called abstraction principles, of which Basic Law V (BLV) ∀C, D (ext(C) = ext(D) ↔︎ ∀x (C(x) ↔︎ D(x))) and HP are the two most famous instances. As is well known, Russell proved that BLV is inconsistent. BLV has been the only example of an abstraction principle from (monadic) concepts to objects giving rise to inconsistency, thereby making it appear as a sort of monster in an otherwise regular universe of abstraction principles free from this pathology. We show that BLV is part of a family of inconsistent abstractions. The main result is a theorem to the effect that second-order logic formally refutes the existence of any function F that sends concepts into objects and satisfies a ‘part-whole’ relation. In addition, we study other properties of abstraction principles that lead to formal refutability in second-order logic.
This is joint work with Benjamin Siskind.
Logic Colloquium
April 27, 2018, 4:10 PM
60 Evans Hall
Vaughan Pratt
Professor Emeritus of
Computer Science, Stanford University
The Class CAT of Locally Small Categories as a Functor-Free Framework for Foundations and Philosophy
Exploiting an evident circularity, we give three elementary properties of the category Set that uniquely determine it up to equivalence within the class CAT of locally small categories. The analogous three properties extend this result to many other categories of interest in algebra, topology, type theory, and philosophy via a framework which structures categories with nothing more than distinguished objects each serving as either a primitive representative of a given sort such as cat or box, or a palette of sorted values for a given property such as color or mass. The notion of property thereby defined is as extensional as the notion of sort. In those categories catering for both sorts and and properties, qualia of sort s for property p arise ambiguously as sort-s members of the palette p and as p-states of the primitive s, for example calico as a possible color for a cat. This framework provides a paradox-free interpretation of C.I. Lewis’s notion of qualia as well as serving the function of the pituitary gland in Descartes’ mind-body dualism. While functors and natural transformations underpin the whole framework we exploit the Yoneda embedding to hide them beneath the framework’s more elementary conceptual language.
Logic Colloquium
April 13, 2018, 4:00 PM
60 Evans
Carlos Eduardo Areces
Professor of
Computer Science, Universidad Nacional de Córdoba, Argentina
Henkin Completeness in Modal Logic
Saul Kripke published in 1959 a proof of completeness for first-order S5 modal logic [7] and in 1963 he extended the method to cover the propositional modal systems T, S4, S5, and B [8]. His method employed a generalization of Beth’s tableaux and completeness was established by showing how to derive a proof from a failed attempt to find a counter model.
In 1966, Kaplan criticized Kripke’s proof in his review of Kripke’s article [6] and suggested a different and, arguably, more elegant approach based on an adaptation of Henkin’s model theoretic completeness proof for first-order logic [5]. (Actually, completeness proof for S5 following Henkin’s ideas had already been published in 1959 in an article by Bayart [2], and other proofs were been published at about the time of Kaplan’s by Makinson in 1966 [9] and Cresswell in 1967 [4].)
Henkin’s proof of completeness for first-order logic used to important ideas: 1) that a consistent set of formulas can be extended to a maximally consistent set of formulas, and 2) that existential quantifiers can be witnessed using constants, which can then be used to form the domain of the model. The first of these two ideas is fundamental in modern completeness proofs for classical propositional modal logics which build a canonical model (satisfying all consistent formulas) that has as domain the (uncountable) set of all maximally consistent sets of formulas. The second idea (the use of witnesses) seemed less useful in the propositional setting — till the arrival of hybrid modal logics [1].
One of the main characteristic of hybrid modal logics is the inclusion of nominals which are special propositional symbols that name particular states in the model. This “naming” is achieved by restricting the interpretation of nominals to be singleton sub- sets of the domain. Nominals can be used as witnesses for the modal existential oper- ators and, in this way, the completeness construction needs to build just a single max- imal, witnessed consistent set of formulas from which a countable model is built (see, e.g., [3, Chap 7.3] and [10]). As a side-effect a particularly strong completeness result can be proved, that establishes that it is possible to give a complete axiomatic system for the basic hybrid logic, which remains complete under the extension with a particular family of axioms w.r.t. the corresponding class of models.
[1] C. Areces and B. ten Cate., Hybrid logics, Handbook of Modal Logics, (P. Blackburn, F. Wolter, and J. van Benthem), Elsevier, 2006, pp. 821–868.
[2] A. Bayart, Quasi-adequation de la logique modale de 2eme ordre, Logique et analyse, vol. 2 (1959), no. 6/7 pp. 99–121.
[3] P. Blackburn, M. De Rijke, and Y. Venema, Modal Logic, Cambridge University Press, Cambridge, 2002.
[4] M. Cresswell, A Henkin completeness for T , Notre Dame Journal of For- mal Logic, vol. 8 (1967), no. 3 pp. 186–190.
[5] L. Henkin, The completeness of the first-order functional calculus, The Journal of Symbolic Logic, vol. 14 (1949), no. 3, pp. 159–166.
[6] D. Kaplan, Review: Saul A. Kripke, semantical analysis of modal logic i. normal modal propositional calculi, The Journal of Symbolic Logic, vol. 31 (1966), no. 1, pp. 120–122.
[7] S. Kripke, A completeness theorem in modal logic, The Journal of Symbolic Logic, vol. 24 (1959), no. 1, pp. 1–14.
[8] S. Kripke, Semantical analysis of modal logic i. normal modal propositional calculi, Zeitschrift fur mathematische Logik und Grundlagen der Mathematik, vol. 9 (1963), no. 5-6, pp. 67–96.
[9] D. Makinson, On some completeness theorems in modal logic, Zeitschrift fur mathematische Logik und Grundlagen der Mathematik, vol. 12 (1966), no. 1, pp. 379–384.
[10] B. ten Cate, Model theory for extended modal languages. PhD thesis, University of Amsterdam, 2005.
Logic Colloquium
March 23, 2018, 4:10 PM
60 Evans Hall
Nam Trang
Visiting Assistant
Professor, Department of Mathematics, University of Califorina,
Irvine
Forcing, elementary embeddings, and determinacy
Forcing and elementary embeddings are central topics in set theo= ry. Most of what set theorists have focused on are the study of forcing and elementary embeddings over models of ZFC.
In this talk, we focus on forcing and elementary embeddings over models of the Axiom of Determinacy (AD). In particular, we focus on answering the following questions: work in V which models AD. Let P be a forcing poset and g ⊆ P be V-generic. 1. Does V[g] model AD? 2. Is there an elementary embedding from V to V[g]?
Regarding question (1), we want to classify what forcings preserve AD. We show that forcings that add Cohen reals, random reals, and many other well-known forcings do not preserve AD. We, however, give an example of a forcing that preserves AD. Regarding question (2), an analogous statement to the famous Kunen’s theorem for models of ZFC, can be shown: suppose V = L(X) for some set X and V models AD, then there is no (nontrivial) elementary embedding from V to itself that moves some ordinal. We conjecture that there are no (nontrivial) elementary embeddings from V to itself. Woodin has shown that there are (P, g) that gives rise to nontrivial elementary embeddings from V to V[g].
We present some of the results discussed above. There is still much work to do to completely answer these questions. This is an ongoing joint work with D. Ikegami.
Alfred Tarski Lectures
March 15, 2018, 4:00 PM
126 Barrows Hall
Hugh Woodin
Professor of Philosophy
and Mathematics, Harvard University
The Ultimate L Conjecture
Alfred Tarski Lectures
March 14, 2018, 4:00 PM
1 LeConte Hall
Hugh Woodin
Professor of Philosophy
and Mathematics, Harvard University
The HOD Dichotomy
Alfred Tarski Lectures
March 13, 2018, 4:00 PM
1 LeConte Hall
Hugh Woodin
Professor of Philosophy
and Mathematics, Harvard University
Ultimate L
Logic Colloquium
March 09, 2018, 4:10 PM
60 Evans Hall
Ray Briggs
Professor of Philosophy,
Stanford University
Real-Life Newcomb Problems
Decision theorists have spilled much ink over the difference between causal and evidential decision rules, whose recommendations differ in puzzle cases. Recently, a class of rivals to causal and evidential rules, aimed at solving more complicated puzzle cases.
But how realistic are the puzzle cases where these decision rules disagree? I argue that such cases are rare, and can often be handled without addressing the difference between decision rules, by changing other parts of the problem description.
Logic Colloquium
February 23, 2018, 4:10 PM
60 Evans Hall
Douglas Marshall
Assistant
Professor of Philosophy, Carleton College
An Approach to Two Puzzles Concerning the Applicability of Mathematics
In its applications to empirical science, mathematics allows us to represent and reason about the physical world. A similar phenomenon occurs internally to mathematics when one branch of mathematics allows us to represent and reason about the subject matter of another. I argue that taking seriously the close analogies between these two kinds of case will help us resolve two philosophical puzzles concerning the applicability of mathematics: 1. If (as some platonists contend) mathematics studies abstract objects, but the empirical sciences study concrete objects, then how is mathematics relevant to the empirical sciences? 2. How can one branch of mathematics be applied to another if (as some empiricists contend) mathematics has no factual content?
Logic Colloquium
February 09, 2018, 4:10 PM
60 Evans Hall
Michael Beeson
Professor Emeritus,
San Jose State University
Proof-checking Euclid
We used computer proof-checking methods to verify the correctness of our proofs of the propositions in Euclid Book I. We used axioms as close as possible to those of Euclid, in a language closely related to that used in Tarski’s formal geometry. We used proofs as close as possible to those given by Euclid, but filling Euclid’s gaps and correcting errors. Then we checked those proofs in the well-known and trusted proof checkers HOL Light and Coq. The talk will contain many geometrical diagrams and discuss both the geometry and the proof-checking. (This is joint work with Julien Narboux and Freek Wiedijk.)
Logic Colloquium
January 26, 2018, 4:10 PM
60 Evans Hall
John MacFarlane
Professor of
Philosophy, University of California, Berkeley
Is Logic a Normative Discipline?
Various philosophers, including Kant, Frege, and more recently Hartry Field and Greg Restall, have held that logic is in some interesting sense a normative discipline. Field, in particular, has argued that we need a normative characterization of logical validity in order to understand what proponents of alternative logics are disagreeing about. I will discuss what one needs to say in order to sustain an interesting version of this claim, and I will consider whether Field gives good reasons for accepting it.
Logic Colloquium
December 01, 2017, 4:10 PM
60 Evans Hall
Marianna Antonutti
Marie
Skłodowska-Curie Postdoctoral Fellow, Munich Center for Mathematical
Philosophy
What is ordinary mathematics?
The term “ordinary mathematics” is used to denote a collection of areas that are, in some sense, central to the practice of most working mathematicians, typically taken to contain such fields as number theory, real and complex analysis, geometry, and algebra. This notion has been taken for granted by both philosophers and logicians; for example, reverse mathematics is often described as the study of what set existence axioms are necessary to prove theorems of ordinary mathematics, and the ability to recover ordinary mathematics has been considered a key measure of the success of a foundational framework at least since Russell. However, the precise extent of this notion has not been explicitly discussed. This talk will propose two ways in which this notion can be made more precise. According to the first, ordinary mathematics is that part of mathematics that concerns countable or countably representable objects. According to the second, ordinary mathematics is that part of mathematics that does not make essential use of intrinsically set theoretic methods or concepts. I will discuss potential counterexamples to both views, and assess the prospects for formulating a precise account of the notion of ordinary mathematics.
This is joint work with Benedict Eastaugh.
Logic Colloquium
November 17, 2017, 4:10 PM
60 Evans Hall
Wesley H. Holliday
Associate
Professor of Philosophy, UC Berkeley
Possibilities for Boolean, Heyting, and modal algebras
In this talk, we describe an alternative to the standard representation theory for Boolean, Heyting, and modal algebras descending from Stone (1934, 1937) and Jónsson and Tarski (1951). While the standard theory leads to the well-known “possible world semantics” in logic, the alternative theory forms the basis of the recently investigated “possibility semantics” in logic. We will give an overview of this line of research, making connections to longstanding open problems and highlighting the contributions of University of California students.
The talk will be based on the following papers:
W. H. Holliday, “Possibility Frames and Forcing for Modal Logic,” 2015.
G. Bezhanishvili and W. H. Holliday, “Locales, Nuclei, and Dragalin Frames,” 2016.
G. Bezhanishvili and W. H. Holliday, “A Semantic Hierarchy for Intuitionistic Logic,” 2017.
N. Bezhanishvili and W. H. Holliday, “Choice-Free Stone Duality,” 2017.
Logic Colloquium
November 03, 2017, 4:10 PM
60 Evans Hall
Lenore Blum
Distinguished Career
Professor of Computer Science, Carnegie Mellon University
Alan Turing and the Other Theory of Computation
Most logicians and theoretical computer scientists are familiar with Alan Turing’s 1936 seminal paper setting the stage for the foundational (discrete) theory of computation. Most however remain unaware of Turing’s 1948 seminal paper introducing the notion of condition, setting the stage for a natural theory of complexity for what I call the “other theory of computation” emanating from the classical tradition of numerical analysis, equation solving and the continuous mathematics of calculus. This talk will recognize Alan Turing’s work in the foundations of numerical computation (in particular, his 1948 paper “Rounding-Off Errors in Matrix Processes”), how it provides a unifying concept for the two major traditions of the Theory of Computation, and its influence on complexity theory today.
Logic Colloquium
October 20, 2017, 4:10 PM
60 Evans Hall
Thomas Icard
Assistant Professor of
Philosophy, Stanford University
Some Philosophical Uses of Randomized Computation
The aim of this talk will be to convey some of the ways that familiar ideas and tools from randomized and probabilistic computation might bear on issues of philosophical interest, focusing especially on questions about cognition, representation, and reasoning. A first question is when it would ever make sense for an agent to employ randomization in the course of decision making. Drawing on ideas from game theory, reinforcement learning, and statistics, we tentatively propose a unified answer to the question. A second set of questions revolves around the idea of characterizing an agent’s implicit causal knowledge of the world by appeal to the explicit causal structure of a probabilistic algorithm. This second topic raises novel questions about the logic of counterfactuals beyond the usual propositional setting. Much of this is work in progress.
Logic Colloquium
October 06, 2017, 4:10 PM
60 Evans Hall
Yann Pequignot
Postdoctoral Fellow,
Department of Mathematics, UCLA
Finite versus infinite: an insufficient shift
Analytic sets enjoy a classical representation theorem based on well-founded relations. I will explain a similar representation theorem for Sigma^1_2 sets due to Marcone which is based on an intriguing topological graph: the Shift Graph. The chromatic number of this graph is 2, but its Borel chromatic number is infinite. We use this representation theorem to show that the Shift Graph is not minimal among the graphs of Borel functions which have infinite Borel chromatic number. While this answers negatively the primary outstanding question from (Kechris, Solecki and Todorcevic; 1999), our proof unfortunately does not construct any explicit example of a Borel function whose graph has infinite Borel chromatic number and admits no homomorphism from the Shift Graph.
Logic Colloquium
September 22, 2017, 4:10 PM
60 Evans Hall
Nadja Hempel
Assistant Adjunct
Professor of Mathematics, UCLA
Mekler construction in generalized stability
Given a so called nice graph (no triangles, no squares, for any choice of two distinct vertices there is a third vertex which is connected to one and not the other), Mekler considered the 2-nilpotent subgroup generated by the vertices of the graph in which two elements given by vertices commute if and only if there is an edge between them. These groups form an interesting collection of examples from a model theoretic point of view. It was shown that such a group is stable if and only if the corresponding graph is stable and Baudisch generalized this fact to the simple theory context. In a joint work with Chernikov, we were able to verify this result for k-dependent and NTP2 theories. This leads to the existence of groups which are (k+1)-dependent but not k-dependent, providing the first algebraic objects witnessing the strictness of these hierarchy.
Logic Colloquium
September 08, 2017, 4:10 PM
60 Evans Hall
Cameron Freer
Founder, Borelian
Corporation and Chief Scientist, Remine LLC
Model theory of invariant probabilistic constructions
There is a long history of studying the logic obtained by assigning probabilities, instead of truth values, to first-order formulas. In a 1964 paper, Gaifman studied probability distributions on countable structures that are invariant under renaming of the underlying set – which he called “symmetric measure-models”, and which are essentially equivalent to what today are known as S∞-invariant measures. In this paper, he asked the question of which first-order theories admit invariant measures concentrated on the models of the theory.
We answer this question of Gaifman, a key first step towards understanding the model theory of these measures, which can be thought of as “probabilistic structures”. In this talk, we will also discuss related questions, such as how many probabilistic structures are models of a given theory, and when probabilistic structures are almost surely isomorphic to a single classical model.
Joint work with Nathanael Ackerman and Rehana Patel.
Logic Colloquium
August 25, 2017, 4:10 PM
60 Evans Hall
Andrew Marks
U.C. Los Angeles
A constructive solution to Tarski’s circle squaring problem
In 1925, Tarski posed the problem of whether a disc in ℝ2 can be partitioned into finitely many pieces which can be rearranged by isometries to form a square of the same area. Unlike the Banach-Tarski paradox in ℝ3, it can be shown that two Lebesgue measurable sets in ℝ2 cannot be equidecomposed by isometries unless they have the same measure. Hence, the disk and square must necessarily be of the same area.
In 1990, Laczkovich showed that Tarski’s circle squaring problem has a positive answer using the axiom of choice. We give a completely constructive solution to the problem and describe an explicit (Borel) way to equidecompose a circle and a square. This answers a question of Wagon.
Our proof has three main ingredients. The first is work of Laczkovich in Diophantine approximation. The second is recent progress in a research program in descriptive set theory to understand how the complexity of a countable group is related to the Borel cardinality of the equivalence relations generated by its Borel actions. The third ingredient is ideas coming from the study of flows in networks.
This is joint work with Spencer Unger.
May 06, 2017, 9:30 AM
North Gate Hall, Room 105
Jeremy Avigad, Johan van Benthem, Ronald
Fagin, Barbara Partee
Logic at UC Berkeley
A two-day conference in mathematical logic and related areas organized by The Group in Logic and the Methodology of Science at UC Berkeley (logic.berkeley.edu). The conference is partly occasioned by the fact that the Group in Logic turns sixty this year.
In 1957, a group of faculty members, most of them from the departments of Mathematics and Philosophy, initiated a pioneering interdisciplinary graduate program leading to the degree of Ph.D. in Logic and the Methodology of Science. The Group has fostered interdisciplinary work in which logic has interacted with mathematics, philosophy, statistics, computer science, linguistics, physics and other disciplines.
While mathematical logic at UC Berkeley cannot be identified only with the Group in Logic, the Group has played a vital role in Berkeley’s worldwide prominence in mathematical logic and significantly contributed to making Berkeley a mecca since the fifties for people interested in mathematical logic and its applications. A full list of all those researchers in logic who taught at UC Berkeley, or studied at UC Berkeley, or visited Berkeley for shorter or longer periods would result in a who’s who of mathematical logic.
While marking an important moment for logic at UC Berkeley, the conference will be forward looking rather than merely celebratory. We have invited eight internationally prominent scholars to talk about the future of mathematical logic in their respective areas of specialization.
The first day of the conference will have four invited speakers in the so-called “foundational” areas: set theory, model theory, recursion theory, and proof theory. The second day will have four invited speakers in areas where mathematical logic plays a prominent role, namely philosophy of logic and mathematics, formal semantics for natural languages, modal logic, and logic in computer science. The line up is given below.
For a map showing the conference lecture rooms, see the red arrows here.
May 5 (141 McCone Hall 9AM-2PM; 3 LeConte Hall 2-5PM)
introductory remarks (9-9:30am): Paolo Mancosu (UC Berkeley), “A brief history of the Group in Logic and the Methodology of Science” (slides)
set theory (9:30-10:45am): John Steel (UC Berkeley), “Absolutely ordinal definable sets” (slides)
model theory (11am-12:15pm): Ehud Hrushovski (Hebrew University, Oxford University), “Reflections on model theory and foundations”
recursion theory (2:15-3:30pm): Denis Hirschfeldt (University of Chicago), “Computability Theory” (slides)
proof theory (3:45-5:00pm): Michael Rathjen (Leeds University), “On relating type theories to (intuitionistic) set theories” (slides)
May 6 (105 North Gate Hall)
philosophy of logic and mathematics (9:30-10:45am): Jeremy Avigad (Carnegie Mellon University), “Modularity of Mathematics” (slides)
formal semantics for natural languages (11am-12:15pm): Barbara Partee (University of Massachusetts at Amherst), “The Intertwining Influences of Logic, Philosophy, and Linguistics in the Development of Formal Semantics and Pragmatics” (slides)
modal logic (2:15-3:30pm): Johan van Benthem (University of Amsterdam, Stanford University, Tsinghua University), “Modal logic, on the cusp of philosophy and mathematics” (handout)
logic in computer science (3:45-5:00pm): Ronald Fagin (IBM Almaden Research Center), “Applying logic to practice in computer science” (slides)
May 05, 2017, 9:00 AM
McCone Hall, Room 141 from 9 AM to 2 PM; LeConte Hall, Room 3 from 2 PM to 5 PM
Denis Hirschfeldt, Ehud Hrushovski, Michael
Rathjen, John Steel
Logic at UC Berkeley
A two-day conference in mathematical logic and related areas organized by The Group in Logic and the Methodology of Science at UC Berkeley (logic.berkeley.edu). The conference is partly occasioned by the fact that the Group in Logic turns sixty this year.
In 1957, a group of faculty members, most of them from the departments of Mathematics and Philosophy, initiated a pioneering interdisciplinary graduate program leading to the degree of Ph.D. in Logic and the Methodology of Science. The Group has fostered interdisciplinary work in which logic has interacted with mathematics, philosophy, statistics, computer science, linguistics, physics and other disciplines.
While mathematical logic at UC Berkeley cannot be identified only with the Group in Logic, the Group has played a vital role in Berkeley’s worldwide prominence in mathematical logic and significantly contributed to making Berkeley a mecca since the fifties for people interested in mathematical logic and its applications. A full list of all those researchers in logic who taught at UC Berkeley, or studied at UC Berkeley, or visited Berkeley for shorter or longer periods would result in a who’s who of mathematical logic.
While marking an important moment for logic at UC Berkeley, the conference will be forward looking rather than merely celebratory. We have invited eight internationally prominent scholars to talk about the future of mathematical logic in their respective areas of specialization.
The first day of the conference will have four invited speakers in the so-called “foundational” areas: set theory, model theory, recursion theory, and proof theory. The second day will have four invited speakers in areas where mathematical logic plays a prominent role, namely philosophy of logic and mathematics, formal semantics for natural languages, modal logic, and logic in computer science. The line up is given below.
For a map showing the conference lecture rooms, see the red arrows here.
May 5 (141 McCone Hall 9AM-2PM; 3 LeConte Hall 2-5PM)
introductory remarks (9-9:30am): Paolo Mancosu (UC Berkeley), “A brief history of the Group in Logic and the Methodology of Science” (slides)
set theory (9:30-10:45am): John Steel (UC Berkeley), “Absolutely ordinal definable sets” (slides)
model theory (11am-12:15pm): Ehud Hrushovski (Hebrew University, Oxford University), “Reflections on model theory and foundations”
recursion theory (2:15-3:30pm): Denis Hirschfeldt (University of Chicago), “Computability Theory” (slides)
proof theory (3:45-5:00pm): Michael Rathjen (Leeds University), “On relating type theories to (intuitionistic) set theories” (slides)
May 6 (105 North Gate Hall)
philosophy of logic and mathematics (9:30-10:45am): Jeremy Avigad (Carnegie Mellon University), “Modularity of Mathematics” (slides)
formal semantics for natural languages (11am-12:15pm): Barbara Partee (University of Massachusetts at Amherst), “The Intertwining Influences of Logic, Philosophy, and Linguistics in the Development of Formal Semantics and Pragmatics” (slides)
modal logic (2:15-3:30pm): Johan van Benthem (University of Amsterdam, Stanford University, Tsinghua University), “Modal logic, on the cusp of philosophy and mathematics” (handout)
logic in computer science (3:45-5:00pm): Ronald Fagin (IBM Almaden Research Center), “Applying logic to practice in computer science” (slides)
Logic Colloquium
April 21, 2017, 4:10 PM
60 Evans Hall
Jana Marikova
Assistant Professor
of Mathematics, Western Illinois University
Convexly valued o-minimal fields
O-minimal structures are ordered structures in which the definable sets behave in a particularly tame fashion. For example, definable one-variable functions are piecewise continuous and differentiable. A prototypical example is the ordered real field in which the definable sets are precisely the semialgebraic sets.
Valuations yield a way of understanding an (expansion of) a field in terms of two, often simpler, associated structures, namely its residue field and value group.
We shall investigate o-minimal fields with valuations that are determined by convex subrings. In particular, we shall see that the class of o-minimal fields with onvex subrings that give rise to o-minimal residue fields have various desirable properties, including quantifier elimination in a suitable language.
Alfred Tarski Lectures
April 14, 2017, 4:10 PM
50 Birge Hall
Lou van den Dries
Professor,
Department of Mathematics, University of Illinois at
Urbana-Champaign
Model Theory of Transseries: Results and Open Problems
The book Asymptotic Differential Algebra and Model Theory of Transseries (arXiv:1509.02588) by Aschenbrenner, van der Hoeven, and myself will soon appear as the Annals of Mathematics Studies, number 195. It contains the main results of twenty years of investigating 𝕋. One of these results is easy to state: The theory of this differential field is completely axiomatized by the requirements of being a Liouville closed H-field with small derivation and the intermediate value property for differential polynomials. I will explain the meaning of these terms, further results on 𝕋 and some consequences for what is definable in 𝕋.
I will also discuss some of the many attractive open problems in this area.
Alfred Tarski Lectures
April 12, 2017, 4:10 PM
50 Birge Hall
Lou van den Dries
Professor,
Department of Mathematics, University of Illinois at
Urbana-Champaign
Orders of Infinity and Transseries
The “orders of Infinity” of the title are du Bois Reymond’s growth rates of functions, put on a firm basis by Hardy and Hausdorff. This led to the notion of a Hardy field (Bourbaki). Other ways of describing these orders of infinity are transseries (formal series containing powers of the variable x as well as exponential and logarithmic terms like ex and log x), and Conway’s surreal numbers. These topics are all closely related, with the differential field 𝕋 of transseries taking center stage.
Alfred Tarski Lectures
April 10, 2017, 4:10 PM
50 Birge Hall
Lou van den Dries
Professor,
Department of Mathematics, University of Illinois at
Urbana-Champaign
Model Theory as a Geography of Mathematics
I like to think of model theory as a geography of mathematics, especially of its “tame” side. Here tame roughly corresponds to geometric as opposed to combinatorial-arithmetic. In this connection I will discuss Tarski’s work on the real field, and the notion of o-minimality that it suggested.
A structure M carries its own mathematical territory with it, via interpretability: its own posets, groups, fields, and so on. Understanding this “world according to M” can be rewarding. Stability-like properties of M forbid certain combinatorial patterns, thus providing highly intrinsic and robust information about this world.
Logic Colloquium
April 07, 2017, 4:10 PM
60 Evans Hall
Pierre Simon
Assistant Professor of
Mathematics, UC Berkeley
On omega-categorical structures and their automorphism groups
The study of omega-categorical structures and their automorphism groups has been for several decades a very active area of research and point of contact between model theory, group theory, combinatorics and descriptive set theory. A countable structure is omega-categorical if its automorphism group acts oligomorphically on it, that is has finitely many orbits on n-tuples, for all n. In this talk, I will introduce this subject by presenting the various angles of approach and talk about a selection of topics, in particular concerning countable subgroups of automorphism groups. I will present a recent work with Itay Kaplan on the existence of finitely generated dense subgroups and ask a number of questions.
Logic Colloquium
March 24, 2017, 4:10 PM
60 Evans Hall
Françoise Point
FNRS-FRS
(UMons)
On expansions of (ℤ, +, 0)
In the special case of the structure (ℤ, +, 0), we will consider the following model-theoretic question: given a well-behaved first-order structure, which kind of predicates can one add and retain model-theoretic properties of the structure one started with, such as stability- like properties, quantifier elimination in a reasonable language, decidability. We will review recent results on stability properties of expansions of (ℤ, +, 0) by a unary predicate and we will make the comparison with former results on the decidability and model-completeness of the corresponding expansions of (ℤ, +, 0, <) (and if time permits, with the corresponding expansions of the field of real numbers).
Logic Colloquium
March 10, 2017, 4:10 PM
60 Evans Hall
Caroline Terry
Brin Postdoctoral
Fellow, University of Maryland, College Park, and Research Fellow,
Simons Institute for the Theory of Computing
Structure and enumeration theorems for hereditary properties of ℒ-structures
The study of structure and enumeration for hereditary graph properties has been a major area of research in extremal combinatorics. Over the years such results have been extended to many combinatorial structures other than graphs. This line of research has developed an informal strategy for how to prove these results in various settings. In this talk we use tools from model theory to formalize this strategy. In particular, we generalize certain definitions, tools, and theorems which appear commonly in approximate structure and enumeration theorems in extremal combinatorics. Our results apply to classes of finite ℒ-structures which are closed under isomorphism and model-theoretic substructure, where ℒ is any finite relational language.
Logic Colloquium
February 24, 2017, 4:10 PM
60 Evans Hall
Sean Walsh
Department of Logic and
Philosophy of Science, University of California, Irvine
Realizability Semantics for Quantified Modal Logic
In 1985, Flagg produced a model of first-order Peano arithmetic and a modal principle known as Epistemic Church’s Thesis, which roughly expresses that any number-theoretic function known to be total is recursive. In some recent work ([1]), this construction was generalized to allow a construction of models of quantified modal logic on top of just about any of the traditional realizability models of various intuitionistic systems, such as fragments of second-order arithmetic and set theory. In this talk, we survey this construction and indicate what is known about the reduct of these structures to the non-modal language.
References: [1] B. G. Rin and S. Walsh. Realizability semantics for quantified modal logic: Generalizing Flagg’s 1985 construction. The Review of Symbolic Logic, 9(4):752–809, 2016.
Logic Colloquium
February 10, 2017, 4:10 PM
60 Evans Hall
Clifton Ealy
Associate Professor of
Mathematics, Western Illinois University
Residue field domination in real closed valued fields
In an algebraically closed valued field, as shown by Haskell, Hrushovski, and Macpherson, the residue field and the value group control the rest of the structure: tp(L/Ck(L)Γ(L) will have a unique extension to M ⊇ C, as long as the residue field and value group of M are independent from those of L. (Here k(L) and Γ(L) denote the residue field and value group, respectively, of L, and C is a maximal field.)
This behaviour is striking, because it is what typically occurs in a stable structure (where types over algebraically closed sets have unique extensions to independent sets) but valued fields are far from stable, due to the order on the value group.
Real closed valued fields are even further from stable since the main sort is ordered. One might expect the analogous theorem about real closed valued fields to be that tp(L/M) is implied by tp(L/Ck(L)Γ(L)) together with the order type of L over M. In fact we show that the order type is unnecessary, that just as in the algebraically closed case, one has that tp(L/M) is implied by tp(L/k(L)Γ(L)). This is joint work with Haskell and Marikova.
No knowledge of value fields will be assumed.
Logic Colloquium
January 27, 2017, 4:10 PM
60 Evans Hall
Ilijas Farah
Professor of
Mathematics, York University
Model theory of operator algebras
Unexpected and deep connections between the theory of operator algebras and logic have been discovered in the past fifteen years. Even more remarkably, diverse areas of logic that are generally (and arguably) thought to have little connection to one another have found natural - and sometimes necessary - applications to operator algebras. In this talk I will emphasize applications of model theory to operator algebras.
Logic Colloquium
December 02, 2016, 4:10 PM
60 Evans Hall
Prakash Panangaden
Professor,
School of Computer Science, McGill University
A Logical Characterization of Probabilistic Bisimulation
Labelled Markov Processes (LMPs) are a combination of traditional labelled transition systems and Markov processes. Discrete versions of such systems have been around for a while and were thoroughly explored by Larsen and Skou in the late 1980s and early 1990s. Our contribution has been to extend this study to systems with continuous state spaces.
The main technical contribution that I will discuss in this talk is a definition of probabilistic bisimulation for such systems and a logical characterization for it. The surprise is that a very simple modal logic with no negative constructs or infinitary conjunctions suffices to characterize bisimulation, even with uncountable branching. This is quite different from the traditional situation with Kripke structures where van Benthem and Hennessy and Milner characterized bisimulation using negation in an essential way.
Logic Colloquium
November 18, 2016, 4:10 PM
60 Evans Hall
Phokion G. Kolaitis
UC Santa Cruz
& IBM Research - Almaden
Dependence Logic vs. Constraint Satisfaction
During the past decade, dependence logic has emerged as a formalism suitable for expressing and analyzing notions of dependence and independence that arise in different scientific areas. The sentences of dependence logic have the same expressive power as those of existential second-order logic, hence dependence logic captures the complexity class NP on the collection of all finite structures over a relational vocabulary. We identify a natural fragment of universal dependence logic and show that, in a precise sense, this fragment captures constraint satisfaction, a ubiquitous algorithmic problem encountered across different areas of computer science. This tight connection between dependence logic and constraint satisfaction contributes to the descriptive complexity of constraint satisfaction and elucidates the expressive power of universal dependence logic.
This is joint work with Lauri Hella, University of Tampere, Finland.
Working Group in the History and Philosophy of Logic, Mathematics, and Science
November 16, 2016, 6:00 PM
234 Moses Hall
Juliette Kennedy
Associate
Professor, Department of Mathematics and Statistics, University of
Helsinki
Squeezing arguments and strong logics
G. Kreisel has suggested that squeezing arguments, originally formulated for the informal concept of first order validity, should be extendable to second order logic, although he points out obvious obstacles. We develop this idea in the light of more recent advances and delineate the difficulties across the spectrum of extensions of first order logics by generalised quantifiers and infinitary logics. This is joint work with Jouko Väänänen.
Logic Colloquium
November 04, 2016, 4:10 PM
60 Evans Hall
Mai Gehrke
Senior Research
Director, CNRS & Université Paris Diderot - Paris 7
Canonical extensions and applications
In 1951-52 Jònsson and Tarski published a two part paper in which they cast Stone duality for Boolean algebras in algebraic form as so-called canonical extensions, and used this formulation to extend Stone duality to Boolean algebras with operators (BAOs). That is, Boolean algebras with additional operations which preserve join in each coordinate. In this way they obtained the relational semantics of modal logic. Since then the theory of canonical extensions, and thus duality for additional structure on lattices and Boolean algebras, has been generalised and better understood. These tools have played a central role in the discovery of the connection between formal language theory and Stone duality. In particular, the fact that the profinite completion of any abstract algebra is the Stone dual of the Boolean algebra of recognisable languages over the algebra equipped with certain residuation operations.
In this talk, I will give an introduction to these ideas and an overview of recent developments.
Logic Colloquium
October 21, 2016, 4:10 PM
60 Evans Hall
Erich Grädel
Professor of
Mathematical Foundations of Computer Science, RWTH Aachen University
Back and Forth Between Team Semantics, Games, and Tarski Semantics
Team semantics, due to Wilfrid Hodges, is the mathematical basis of modern logics of dependence and independence in which, following a proposal by Väänänen, dependencies are considered as atomic statements, and not as annotations of quantifiers. In team semantics, a formula is evaluated not against a single assignment of values to the free variables, but against a set of such assignments.
Logics with team semantics have strong expressive power and some surprising properties. They can be analyzed in several different ways. First-order formulae with team semantics and dependencies can be translated into sentences of existential second-order logic, with an additional predicate for the team (and with classical Tarski semantics). We can thus understand the power of a specific logic of dependence or independence by identifying the fragment of existential second-order logic to which it corresponds. Further, logics can be understood through their games. In general the model-checking games for logics with team semantics turn out to be second-order reachability games.
We shall then have a closer look at the specific case of logics with inclusion dependencies, and reveal their connection with safety games and logics with greatest fixed-points.
Logic Colloquium
October 07, 2016, 4:10 PM
60 Evans Hall
Dag Westerståhl
Professor of
Theoretical Philosophy and Logic, Stockholm University
A Carnapian approach to the meaning of logical constants: the case of modal logic
Does a consequence relation in a language L, as a syntactic relation between sets of L-sentences and L-sentences, fix the meaning of the logical constants in L? In his 1943 book The Formalization of Logic, Carnap worried that this seems to fail even for classical propositional logic CL. However, by applying the viewpoint of modern formal semantics, which in particular requires meaning assignment to be compositional, Carnap’s worries about CL can be allayed. More importantly, it provides a precise framework for asking Carnap’s question about any logic. (To what extent) does classical first-order consequence determine the meaning of ∀? What about other (generalized) quantifiers? What about the intuitionistic meaning of the connectives? Or □ in (classical) modal logic? [1] answers the first of these questions. In this talk I focus on the last one. Despite obvious similarities between ∀ and □, there are important differences. Roughly, while there is essentially just one first-order logic, there are innumerable modal logics. That makes answers to Carnap style questions about the meaning of □ more intricate, and more interesting, than in the case of ∀. Although we in a sense cover extremely familiar ground, the perspective one gets, in particular of neighborhood semantics for modal logic, seems novel. This is joint work with Denis Bonnay.
[1] D. Bonnay and D. Westerståhl, ‘Compositionality solves Carnap’s Problem’, Erkenntnis 81 (4), 2016 (721–739).
Logic Colloquium
September 23, 2016, 4:10 PM
60 Evans Hall
Andreas Blass
Professor of
Mathematics, University of Michigan, and Visiting Scientist, Simons
Institute for the Theory of Computing
Between countable and continuum
I plan to give a general introduction to cardinal characteristics of the continuum, including several examples of characteristics and of the sorts of results that the theory obtains. Afterward, if time permits, I’ll describe some recent work on a specific characteristic suggested by a familiar theorem from calculus. The talk will be aimed at people who are not set theorists.
Logic Colloquium
September 09, 2016, 4:10 PM
60 Evans Hall
Sam Buss
Professor of Mathematics
and Computer Science, UC San Diego
Frege Proofs, Extended Frege Proofs, and Total NP Search Problems
This talk will discuss some of the fundamental problems of proof complexity and their connection to computational complexity. The talk focuses on the proof complexity of the Frege and extended Frege proof systems for propositional logic. Recent work includes new proofs for the pigeonhole principle (!), Frankl’s theorem, the AB=I theorem, and the Kneser-Lovasz theorem on chromatic numbers. These problems are closely related to Total NP Search Problems (TFNP), a complexity class that lies midway between P and NP. The consistency statements for exponentially large Frege and extended Frege proofs can be used to characterize the provably total functions of second-order theories of bounded arithmetic for polynomial space and exponential time.
Logic Colloquium
August 26, 2016, 4:10 PM
60 Evans Hall
Martin Otto
Professor in
Mathematical Logic, Department of Mathematics, Technische Universitaet
Darmstadt (currently visiting the Simons Institute for Theory of
Computing, Berkeley)
Back & Forth Between Malleable Finite Models
This talk attempts to survey some model-theoretic techniques, model transformations and constructions - especially for logics of a modal/guarded flavour - that can replace classical compactness arguments, which are not available in restriction to finite models or other non-elementary classes of interest. Even where these algebraic-combinatorial and game-oriented techniques concur with the usually smoother infinitary methods of classical model theory they occasionally offer additional constructive or quantitative information. Some of the underlying combinatorial constructions to be discussed involve rather generic finite coverings for graphs and hypergraphs and are of independent interest. They can also be used towards alternative proofs of powerful extension properties for partial isomorphisms, which also have applications for the finite model theory of related logics.
Special Logic Seminar
May 10, 2016, 4:10 PM
234 Moses Hall
Johan van Benthem
Amsterdam and
Stanford
Decidable Versions of First-Order Predicate Logic
There are two ways of finding decidability inside first-order logic: one is by restricting attention to language fragments, the other is by generalizing the usual semantics. In this talk, I explain a recent generalized semantics by Aldo Antonelli, which leads to a decidable version of predicate logic that induces an effective translation into the Guarded Fragment. I will discuss this proposal and the resulting program.
Also, I make a comparison with existing decidable first-order semantics via general assignment models and via games, and with the move in modal logic from relational semantics to neighborhood semantics. Finally, I discuss the resulting landscape of weak decidable first-order logics, and its links with decidable fragments of the first-order language. This suggests many open problems that I will flag throughout.
Ref. H. Andréka, J. van Benthem & I. Németi, 2016, ‘On a New Semantics for First‐Order Predicate Logic’, Journal of Philosophical Logic, to appear.
Logic Colloquium
May 06, 2016, 4:10 PM
60 Evans Hall
Omer Ben Neria
Hedrick Assistant
Adjunct Professor of Mathematics, UCLA
The distance between HOD and V
The pursuit of better understanding of the universe of set theory V
motivated an extensive study of definable inner models M whose goal is
to serve as good approximations to V.
A common property of these inner models is that they are contained in
HOD, the universe of hereditarily ordinal definable sets. Motivated by
the question of how “close” HOD is to V, we consider various related
forcing methods and survey known and new results. This is a joint work
with Spencer Unger.
Logic Colloquium
April 22, 2016, 4:10 PM
60 Evans Hall
Kai Wehmeier
Professor of Logic and
Philosophy of Science, UC Irvine
The Role of Variables in Predicate Logic: Frege vs. Tarski
Standard presentations of predicate logic, following Tarski, segment a quantified statement such as ∃x(P(x) ∧ R(x, a)) into the constituents ∃x and P(x) ∧ R(x, a). At least on the face of it, Frege held a different view about the logical form of quantification; for him the same sentence decomposes into the constituents ∃x…x…x… and P( ) ∧ R( , a). In order to adjudicate between these views, I will develop some of the details of Fregean predicate logic, with special emphasis on the compositionality of its semantics. I will then compare Tarskian and Fregean treatments of variables and draw some philosophical conclusions from this comparison. In particular, I will address the extensionality of predicate logic and Kit Fine’s so-called antinomy of the variable.
Alfred Tarski Lectures
April 08, 2016, 4:10 PM
20 Barrows
William W. Tait
Professor Emeritus,
Department of Philosophy and CHSS, University of Chicago
Cut-Elimination for Subsystems of Classical Second-Order Number Theory: Cut-Elimination for Π11 − CA with the ω-Rule—and Beyond(?)
I will present a simplification of the proof of cut-eliminability of Gaisi Takeuti and Mariko Yasugi for this system. In particular, I will avoid the use of Takeuti’s ordinal diagrams. (I do use the finite part of the Veblen hierarchy, though.) Given time and sufficient conviction, I may speak about the possibility of extending the result to Π21 − CA and beyond.
Alfred Tarski Lectures
April 06, 2016, 4:10 PM
20 Barrows
William W. Tait
Professor Emeritus,
Department of Philosophy and CHSS, University of Chicago
Cut-Elimination for Subsystems of Classical Second-Order Number Theory: The Predicative Case
I will present the classical results of Gentzen and Schuette for first-order and ramified second-order number theory.
Alfred Tarski Lectures
April 04, 2016, 4:10 PM
20 Barrows
William W. Tait
Professor Emeritus,
Department of Philosophy and CHSS, University of Chicago
On Skepticism about the Ideal
There is a historic skepticism about mathematics that hangs on the fact that the objects of mathematics, structures and their elements—numbers, functions, sets, etc., are ideal, i.e. that empirical facts, facts about the natural world, have no relevance to the truth of propositions about them.
Of course, the view that ‘existence’ simply means empirical existence, so that the term is misused as applied to ideal things can be countered only on pragmatic grounds, that we use the term in other contexts and that it is very useful there. The rejection of ideal existence becomes meaningful only if one has a transcendental ground on which to stand and judge applications of the term. I believe that the ground, at least implicitly, has been a wrong thesis about how language works, namely the view that genuine reference to objects presupposes a non-linguistic interaction with them. And that is what I want to talk about. My argument draws on a reading of Wittgenstein’s Philosophical Investigations, a reading which is more positive than a more common one according to which he, himself, was a skeptic.
Logic Colloquium
March 18, 2016, 4:10 PM
60 Evans Hall
Guram Bezhanishvili
Professor of
Mathematical Sciences, New Mexico State University
The algebra of topology: Tarski’s program 70 years later
In a slogan, Tarski’s program can be described as “creating an algebraic apparatus adequate for the treatment of portions of point-set topology” (McKinsey-Tarski, 1944). The “algebraic apparatus” can be developed in several ways. For example, we can look at the algebra of open sets of a topological space X or the powerset algebra of X equipped with topological closure (or interior). Both of these approaches are not only closely related, but are also widely used in logic. The first approach yields a topological representation of Heyting algebras, and hence provides topological completeness of intuitionistic logic (one of the first completeness results for intuitionistic logic, established by Tarski in the 1930s). On the other hand, the second approach provides a topological representation of closure algebras, thus yielding topological completeness of Lewis’s modal system S4. McKinsey and Tarski observed a close connection between these two approaches, which allowed them to prove that Gödel’s translation of intuitionistic logic into S4 is full and faithful. I will review this line of research of Tarski and his collaborators, discuss further advances in this direction, obtained in the second half of the twentieth century, and finish with the latest new developments.
Logic Colloquium
March 04, 2016, 4:10 PM
60 Evans Hall
Katrin Tent
Professor of
Mathematics and Mathematical Logic, Universität Münster
Describing finite groups by short sentences
We say that a class of finite structures for a finite first-order signature is r-compressible for an unbounded function r : N → N+ if each structure G in the class has a first-order description of size at most O(r(|G|)). We show that the class of finite simple groups is log-compressible, and the class of all finite groups is log3-compressible. The result relies on the classification of finite simple groups, the bi-interpretability of the twisted Ree groups with finite difference fields, the existence of profinite presentations with few relators for finite groups, and group cohomology.
Logic Colloquium
February 19, 2016, 4:10 PM
60 Evans Hall
Christoph Benzmüller
Stanford
University and Freie Universität Berlin
A Success Story of Higher-Order Theorem Proving in Computational Metaphysics
I will report on the discovery and verification of the inconsistency in Gödel’s ontological argument with reasoning tools for higher-order logic. Despite the popularity of the argument since the appearance of Gödel’s manuscript in the early 70’s, the inconsistency of the axioms used in the argument remained unnoticed until 2013, when it was detected automatically by the higher-order theorem prover LEO-II.
Understanding and verifying the refutation generated by LEO-II turned out to be a time-consuming task. Its completion required the reconstruction of the refutation in the Isabelle/HOL proof assistant. This effort also led to a more efficient way of automating higher-order modal logic S5 with a universal accessibility relation within the logic embedding technique we utilize in our work.
The development of an improved syntactical hiding for the logic embedding technique allows the refutation to be presented in a human-friendly way, suitable for non-experts in the technicalities of higher-order theorem proving. This should bring us a step closer to wider adoption of logic-based artificial intelligence tools by philosophers.
If time permits, I will also point to alternative, ongoing applications of the logic embeddings technique, such as the encoding of Zalta’s theory of abstract objects or the mechanization of Scott’s free logic.
Logic Colloquium
February 05, 2016, 4:10 PM
60 Evans Hall
Vijay D’Silva
Google Research
Interpolant Construction and Applications
In the last decade, algorithms for the construction of interpolants from proofs have found numerous practical applications such as the analysis of circuits, generation of Floyd/Hoare proofs of program correctness, and parallelization of constraint solvers. The first part of this talk will trace how interpolation has spread from mathematical logic to commercial software via computational complexity theory and computer aided verification. The second part will delve into details of the algorithms for construction of interpolants in logics and theories. In the propositional case, the space of interpolant constructions has a semi-lattice structure that determines the logical strength and size of the interpolants obtained. In the case of first-order theories, this structure is still being unravelled. The talk will conclude with a review of open theoretical and empirical problems concerning interpolant construction.
Model Theory Seminar
January 29, 2016, 4:10 PM
60 Evans Hall
Pierre Simon
CNRS, Université
Claude Bernard - Lyon 1
Decomposition of dependent structures
A structure is a set equipped with a family of predicates and functions on it, for example a group or a field. A structure is said to be dependent if it is not as complicated as a random graph (in some precise sense). The fields of complex numbers, real numbers and p-adic numbers are examples of dependent structures. If a dependent structure contains no definable order, then it is stable: a much stronger property that is now very well understood. Stability can be thought of as an abstraction of algebraic geometry. At the other extreme, are dependent structures which are very much controlled by linear orders. We call such structures distal. Distality is meant to serve as a general model for semi-algebraic (or analytic) geometry. In this talk, I will explain those notions and state a recent theorem saying that any dependent structure can be locally decomposed into a stable-like and a distal-like part.
Logic Colloquium
January 22, 2016, 4:10 PM
60 Evans Hall
John R. Steel
Professor of
Mathematics, UC Berkeley
Ordinality Definability in Models of the Axiom of Determinacy
Let HOD be the class of all hereditarily ordinal definable sets, and let M be a model of the Axiom of Determinacy. The inner model HOD^M consisting of all sets that are hereditarily ordinal definable in the sense of M is of great interest, for a variety of reasons. In the first part of the talk, we shall give a general introduction that explains some of these reasons.
We shall then describe a general Comparison Lemma for “hod mice” (structures that approximate HOD^M). Modulo still-open conjectures on the existence of iteration strategies, this lemma yields models M of the Axiom of Determinacy such that HOD^M can be analyzed fine structurally (for example, satisfies the GCH), and yet satisfies very strong large cardinal hypotheses (for example, that there are superstrong cardinals).
Logic Colloquium
December 04, 2015, 4:10 PM
60 Evans Hall
Natarajan Shankar
Computer
Scientist, SRI International
PVS and the Pragmatics of Formal Proof
A formal system establishes a trinity between language, meaning, and inference. Many domains of thought can be captured using formal systems so that sound conclusions can be drawn through correct inferences. Philosophers have long speculated about machines that can distinguish sound arguments from flawed ones, and with modern computing, we can realize this dream to a significant extent. A research program to mechanize formal proof inevitably confronts a range of foundational and pragmatic choices. Should the foundations be classical or constructive, and should they be based on set theory, type theory, or category theory? What definitional principles should the language admit? How can formalizations be constructed from reusable modules? Should proofs be represented in a Hilbert calculus, natural deduction, or sequent calculus? Which inference steps in the proof calculus can and should be effectively mechanized in order to close the gap between informal arguments and their formal counterparts?
SRI’s Prototype Verification System (PVS) is an interactive proof assistant aimed at capturing the pragmatic features of mathematical expression and argumentation. The PVS language augments a simply typed higher-order logic with subtypes, dependent types, and algebraic datatypes. The PVS interactive proof checker integrates a number of decision procedures into the sequent calculus proof system. We examine the implications of these choices for the original goal of constructing mathematically sound arguments.
Logic Colloquium
November 20, 2015, 4:10 PM
60 Evans Hall
Edward Zalta
Senior Research
Scholar, Center for the Study of Language and Information, Stanford
University
Reflections on Foundations for Mathematical Structuralism
In a paper recently published in Mind (“Foundations for Mathematical Structuralism”), Nodelman and I offer a rigorous formulation of mathematical structuralism. A mathematics-free theory of structures and structural objects is presented in a background formalism and then the whole is applied to the analysis of mathematics. While the principal structuralist intuitions about structures and the elements of structures are preserved, distinctions formalized in the theory undermine a variety of arguments that have been put forward concerning the nature of structural elements. Arguments from Russell, Dedekind, Benacerraf, Shapiro, Linnebo, Awodey, Keranen, and others, are considered. For those interested in reading the paper in advance, a preprint is available at: http://mally.stanford.edu/Papers/structuralism.pdf
Logic Colloquium
November 06, 2015, 4:10 PM
60 Evans Hall
Silvain Rideau
Morrey Visiting
Assistant Professor, Department of Mathematics, UC Berkeley
Transferring imaginaries
Since it was first formulated by Poizat thirty years ago, the question of eliminating imaginaries, has become an important, if not necessary, step in understanding the model theory of given algebraic structures; it consists in classifying all the definable equivalence relations in a structure. An ever growing collection of such results is now available but this question remains open in many otherwise well understood structures.
Some of the recent elimination of imaginaries results are actually reductions to previously known elimination of imaginaries results. The goal of this talk will be to consider these reductions from an abstract standpoint and provide various settings in which one might derive elimination of imaginaires in a theory from elimination of imaginaries in some other theory.
Logic Colloquium
October 23, 2015, 4:10 PM
60 Evans Hall
Larry Moss
Professor of
Mathematics, Indiana University, Bloomington
Natural Logic
Much of modern logic originates in work on the foundations of mathematics. My talk reports on work in logic that has a different goal, the study of inference in language. This study leads to what I will call “natural logic”, the enterprise of studying logical inference in languages that look more like natural language than standard logical systems. An early paper in this field is George Lakoff’s 1970 paper “Linguistics and Natural Logic.”
By now there is a growing body of work which presents logical systems that differ from first-order logic in various ways. Most of the systems are complete and decidable. Some are modern versions of syllogistic logic, but with additional features not present in syllogistic logics. And then there are flavors of logic which look rather far from either traditional or modern logic.
The talk will be programmatic and far-ranging rather than detailed. I hope to touch on computer implementations of natural logics, teaching materials on this topic, and interactions of logic and cognitive science. And keeping with most other talks in the Logic Colloquium, I’ll show that the whole subject of natural logic is full of questions with mathematical interest.
Logic Colloquium
October 09, 2015, 4:10 PM
60 Evans Hall
Takayuki Kihara
JSPS Postdoctoral
Fellow, Department of Mathematics, UC Berkeley
Recursion Theoretic Methods in Descriptive Set Theory and Infinite Dimensional Topology
The notion of degree spectrum of a structure in computable model theory is defined as the collection of all Turing degrees of presentations of the structure. We introduce the degree spectrum of a represented space as the class of collections of all Turing degrees of presentations of points in the space. The notion of point degree spectrum creates a connection among various areas of mathematics including computability theory, descriptive set theory, infinite dimensional topology and Banach space theory.
Through this new connection, for instance, we construct a family of continuum many infinite dimensional Cantor manifolds with property C in the sense of H aver/Addis-Gresham whose Borel structures at an arbitrary finite rank are mutually non-isomorphic. This provides new examples of Banach algebras of real valued Baire class two functions on metrizable compacta, and strengthen various theorems in infinite dimensional topology such as Pol’s solution to Alexandrov’s old problem.
To prove our main theorem, an invariant which we call degree co-spectrum, a collection of Turing ideals realized as lower Turing cones of points of a Polish space, plays a key role. The key idea is measuring the quantity of all possible Scott ideals (omega-models of WKL) realized within the degree co-spectrum (on a cone) of a given space. This is joint work with Arno Pauly (University of Cambridge, UK).
Logic Colloquium
September 25, 2015, 4:10 PM
60 Evans Hall
Dana Scott
University Professor,
Emeritus, Carnegie Mellon University; Visiting Scholar, UC Berkeley
Can Modalities Save Naive Set Theory?
The late “Grisha” Mints once asked the speaker whether a naive set theory could be consistent in modal logic. Specifically he asked whether restricting the comprehension scheme to necessary properties was safe. Scott was working on a set theory in the Lewis system S4 of modal logic and Mints was happy to position his question in the same modal system. Obviously a very, very weak modality can avoid paradoxes, but such results are not especially interesting. At that time (2009) Scott could not answer the consistency question, and neither could Mints. Last November Scott noted that CMU Philosophy was hosting a seminar on a naive set theory by Harvey Lederman. Scott wrote him for his paper and said, “By the way, there is this question of Grisha Mints, and I wonder if you have an opinion?” Lederman sent back a sketch of a proof of the inconsistency of a strengthened version of comprehension. That proof at first did not quite work out, but was repaired in correspondence. Lederman mentioned the questions to two of his colleagues, and in March of 2015 Tiankai Liu suggested a possible model of a weaker comprehension scheme, which after a small correction gave a consistency proof. A few days later, Peter Fritz came up with essentially the same model. A paper has now been submitted for publication jointly by Fritz, Lederman, Liu, and Scott.
Logic Colloquium
September 11, 2015, 4:10 PM
60 Evans Hall
Thomas Icard
Assistant Professor of
Philosophy and Symbolic Systems, Stanford University
A Survey on Topological Semantics for Provability Logics
This will be a survey of ideas and results on topological interpretations of provability logics, especially polymodal provability logics. Esakia first proved completeness for the basic Gödel-Löb logic of provability with respect to scattered spaces. Abashidze and Blass (independently) proved completeness w.r.t. a particular scattered space defined on the ordinal omega^omega. I will discuss older work of my own that extended the Abashidze-Blass result to a polymodal provability logic, which is complete with respect to a polytopological space on epsilon_0. I will then discuss more recent developments, including work by Beklemishev, Fernandez, Joosten, Gabelaia, Aguilera, and others, generalizing and improving upon much of this.
Logic Colloquium
August 28, 2015, 4:10 PM
60 Evans Hall
Antonio Montalban
Associate
Professor of Mathematics, UC Berkeley
Analytic Equivalence Relations with ℵ1-many Classes
We will survey the new results connecting Vaught’s conjecture and computability theory. We will look at some of these result from the more general viewpoint of analytic equivalence relations.
There are two types of results we will concentrate on. On the one hand are the ones about equivalence relations satisfying hyperarithmetic-is-recursive, which provide a purely computability theoretic equivalent to Vaught’s conjecture. On the other hand are the ones about equivalence relations on the natural numbers that are “intermediate,” which I hope will eventually provide another equivalent statement.
Logic Colloquium
May 01, 2015, 4:10 PM
60 Evans Hall
Stuart Russell
Professor of
Computer Science and Smith-Zadeh Professor in Engineering, University of
California, Berkeley
Unifying logic and probability: The BLOG language
Logic and probability are ancient subjects whose unification holds significant potential for the field of artificial intelligence. The BLOG (Bayesian LOGic) language provides a way to write probability models using syntactic and semantic devices from first-order logic. In modern parlance, it is a relational, open-universe probabilistic programming language that allows one to define probability distributions over the entire space of first-order model structures that can be constructed given the constant, function, and predicate symbols of the program. I will describe the language mainly through examples and cover its application to monitoring the Comprehensive Nuclear-Test-Ban Treaty.
Logic Colloquium
April 17, 2015, 4:10 PM
60 Evans Hall
Tadeusz Litak
FAU
Erlangen-Nürnberg
Whence Cometh Incompleteness?
I will explain the background of my recent collaboration with Wes Holliday. It is well known among modal logicians that while “naturally” obtained modal logics tend to be Kripke complete, things look very differently when one looks at the class of all modal logics. In the 1970’s, S.K. Thomason found a fundamental reason why this phenomenon is unavoidable: modal consequence over Kripke frames is every bit as complex as full second-order consequence. Shortly afterward, Wim Blok proved another amazing result: even those logics that are complete tend to be surrounded by uncountably many incomplete ones, which are sound wrt precisely the same class of Kripke structures. Kripke incompleteness turned out to be much more of a norm than anybody had expected; and in a very precise sense, the problem is irreparable, at least as long as the notion of proof remains decidable.
A question then arises if some sensible generalization of Kripke semantics would make the phenomenon of incompleteness disappear. In a way, the answer was known even before Kripke frames dominated the picture: every modal logic is complete wrt Boolean Algebras with Operators (BAOs). As it turns out, this is the best general completeness result we can hope for. It is possible to find naturally motivated semantics properly contained between Kripke semantics and algebraic semantics which do not suffer from the “Thomason effect”: the induced notion of consequence can be axiomatized. However, as soon as these notions nontrivially depart from the full generality of algebraic semantics and try to capture at least some of the defining properties of Kripke frames, not only can one still find examples of incomplete logics, but the limitative result of Blok survives.
These questions were investigated in my PhD dissertation a decade ago, building on several decades of work in the modal community. By the time the dissertation was completed, there was still one possible escape route left: one of the three defining properties of (dual algebras of) Kripke frames could perhaps allow a general completeness result. More recently, Wes Holliday’s work on “possibility semantics” for modal logic rekindled the interest in that particular notion of completeness. Now we have found that even this escape route is blocked. Furthermore, the story turns out to involve an unjustly forgotten example, paper and program of Johan van Benthem.
Alfred Tarski Lectures
April 03, 2015, 4:10 PM
2 LeConte Hall
Julia F. Knight
Charles L. Huisking
Professor of Mathematics, University of Notre Dame
Computability and complexity of uncountable structures
Alfred Tarski Lectures
April 01, 2015, 4:10 PM
2 LeConte Hall
Julia F. Knight
Charles L. Huisking
Professor of Mathematics, University of Notre Dame
Comparing classes of countable structures
Alfred Tarski Lectures
March 30, 2015, 4:10 PM
2 LeConte Hall
Julia F. Knight
Charles L. Huisking
Professor of Mathematics, University of Notre Dame
Computability and complexity of mathematical structures
Logic Colloquium
March 20, 2015, 4:10 PM
60 Evans Hall
Keita Yokoyama
Assistant Professor,
School of Information Science, Japan Advanced Institute of Science and
Technology
Reverse mathematics and program termination
It is well-known that proof theory and computer science have a good connection. In this talk, I will focus on the relation between reverse mathematics and the study of program termination. In the study of reverse mathematics, Ramsey’s theorem is one of the most important topics and the strength of Ramsey’s theorem is well-investigated from both of the proof-theoretic view point and the computability theoretic view point. On the other hand, Ramsey’s theorem is often used in computer science. In the study of termination analysis, Podelski/Rybalchenko used Ramsey’s theorem for pairs to introduce a new algorithm for termination checking. Looking at the above connection with Ramsey’s theorem, we study the relation between reverse mathematics and termination analysis.
This is a joint work with Stefano Berardi and Silvia Steila.
Logic Colloquium
March 06, 2015, 4:10 PM
60 Evans Hall
Valeria de Paiva
AI and NLU
Sunnyvale Lab, Nuance Communications
Constructive Modal Logics: 15 years later
This talk will review the program of intuitionistic modal logic in its two original sources: modal logicians interested in Intuitionism and functional programmers/type theorists interested in (computational) modalities. I will discuss some trends of this kind of work since the first Intuitionistic Modal Logic and Applications (IMLA) in Trento in 1999 and the tendencies that we can see in this line of work today. http://www.cs.bham.ac.uk/~vdp/
Logic Colloquium
February 20, 2015, 4:10 PM
60 Evans Hall
Andrew Marks
NSF Postdoctoral
Fellow, Department of Mathematics, California Institute of
Technology
Baire measurable paradoxical decompositions and a Baire category solution to the dynamical von Neumann-Day problem
The Banach-Tarski paradox states that the unit ball in ℝ3 is equidecomposable with two unit balls in ℝ3 by rigid motions. In 1930, Marczewski asked whether there is such an equidecomposition where each piece has the Baire property. Using an intricate construction, Dougherty and Foreman gave a positive answer to this question.
We generalize Dougherty and Foreman’s result to completely characterize which Borel group actions have Baire measurable paradoxical decompositions. We show that if a group acting by Borel automorphisms on a Polish space has a paradoxical decomposition, then it admits a paradoxical decomposition using pieces having the Baire property. We also obtain a Baire category solution to the dynamical von Neumann-Day problem, in the spirit of Whyte and Gaboriau-Lyons: if a is a nonamenable action of a group on a Polish space X by Borel automorphisms, then there is a free Baire measurable action of 𝔽2 on X which is Lipschitz with respect to a. The main tool we use to prove these theorems is a version of Hall’s matching theorem for Borel graphs.
This is joint work with Spencer Unger at UCLA.
Logic Colloquium
February 06, 2015, 4:10 PM
60 Evans Hall
Pierre Simon
CNRS, Université
Claude Bernard - Lyon 1
Around finite VC-dimension and the (p,k)-theorem
Let N points be chosen at random in some finite domain of the plane. With high probability, for every circle, the proportion of points inside the circle is very close to the area of the circle. (Hence for example a computer can approximately reconstruct a disc given random points inside and outside of it.) This works because the family of discs satisfies a combinatorial property known as finite VC-dimension. This notion originates in machine learning theory: the property described implies that a class of finite VC-dimension is “learnable”. In the first part of my talk, I will present this and state a measure-free analog of this learnability property called the (p,k)-theorem (due to Alon-Kleitman and Matousek).
The definition of finite VC-dimension also appeared independently in model theory under the name NIP (negation of the Independence Property). A formula f(x,y) is NIP if the family of definable sets it defines has finite VC-dimension. In the second part of the talk, I will discuss some interesting consequences of the (p,k)-theorem in model theory and mention open problems.
Logic Colloquium
January 23, 2015, 4:10 PM
60 Evans Hall
Samson Abramsky
Christopher
Strachey Professor of Computing, Oxford University
Contextual Semantics: From Quantum Mechanics to Logic, Databases, Constraints and Natural Language Semantics
Quantum Mechanics presents a radically different perspective on physical reality compared with the world of classical physics. In particular, results such as the Bell and Kochen-Specker theorems highlight the essentially non-local and contextual nature of quantum mechanics. The rapidly developing field of quantum information seeks to exploit these non-classical features of quantum physics to transcend classical bounds on information processing tasks.
In this talk, we shall explore the rich mathematical structures underlying these results. The study of non-locality and contextuality can be expressed in a unified and generalised form in the language of sheaves or bundles, in terms of obstructions to global sections. There are also strong connections with logic. For example, Bell inequalities, one of the major tools of quantum information and foundations, arise systematically from logical consistency conditions.
These general mathematical characterisations of non-locality and contextuality also allow precise connections to be made with a number of seemingly unrelated topics, in classical computation, logic, and natural language semantics. By varying the semiring in which distributions are valued, the same structures and results can be recognised in databases and constraint satisfaction as in probability models arising from quantum mechanics. A rich field of contextual semantics, applicable to many of the situations where the pervasive phenomenon of contextuality arises, promises to emerge.
Logic Colloquium
December 12, 2014, 4:10 PM
60 Evans Hall
Christos Papadimitriou
C. Lester
Hogan Professor of Electrical Engineering and Computer Science, UC
Berkeley
Exponential existence proofs and complexity
Many computational problems in NP have the property that a solution is guaranteed to exist for every instance, and yet these problems do not seem to belong in P. Examples are factoring, Nash equilibrium, finding a short vector in a lattice, or a prime in [n, 2n], among many others. Such problems can be categorized by the corresponding existence proof. Rather surprisingly, it turns out that, in all known cases, this existence proof contains a simple graph-theoretic argument (such as “every graph has an even number of odd nodes,” or “every directed acyclic graph has a sink” or the pigeonhole principle) albeit applied to an exponentially large graph. Several natural problems are known to be complete for the corresponding classes. I shall introduce and discuss this intriguing aspect of complexity theory, including certain recent developments.
Logic Colloquium
November 21, 2014, 4:10 PM
60 Evans Hall
Juliet Floyd
Professor of
Philosophy, Boston University
The Sheffer Box
In May 2012, with the help of Bernard Linsky, I discovered a missing box of original documents and research materials pertaining to the work of H.M. Sheffer (1882-1964), the logician known for his strokes. The box had been assembled and saved by Burton Dreben, with whom I worked on Sheffer in 1988. This talk will report on my assembly and analysis of these primary resource materials.
Sheffer is of interest for our understanding of the reception of
logic in the United States: the work of Frege, Russell, Zermelo, the
early Wittgenstein, and others. A student of Royce, William James,
Huntington, and Russell, he made some contributions to the algebra of
logic tradition, though his own “general theory of notational
relativity” never reached mathematical fruition. He influenced many,
including C.H. Langford, and, slightly and indirectly, Zermelo.
In addition, however, the philosophical strands of his thinking form an
interesting comparison and contrast to those of his contemporaries
Wittgenstein and C.I. Lewis, on whom he lectured at Harvard, and those
of Emil Post, with whom Sheffer corresponded about his “general theory
of notational relativity”.
Sheffer took down the only extant notes of Bertrand Russell’s 1910 lectures at Cambridge University, and these, now being edited by Bernard Linsky, are among the materials I shall discuss. Other materials in the box included correspondence with Russell in which Sheffer describes meeting with Frege, Peano, Hadamard and others in 1911; notes of Sheffer’s spring 1922 logic lectures at Harvard, taken down by Marvin Farber; Farber’s correspondence with Sheffer concerning his work with Zermelo in 1923, as well as other lecture notes of relevance to our understanding of the history and presentation of logic in the United States the 1910s and 1920s.
Logic Colloquium
November 07, 2014, 4:10 PM
60 Evans Hall
Balder ten Cate
Computer Scientist
at LogicBlox and Associate Adjunct Professor of Computer Science, UC
Santa Cruz
Craig interpolation theorems and database applications
The Craig interpolation theorem, proved by William Craig in the 1950s, states that every valid first-order implication A |= B has an interpolant C such that A |= C and C |= B and all non-logical symbols occurring in C occur both in A and in B. Moreover, such an interpolant C can be effectively constructed from a proof of the implication A |= B.
The Craig interpolation theorem is considered one of the cornerstones in our understanding of first-order model theory and proof theory.
Over the last few years, we have seen the emergence of several new applications of Craig interpolation in the area of data management. In particular, interpolation-based techniques have been proposed for querying under access restrictions and semantic query optimization.
These applications, in turn, have led to the study of new variants of Craig interpolation. I will present two such new theorems, namely an interpolation theorem for first-order formulas with relational access restrictions, and an effective interpolation theorem for the guarded-negation fragment of first-order logic.
Logic Colloquium
October 24, 2014, 4:10 PM
60 Evans Hall
Thomas Scanlon
Professor of
Mathematics, University of California, Berkeley
Differential algebraic equations from definability
Many valued “functions” play an important role in complex analysis. In many cases of interest, these may converted to honest functions through the application of a differential operator. For example, the complex logarithm is only well defined up to the addition of an integral multiple of twice pi times the square root of negative one, but the logarithmic derivative which takes a function f(z) to the derivative of log(f(z)) is a well-defined function from (nowhere zero, differentiable, complex valued) functions to functions. Curiously, the logarithmic derivative is actually a differential rational function. That is, it is given by an expression which is a rational function of its argument and some of the derivatives of the argument. Other functions obtained by applying differential operators to eliminate the ambiguity of multivalued functions share this feature. I will describe a general context in which multivalued functions arising as inverses to certain analytic covering maps may be converted to well-defined differential algebraic functions by composing them with a well chosen differential operator.
While the theorem itself has some nice consequences in the theories of functional transcendence and differential algebra, in this lecture I will focus on two ideas from logic underlying the proof. First, the notion of elimination of imaginaries in the theories of algebraically closed and differentially closed fields will supply the requisite differential operator. Secondly, a beautiful theorem of Peterzil and Starchenko on definable complex analytic sets will be used to show that the classical analytic construction of such functions is necessarily algebraic.
Logic Colloquium
October 17, 2014, 4:10 PM
60 Evans Hall
Dominic Hughes
Visiting Scholar,
Department of Mathematics, Stanford University
First-order proofs without syntax
Proofs of first-order logic are traditionally syntactic, built inductively from symbolic inference rules. This talk reformulates classical first-order logic (predicate calculus) with proofs which are combinatorial rather than syntactic. A combinatorial proof is defined by graph-theoretic conditions which can be verified easily (linear time complexity).
To be accessible to a broad audience (mathematicians, computer scientists, logicians and philosophers—including undergraduates) the lecture uses many colorful pictorial examples. Niche technicalities (such as relationships with Gentzen’s sharpened Hauptsatz for LK and Herbrand’s theorem) are deferred to the end.
The work extends ‘Proofs Without Syntax’ [Annals of Mathematics, 2006], which treated the propositional case.
Logic Colloquium
October 03, 2014, 4:10 PM
60 Evans Hall
Sanjit Seshia
Associate Professor
of Electrical Engineering and Computer Science, University of
California, Berkeley
The Logic of Cars: Reasoning about Cyber-Physical Systems with Computational Logic
Cyber-physical systems (CPS) are those that tightly integrate computational processes with the physical world. Examples include medical devices and systems, automotive systems, intelligent power systems, and robots. In this talk, we explore how computational logic can be an effective way to model, design, and verify cyber-physical systems. Using today’s automobiles as a motivating application, we will explore the use of logic, especially temporal logic, in reasoning about such systems. The talk will survey some core theoretical problems along with current approaches to solve them, and sketch directions for future work.
Logic Colloquium
September 05, 2014, 4:10 PM
60 Evans Hall
Paolo Mancosu
Professor of
Philosophy, University of California, Berkeley
In good company? On Hume’s principle and the assignment of numbers to infinite concepts
In a recent article, I have explored the historical, mathematical, and philosophical issues related to the new theory of numerosities. The theory of numerosities provides a context in which to assign numerosities to infinite sets of natural numbers in such a way as to preserve the part-whole principle, namely if a set A is properly included in B then the numerosity of A is strictly less than the numerosity of B. Numerosities assignments differ from the standard assignment of size provided by Cantor’s cardinality assignments. In this talk I generalize some specific worries, raised by Richard Heck, emerging from the theory of numerosities to a line of thought resulting in what I call a ‘good company’ objection to Hume’s principle (HP). The talk has four main parts. The first takes a historical look at nineteenth-century attributions of equality of numbers in terms of one-one correlation and argues that there was no agreement as to how to extend such determinations to infinite sets of objects. This leads to the second part where I show that there are countably infinite many abstraction principles that are ‘good’, in the sense that they share the same virtues of HP and from which we can derive the axioms of second-order arithmetic. All the principles I present agree with HP in the assignment of numbers to finite concepts but diverge from it in the assignment of numbers to infinite concepts. The third part connects this material to a debate on Finite Hume Principle between Heck and MacBride and states the ‘good company’ objection as a generalization of Heck’s objection to the analyticity of HP based on the theory of numerosities. I then give a taxonomy of possible neo-logicist responses to the ‘good company’ objection.
Logic Colloquium
April 25, 2014, 4:10 PM
60 Evans Hall
Lou van den Dries
Professor of
Mathematics, University of Illinois (Urbana-Champaign); Research
Professor, MSRI (Spring 2014)
The Differential Field of Transseries
Transseries originated some 25 years ago, mainly in Ecalle’s work, but also independently in the study of Tarski’s problem on the field of reals with exponentiation. Some 20 years ago I formulated some conjectures about the differential field of transseries, and started exploring its model-theoretic properties, first jointly with Matthias Aschenbrenner, and a few years later also in collaboration with Joris van der Hoeven. Around 2011 we finally saw a clear path towards a proof, sharpening the conjectures in the meantime. In the last few weeks we finished the job, with one proviso: as part of a division of labor and for lack of time, some final details have only been checked by some of us, but not yet by all three of us. In any case, I will motivate these conjectures, and give an account of their status.
Logic Colloquium
April 12, 2014, 4:10 PM
60 Evans Hall
Zoe Chatzidakis
Senior Researcher,
CNRS, Paris; Research Professor, MSRI, Berkeley (Spring 2014)
Model Theory of Difference Fields and Applications
I will present classical results on the model theory of difference fields, and some recent applications by Hrushovski and myself.
Alfred Tarski Lectures
April 04, 2014, 4:10 PM
3 LeConte Hall
Stevo Todorcevic
University of
Toronto and CNRS, Paris
Combinatorial and Set-theoretic Forcing
Alfred Tarski Lectures
April 02, 2014, 4:10 PM
3 LeConte Hall
Stevo Todorcevic
University of
Toronto and CNRS, Paris
Chain-conditions of Horn-Tarski
Alfred Tarski Lectures
March 31, 2014, 4:10 PM
3 LeConte Hall
Stevo Todorcevic
University of
Toronto and CNRS, Paris
The Measurability Problem for Boolean Algebras
Logic Colloquium
March 14, 2014, 4:10 PM
60 Evans Hall
G. Aldo Antonelli
Professor of
Philosophy, University of California, Davis
General First-Order Models: Concepts and Results
In his 1950 dissertation, Leon Henkin showed how to provide higher- order quantifiers with non-standard, or “general” interpretations, on which, for instance, second-order quantifiers are taken to range over collections of subsets of the domain that may fall short of the full power-set. In contrast, first-order quantifiers are usually regarded as immune to this sort of non- standard interpretations, since their semantics is ordinarily taken to be completely determined once a first-order domain of objects is selected.
The asymmetry is particularly evident from the point of view of the modern theory of generalized quantifiers, according to which a first-order quantifier is construed as a predicate of subsets of the domain. But the generalized conception still views first-order quantifiers as predicates over the full power-set. Accordingly, the possibility that they, similarly to their second-order counterparts, might denote arbitrary collections of subsets has not been pursued in full generality.
This talk introduces a Henkin-style semantics for arbitrary first-order quantifiers, exploring some of the resulting properties, and emphasizing the effects of imposing various further closure conditions on the second-order component of the interpretation. Among other results, we show by a model- theoretic argument that in certain cases the notion of validity relative to models satisfying the closure conditions is axiomatizable.
Logic Colloquium
February 28, 2014, 4:10 PM
60 Evans Hall
Alexander Melnikov
Lecturer, Massey
University, Auckland, New Zealand, and Postdoctoral Scholar, University
of California, Berkeley
Infinitely Generated Abelian Groups with Solvable Word Problem
Logic Colloquium
February 14, 2014, 4:10 PM
60 Evans Hall
Sherrilyn Roush
Professor of
Philosophy, University of California, Berkeley
The Difference Between Knowledge and Understanding
In the classic Gettier problem, cases are imagined where a person has very good reasons to believe a proposition p, and p is true, but many people think the person still doesn’t have knowledge. I characterize what is missing in these cases probabilistically, as a failure of what I call relevance matching. Relevance matching is best interpreted as understanding why p is true, and it is deterministically related to but distinct from the tracking conditions that in my view define knowledge. This view of understanding makes it a simulation among your mental states of the dispositions of factors probabilistically relevant to p’s being true, and as such avoids requiring unrealistic information- and computation-intensive beliefs and inferences about why p is true in order to understand.
Logic Colloquium
January 31, 2014, 4:10 PM
60 Evans Hall
Jan Reimann
Assistant Professor of
Mathematics, The Pennsylvania State University
Effective Multifractal Spectra
Multifractal measures play an important role in the study of point processes and strange attractors. A central component of the theory is the multifractal formalism, which connects local properties of a measure (pointwise dimensions) with its global properties (average scaling behavior).
In this talk I will introduce a new, effective multifractal spectrum, where we replace pointwise dimension by asymptotic compression ratio. It turns out that the underlying measure can be seen as a universal object for the multifractal analysis of computable measures. The multifractal spectrum of a computable measure can be expressed as a “deficiency of multifractality” spectrum with respect to the universal measure. This in turn allows for developing a quantitative theory of dimension estimators based on Kolmogorov complexity. I will discuss some applications to seismological dynamics.
Logic Colloquium
December 06, 2013, 4:10 PM
60 Evans Hall
Alexander S. Kechris
Professor of
Mathematics, California Institute of Technology
Topological Dynamics and Ergodic Theory of Automorphism Groups of Countable Structures
I will discuss some aspects of the topological dynamics and ergodic theory of automorphism groups of countable first-order structures and their connections with logic, finite combinatorics, and probability theory. This is joint work with Omer Angel and Russell Lyons.
Logic Colloquium
November 15, 2013, 4:10 PM
60 Evans Hall
Dana S. Scott
University Professor,
Emeritus, Carnegie Mellon University; Visiting Scholar, University of
California, Berkeley
Stochastic Lambda-Calculi
Many authors have suggested ways of adding random elements an probability assessments to versions of Church’s Lambda-Calculus. When asked recently about models, the speaker realized that the so-called Graph Model (based on using enumeration operators acting on the power set of the integers) could easily be expanded to allow random variables taking values in the model. Other models can be treated similarly. The talk will discuss randomizing algorithms and how a continuation-passing semantics can be used for modeling a branching combinator using random coin tossing. These ideas can also be employed for introducing many other kinds of random combinators.
Logic Colloquium
November 01, 2013, 4:10 PM
60 Evans Hall
Lara Buchak
Assistant Professor of
Philosophy, University of California, Berkeley
Risk and Inequality
Decision theory concerns the evaluation of gambles. When choosing among gambles, individuals are forced to consider how they will turn out under various circumstances, and decide how to trade off the possibility that a gamble will turn out well against the possibility that it will turn out poorly. How should we aggregate the values one might get in different possible circumstances, in order to arrive at a single value for a gamble? The orthodox view is that there is only one acceptable way to do this: rational individuals must maximize expected (i.e. average) utility. The contention of my recent book, however, is that the orthodox theory (expected utility theory) dictates an overly narrow way in which considerations about risk can play a role in an individual’s choices. There, I argued for an alternative, more permissive, theory of decision-making: risk-weighted expected utility theory (REU theory). This theory allows individuals to pay proportionally more attention to the worst-case scenario than the best-case scenario.
Social choice theory concerns the evaluation of social distributions: distributions of goods or outcomes to individuals. To determine which social distribution is better, we must consider how distributions go for various individuals, and decide how to trade off the fact that one distribution is better for some people against the fact that is it worse for others. How should we aggregate the values that go to each person, in order to arrive as a single value for a social distribution and to say which of two social distributions is better? A traditional answer to this question is that the value of a social distribution is the average of the utility values that each individual in the society receives (utilitarianism). And one traditional justification of utilitarianism relies on the assumption that expected utility theory is the correct decision theory.
There are two ways in which decision theory has been used to justify a particular aggregation method in social choice theory. The first is to propose that facts about the values of social distributions are determined or discerned from individuals’ preferences about gambles. This is the method employed by John Rawls and John Harsanyi, for example: both consider individual preferences about gambles over social distributions in which one does not know which place one will occupy in society, and use these to determine the aggregative social welfare function. The second way to use decision theory to justify a particular aggregation method in social choice theory is to examine the conditions on the preference relation in decision theory, and explore whether analogous conditions might hold of the betterness relation. This is the rough strategy behind John Broome’s justification of utilitarianism in his book Weighing Goods.
Existing philosophical versions of these justifications are based on views about decision-making that I’ve argued are incorrect: thus, I claim, their conclusions for social choice theory are unjustified.
In this work-in-progress talk, I explore the prospects for using REU theory to justify an alternative evaluation of social distributions, one that allows us to pay special attention (but not exclusive attention) to the worst-off person as opposed to the best-off person.
Logic Colloquium
October 18, 2013, 4:10 PM
60 Evans Hall
James Freitag
National Science
Foundation Postdoctoral Fellow, University of California, Berkeley
Differential Algebra and the Model Theory of Groups
Sacks called differential fields the “least misleading” examples of omega-stable theories. This talk will reinforce that philosophy, because the theorems we will discuss fall into two general situations: (1) theorems from model theory which inspired results in differential algebra or (2) theorems from differential algebra which were generalized to the model-theoretic setting. Specifically, we will be discussing theorems regarding differential algebraic groups and groups definable in superstable theories (a generalization of omega-stability). We will discuss three types of theorems: (1) Jordan-Hölder theorems, (2) indecomposability theorems, and (3) results on central extensions. To open the talk, we will give a survey of the model theory of differential fields.
Logic Colloquium
October 04, 2013, 4:10 PM
60 Evans Hall
Richard Tieszen
Professor of
Philosophy, San Jose State University
Monads and Mathematics: Gödel, Leibniz, and Husserl
On the basis of his discussions with Kurt Gödel, Hao Wang (A Logical Journey: From Gödel to Philosophy, p. 166) tells us that “Gödel’s own main aim in philosophy was to develop metaphysics – specifically, something like the monadology of Leibniz transformed into exact theory – with the help of phenomenology”. Gödel began to study Edmund Husserl’s phenomenology in 1959. In 1928 Husserl (“Phenomenology”, Encyclopedia Britannica draft) wrote that “The ideal of the future is essentially that of phenomenologically based (”philosophical”) sciences, in unitary relation to an absolute theory of monads”. In the Cartesian Meditations and other works Husserl identifies ‘monads’ (in his sense) with ‘transcendental egos in their full concreteness’. In the first part of my talk I explore some prospects for a Gödelian monadology that result from this identification, with reference to texts of Gödel, Wang’s reports, and aspects of Leibniz’s original monadology. The latter part of the talk will be on human monads, the incompleteness theorems, and (Turing) machines. (For background see my recent book, After Gödel: Platonism and Rationalism in Mathematics and Logic, Oxford University Press.)
Logic Colloquium
September 20, 2013, 4:10 PM
60 Evans Hall
Lotfi A. Zadeh
Professor Emeritus
of Electrical Engineering and Computer Sciences, Director, Berkeley
Initiative in Soft Computing (BISC), University of California,
Berkeley
Truth and Meaning
The theory which is outlined in my lecture, call it RCT for short, is a departure from traditional theories of truth and meaning, including correspondence theory, coherence theory, possible-world semantics and truth-conditional semantics. The principal objective of RCT is construction of a procedure which on application to a proposition, p, drawn from a natural language leads to: (a) a mathematically well-defined meaning of p; and (b) truth value of p.
The centerpiece of RCT is the concept of a restriction. Informally, a restriction, R(X), on a variable, X, is a limitation on the values which X can take. Typically, a restriction is described in a natural language. Simple example. Usually X is significantly larger than a, where a is a real number. The canonical form of a restriction is: X isr R, where X is the restricted variable, R is the restricting relation, and r is an indexical variable which defines the way in which R restricts X.
There are two key postulates in RCT. First, the meaning postulate, MP. MP asserts that the meaning of p is a restriction, X isr R, in which X, R and r are implicit in p. This restriction is referred to as the canonical form of p. Second, the truth postulate, TP, which asserts that the truth value of p is the degree to which X satisfies R.
In RCT, a proposition, p, is associated with two truth values—internal truth value and external truth value. The internal truth value modifies the meaning of p. The external truth value relates to the degree of agreement of p with factual information. In the definition of truth value which was stated earlier, the truth value is internal. In RCT, truth values are numerical, taking values in the unit interval, or linguistic, e.g., very true, not quite true, more or less true, usually true, etc.
Logic Colloquium
September 06, 2013, 4:10 PM
60 Evans Hall
Theodore A. Slaman
Professor of
Mathematics, University of California, Berkeley
On Normal Numbers
A real number is simply normal in base b if in its base-b expansion each digit appears with asymptotic frequency 1/b. It is normal in base b if it is simply normal in all powers of b, and absolutely normal if it is simply normal in every integer base. By a theorem of E. Borel, almost every real number is absolutely normal. We will present three main results. We give an efficient algorithm, which runs in nearly quadratic time, to compute the binary expansion of an absolutely normal number. We demonstrate the full logical independence between normality in one base and another. We will give a necessary and sufficient condition on a set of natural numbers M for there to exist a real number X such that X is simply normal to base b if and only if b is an element of M.
Tarski Lectures
April 26, 2013, 4:10 PM
60 Evans Hall
Jonathan Pila
Reader in
Mathematical Logic, University of Oxford
The Zilber-Pink Conjecture
Tarski Lectures
April 24, 2013, 4:10 PM
TBA
Jonathan Pila
Reader in
Mathematical Logic, University of Oxford
Special Points and Ax-Lindemann
Tarski Lectures
April 22, 2013, 4:10 PM
TBA
Jonathan Pila
Reader in
Mathematical Logic, University of Oxford
Rational Points of Definable Sets and Diophantine Problems
Logic Colloquium
April 19, 2013, 4:10 PM
60 Evans Hall
W. Hugh Woodin
Professor of
Mathematics, University of California, Berkeley
Ultimate Truth or Ultimate Chaos?
Set Theory is arguably reaching a tipping point. But which way will it tip?
Logic Colloquium
April 05, 2013, 4:10 PM
60 Evans Hall
C. Ward Henson
Professor Emeritus
of Mathematics, University of Illinois at Urbana-Champaign
Continuous Model Theory and Gurarij’s Universal, Homogeneous, Separable Banach Space
Gurarij’s Banach space was constructed in the 1960s using a metric version of a Fraisse construction; it is universal isometrically (for separable Banach spaces) and is homogeneous in an almost-isometric sense relative to its finite-dimensional subspaces. It is the analogue (for Banach spaces) of such structures as the random graph and Urysohn’s metric space. General results in Banach space theory from the 1960s show that its dual space is of the form L1(μ) for some measure μ, so it falls into the important class of “classical Banach spaces”, a fact that is far from obvious based on the original construction. Wolfgang Lusky showed in the 1970s that the Gurarij space is isometrically unique, a surprising result. He also indicated that the set of smooth points of norm 1 is an orbit of its automorphism group. In this talk it will be shown how these results can be seen and extended using continuous model theory. In particular, the class of separable Gurarij spaces can be realized as the class of separable models of a certain continuous theory T (of unit balls of Banach spaces); this theory has quantifier elimination and is the model completion of the theory of all Banach spaces. An optimal amalgamation result yields a simple formula for the induced metric on the type spaces of T over sets of parameters, which is a key to the applications that will be discussed in this talk. A highlight of recent research is the following: let X be Gurarij’s space and let E be any finite dimensional space. Then there is anisometric linear embedding J of E into X such that J (E) has the unique Hahn-Banach extension property in X; moreover, the set of all such embeddings forms a full orbit under the action of the automorphism group of X. Model-theoretically this situation is equivalent to saying that X expanded by naming all elements of J(E) is an atomic model of its theory. This is joint work with Itai Ben Yaacov, and our paper is available at the arXiv. [Note: Earlier Gurarij’s name was transliterated “Gurarii” and in the west it is always pronounced “Ger-ar-eee”.]
Logic Colloquiium
March 15, 2013, 4:10 PM
Farmer Schlutzenberg
Visiting
Scholar in Mathematics, University of California, Berkeley
Jónsson Cardinals in L(R)
The standard axioms of set theory leave many natural questions undecided. Determinacy axioms and large cardinal axioms give a much more complete picture, one that is natural and compelling. Strong connections are known to exist between these two families of axioms.
The Axiom of Determinacy states that in any infinite two- player game of perfect information on the natural numbers one player has a winning strategy. AD implies large cardinal axioms. It also contradicts the Axiom of Choice, and even implies all sets of reals are Lebesgue measurable. Therefore AD fails in the full universe of sets, where Choice holds. However, there are important subuniverses where AD may hold, among them the universe L(R) of all sets constructible from the reals. Assuming AD holds in L(R), a detailed analysis of L(R) is possible, extending classical results of descriptive set theory on analytic and Borel sets.
A cardinal is Jónsson if every structure of that size has a proper elementary substructure of that size. They were introduced by Bjarni Jónsson in 1972. We will survey the above background and sketch a proof that, given AD, there are many Jónsson and Rowbottom cardinals in L(R). Both results use the same key argument. (The result for Jónsson cardinals was first announced by Woodin.)
Logic Colloquium
March 01, 2013, 4:10 PM
Wesley H. Holliday
Assistant
Professor of Philosophy, University of California, Berkeley
Recent Work in Epistemic Logic
Epistemic logic provides a formal framework for modeling the knowledge of agents. Used by philosophers, theoretical computer scientists, AI researchers, game theorists, and others, epistemic logic has become one of the main application areas for modal logic. In this talk, I will survey some recent work in epistemic logic related to my research. The focus will be on how the formalization of philosophical ideas has led to interesting logical issues. Examples will include new kinds of preservation theorems and axiomatization results, as well as open problems. Some of the papers mentioned in the talk are available at http://philosophy.berkeley.edu/people/page/128.
Logic Colloquium
February 15, 2013, 4:10 PM
Itay Neeman
Professor of
Mathematics, University of California, Los Angeles
Forcing Axioms
Forcing axioms are statements about the existence of filters over partially ordered sets, meeting given collections of dense open subsets. They have been used for several decades as center points for consistency proofs. A forcing axiom is shown consistent, typically through an iterated forcing construction, and then the consistency of other statements can be established by deriving them from the axiom.
In my talk I will begin with background on forcing and on some of the most important forcing axioms, including Martin’s Axiom (MA), and the Proper Forcing Axiom (PFA). I will discuss some key points in the axioms’ consistency proofs, and difficulties in adapting the proofs to obtain analogues of PFA that involve meeting more than aleph-sub-one dense open sets. I will end with recent work on new consistency proofs, and higher analogues of PFA.
Logic Colloquium
February 01, 2013, 4:10 PM
John MacFarlane
Professor of
Philosophy, University of California, Berkeley
On the Motivation for the Medieval Distinction between Formal and Material Consequence
Fourteenth-century logicians define formal consequences as consequences that remain valid under uniform substitutions of categorematic terms, but they say little about the significance of this distinction. Why does it matter whether a consequence is formal or material? One possible answer is that, whereas the validity of a material consequence depends on both its structure and “the nature of things”, the validity of a formal consequence depends on its structure alone. But this claim does not follow from the definition of formal consequence by itself, and the fourteenth- century logicians do not give an argument for it. For that we must turn to Abelard, who argues explicitly that consequences that hold under uniform substitutions of their terms take their validity from their construction alone, and not from “the nature of things”. I will look at Abelard’s argument in its historical context. If my reconstruction is correct, the argument, and hence also the significance of the distinction between formal and material consequence, depends on a conception of “the nature of things” that we can no longer accept. This should give pause to contemporary thinkers who look to medieval notions of formal consequence as antecedents of their own.
Logic Colloquium
November 30, 2012, 4:10 PM
Martin Davis
Professor Emeritus of
Mathematics and of Computer Science, Courant Institute, New York
University, and Visiting Scholar in Mathematics, University of
California, Berkeley
Pragmatic Platonism
It will be argued that Platonism in mathematics comes from the empirical fact of the success of mathematicians in dealing with the infinite and not from metaphysical speculation. Thus, absolute certainty is no more to be expected of mathematical knowledge than it is of empirical science.
Logic Colloquium
November 16, 2012, 4:10 PM
60 Evans Hall
Grigori Mints
Professor of
Philosophy and Mathematics, Stanford University; Senior Computer
Scientist, SRI International, Menlo Park (July 2012 - April 2013)
Failure of Interpolation for Intuitionistic Logic of Constant Domains
The intuitionistic logic CD of constant domains is axiomatized by adding the schema (∀x)(C ∨ A(x)) → C ∨ (∀x)A(x) (x is not free in C) to the familiar axiomatization of intuitionistic predicate logic. CD is sound and complete for Kripke semantics with constant individual domains. The interpolation theorem says that for any provable implication A → B there is an interpolant I such that A → I and I → B are both provable, and I contains only predicates common to A and B. There are at least two claimed proofs of the interpolation theorem for CD, both published in The Journal of Symbolic Logic (v. 42, 1977 and v. 46, 1981). We prove that in fact the interpolation theorem fails for CD: the provable implication ((∀x)(∃y)(Py&(Qy → Rx))&¬(∀x)Rx) → ((∀x)(Px → (Qx ∨ r)) → r) does not have an interpolant. Posted as http://arxiv.org/abs/1202.3519. This is a joint work with Grigory Olkhovikov (Ural State University) and Alasdair Urquhart (University of Toronto).
Logic Colloquium
November 02, 2012, 4:10 PM
60 Evans Hall
Joseph Mileti
Assistant Professor
of Mathematics and Statistics, Grinnell College
Computable Combinatorics
We survey the history and some recent results analyzing theorems about trees, graph colorings, and other combinatorial structures from the viewpoint of computable mathematics. By analyzing the complexity of solutions to computable instances of these problems, we discover some striking connections across different theorems and gain insight into the techniques required in every possible proof.
Logic Colloquium
October 19, 2012, 4:10 PM
60 Evans Hall
Thomas Scanlon
Professor of
Mathematics, University of California, Berkeley
A Logic for General Differential Equations
Differential and difference equations have been studied from the point of view of model theory through the theories of model-complete differential and difference fields. The known theorems are similar, but the theories have been developed in parallel. I will report on joint work with Rahim Moosa in which we propose a general theory of D-fields, proving the existence of model companions, describing the definable sets, and establishing the fine-structural Zilber trichotomy principle uniformly. Our theory also grounds a general Galois theory for difference/differential equations and provides a rigorous framework for understanding the confluence from difference to differential equations.
Logic Colloquium
October 05, 2012, 4:10 PM
60 Evans Hall
Seth Yalcin
Associate Professor of
Philosophy, University of California, Berkeley
Some Aspects of the Logical Structure of Conversation
Normally, uttering a sentence changes the state of a conversation. Suppose we thought a natural language as determining a state transition system, where each state corresponds to a state of a possible conversation, and each sentence corresponds to a label inducing a state transition. Then we can ask: what properties are characteristic of the state transition systems appropriate to modeling conversation in natural language (or various fragments thereof)? A leading idea in the philosophy of language, and in linguistic semantics and pragmatics, is that the central effect of uttering a declarative sentence in conversation is to add a proposition to the common background assumptions of the interlocutors. We give a representation theorem indicating what formal properties are characteristic of this effect from a state transition system perspective. We then go on to discuss some fragments of language that appear to lack these properties. I connect the results to the debate about how ‘dynamic’ natural language is, and to debates about what a compositional theory of meaning should look like. (This work is joint with Daniel Rothschild, All Souls College, Oxford.)
Logic Colloquium
September 21, 2012, 4:10 PM
60 Evans Hall
Antonio Montalban
Associate
Professor of Mathematics, University of California, Berkeley
A Computability-Theoretic Equivalent to Vaught’s Conjecture
We find two computability-theoretic properties on the models of a theory T which hold if and only if T is a counterexample to Vaught’s conjecture.
Logic Colloquium
September 07, 2012, 4:10 PM
60 Evans Hall
Maryanthe Malliaris
L. E. Dickson
Instructor in Mathematics, University of Chicago
Saturation of Ultrapowers and the Structure of Unstable Theories
The talk will be about some very recent progress on Keisler’s order, a far-reaching program of understanding basic model-theoretic structure through the lens of regular ultrapowers. The focus of the talk will be on classification of unstable theories; some applications to problems in set theory/general topology will also be mentioned, including a recent result of Malliaris and Shelah solving the oldest problem on cardinal invariants of the continuum.
I will assume basic familiarity with ultraproducts, Los’s theorem, and saturation, and plan to give most other relevant definitions. Papers mentioned in the talk are available at http://math.uchicago.edu/~mem.
Logic Colloquium
August 24, 2012, 4:10 PM
60 Evans Hall
Dana Scott
University Professor
Emeritus, Carnegie Mellon University; Visiting Scholar, University of
California, Berkeley
Lambda Calculus: Then and Now
A very fast development in the early 1930’s following Hilbert’s codification of Mathematical Logic led to the Incompleteness Theorems, Computable Functions, Undecidability Theorems, and the general formulation of Recursive Function Theory. The so-called Lambda Calculus played a key role. The history of these developments will be traced, and the much later place of Lambda Calculus in Mathematics and Programming-Language Theory will be outlined.
Logic Colloquium
April 20, 2012, 4:10 PM
60 Evans Hall
Umesh Vazirani
Roger A.Strauch
Professor of Electrical Engineering and Computer Sciences Director,
Berkeley Quantum Computation Center, University of California,
Berkeley
A Turing Test for Randomness
Is it possible to certify that the n-bit output of a random number generator is “really random”? A possible approach to this question is via the theory of algorithmic randomness due to Kolmogorov, Chaitin, and Solomonoff, which identifies randomness with uncompressibility by any Turing Machine. Unfortunately this definition does not result in an efficient test for randomness. Indeed, in the classical World it seems impossible to provide such a test.
Quantum mechanics allows for a remarkable random number generator: its output is certifiably random in the sense that if the output passes a simple statistical test, and there is no information communicated between the two boxes in the randomness generating device (based, say, on the speed of light limit imposed by special relativity), then the output is certifiably random. Moreover, the proof that the output is truly random does not even depend upon the correctness of quantum mechanics! Based on joint work with Thomas Vidick.
Logic Colloquium
April 06, 2012, 4:10 PM
60 Evans Hall
Jouko Väänänen
Professor of
Mathematics, University of Helsinki, and Professor of Foundations of
Mathematics, University of Amsterdam
Second-Order Logic and Model Theory
Second-order logic is so strong that it makes sense to ask, are complete finitely axiomatized second-order theories always categorical? Carnap claimed that the answer is yes, but his proof did not work. Ajtai and Solovay showed that the answer depends on set theory. We present some recent new results in this field. This is joint work with Hyttinen and Kangas.
Logic Colloquium
March 16, 2012, 4:10 PM
60 Evans Hall
George M. Bergman
Professor
Emeritus of Mathematics, University of California, Berkeley
Linear Maps and Ultrafilters
Most of my talk will be devoted to showing that if k is an infinite field, I an infinite set, and g : kI → V a linear map to a finite-dimensional k-vector-space, then g has elements where “almost all’’ is made precise in terms of a finite set of card(k)+-complete ultrafilters on I. In particular, if I is not enormous,”almost all’’ means “all but finitely many’’.
Before proving the above, I will show why such results are of value in studying homomorphisms from an infinite direct product of algebras (e.g., of associative algebras, or Lie algebras) to another algebra, and sketch the proofs of some easier results on linear maps kI → V.
(The result of the first paragraph above is Lemma 7 of a paper with N.Nahlus, which can be read in preprint form at http://math.berkeley.edu/~gbergman/papers/prod_Lie2.pdf or in published form at http://dx.doi.org/10.1016/j.jalgebra.2012.01.004. The discussion of how such results are used, and proofs of easier results of the same nature, can be found in §§1-2 of that paper.)
Logic Colloquium
March 02, 2012, 4:10 PM
60 Evans Hall
Alf Onshuus
Professor Titular,
University of the Andes
VC-Density and dp-Rank in Model Theory
VC-dimension was introduced by Vapnik and Chervonenkis to measure, in some way, the complexity of a family of subsets of a given universe. This notion has had important applications in statistics an learning theory, usually in the framework where one looks at a definable family of definable sets (usually in a cartesian power of the real or the real exponential field). Now, in this context the model theory becomes clearly related and in fact any model where all uniform definable families of sets have finite VC-dimension is precisely the model of a theory with the “independence property” introduced by Shelah to define dependent theories (also known as theories with NIP). This setting of “VC-theory” of a given uniform definable family of definable sets in a fixed structure and the relation with notions in model theory around it is the setting for the talk.
I will survey some of the results on VC-density (a very important notion in VC-theory) that arise from the model-theoretic analysis, introduce and show some results about dp-rank (a rank introduced by Shelah in the context of dependent theories), and show some evidence of a possible strong relation between the two notions.
Alfred Tarski Lectures
February 24, 2012, 4:10 PM
60 Evans Hall
Per Martin-Löf
Emeritus Professor
of Logic, Departments of Mathematics and Philosophy, Stockhold
University
Tarski’s Metamathematical Reconstruction of the Notions of Truth and Logical Consequence
Alfred Tarski Lectures
February 22, 2012, 4:10 PM
TBA
Per Martin-Löf
Emeritus Professor
of Logic, Departments of Mathematics and Philosophy, Stockhold
University
Propositions, Truth and Consequence
Alfred Tarski Lectures
February 21, 2012, 4:10 PM
TBA
Per Martin-Löf
Emeritus Professor
of Logic, Departments of Mathematics and Philosophy, Stockhold
University
Assertion and Inference
Logic Colloquium
February 03, 2012, 4:10 PM
60 Evans Hall
Leo Harrington
Professor of
Mathematics, University of California, Berkeley
Is There a Philosophical View Already in Mathematical Logic?
This talk will suggest that a certain “view”, readily available from basic ingredients of mathematical logic, bears a resemblance with views found respectively in the philosophers Parmenides, Hegel, and Heidegger. And this talk will attempt to indicate the possibility that, while staying inside mathematical logic, this view can be elaborated to maintain such a resemblance with a philosophical view found in common in the works of these philosophers.
Logic Colloquium
January 20, 2012, 4:10 PM
60 Evans Hall
Sherrilyn Roush
Associate Professor
of Philosophy and Chair, Group in Logic and the Methodology of Science,
University of California, Berkeley
Reasoning and the Growth of Error
I address several questions about controlling the growth of error introduced by reasoning. For example, if every step of reasoning introduces a new source of error, how can we possibly improve our reliability – as we think we do – by proof-checking, filling in gaps in proofs, or consulting an expert, all of which add steps to our reasoning? I will devote most attention to the topic of closure of knowledge under known implication. How can knowledge be closed if in one step of valid deductive reasoning I can go from knowledge that my car is parked on Main Street to a belief that it has not been stolen since I parked it, which it seems I do not know? I argue that this closure problem is entirely a matter of how fast false positive error in the conclusion grows over deductive reasoning from the premise, and develop a graded model with upper bounds on this growth. It turns out that problem examples like the car theft case, and more radical skeptical cases, can be generated only by committing an error of reasoning.
Logic Colloquium
December 02, 2011, 4:10 PM
60 Evans Hall
Lara Buchak
Assistant Professor of
Philosophy, University of California, Berkeley
Risk and Tradeoffs
The prevailing view is that subjective expected utility theory is the correct theory of instrumental rationality. Subjective expected utility theory (hereafter, EU theory) is thought to characterize the preferences of all rational decision makers. And yet, there are some preferences that violate EU theory that seem both intuitively appealing and prima facie consistent. An important group of these preferences stem from how ordinary decision makers take risk into account: ordinary decision makers seem to care about “global” properties of gambles, but EU theory rules out their doing so.
EU theory allows agents to subjectively determine how much they value outcomes (their utility function) and how likely they think various states are to obtain (their probability function). However, there is arguably a third subjective component of instrumental rationality, namely one’s norm for translating these two values into preferences among risky acts. On EU theory, every rational agent must use the same norm, a norm that is insensitive to global properties. I propose a theory on which decision makers subjectively determine how they want to take risk into account, and thus subjectively determine all three components of instrumental rationality. By providing a “representation theorem,” I show that the third component corresponds at the level of preferences to how an individual values tradeoffs in different structural parts of the gamble, e.g., to whether he cares more about what happens in the worst-case scenario or the best-case scenario. And I therefore show how non-EU maximizers can be seen as instrumentally rational: as taking the means to their ends.
Logic Colloquium
November 18, 2011, 4:10 PM
60 Evans Hall
Theodore A. Slaman
Professor and
Chair of the Department of Mathematics, University of California,
Berkeley
The First-Order Consequences of the Existence of an Infinite Random Sequence
We will discuss the question, “What first-order statements follow from the existence of an infinite random sequence by effective means?” The answer depends on the degree of randomness in the infinite source. In one case, it isolates a mysterious subtheory of arithmetic strictly between Σ1-Induction and Σ2-Bounding.
Logic Colloquium
October 28, 2011, 4:10 PM
60 Evans Hall
Tomasz Placek
Professor of
Philosophy, Jagiellonian University, Krakow, Poland; Visiting Scholar in
Philosophy, University of California, Berkeley
Possibilites without Possible Worlds/Histories
Possible worlds have turned out to be a particularly useful tool of modal metaphysics, although their global character makes them philosophically suspect. Hence, it would be desirable to arrive at some local modal notions that could be used instead of possible worlds. In this talk I will focus on what is known as historical (or real) modalities, an example of which is tomorrow’s sea-battle. The modalities involved in this example are local since they refer to relatively small chunks of our world: a gathering of inimical fleets on a bay near-by has two alternative possible future continuations: one with a sea-battle and the other with no-sea-battle. The objective of this talk is to sketch a theory of such modalities that is framed in terms of possible continuations rather than possible worlds or possible histories. The proposal will be tested as a semantic theory for a language with historical modalities, tenses, and indexicals. The talk builds on my JPL paper (cf. http://www.springerlink.com/content/q2423241l6525063/).
Logic Colloquium
October 14, 2011, 4:10 PM
60 Evans Hall
Daisuke Ikegami
Japan Society for
the Promotion of Science Postdoctoral Fellow, University of California,
Berkeley
Gale-Stewart Games and Blackwell Games
In 1953, Gale and Stewart developed the general theory of infinite games (the so-called Gale-Stewart games) which are two- player zero-sum infinite games with perfect information. The theory of Gale-Stewart games has been deeply investigated by many logicians and it has been one of the main topics in set theory while having connections with other topics of set theory as well as model theory and computer science.
In 1969, Blackwell proved the extension of von Neumann’s mini-max theorem where he introduced infinite games with imperfect information – nowadays called Blackwell games. Although Blackwell games have not been as much investigated as Gale-Stewart games, in 1998, Martin proved that the Axiom of Determinacy (AD) implies the Axiom of Blackwell Determinacy (Bl-AD) and conjectured the converse, which is still not known to be true.
In this talk, we introduce Blackwell games and their determinacy and talk about the connection between the determinacy of Gale-Stewart games and that of Blackwell games. A part of the work is with Benedikt Löwe & Devid de Kloet and another part is with Hugh Woodin.
Logic Colloquium
September 30, 2011, 4:10 PM
60 Evans Hall
Lotfi A. Zadeh
Professor in the
Graduate School, University of California, Berkeley; Director, Berkeley
Initiative in Soft Computing
Can Mathematics Deal with Computational Problems Which Are Stated in a Natural Language?
Here are a few very simple examples of computational problems which are stated in a natural language. (a) Most Swedes are tall. What is the average height of Swedes? (b) Probably John is tall. What is the probability that John is short? What is the probability that John is very short? What is the probability that John is not very tall? (c) Usually Robert leaves his office at about 5 pm. Usually it takes Robert about an hour to get home from work. At what time does John get home? (d) X is a real-valued random variable. Usually X is much larger than approximately a. Usually X is much smaller than approximately b. What is the probability that X is approximately c, where c is a number between a and b? (e) A and B are boxes, each containing 20 balls of various sizes. Most of the balls in A are large, a few are medium, and a few are small. Most of the balls in B are small, a few are medium, and a few are large. The balls in A and B are put into a box C. What is the number of balls in C which are neither large nor small? For convenience, such problems will be referred to as CNL problems.
It is a long-standing tradition in mathematics to view computational problems which are stated in a natural language as being outside the purview of mathematics. Such problems are dismissed as ill-posed and not worthy of attention. In the instance of CNL problems, mathematics has nothing constructive to say. In my lecture, this tradition is questioned and a system of computation is suggested which opens the door to construction of mathematical solutions of CNL problems. The system draws on the fuzzy-logic-based formalism of computing with words (CW). (Zadeh 2006) A concept which plays a pivotal role in CW is that of precisiation of meaning. More concretely, precisiation involves translation of natural language into a mathematical language in which the objects of computation are well-defined — though not conventional — mathematical constructs.
A key idea involves representation of the meaning of a proposition, p, drawn from a natural language, as a restriction on the values which a variable, X, can take. Generally, X is a variable which is implicit in p. The restriction is represented as an expression of the form X isr R, where X is the restricted variable, R is the restricting relation and r is an indexical variable which defines the way in which R restricts X. This expression is referred to as the canonical form of p. Canonical forms of propositions in a natural language statement of a computational problem serve as objects of computation in CW.
Construction and use of mathematical solutions of CNL problems is an unexplored domain in mathematics. The importance of this domain derives from the fact that much of human knowledge, and particularly world knowledge, is described in natural language.
Logic Colloquium
September 16, 2011, 4:10 PM
60 Evans Hall
Adam Day
Miller Research Fellow,
University of California, Berkeley
Randomness and Neutral Measures
Algorithmic randomness provides a way to define a random outcome. The underlying idea is that a random outcome should have no ‘rare’ mathematical properties. A combination of measure theory and recursion theory provides a framework for defining what a rare property is and consequently for defining randomness itself.
Most research in algorithmic randomness has been conducted using computable measures. However, there are some interesting results that have come from considering noncomputable measures. In a surprising result, Levin established the existence of probability measures for which all in^Lfinite binary sequences are random. These measures are termed neutral measures. In this talk I will introduce Levin’s neutral measures. I will also outline some recent joint work with Joseph Miller in which we determine some key properties of neutral measures.
Logic Colloquium
September 02, 2011, 4:10 PM
60 Evans Hall
Paolo Mancosu
Professor and Chair
of Philosophy, University of California, Berkeley
Axiomatics and purity of methods: On the relationship between plane and solid geometry
Traditional geometry concerns itself with planimetric and stereometric considerations, which are at the root of the division between plane and solid geometry. To raise the issue of the relation between these two areas brings with it a host of different problems that pertain to mathematical practice, epistemology, semantics, ontology, methodology, and logic. In addition, issues of psychology and pedagogy are also important here.
In this talk (which is based on joint work with Andy Arana), my major concern is with methodological issues of purity. In the first part I will give a rough sketch of some key episodes in mathematical practice that relate to the interaction between plane and solid geometry. In the second part, I will look at a late nineteenth century debate (on “fusionism”) in which for the first time methodological and foundational issues related to aspects of the mathematical practice covered in the first part of the paper came to the fore. I conclude this part of the talk by remarking that only through an axiomatic and analytical effort could the issues raised by the debate on “fusionism” be made precise. The third part of the talk focuses on Hilbert’s axiomatic and foundational analysis of the plane version of Desargues’ theorem on homological triangles and its implications for the relationship between plane and solid geometry. Finally, building on the foundational case study analyzed in the third section, in the fourth section I point the way to the analytic work necessary for exploring various important claims on “purity”, “content”, and other relevant notions.
Logic Colloquium
April 22, 2011, 4:10 PM
60 Evans Hall
Matthew D. Foreman
Professor of
Mathematics, University of California, Irvine
Classifying Measure Preserving Diffeomorphisms of the Torus
The statistical behavior of deterministic dynamical systems can be completely random. The ergodic theorem tells us that the statistical behavior of a system carrying an invariant ergodic measure is captured by the isomorphism class of that measure. Motivated by this type of phenomenon, in 1932 von Neumann proposed the program of classifying the ergodic measure preserving systems.
In this talk we will describe applications of the tools of descriptive set theory to show that this program is impossible. In earlier joint work with Weiss, it was shown that the isomorphism relation is “turbulent” and hence no generic class can be reduced to an S-infinity action. With Rudolph and Weiss it was shown that the isomorphism relation is a complete analytic set.
But what about concrete systems: C-infinity measure preserving transformations of the torus? Recent work with Weiss extends these results to such diffeomorphisms. Finally a very recent result shows that the “graph-isomorphism problem” can be reduced to isomorphism of measure preserving diffeomorphisms. As a corollary the isomorphism relation is complete for S-infinity actions.
Logic Colloquium
April 08, 2011, 4:10 PM
60 Evans Hall
Ronald Fagin
Manager, Foundations
of Computer Science, IBM Almaden Research Center
P vs. NP, and Community Refereeing in the Web Era
P and NP are “complexity classes”. The problem of whether they are the same is renowned in computer science and mathematics, and a solution yields a Clay Millennium Prize of one million dollars, mathematical immortality, and deep insight into efficient computation. In August of 2010 there was a lot of excitement over the announcement of a possible solution: a claimed mathematical proof that P and NP are different, using ideas from mathematical logic and statistical physics. We will describe the P vs. NP problem at a high level for a general scientific audience. What is this problem? How is it important? We will also discuss the fascinating sociological phenomenon of mathematical refereeing in “internet time”. Finally, we will discuss the idea of the attempted proof, and issues that were raised during the “internet refereeing” process.
Herbert Enderton Memorial Gathering
March 15, 2011, 11:00 AM
Berkeley City Club, 2315 Durant Avenue, Berkeley, California
You are invited to join us to celebrate Herbert’s life on April 15, his 75th birthday. The gathering will be in Berkeley, CA at the Berkeley City Club from 11 am to 3 pm. Lunch will be served.
Whether or not you can attend, we would love to have you send us short written stories and memories of Herbert that we can draw on for the eulogy and share with his loved ones.
Warmly, Cathy, Eric, and Bert
(Please forward this note to anyone whose address we might not have. All who wish to attend are welcome; we just need a head count. RSVP: cenderton@hotmail.com)
(A group rate is available for lodging at the Berkeley City Club. (510) 848-7800.)
Logic Colloquium
March 11, 2011, 4:10 PM
60 Evans Hall
Bjørn Kjos-Hanssen
Associate
Professor of Mathematics, University of Hawaii at Manoa, and, Visiting
Associate Professor of Mathematics, University of California, San
Diego
Discerning Randomness from an Asymptotic Hamming Distance
A notion of asymptotic Hamming distance suitable for the study of algorithmic randomness of infinite binary sequences is developed. As an application, it is shown that there is no fixed procedure that computes a (weakly) Mises-Wald-Church stochastic sequence from an arbitrary complex sequence. Here a sequence is complex if its fixes have Kolmogorov complexity bounded below by an unbounded, nondecreasing computable function.
Logic Colloquium
February 25, 2011, 4:10 PM
60 Evans Hall
Phokion G. Kolaitis
Professor of
Computer Science, University of California Santa Cruz, and, Research
Staff Member, IBM Almaden Research Center
Random Graphs and the Parity Quantifier [Slides]
The classical zero-one law for first-order logic asserts that the asymptotic probability of every first-order definable property of finite graphs always exists and is either zero or one. Over the years zero-one laws have been established for numerous extensions of first-order logic, including least-fixed point logic, finite-variable infinitary logics, and certain fragments of existential second-order logic. The zero-one law, however, fails to hold for any logical formalism that is powerful enough to express parity, that is, the property “there is an odd number of elements”; in fact, for such logical formalisms, even the convergence law fails to hold. In this work, we turn the parity barrier into a feature and systematically investigate the asymptotic probabilities of properties of finite graphs expressible in first-order logic augmented with the parity quantifier. Our main result is a “modular convergence law” that captures the limiting behavior of properties expressible in this extension of first-order logic on finite graphs. This is joint work with Swastik Kopparty, Institute for Advanced Study.
Logic Colloquium
February 11, 2011, 4:10 PM
60 Evans Hall
Grigori Mints
Professor of
Philosophy and of Mathematics, Stanford University
Extension of Epsilon Substitution to Second-Order Systems
The problem of extending the epsilon substitution method to second-order systems was posed by Hilbert in 1928 and repeated in “Grundlagen der Mathematik” (1939). In this talk I will describe the first-order formulation (due to Hilbert, Ackermann, von Neumann, and Bernays) and outline possible extensions to the second order, progress achieved up to now, and obstacles to a termination proof.
Logic Colloquium
January 28, 2011, 4:10 PM
60 Evans Hall
Thomas Scanlon
Professor of
Mathematics, University of California, Berkeley
Number-theoretic Consequences of the Model Theory of Real Geometry
We say that a first-order structure (R, <, …) in some language expanding the language of orders is o-minimal if < linearly orders R and every definable subset of R is a finite union of points and intervals. Tarski’s quantifier elimination theorem for the real field implies that the real numbers considered as an ordered ring is o-minimal and subsequent work by various authors has established the o-minimality of more sophisticated structures on the real numbers. The definable sets in an o-minimal structure (in any number of variables) admit a very tame geometric structure theory, very much at odds with the Gödel phenomena in arithmetic. Nevertheless, Pila and Wilkie bridged the divide between geometry and arithmetic by establishing bounds on the number of rational points in sets definable in o-minimal expansions of the real numbers. In recent work, Pila employed this counting theorem to prove the celebrated André-Oort conjecture on algebraic relations on special points of modular curves. In this talk, I will describe the Pila-Wilkie counting theorem and some of my own work relating arithmetic questions about algebraic dynamical systems to o-minimality and the Pila-Wilkie bounds.
Logic Colloquium
December 03, 2010, 4:10 PM
60 Evans Hall
John R. Steel
Professor of
Mathematics, University of California, Berkeley
The Triple Helix
Many different foundational theories have been developed and studied by set theorists. Remarkably, no divergence at the level of statements about “concrete” objects (for example, natural numbers) has emerged, and indeed, these theories are linearly ordered by the inclusion order on their sets of arithmetic consequences. The place of a theory in this order is known as its consistency strength. Our methods for comparing the consistency strengths of set theories center on three intertwined canonical hierarchies of sets. In this talk we shall describe those three hierarchies in very general terms. We shall focus on the philosophical significance of the results and open questions concerning them.
Logic Colloquium
November 12, 2010, 4:10 PM
60 Evans Hall
G. Aldo Antonelli
Professor of
Philosophy, University of California, Davis
The Abstraction Mystique
The logical status of abstraction principles, and especially Hume’s Principle, has been long debated, but the best currently available tool for explicating a notion’s logical character – the notion of permutation invariance introduced by Tarski in 1966 – has not received a lot of attention in this debate. This talk aims to fill this gap. After characterizing abstraction principles as particular classifiers, i.e., mappings from the subsets of a domain into that domain and exploring some of their properties, the paper introduces several distinct notions of permutation invariance for such principles, assessing the philosophical significance of each.
Logic Colloquium
October 29, 2010, 4:10 PM
60 Evans Hall
Isaac Goldbring
Hedrick Assistant
Adjunct Professor of Mathematics, University of California, Los
Angeles
A Survey of the Model Theory of Urysohn’s Metric Space
Urysohn’s metric space U is the unique (up to isometry) Polish (i.e. complete, separable) metric space which is universal, that is it contains an isometric copy of all Polish metric spaces, and ultrahomogeneous, that is any isometry between finite subspaces of U extends to an isometry of U. Urysohn’s metric space (and its isometry group) has been studied by topologists and descriptive set theorists for a plethora of reasons. In this talk, I will outline much of what is known on the model theory of Urysohn’s metric space in the context of model theory for metric structures, which I will introduce at the beginning of my talk. I will discuss matters such as axiomatizability, quantifier elimination, independence relations, and definability.
Logic Colloquium
October 08, 2010, 4:10 PM
60 Evans Hall
Henry Towsner
Hedrick Assistant
Adjunct Professor of Mathematics, University of California, Los
Angeles
The Infinitary Approach to Finite Combinatorics
The “Correspondence Principle” makes it possible to take combinatorial problems about finite structures and translate them into questions about infinitary, measure-theoretic structures. While there are fairly elementary combinatorial approaches, it has recently become clear that the model-theoretic approach creates a particularly rich and flexible infinitary structure. We will outline the model-theoretic proof of the correspondence principle and point out some of its new applications. To illustrate the flexibility of this method, we will give a short proof of the Szemeredi regularity lemma, an important tool in the study of finite graphs.
Logic Colloquium
October 01, 2010, 4:10 PM
60 Evans Hall
Solomon Feferman
Professor Emeritus
of Mathematics and Philosophy, Stanford University
What’s Definite? What’s not?
Most axiomatizations of set theory that have been treated metamathematically have been based either entirely on classical logic or entirely on intuitionistic logic. But a natural conception of the set-theoretic universe is as an indefinite (or “potential”) totality, to which intuitionistic logic is more appropriately applied, while each set is taken to be a definite (“or completed”) totality, for which classical logic is appropriate; so on that view, set theory should be axiomatized on some correspondingly mixed basis. Similarly, in the case of predicative analysis, the natural numbers are conceived to form a definite totality, while the universe of sets (or functions) of natural numbers are viewed as an indefinite totality, so that, again, a mixed semi-constructive logic should be the appropriate one to treat the two together. In the first part of the talk I will present ways of formulating such semi-constructive systems of analysis and set theory and survey some results characterizing their proof-theoretic strength. Interestingly, though the logic is weakened, one can usefully strengthen certain principles. In the last part of the talk I will relate this work to the controversial discussion as to whether certain statements in the language of set theory such as the Continuum Hypothesis express definite or indefinite problems.
Logic Colloquium
September 17, 2010, 4:10 PM
60 Evans Hall
Alice Medvedev
Visiting Assistant
Professor of Mathematics, University of California Berkeley
The Recursive Spectrum of a Strongly Minimal Modular Theory of Groups in a Finite Language
The recursive spectrum of a fixed strongly minimal theory T is the set of integers n such that the model of T of dimension n admits a recursive presentation. How complicated can this set be? Some interesting examples can be constructed by coding complicatedness into an infinite language, or by using Hrushovski constructions. We are working towards showing that the recursive spectrum of a strongly minimal locally modular theory in finite language must be nothing; everything; or only the prime model. The title describes the current state of this joint work with Uri Andrews.
Logic Colloquium
September 03, 2010, 4:10 PM
60 Evans Hall
W. Hugh Woodin
Professor of
Mathematics, University of California Berkeley
Strong Axioms of Infinity and the Search for V
Professor Woodin plans to give a broad and nontechnical overview of the prospects of the existence of an ultimate version of Gödel’s Axiom of Constructibility. The talk will be essentially a repeat of the Plenary Lecture in Logic at the International Congress of Mathematicians in Hyderabad, India, which he delivered on Thursday, August 26.
Logic Colloquium
April 23, 2010, 4:10 PM
60 Evans Hall
Richard A. Shore
Professor of
Mathematics, Cornell University
The Limits of Determinacy in Second-Order Arithmetic
Logic Colloquium
April 16, 2010, 4:10 PM
60 Evans Hall
Umesh V. Vazirani
Strauch Professor
of Electrical Engineering and Computer Sciences, University of
California, Berkeley, and Director of the Berkeley Quantum Computation
Center
Is There a Quantum Analog of the PCP (Probabilistically Checkable Proofs) Theorem?
Logic Colloquium
April 12, 2010, 4:10 PM
60 Evans Hall
Grigor Sargsyan
National Science
Foundation Postdoctoral Fellow, and Assistant Adjunct Professor,
University of California, Los Angeles
Descriptive Inner Model Theory
It is by now a well-known theorem of Martin, Steel, and Woodin that if there is a measurable cardinal above omega-many Woodin cardinals then AD holds in L(R). Since this theorem was proved, many theorems of a similar nature have appeared in set theory. For instance, Steel recently showed that PFA, the proper forcing axiom, implies that AD holds in L(R). Many results that followed the Martin-Steel-Woodin theorem, including Steel’s aforementioned result, used a substantial amount of inner model theory to prove that all sets in a certain “canonical” collection of sets of reals are determined. In this talk we will take canonical stand for Universally Baire. Letting then Gamma be the collection of Universally Baire sets, many of the theorems that have been proven after the Martin-Steel-Woodin result are about the size of Gamma, and in particular, the earlier results just show that under certain hypotheses all sets of reals in L(R) are in Gamma. Descriptive inner model theory then can be defined as the the study of the properties of Gamma in various mathematically rich extensions of ZFC, using methods from inner model theory. Inner model theory enters into the study of the properties of Gamma through a conjecture known as the Mouse Set Conjecture, which allows one to translate questions about sets of reals into questions about iteration strategies for a certain kind of mice known as hod mice. In this talk, we will explain all this terminology and will outline some of the more recent theorems, including the following.
Theorem (S.) If there is a Woodin limit of Woodins then there is Gamma^* which is a subset of Gamma and such that and L(Gamma^*, R) satisfies AD_R + Theta is regular.
We will also explain some applications in calibrating lower bounds of consistency strength and in the inner model problem.
Logic Colloquium
April 02, 2010, 4:10 PM
60 Evans Hall
Seth Yalcin
Assistant Professor of
Philosophy, University of California, Berkeley
Modus Tollens Denied
I present and defend a number of counterexamples to Modus Tollens, construing Modus Tollens as a pattern of inference expressible in ordinary language. I ask what these counterexamples teach us about the meaning of conditional sentences and about the concept of logical consequence appropriate to understanding natural language inference. Since under classical assumptions, Modus Tollens can be proved from Modus Ponens, I ask in particular whether we should abandon Modus Ponens (as Vann McGee has independently argued we should), or give up aspects of the classical reasoning that allows us to prove Modus Tollens from Modus Ponens. I develop the latter approach.
Logic Colloquium
February 26, 2010, 4:10 PM
60 Evans Hall
Lotfi A. Zadeh
Professor in the
Graduate School and Director of the Berkeley Initiative in Soft
Computing, University of California, Berkeley
Toward a Logic of Everyday Reasoning
In one form or another, attempts to construct a logic of everyday reasoning go back to antiquity. The classical, Aristotelean, bivalent logic may be viewed as a product of such attempts. However, Aristotelean logic does not qualify as a logic of everyday reasoning because it does not come to grips with the core issue — the intrinsic imprecision of everyday reasoning. In modern times, logical systems — driven by a quest for the ultimate in precision, rigor and depth — have become more and more estranged from everyday reasoning. But as we move further into the age of machine intelligence and automated decision-making, the need for a logic of everyday reasoning and natural language understanding becomes increasingly apparent.
Today, a logic of everyday reasoning is not in existence. In my lecture, I sketch two attempts to construct one. The first attempt relates to what may be called precisiable everyday reasoning. A very simple example is: Icy roads are slippery; slippery roads are dangerous; therefore icy roads are dangerous. In this case, it can be shown that the seemingly reasonable conclusion is incorrect, and that a correct conclusion can be inferred. A simple example of unprecisiable reasoning is the following. I hail a cab and ask the driver to take me to address A the fastest way. The driver takes me to A but it is not possible to prove or disprove that the route which he chose is the fastest way. Construction of a logic of unprecisiable everyday reasoning requires a radical departure from traditional approaches to logical reasoning.
Logic Colloquium
February 12, 2010, 4:10 PM
60 Evans Hall
Jennifer Chubb
Assistant Professor
of Mathematics, University of San Francisco
Degree Spectra of Relations
We consider algorithmic properties of additional relations on computable structures. For example, for a computable linear ordering we may consider the successor relation, which does not have to be computable. The degree spectrum of a relation on a computable structure is the set of degrees (Turing or other) of images of that relation under isomorphism to another computable copy of the structure. We’ll see a variety of examples of degree spectra of definable relations, and will use algorithmic information theory to analyze the strong degree spectra of the omega-type initial segment of computable linear orderings of type omega + omega*. I will discuss some general results, and present some examples from recent collaborative projects.
Logic Colloquium
January 29, 2010, 4:10 PM
60 Evans Hall
John MacFarlane
Associate Professor
of Philosophy, University of California, Berkeley
Fuzzy Epistemicism
It is taken for granted in much of the literature on vagueness that semantic and epistemic approaches to vagueness are fundamentally at odds. If we can analyze borderline cases and the sorites paradox in terms of degrees of truth, then we don’t need an epistemic explanation. Conversely, if an epistemic explanation suffices, then there is no reason to depart from the familiar simplicity of classical bivalent semantics. I question this assumption, showing that there is an intelligible motivation for adopting a many-valued semantics even if one accepts a form of epistemicism. The resulting hybrid view has advantages over both classical epistemicism and traditional many-valued approaches.
Logic Colloquium
December 04, 2009, 4:10 PM
60 Evans Hall
Sherrilyn Roush
Associate Professor
of Philosophy, University of California, Berkeley
Rational Self-Doubt
If one is highly confident that Rosco is the murderer from having seen the crime, and then learns of the substantial experimental psychology evidence that human beings are very unreliable at eyewitness testimony, is one thereby obligated to reduce one’s confidence about Rosco? Current Bayesian rationality theory forbids any such revision because of idealizations that I argue are unnecessary. I develop a more general probabilistic framework that shows how we can learn about the world from information about our own unreliability, through higher-order updating that re-calibrates us. It shows why taking doubt about one’s own judgment seriously does not end up in incoherence or runaway skepticism, and what the added value of this kind of evidence is.
Logic Colloquium
November 20, 2009, 4:10 PM
60 Evans Hall
Denis R. Hirschfeldt
Professor of
Mathematics, University of Chicago
Lowness Properties and Cost Functions
One of the byproduts of recent developments in the theory of algorithmic randomness is an increased interest in notions of computability-theoretic lowness. We will focus on two such notions: K-triviality, which is a natural notion of “randomness-theoretic weakness” with a number of interesting characterizations, and strong jump traceability, which has a more combinatorial definition but is also strongly connected to algorithmic randomness. In particular, we will see how these notions can be characterized in terms of two other powerful ideas arising from the theory of algorithmic randomness: cost functions and diamond classes.
We will use these themes to explain some of the issues of current interest in the rapidly developing area of algorithmic randomness, including a discussion of some major open questions.
Logic Colloquium
November 06, 2009, 4:10 PM
60 Evans Hall
Martin Davis
Professor Emeritus of
Mathematics and of Computer Science, Courant Institute, New York
University, and Visiting Scholar in Mathematics, University of
California, Berkeley
Reflections on Hilbert’s 10th Problem with a New Conjecture
Hilbert’s 10th Problem asked for an algorithm to determine whether a given polynomial equation with integer coefficients has an integer solution. In this talk I will discuss the negative solution to the problem as well as various extensions.
Logic Colloquium
October 23, 2009, 4:10 PM
60 Evans Hall
Lara Buchak
Assistant Professor of
Philosophy, University of California, Berkeley
Different Decision or Different Decision Theory?: Modeling Risk Aversion
Decision theory is supposed to accommodate any preferences that a rational decision maker might have. I show that standard decision theory – expected utility theory – cannot accommodate common preferences that stem from the way apparently rational decision makers account for risk. I propose a generalization of expected utility theory that can accommodate these preferences, and show that it has all the theoretical power of the standard theory: in particular, it has a representation theorem that allows us to derive an agent’s beliefs, desires, and attitudes towards risk from his preferences. I then explore a classic strategy for responding to counterexamples of the type I present: re-characterizing the choice problem facing an agent, by individuating outcomes more finely. This introduces an important question in the methodology of decision theory: under what circumstances should we interpret the agent as facing a different choice problem rather than adopt a different theory to describe his behavior? I approach this question formally, and show that we cannot individuate the outcomes to make his behavior compatible with expected utility theory without also making his behavior compatible with many other opposing theories: thus, I suggest, we cannot save the standard theory.
Logic Colloquium
October 09, 2009, 4:10 PM
60 Evans Hall
Nate Ackerman
Visiting Scholar in
Mathematics, University of California, Berkeley
Trees, Sheaves, and Definition by Recursion
We will show there is a topological space for which presheaves are the same thing as trees. We will further show that there is a sheaf on this topological space which has an important relationship with Baire space.
We will then use these connections to show how a definition by transfinite recursion can be thought of as an operation on sheaves, and how the well-definedness of such a definition can be thought of as a property of the sheaf we are working on.
This will then allow us to define a second-order tree as a sheaf on a tree and to expand our notion of definition by transfinite recursion to all well-founded second-order trees (even those which are ill-founded as normal trees).
We will then mention how these techniques can be used to prove a variant of the Suslin-Kleene Separation theorem.
Logic Colloquium
September 25, 2009, 4:10 PM
60 Evans Hall
Dana S. Scott
University Professor,
Emeritus, Carnegie Mellon University, and Visiting Scholar, University
of California, Berkeley
Mixing Modality and Probability
Orlov first [1928] and Gödel later [1933] pointed out the connection between the Lewis System S4 and Intuitionistic Logic. McKinsey and Tarski gave an algebraic formulation and proved completeness theorems for propositional systems using as models topological spaces with the interior operator corresponding to the necessitation modality. Earlier, Tarski and Stone had each shown that the lattice of open subsets of a topological space models intuitionistic propositional logic. Expanding on a suggestion of Mostowski about interpreting quantifiers, Rasiowa and Sikorski used the topological models to model many first-order logics. After the advent of Solovay’s recasting of Cohen’s independence proofs as using Boolean-valued models, topological models for modal higher-order logic were studied by Gallin and others. For Boolean-valued logic, the complete Boolean algebra M = Meas([0,1])/Null of measurable subsets of the unit interval modulo sets of measure zero gives every proposition a probability. Perhaps not as well known is the observation that the measure algebra also carries a nontrivial S4 modality defined with the aid of the sublattice Open([0,1])/Null of open sets modulo null sets. This sublattice is closed under arbitrary joins and finite meets in the measure algebra, but it is not the whole of the measure algebra. By working by analogy to the construction of Boolean-valued models for ZF, we can construct over M a model for a modal ZF (MZF) where membership and equality predicates have interesting and natural modal properties. In such a universe the real numbers correspond to random variables, and – following a suggestion of Alex Simpson (Edinburgh) – there is also a well-motivated modeling of random reals. A modal set theory, however, requires a reexamination of comprehension principles, and much work remains to be done to organize methods of proof to take account of the new distinctions encountered. It is also possible to recast well-known theorems of Ergodic Theory as principles about this modal universe, and the question to be considered is whether the new perspective can lead to some new results.
Logic Colloquium
September 11, 2009, 4:10 PM
60 Evans Hall
Theodore A. Slaman
Professor and
Chair of Mathematics, University of California, Berkeley
Degree Invariant Functions
We will discuss the issues surrounding Martin’s conjectured characterization of the functions on reals which are invariant with respect to Turing degree. We will give a recent application, obtained jointly with Montalban and Reimann, to the question of whether Turing equivalence is universal among countable Borel equivalence relations.
Logic Colloquium
August 28, 2009, 4:10 PM
60 Evans Hall
Paolo Mancosu
Professor of
Philosophy, University of California, Berkeley
Measuring the size of infinite collections of natural numbers: Was Cantor’s theory of infinite number inevitable?
Cantor’s theory of cardinal numbers offers a way to generalize arithmetic from finite sets to infinite sets using the notion of one-to-one correspondence between two sets. As is well known, all countably infinite sets have the same ‘size’ in this account, namely that of the cardinality of the natural numbers. However, throughout the history of reflections on infinity another powerful intuition has played a major role: if a collection A is properly included in a collection B than the ‘size’ of A should be less than the ‘size’ of B (part-whole principle). This second intuition was not developed mathematically in a satisfactory way until quite recently. In this talk I begin by reviewing the contributions of some thinkers who argued in favor of the assignment of different sizes to infinite collections of natural numbers. Then, I present some recent mathematical developments that generalize the part-whole principle to infinite sets in a coherent fashion. Finally, I show how these new developments are important for a proper evaluation of a number of positions in philosophy of mathematics which argue either for the inevitability of the Cantorian notion of infinite number (Gödel) or for the rational nature of the Cantorian generalization as opposed to that, offered by Bolzano, based on the part-whole principle (Kitcher).
Alfred Tarski Lectures
April 10, 2009, 4:10 PM
60 Evans Hall
Anand Pillay
Professor of
Mathematical Logic, University of Leeds, Swanlund Chair and Professor
Emeritus, University of Illinois at Urbana-Champaign
Measures and Domination
We introduce Keisler measures on definable sets, and in suitable contexts generalize uniqeness of Haar measure on compact groups, to definable groups, using the Vapnik-Chervonenkis theorem from learning theory/probability theory.
Alfred Tarski Lectures
April 08, 2009, 4:10 PM
TBA
Anand Pillay
Professor of
Mathematical Logic, University of Leeds, Swanlund Chair and Professor
Emeritus, University of Illinois at Urbana-Champaign
Lie Groups from Nonstandard Models
We discuss various kinds of standard part or reduction maps in the light of the logic topology, and relate model-theoretic standard part maps for definable groups to algebraic geometric reduction maps in the case of elliptic curves.
Alfred Tarski Lectures
April 06, 2009, 4:10 PM
60 Evans Hall
Anand Pillay
Professor of
Mathematical Logic, University of Leeds, Swanlund Chair and Professor
Emeritus, University of Illinois at Urbana-Champaign
The Logic Topology
We introduce first-order theories, concentrating on their “type spaces”. We discuss various notions of definable set, and define the “logic topology” on suitable definable sets.