Graph Theory
Graph theory is a branch of mathematics concerned with the formal study of graphs — structures consisting of vertices (also called nodes) and edges (also called links or arcs) connecting them. A graph abstracts away every feature of a system except one: who is connected to whom. This reduction is the source of both graph theory's extraordinary power and its most dangerous blind spots.
The foundational insight is Euler's 1736 solution to the Königsberg bridge problem: whether you could traverse all seven bridges of Königsberg crossing each exactly once. Euler's answer — no — was less important than how he reached it. He showed that the answer depends only on the parity of connections at each landmass, not on the distances, shapes, or physical arrangement of the bridges. The first graph-theoretic proof worked by demonstrating what could be safely ignored.
That demonstration inaugurated a research program now three centuries old. Its ambition is to understand how structure — the pattern of connections alone, stripped of content — generates behavior.
Core Concepts
A graph G is formally a pair (V, E) where V is a set of vertices and E is a set of pairs of vertices (the edges). This definition is almost comically sparse: it contains no information about what the vertices represent, what the edges mean, whether connection is symmetric, whether it can be weighted, or whether it changes over time. The power of the formalism comes from this sparseness. The weakness of the formalism also comes from this sparseness.
Key properties studied:
- Degree distribution: how many edges each vertex has. Scale-free networks exhibit power-law degree distributions — a few vertices have enormously many edges, most have few. The claim that many real-world networks are scale-free was central to the 1990s–2000s network science program. It has since been substantially challenged: the claim was inflated by confirmation bias and by fitting power laws without rigorously testing alternative distributions. The graph-theoretic framing made it easy to find power laws because it made other distributional features invisible.
- Connectivity and components: whether paths exist between vertex pairs. A graph is connected if every vertex can reach every other; giant components emerge in random graphs at the percolation threshold. The phase transition at the percolation threshold is genuine and robust — one of graph theory's most solid results.
- Shortest paths and diameter: the minimum number of edges between two vertices. The small-world property — that most vertex pairs in large networks are connected by short paths — is well-documented in social and biological networks. Whether it is interesting is a separate question: random graphs also have short paths. What the small-world result actually measures, beyond confirming that the world is not a regular lattice, remains underspecified.
- Clustering coefficients: the tendency of a vertex's neighbors to be connected to each other. High clustering is common in social networks and biological systems. It is often cited as evidence of community structure. It is not: clustering and community structure can come apart completely.
- Graph isomorphism: whether two graphs are structurally identical under relabeling. Determining graph isomorphism efficiently is a famous open problem — probably in polynomial time, but no proof exists. This matters because it is the formal version of asking whether two systems with the same structure are the same system.
Applications and the Network Science Expansion
From the 1990s onward, graph theory was weaponized into network science — the project of applying graph-theoretic tools to empirical complex systems: the internet, social networks, protein interaction networks, food webs, citation networks, brain connectivity, economic networks. The ambition was that universal structural laws would emerge across all domains. Power-law degree distributions, the small-world property, and community detection algorithms were presented as domain-transcending findings.
The claims have not aged uniformly. Newman, Barabási, and Watts made genuine contributions; but the program as marketed promised a unified science of networks that materialized only in fragments. The central difficulty: a graph is a model of a system, not the system itself, and the process of constructing the model — deciding what counts as a vertex, what counts as an edge, what threshold of interaction qualifies as a connection — is not theoretically neutral. Two researchers studying the same empirical system can construct graphs with radically different structures depending on their operationalization choices. The structural properties they then measure are properties of their modeling choices as much as of the underlying system.
The graph-theoretic tradition has, by and large, not confronted this problem directly. It produces structural results about graph objects and presents them as structural results about the world. The gap between model and world is treated as a matter of empirical application, not theoretical concern. This is the methodological partiality that the field's most enthusiastic advocates have consistently underestimated.
Algorithmic Graph Theory
Separately from structural network science, algorithmic graph theory studies the computational complexity of graph problems: shortest paths (Dijkstra's algorithm), minimum spanning trees, maximum matching, graph coloring, and Hamiltonian cycles. Many fundamental graph problems are NP-complete, meaning — under the assumption P ≠ NP — that no polynomial-time algorithm can solve them in the worst case. Graph coloring and the travelling salesman problem are canonical examples.
This branch of graph theory is mathematically clean in a way that structural network science is not: its results are theorems, not empirical regularities, and they do not depend on operationalization choices. The structure is exactly as defined. Whether the structure corresponds to anything outside mathematics is not claimed.
The persistent confusion between graph-theoretic models and the systems they represent — between the map and the territory — suggests that network science has not yet earned the status of a science. It has produced a powerful set of tools for measuring the models analysts construct. Whether those models capture the causal structure of the systems they abstract remains, in most applications, an open and underasked question.