“Peak Data” or “The Capacity Crunch” [From Science]

From this week’s issue of Science comes Filling the Light Pipe by David J. Richardson. This is an important article highlighting a serious challenge facing the both the scientific and policy community.

From the abstract: “It has been a landmark year for the field of optical telecommunications, with the award of the 2009 Nobel Prize to Charles Kao for his insight in the mid-1960s that the future of communications lay in single-mode silica-based optical fibers (1) as well as the 50th anniversary of the first demonstration of the laser—both key technologies responsible for the development of the global-scale communications networks of today (2). Recently, however, a growing realization has emerged within the telecommunications industry that the end of the phenomenal growth in optical fiber communication capacity is within sight. At this year’s Optical Fiber Communication Conference (OFC 2010), several groups reported results within a factor of ~2 of the ultimate capacity limits of existing optical fiber technology. Without radical innovation in our physical network infrastructure—that is, improvements in the key physical properties of transmission fibers and the optical amplifiers that we rely on to transmit data over long distances—we face what has been widely referred to as a “capacity crunch” that could severely constrain future Internet growth, as well as having social and political ramifications.”

[HT to Paul Kedrosky]

How Long is the Coastline of the Law: Additional Thoughts on the Fractal Nature of Legal Systems

Fractal Nature of Legal Systems

Do legal systems have physical properties? Considered in the aggregate, do the distinctions upon distinctions developed by common law judges self-organize in a manner that can be said to have definable physical property (at least at a broad level of abstraction)? The answer might lie in fractal geometry.

Fractal geometry was developed in a set of classic papers by mathematician Benoît Mandelbrot. The original paper in the field How Long is the Coastline of Britain describes the coastline measurement problem.  In short form, the length of the coast line is a function of the size of measurement one employs.  As shown below, as the unit of measurement decreases the length of the coastline increases.  The ideas expressed in this and subsequent papers have been applied to a wide class of substantive questions. In particular, the application to economic systems has been particularly illuminating. Given recent economic events, we agree with views of the Everyday Economist arguing the applied economic theory built upon his work should earn Mandelbrot a share of the Nobel Prize.

Coastline of BritainA more abstract fractal is the simple version of the Sierpinski triangle displayed at the top of this post. Here, there exists self similarity at all levels. Specifically, at each iteration of the model, the triangles at the tip of each of the lines replicate into self similar versions of the original triangle. If you click on the visual above, you can run the applet (provided you have java installed on your computer). {Side note: those of you NKS Wolfram fans out there will know the Sierpinski triangle can be generated using cellular automata Rule 90.}

For those who are interested in another demonstration consider the Koch Snowflake — a fractal which also offers a view of the relevant properties.  The Koch Snowflake is a curve with infinite length (i.e. there is no convergence even though it is located in a bounded region around the original triangle).  Click here to view an online demo of the Koch Snowflake.

So, you might be wondering … what is the law analog to fractals? As a first-order description of one important dynamic of the common law, we believe significant progress can be made by considering the conditions under which legal systems behave in a manner similar to fractals. For those interested, a number of important papers have discussed the fractal nature of legal systems.  While discussing legal argumentation, the original idea is outlined in two important early papers The Crystalline Structure of Legal Thought and  The Promise of Legal Semiotics both by Jack Balkin.  The empirical case began more than ten years ago in the important paper How Long is the Coastline of the Law? Thoughts on the Fractal Nature of Legal Systems by David G. Post & Michael B. Eisen. It continues in more recent scholarship such as The Web of the Law by Thomas Smith.

In our view, the utility of this research is not to adjudicate the common law to be a fractal. Indeed, there exist mechanisms which likely prevent legal systems from actually behaving as unbounded fractal.  The purpose of the discussion is determine whether describing law as a fractal is a reasonable first-order description of at least one dynamic within this complex adaptive system. While full adjudication of these questions is still an area of active research, we highlight these ideas for their important potential contribution to positive legal theory.

One thing we want to flag is the important relationship between the power law distributions we discussed in these prior posts (here and here) and the original work of  Benoît Mandelbrot. The mapping of the power law like properties displayed by the common law and its constitutive institutions is part of the larger empirical case for the fractal nature of legal systems. Building upon the prior work, in two recent papers, which are available on SSRN here and here, we mapped this property of self organization among two sets of legal elites — judges and law professors.

Twitter Mood Predicts the Stock Market

From the physics arXiv comes the interesting paper entitled Twitter mood predicts the stock market. Mike has additional information on the paper over at ETF Central.  However — for those who might be interested — here is the abstract:

“Behavioral economics tells us that emotions can profoundly affect individual behavior and decision-making. Does this also apply to societies at large, i.e., can societies experience mood states that affect their collective decision making? By extension is the public mood correlated or even predictive of economic indicators? Here we investigate whether measurements of collective mood states derived from large-scale Twitter feeds are correlated to the value of the Dow Jones Industrial Average (DJIA) over time. We analyze the text content of daily Twitter feeds by two mood tracking tools, namely OpinionFinder that measures positive vs. negative mood and Google-Profile of Mood States (GPOMS) that measures mood in terms of 6 dimensions (Calm, Alert, Sure, Vital, Kind, and Happy). We cross-validate the resulting mood time series by comparing their ability to detect the public’s response to the presidential election and Thanksgiving day in 2008. A Granger causality analysis and a Self-Organizing Fuzzy Neural Network are then used to investigate the hypothesis that public mood states, as measured by the OpinionFinder and GPOMS mood time series, are predictive of changes in DJIA closing values. Our results indicate that the accuracy of DJIA predictions can be significantly improved by the inclusion of specific public mood dimensions but not others. We find an accuracy of 87.6% in predicting the daily up and down changes in the closing values of the DJIA and a reduction of the Mean Average Percentage Error by more than 6%.”

Measuring the Complexity of the Law : The United States Code [Repost]

Understanding the sources of complexity in legal systems is a matter long considered by legal commentators. In tackling the question, scholars have applied various approaches including descriptive, theoretical and, in some cases, empirical analysis. The list is long but would certainly include work such as Long & Swingen (1987), Schuck (1992), White (1992), Kaplow (1995), Epstein (1997), Kades (1997), Wright (2000), Holz (2007) and Bourcier & Mazzega (2007). Notwithstanding the significant contributions made by these and other scholars, we argue that an extensive empirical inquiry into the complexity of the law still remains to be undertaken.

While certainly just a slice of the broader legal universe, the United States Code represents a substantively important body of law familiar to both legal scholars and laypersons. In published form, the Code spans many volumes. Those volumes feature hundreds of thousands of provisions and tens of millions of words. The United States Code is obviously complicated, however, measuring its size and complexity has proven be non-trivial.

In our paper entitled, A Mathematical Approach to the Study of the United States Code we hope to contribute to the effort by formalizing the United States Code as a mathematical object with ahierarchical structure, a citation network and an associated text function that projects language onto specific vertices.

In the visualization above, Figure (a) is the full United States Code visualized to the section level. In other words, each ring is a layer of a hierarchical tree that halts at the section level. Of course, many sections feature a variety of nested sub-sections, etc. For example, the well known 26 U.S.C. 501(c)(3) is only shown above at the depth of Section 501.  If we added all of these layers there would simply be additional rings. For those interested in the visualization of specific Titles of the United States Code … we have previously created fully zoomable visualizations of Title 17 (Copyright)Title 11 (Bankruptcy), Title 26 (Tax) [at section depth]Title 26 (Tax) [Capital Gains & Losses] as well as specific pieces of legislation such as the original Health Care Bill — HR 3962.

In the visualization above, Figure (b) combines this hierarchical structure together with a citation network.  We have previously visualized the United States Code citation network and have a working paper entitled Properties of the United States Code Citation Network. Figure (b) is thus a realization of the full United States Code through the section level.

With this representation in place, it is possible to measure the size of the Code using its various structural features such as vertices V and its edges E.  It is possible to measure the full Code at various time snapshots and consider whether the Code is growing or shrinking. Using a limited window of data, we observe growth not only in the size of the code but also its network of dependancies (i.e. its citation network).

Of course, growth in the size United States Code alone is not necessarily analogous to an increase in complexity.  Indeed, while we believe in general the size of the code tends to contribute to “complexity,” some additional measures are needed.  Thus, our paper features structural measurements such as number of sections, section sizes, etc.

In addition, we apply the well known Shannon Entropy measure (borrowed from Information Theory) to evaluate the “complexity” of the message passing / language contained therein. Shannon Entropy has a long intellectual history and has been used as a measure of complexity by many scholars.  Here is the formula for Shannon entropy:

For those interested in reviewing the full paper, it is forthcoming in Physica A: Statistical Mechanics and its Applications. For those not familiar, Physica A is a journal published by Elsevier and is a popular outlet for Econophysics and Quantitative Finance. A current draft of the paper is available on the SSRN and the physics arXiv.

We are currently working on a follow up paper that is longer, more detailed and designed for a general audience.  Even if you have little or no interest in the analysis of the United States Code, we hope principles such as entropy, structure, etc. will prove useful in the measurement of other classes of legal documents including contracts, treaties, administrative regulations, etc.

Sebastian Seung: I Am My Connectome [ TED 2010 ]

Sebastian Seung is mapping a massively ambitious new model of the brain that focuses on the connections between each neuron. He calls it our “connectome,” and it’s as individual as our genome — and understanding it could open a new way to understand our brains and our minds.

Public Resource.org / Law.gov Win 10^100 Google Project Funding

Congratulations to Carl Malamud and publicresource.org for being selected as a Google 10^100 project winner! Among other things, the $2 million grant “will support the Law.Gov initiative, which aims to make all primary legal materials in the United States available to all.” I had the pleasure of presenting at the UC-Boulder Law.gov meeting and Texas Law.gov meeting earlier this year. Thus, I am very excited about Google’s decision to fund this very worthy project. For those who are interested in reading more, O’Reilly Radar has additional information on the law.gov / google 10^100 announcement here.

Law as a Complex Adaptive System: An Updated Reading List / Syllabus

As a new semester is here at Michigan CSCS, I have made several revisions to the content of our global reading list for the Computational Legal Studies Working Group. The content of this interdisciplinary reading list features work from economics, physics, sociology, biology, computer science, political science, public policy, theoretical and empirical legal studies and applied math. I wanted to highlight this reading list for anyone who is interesting in learning more about the state of the literature in this interdisciplinary space.  Also, for those interested in learning model implementation, please consult my my slides from the 2010 ICPSR Course Introduction to Computing for Complex Systems. Feel free to email me if you have any questions.

9 Weeks to Go — House and Senate Control as Measured by the Iowa Electronic Market

With nine weeks to go before the 2010 Midterm Elections, it is worth checking in with Iowa Electronic Markets to see where things stand. “The IEM 2010 Congressional Election Markets are real-money futures markets where contract payoffs will be determined by the votes cast in the 2010 U.S. Congressional Elections. “Congress10” (plotted above) is based on the composition of both houses of Congress.”

Take a look at the plot above. You will notice there has been significant movement in the past few weeks. Consistent with the beliefs of a number of pundits, the dominant scenario for 2010 is split control “RH_DS10” (i.e Republican House and Democrat Senate). Whether you view this outcome as good or bad, it is important to emphasize there is still time left and these trends could reverse.