Inspiration: Essential Magic or a Load of Hooey?

Ah, Sweet Panic!

panic!Cartoonist Bill Watterson cranked out one brilliant Calvin and Hobbes comic strip after another for about ten years. Even if (bizarrely) you aren’t a fan of Calvin and Hobbes, it’s clear Watterson knew how to create art that spoke to a lot of people in a clever, funny, and meaningful way. Here’s a conversation his two main characters  had about inspiration.

HOBBES: Do you have an idea for your project yet?

CALVIN: No, I’m waiting for inspiration. You can’t just turn on creativity like a faucet. You have to be in the right mood.

HOBBES: What mood is that?

CALVIN: Last-minute panic.

Continue reading Inspiration: Essential Magic or a Load of Hooey?

Cosmological constant not optimal after all

Interesting news from that weird and wonderful intellectual space where physics and theology trade slow, dignified blows; new research into the effects of varying the cosmological constant swings out like a haymaker from the atheist corner and knocks at least one God-of-the-gaps out of the ring [via SlashDot].

… although positive, the cosmological constant is tiny, some 122 orders of magnitude smaller than Planck’s constant, which itself is a small number.

So Page and others have examined the effects of changing this constant. It’s straightforward to show that if the the constant were any larger, matter would not form into galaxies and stars meaning that life could not form, at least not in the form we know it,.

So what value of the cosmological constant best encourages galaxy and star formation, and therefore the evolution of life? Page says that a slightly negative value of the constant would maximise this process. And since life is some small fraction of the amount of matter in galaxies, then this is the value that an omnipotent being would choose.

In fact, he says that any positive value of the constant would tend to decrease the fraction of matter that forms into galaxies, reducing the amount available for life.

Therefore the measured value of the cosmological constant, which is positive, is evidence against the idea that the constants have been fine-tuned for life.

I guess the obvious theist retort would be that God’s ineffable decision to use a sub-optimal value for the constant is a test of our faith… hi-ho, anthropic principle!

This is my genome. There are many others like it, but this one is mine.

With the increasing difficulty of getting people to actually sign up for military service in the first place, you’d think the Pentagon would make more of an effort to not treat its soldiery as disposable meatbags. Or at least I’d think that… which is one more reason to add to the list of reasons that I’m not a five-star general, I guess.

Aaaaaanyway, here’s the skinny on a Pentagon report that recommends the Department of Defense get some more mileage out of their human resources by collecting and sequencing the DNA of their soldiers en masse [via grinding.be]:

According to the report, the Department of Defense (DoD) and the Veteran’s Administration (VA) “may be uniquely positioned to make great advances in this space. DoD has a large population of possible participants that can provide quality information on phenotype and the necessary DNA samples. The VA has enormous reach-back potential, wherein archived medical records and DNA samples could allow immediate longitudinal studies to be conducted.”

Specifically, the report recommends that the Pentagon begin collecting sequencing soldiers’ DNA for “diagnostic and predictive applications.” It recommends that the military begin seeking correlations between soldiers’ genotypes and phenotypes (outward characteristics) “of relevance to the military” in order to correlate the two. And the report says — without offering details — that both “offensive and defensive military operations” could be affected.

That HuffPo piece leads off with the privacy angle, and wanders onto the more interesting (if potentially nasty) territory of promotional assessment based on genetic factors – a little like like a version of Gattaca where your perfection entitles you to use bigger and better guns. (Or, if you’re lucky, a job in the generals’ tent instead of the trenches.) More interesting still is the news that the DoD already has over 3 million DNA samples on file…

HuffPo being HuffPo, the piece ends with a blustering condemnation of the report:

Soldiers, having signed away many of their rights upon enlistment, should not be used for research that would not otherwise comport with our values, just because they are conveniently available.

Our enormous military establishment is a whole world unto itself, and there is no good reason why that world should depart from the standards that Congress so definitively banned in the rest of the employment world. Congress should prohibit the military from spending money on sequencing individual soldiers’ genomes (without individualized medical or forensic cause) or carrying out large-scale research on soldiers’ DNA.

Yeah, good luck with that. Frankly, I’d have thought a cheaper and more effective option for selecting the optimum soldierly phenotypes would be taking a more honest approach at the recruitment screening phase…

Rebellious robots: how likely is the Terminator scenario?

Via George Dvorsky, Popular Science ponders the possibility of military robots going rogue:

We are surprisingly far along in this radical reordering of the military’s ranks, yet neither the U.S. nor any other country has fashioned anything like a robot doctrine or even a clear policy on military machines. As quickly as countries build these systems, they want to deploy them, says Noel Sharkey, a professor of artificial intelligence and robotics at the University of Sheffield in England: “There’s been absolutely no international discussion. It’s all going forward without anyone talking to one another.” In his recent book Wired for War: The Robotics Revolution and Conflict in the 21st Century, Brookings Institution fellow P.W. Singer argues that robots and remotely operated weapons are transforming wars and the wider world in much the way gunpowder, mechanization and the atomic bomb did in previous generations. But Singer sees significant differences as well. “We’re experiencing Moore’s Law,” he told me, citing the axiom that computer processing power will double every two years, “but we haven’t got past Murphy’s Law.” Robots will come to possess far greater intelligence, with more ability to reason and self- adapt, and they will also of course acquire ever greater destructive power.

[…]

It turns out that it’s easier to design intelligent robots with greater independence than it is to prove that they will always operate safely. The “Technology Horizons” report emphasizes “the relative ease with which autonomous systems can be developed, in contrast to the burden of developing V&V [verification and validation] measures,” and the document affirms that “developing methods for establishing ‘certifiable trust in autonomous systems’ is the single greatest technical barrier that must be overcome to obtain the capability advantages that are achievable by increasing use of autonomous systems.” Ground and flight tests are one method of showing that machines work correctly, but they are expensive and extremely limited in the variables they can check. Software simulations can run through a vast number of scenarios cheaply, but there is no way to know for sure how the literal-minded machine will react when on missions in the messy real world. Daniel Thompson, the technical adviser to the Control Sciences Division at the Air Force research lab, told me that as machine autonomy evolves from autopilot to adaptive flight control and all the way to advanced learning systems, certifying that machines are doing what they’re supposed to becomes much more difficult. “We still need to develop the tools that would allow us to handle this exponential growth,” he says. “What we’re talking about here are things that are very complex.”

Of course, the easiest way to avoid rogue killer robots would be to build less of them.

*tumbleweed*

[$mind]!=[$computer]: why uploading your brain probably won’t happen

Via Science Not Fiction, here’s one Timothy B Lee taking down that cornerstone of Singularitarianism, the uploading of minds to digital substrates. How can we hope to reverse-engineer something that wasn’t engineered in the first place?

You can’t emulate a natural system because natural systems don’t have designers, and therefore weren’t built to conform to any particular mathematical model. Modeling natural systems is much more difficult—indeed, so difficult that we use a different word, “simulation” to describe the process. Creating a simulation of a natural system inherently means means making judgment calls about which aspects of a physical system are the most important. And because there’s no underlying blueprint, these guesses are never perfect: it will always be necessary to leave out some details that affect the behavior of the overall system, which means that simulations are never more than approximately right. Weather simulations, for example, are never going to be able to predict precisely where each raindrop will fall, they only predict general large-scale trends, and only for a limited period of time. This is different than an emulator, which (if implemented well) can be expected to behave exactly like the system it is emulating, for as long as you care to run it.

Hanson’s fundamental mistake is to treat the brain like a human-designed system we could conceivably reverse-engineer rather than a natural system we can only simulate. We may have relatively good models for the operation of nerves, but these models are simplifications, and therefore they will differ in subtle ways from the operation of actual nerves. And these subtle micro-level inaccuracies will snowball into large-scale errors when we try to simulate an entire brain, in precisely the same way that small micro-level imperfections in weather models accumulate to make accurate long-range forecasting inaccurate.

As discussed before, I rather think that mind simulation – much like its related discipline, general artificial intelligence – is one of those things whose possibility will only be resolved by its achievement (or lack thereof). Which, come to think of it, might explain the somewhat theological flavour of the discourse around it…