Wednesday, November 28, 2018

More perspective on the atomic age mindset.



In an earlier post, we discussed Willy Ley's observation that, from a 1930s standpoint, a successful moon landing seemed far more of a reach than an atomic bomb, suggesting that the modern usage of "moonshot" – – committing yourself to the an ambitious bordering on impossible objective – – would actually apply better to the Manhattan project.

It's useful at this point to consider just how rapidly this field was advancing.

From Wikipedia (pay close attention to the dates):
In 1932 physicist Ernest Rutherford discovered that when lithium atoms were "split" by protons from a proton accelerator, immense amounts of energy were released in accordance with the principle of mass–energy equivalence. However, he and other nuclear physics pioneers Niels Bohr and Albert Einstein believed harnessing the power of the atom for practical purposes anytime in the near future was unlikely, with Rutherford labeling such expectations "moonshine."

The same year, his doctoral student James Chadwick discovered the neutron, which was immediately recognized as a potential tool for nuclear experimentation because of its lack of an electric charge. Experimentation with bombardment of materials with neutrons led Frédéric and Irène Joliot-Curie to discover induced radioactivity in 1934, which allowed the creation of radium-like elements at much less the price of natural radium. Further work by Enrico Fermi in the 1930s focused on using slow neutrons to increase the effectiveness of induced radioactivity. Experiments bombarding uranium with neutrons led Fermi to believe he had created a new, transuranic element, which was dubbed hesperium.

In 1938, German chemists Otto Hahn and Fritz Strassmann, along with Austrian physicist Lise Meitner and Meitner's nephew, Otto Robert Frisch, conducted experiments with the products of neutron-bombarded uranium, as a means of further investigating Fermi's claims. They determined that the relatively tiny neutron split the nucleus of the massive uranium atoms into two roughly equal pieces, contradicting Fermi. This was an extremely surprising result: all other forms of nuclear decay involved only small changes to the mass of the nucleus, whereas this process—dubbed "fission" as a reference to biology—involved a complete rupture of the nucleus. Numerous scientists, including Leó Szilárd, who was one of the first, recognized that if fission reactions released additional neutrons, a self-sustaining nuclear chain reaction could result. Once this was experimentally confirmed and announced by Frédéric Joliot-Curie in 1939, scientists in many countries (including the United States, the United Kingdom, France, Germany, and the Soviet Union) petitioned their governments for support of nuclear fission research, just on the cusp of World War II, for the development of a nuclear weapon.

First nuclear reactor

In the United States, where Fermi and Szilárd had both emigrated, the discovery of the nuclear chain reaction led to the creation of the first man-made reactor, known as Chicago Pile-1, which achieved criticality on December 2, 1942. This work became part of the Manhattan Project, a massive secret U.S. government military project to make enriched uranium and by building large production reactors to produce (breed) plutonium for use in the first nuclear weapons. The United States would test an atom bomb in July 1945 with the Trinity test, and eventually two such weapons were used in the atomic bombings of Hiroshima and Nagasaki. 


From the perspective of well over a half-century later, the advances in nuclear energy obviously represent a very sharp S curve. At the time, though, there was an entirely natural impulse to extrapolate along a linear or even exponential path.

In August 1945, the first widely distributed account of nuclear energy, in the form of the pocketbook The Atomic Age, discussed the peaceful future uses of nuclear energy and depicted a future where fossil fuels would go unused. Nobel laureate Glenn Seaborg, who later chaired the Atomic Energy Commission, is quoted as saying "there will be nuclear powered earth-to-moon shuttles, nuclear powered artificial hearts, plutonium heated swimming pools for SCUBA divers, and much more".

Tuesday, November 27, 2018

Space exploration is hard.

Yes, I realize that's probably not the most controversial claim I'll make this week, but in this age of hype and bullshit, it's important to occasionally remind ourselves of these basic facts. This is what we go through to put an unoccupied payload roughly the size of a minivan on Mars.


NASA's Mars probe lands Monday after 'seven minutes of terror'


For the eighth time ever, humanity has achieved one of the toughest tasks in the solar system: landing a spacecraft on Mars.

The InSight lander, operated by NASA and built by scientists in the United States, France and Germany, touched down in the vast, red expanse of Mars’ Elysium Planitia just before 3 p.m. Eastern on Monday.

...

The interminable stretch from the moment a spacecraft hits the Martian atmosphere to the second it touches down on the Red Planet’s rusty surface is what scientists call “the seven minutes of terror."

More than half of all missions don’t make it safely to the surface. Because it takes more than eight minutes for light signals to travel 100 million miles to Earth, scientists have no control over the process. All they can do is program the spacecraft with their best technology and wait.

“Every milestone is something that happened 8 minutes ago,” Bridenstine said. “It’s already history.”

The tension was palpable Monday morning in the control room at JPL, where InSight was built and will be operated. At watch parties around the globe — NASA’s headquarters in Washington, the Nasdaq tower in Times Square, the grand hall of the Museum of Sciences and Industry in Paris, a public library in Haines, Alaska — legs jiggled and fingers were crossed as minutes ticked toward the beginning of entry, descent and landing.

At about 11:47 a.m., engineers received a signal indicating that InSight had entered the Martian atmosphere. The spacecraft plummeted to the planet’s surface at a pace of 12,300 mph. Within two minutes, the friction roasted InSight’s heat shield to a blistering 2,700 degrees.

Grover released a deep breath: “That’s hot.”

In another two minutes, a supersonic parachute deployed to help slow down the spacecraft. Radar was powered on.

From there, the most critical descent checklist unfolded at a rapid clip: 15 seconds to separate the heat shield. Ten seconds to deploy the legs. Activate the radar. Jettison the back shell. Fire the retrorockets. Orient for landing.

One of the engineers leaned toward her computer, hands clasped in front of her face, elbows on her desk.

“400 meters,” came a voice over the radio at mission control. “300 meters. 80 meters. 30 meters. Constant velocity."

Engineer Kris Bruvold’s eyes widened. His mouth opened in an “o.” He bounced in his seat.

“Touchdown confirmed.”




Saturday, November 24, 2018

Kevin Drum makes a good point

This is Joseph.

There has been a lot of concern about recent comments by Hillary Clinton about Europe curbing refugee admissions.  Kevin Drum looked at just how many refugees Europe is actually taking and compared it to a reader survey about how many refugees the US should take in:

I don’t want anyone to take my survey too seriously. It’s obviously just a casual thing. However, I think it’s fair to say that the responses are almost entirely from a left-leaning readership, and even at that a solid majority thought the US shouldn’t take in more than half a million refugees in a single year. Adjusted for population, Germany took in nearly ten times that many.
This is a growing problem with mass population displacement.  It strains any system to take in a lot of refugees.  Wanting to be compassionate is very important and we should not allow xenophobia to interfere with saving people who need to be saved.  But it opens up a very important conversation about how one deals with extremely large population displacement and, in a democracy, there may be a limit to the rate that the populace is comfortable with integrating at once.  If climate change drives a longer term issue here, then we need to think about ways to smooth out the process.

Friday, November 23, 2018

Roy Clark and friend

Seems like an appropriate way to kick off the weekend.














Thursday, November 22, 2018

"As God as my witness..." is my second favorite Thanksgiving episode line [Repost]





If you watch this and you could swear you remember Johnny and Mr. Carlson discussing Pink Floyd, you're not imagining things. Hulu uses the DVD edit which cuts out almost all of the copyrighted music. [The original link has gone dead, but I was able to find the relevant clip.]

As for my favorite line, it comes from the Buffy episode "Pangs" and it requires a bit of a set up (which is a pain because it makes it next to impossible to work into a conversation).

Buffy's luckless friend Xander had accidentally violated a native American grave yard and, in addition to freeing a vengeful spirit, was been cursed with all of the diseases Europeans brought to the Americas.

Spike: I just can't take all this mamby-pamby boo-hooing about the bloody Indians.
Willow: Uh, the preferred term is...
Spike: You won. All right? You came in and you killed them and you took their land. That's what conquering nations do. It's what Caesar did, and he's not goin' around saying, "I came, I conquered, I felt really bad about it." The history of the world is not people making friends. You had better weapons, and you massacred them. End of story.
Buffy: Well, I think the Spaniards actually did a lot of - Not that I don't like Spaniards.
Spike: Listen to you. How you gonna fight anyone with that attitude?
Willow: We don't wanna fight anyone.
Buffy: I just wanna have Thanksgiving.
Spike: Heh heh. Yeah... Good luck.
Willow: Well, if we could talk to him...
Spike: You exterminated his race. What could you possibly say that would make him feel better? It's kill or be killed here. Take your bloody pick.
Xander: Maybe it's the syphilis talking, but, some of that made sense.



Wednesday, November 21, 2018

Fifty years late and the Russians are the ones doing it, but otherwise...

That was fortuitous timing. We just ran a post on Willy Ley (circa 1959) discussing the possibility of using nuclear powered rockets for space exploration. Now we get thee following announcement.
Speaking with reporters, Vladimir Koshlakov explained that Elon Musk and SpaceX pose no real threat to the group’s plans. Musk, Koshlakov says, is relying on technology that will soon be antiquated, while Russia is looking towards shaping the future of spaceflight.

The Russian researchers say that their nuclear-powered rocket platform will be able to make it to Mars seven months after launch, and that its reusable rocket stages can be put back into service after just 48 hours.

“Reusability is the priority,” Koshlakov reportedly said. “We must develop engines that do not need to be fine-tuned or repaired more than once every ten flights. Also, 48 hours after the rocket returns from space, it must be ready to go again. This is what the market demands.”

...

“Elon Musk is using the existing tech, developed a long time ago,” he noted. “He is a businessman: he took a solution that was already there, and applied it successfully. Notably, he is also doing his work with help from the government.”

That last paragraph is a bit of Musk-trolling but it's consistent with a point I've heard repeatedly from engineers in the field. While SpaceX has made some serious advances, the underlying tech is decades-old, dating back at least to the lunar lander.

Tuesday, November 20, 2018

The point that no one wants to make about the bad news from Facebook.



This was the year that a lot of people woke up to what Josh Marshall among others had been pointing out for a long time, that while all of the tech giants have accumulated and sometimes abused an extraordinary amount of power, Facebook stood alone as a genuinely bad actor doing a great deal of damage to a wide array of stakeholders.

What's notably absent from all of these analyses is an acknowledgment of the role that the press played in building and maintaining the myths of Zuckerberg and others as tech messiahs. Major news outlets and venerable publications, particularly the New York Times, willingly helped spread the bullshit. We should never forget that when Silicon Valley billionaires went after their toughest (and, in retrospect, most clear eyed) critic, Gawker, the NYT not only failed to stand up for journalism, they actually gave an op-ed spot to Peter Thiel so he could better spin his side of the story.

As you can see, we've been on this beat for a long time.

Wednesday, June 15, 2011

"How To Party Your Way Into a Multi-Million Dollar Facebook Job" -- the sad state of business journalism

Andrew Gelman (before his virtual sabbatical) linked to this fascinating Gawker article by Ryan Tate:

If you want Facebook to spend millions of dollars hiring you, it helps to be a talented engineer, as the New York Times today [18 May 2011] suggests. But it also helps to carouse with Facebook honchos, invite them to your dad's Mediterranean party palace, and get them introduced to your father's venture capital pals, like Sam Lessin did. Lessin is the poster boy for today's Times story on Facebook "talent acquisitions." Facebook spent several million dollars to buy Lessin's drop.io, only to shut it down and put Lessin to work on internal projects. To the Times, Lessin is an example of how "the best talent" fetches tons of money these days. "Engineers are worth half a million to one million," a Facebook executive told the paper.
We'll let you in on a few things the Times left out: Lessin is not an engineer, but a Harvard social studies major and a former Bain consultant. His file-sharing startup drop.io was an also-ran competitor to the much more popular Dropbox, and was funded by a chum from Lessin's very rich childhood. Lessin's wealthy investment banker dad provided Facebook founder Mark Zuckerberg crucial access to venture capitalists in Facebook's early days. And Lessin had made a habit of wining and dining with Facebook executives for years before he finally scored a deal, including at a famous party he threw at his father's vacation home in Cyprus with girlfriend and Wall Street Journal tech reporter Jessica Vascellaro. (Lessin is well connected in media, too.) . . .
To get the full impact, you have to read the original New York Times piece by Miguel Helft. It's an almost perfect example modern business reporting, gushing and wide-eyed, eager to repeat conventional narratives about the next big thing, and showing no interest in digging for the truth.
It is not just that Helft failed to do even the most rudimentary of fact-checking (twenty minutes on Google would have uncovered a number of major holes); it is that he failed to check an unconvincing story that blatantly served the interests of the people telling it.

Let's start with the credibility of the story. While computer science may well be the top deck of the Titanic in this economy, has the industry really been driven to cannibalization by the dearth of talented people? There are certainly plenty of people in related fields with overlapping skill sets who are looking for work and there's no sign that the companies like Facebook are making a big push to mine these rich pools of labor. Nor have I seen any extraordinary efforts to go beyond the standard recruiting practices in comp sci departments.

How about self-interest? From a PR standpoint, this is the kind of story these companies want told. It depicts the people behind these companies as strong and decisive, the kind of leaders you'd want when you expect to encounter a large number of Gordian Knots. When the NYT quotes Zuckerberg saying “Someone who is exceptional in their role is not just a little better than someone who is pretty good. They are 100 times better,” they are helping him build a do-what-it-takes-to-be-the-best image.

The dude-throws-awesome-parties criteria for hiring tends to undermine that image, as does the quid pro quo aspect of Facebook's deals with Lessin's father.

Of course, there's more at stake here than corporate vanity. Tech companies have spent a great deal of time and money trying to persuade Congress that the country must increase the number of H-1Bs we issue in order to have a viable Tech industry. Without getting into the merits of the case (for that you can check out my reply to Noah Smith on the subject), this article proves once again that one easily impressed NYT reporter is worth any number of highly paid K Street lobbyists.

The New York Times is still, for many people, the paper. I've argued before that I didn't feel the paper deserved its reputation, that you can find better journalism and better newspapers out there, but there's no denying that the paper does have a tremendous brand. People believe things they read in the New York Times. It would be nice if the paper looked at this as an obligation to live up to rather than laurels to rest on.

Monday, November 19, 2018

"The Case Against Quantum Computing"


I am approaching this one cautiously both out of concern for confirmation bias and because I know so little about the subject, but this pessimistic take by Mikhail Dyakonov on the short-term prospects of quantum computing raises troubling questions about the coverage of this field and about the way hype undermines the allocation of resources.

The pattern here is disturbingly familiar. We've seen it with AI, fusion reactors, maglev vactrains, subliminal framing, just to name a few. Credulous reporters seek out optimistic sources. Theoretical possibilities are treated as just-around-the-corner developments. Decades of slow progress, false starts, and sometimes outright failure are ignored.

Those who can claim some association with the next big thing are richly rewarded. Entrepreneurs get enormous piles of venture capital. Business lines and academic departments get generous funding. Researchers who can pull off a slick TED Talk get six-figure book deals and fawning celebrity treatment.

Just to be clear, Dyakonov's is not the consensus opinion. Lots of his colleagues are very optimistic, but these concerns do seem to be valid. The fact that almost all of the coverage glosses over that part of the picture tells us something about the state of science journalism.

From The Case Against Quantum Computing [emphasis added]
Quantum computing is all the rage. It seems like hardly a day goes by without some news outlet describing the extraordinary things this technology promises. Most commentators forget, or just gloss over, the fact that people have been working on quantum computing for decades—and without any practical results to show for it.

We’ve been told that quantum computers could “provide breakthroughs in many disciplines, including materials and drug discovery, the optimization of complex manmade systems, and artificial intelligence.” We’ve been assured that quantum computers will “forever alter our economic, industrial, academic, and societal landscape.” We’ve even been told that “the encryption that protects the world’s most sensitive data may soon be broken” by quantum computers. It has gotten to the point where many researchers in various fields of physics feel obliged to justify whatever work they are doing by claiming that it has some relevance to quantum computing.

Meanwhile, government research agencies, academic departments (many of them funded by government agencies), and corporate laboratories are spending billions of dollars a year developing quantum computers. On Wall Street, Morgan Stanley and other financial giants expect quantum computing to mature soon and are keen to figure out how this technology can help them.

It’s become something of a self-perpetuating arms race, with many organizations seemingly staying in the race if only to avoid being left behind. Some of the world’s top technical talent, at places like Google, IBM, and Microsoft, are working hard, and with lavish resources in state-of-the-art laboratories, to realize their vision of a quantum-computing future.

In light of all this, it’s natural to wonder: When will useful quantum computers be constructed? The most optimistic experts estimate it will take 5 to 10 years. More cautious ones predict 20 to 30 years. (Similar predictions have been voiced, by the way, for the last 20 years.) I belong to a tiny minority that answers, “Not in the foreseeable future.” Having spent decades conducting research in quantum and condensed-matter physics, I’ve developed my very pessimistic view. It’s based on an understanding of the gargantuan technical challenges that would have to be overcome to ever make quantum computing work.



In the early 2000s, at the request of the Advanced Research and Development Activity (a funding agency of the U.S. intelligence community that is now part of Intelligence Advanced Research Projects Activity), a team of distinguished experts in quantum information established a road map for quantum computing. It had a goal for 2012 that “requires on the order of 50 physical qubits” and “exercises multiple logical qubits through the full range of operations required for fault-tolerant [quantum computation] in order to perform a simple instance of a relevant quantum algorithm….” It’s now the end of 2018, and that ability has still not been demonstrated.

Friday, November 16, 2018

This John Oliver segment has a way of popping back into the relevant category

In case you haven't been following the news.















Thursday, November 15, 2018

Launching the USS Holland -- my big regret is that I couldn't work in a reference to the dynamite gun


Perhaps even more than the airplane, submarines are the perfect example of how a wave of enabling technologies at the end of the 19th century suddenly made the long dreamed of both possible and practical. Experiments in the field went back literally hundreds of years.

But it wasn't until the last third of the 19th century that a set of four advances – – one revolutionary in the field of naval warfare, the other three revolutionary period – – would make submarines a major military factor. Whitehead torpedoes, Bessemer steel, electric batteries and motors, and internal combustion made the modern version of the craft possible.

The models being developed by most of the major powers around 1900 were, in broad strokes, the same basic configuration as those that would patrol the oceans for more than 50 years until the launch of the Nautilus. There would, of course, be great progress. The subs of World War I would be far more sophisticated than those of 15 years earlier, just as the subs of World War II would surpass those of the previous generation, but the underlying approach would remain fundamentally the same.

The following article, complete with very cool illustrations, comes from Scientific American (December 28, 1901). Just to give you an idea how quickly things were moving at the time, the same issue has two news items on major advances in wireless telegraphy including Marconi's announcement of the first successful transatlantic radio transmission, accepted as authentic by "Mr. Edison" and prompting a cable of congratulations from "Prof. Bell" who graciously offered his house on the coast of Nova Scotia as a site for future experiments.











Wednesday, November 14, 2018

There's still nothing there (and other lessons journalists refuse to learn about Elon Musk)


[See comments]

From Ars Technica

Similarly, Musk told mayors on Thursday that he wants The Boring Company to dig sewers, water transport, and electrical tunnels under cities, in addition to the transportation-focused tunnels he hopes to dig to house electric skate systems.

Musk mentioned this alternate use for his boring machines at the National League of Cities' City Summit, during a "fireside chat" with Los Angeles mayor Eric Garcetti. According to Forbes, Musk told the audience, "The Boring Company is also going to do tunneling for, like, water transport, sewage, electrical. We're not going to turn our noses up at sewage tunnels. We're happy to do that too."

The Boring Company is built on the premise that tunneling technology has not been adequately developed. Musk claims that his boring machines will tunnel faster than the industry's best machines.

Elon Musk is good copy. Perhaps more than anything else, that is the one thing you need to keep reminding yourself of when trying to make sense of why reporters remain so hopelessly credulous on this story. As Upton Sinclair might have told you, the press is remarkably willing to accept dubious claims when they drive traffic and reinforce rather than challenge the standard narrative.

In this case, "reinforce" is far too weak a term. Elon Musk has fashioned himself to personify the cherished tech messiahs narrative. Like Abraham Lincoln in the old Bob Newhart monologue, if he hadn't existed, they would've had to invent him.

Musk is, to his credit, an exceptionally gifted promoter, particularly adept at the art of misdirection. No one is better at distraction, dramatically changing the focus of attention just long enough for goalposts to be moved, promises to be forgotten, and "I'll address that later" to become "we've already covered that."

These distractions are nested. Less like a "real life Tony Stark" and more like a modern day Scheherazade, Elon Musk tells stories within stories, constantly shifting back and forth so that all but the most careful and critical listener will lose the thread and get swept up in the fantasy. When it becomes increasingly obvious that Tesla is unlikely to ever justify its stock price, he announces that construction will soon begin on a long-distance maglev vactrains running along the East Coast. When the buzz fades from that, he very publicly launches a company that claims to be able to increase tunneling speed and decreased costs by an order of magnitude. When the lack of actual breakthroughs start to become noticeable, he releases cool CGI videos of giant slot cars racing underneath Los Angeles.

A key part of this magic show is the ability to make the ordinary seem wondrous. People have been digging tunnels for thousands of years and there is no reason at this point for us to believe that the excavation which is about to be announced with such fanfare employed methods in any way more sophisticated than those used on construction projects around the world.

The press has become so docile on this point that Musk doesn't even have to lie about having made some major advance in the technology. He can just pretend that the enormous superiority of his system was a proven fact, confident in the assumption that no reporter will point out the truth. As far as I can tell (and I've read all that I had time and stomach for), the few specifics he has provided have been either meaninglessly vague (blah blah blah automation blah blah blah) or have displayed a fundamental lack of understanding about engineering and infrastructure (making projects cheaper by making tunnels smaller in situations where the reduction in capacity would actually drive up costs).

Tuesday, November 13, 2018

Another reason to have mixed feelings about the franchise model

From Bloomberg:
All’s fair in the bitter, protracted war between 7-Eleven and its franchisees. The tensions have built steadily in the years since DePinto, a West Point-educated veteran, took charge and began demanding more of franchisees—more inventory, more money, more adherence in matters large and small. Some franchisees have responded by organizing and complaining and sometimes suing.

As detailed in a series of lawsuits and court cases, the company has plotted for much of DePinto’s tenure to purge certain underperformers and troublemakers. It’s targeted store owners and spent millions on an investigative force to go after them. The corporate investigators have used tactics including tailing franchisees in unmarked vehicles, planting hidden cameras and listening devices, and deploying a surveillance van disguised as a plumber’s truck. The company has also given the names of franchisees to the government, which in some cases has led immigration authorities to inspect their stores, according to three officials with Homeland Security Investigations, which like ICE is under the jurisdiction of the Department of Homeland Security.

Monday, November 12, 2018

Some cool old pictures to start the week


From Scientific American 1867/03/30















Friday, November 9, 2018

If Elon Musk had a radium drill, he could really go to town

Not a good movie, but a sometimes interesting look at attitudes toward the future in the first part of the 20th Century, based on a popular 1913 book. It's worth noting that advances in radiation and (more importantly) metallurgy -- particularly the development of Bessemer steel -- had been major parts of the late 19th Century spike of progress.




From Wikipedia:

A group of wealthy industrialists gather in the home of Mr. Lloyd, a millionaire who introduces them to Richard "Mack" McAllan, the engineer who successfully spearheaded the construction of the Channel Tunnel (the story takes place in the unspecified near future, though it is noted in the film that the Channel Tunnel is built "in 1940"). McAllan informs the group that the "Allanite steel" he developed, along with a "radium drill" developed by his friend Frederick "Robbie" Robbins, makes it possible to construct an undersea tunnel linking England with the United States. Though the group is initially sceptical, the backing of Lloyd and his associate Mostyn convinces the group to buy shares in the project.











Thursday, November 8, 2018

A few points on Willy Ley and "the Conquest of Space"

To understand the 21st century narrative around technology and progress, you need to go back to two eras of extraordinary advances, the late 19th/early 20th centuries and the postwar era. Virtually all of the frameworks, assumptions, imagery, language, and iconography we use to discuss and think about the future can be traced back to these two periods.

The essential popularizer of science in the latter era was Willy Ley. In terms of influence and popularity, it is difficult to think of a comparable figure. Carl Sagan and Neil Degrasse Tyson hold somewhat analogous positions, but neither can claim anywhere near the impact. When you add in Ley's close association with Werner von Braun, it is entirely reasonable to use his books as indicators of what serious people in the field of aerospace were thinking at the time. The excerpt below comes with a 1949 copyright and gives us an excellent idea of what seemed feasible 70 years ago.

There is a lot to digest here, but I want to highlight two points in particular.

First is the widespread assumption at the time that atomic energy would play a comparable role in the remainder of the 20th century to that of hydrocarbons in the previous century and a half, certainly for power generation and large-scale transportation. Keep in mind that it took a mere decade to go from Hiroshima to the launch of the Nautilus and there was serious research (including limited prototypes) into nuclear powered aircraft. Even if fusion reactors remained out of reach, a world where all large vehicles were powered by the atom seemed, if anything, likely.

Second, check out Ley's description of the less sophisticated, non-atomic option and compare it to the actual approach taken by the Apollo program 20 years later.

I think we have reversed the symbolic meaning of a Manhattan project and a moonshot. The former has come to mean a large, focus, and dedicated commitment to rapidly addressing a challenging but solvable problem. The second has come to mean trying to do something so fantastic it seems impossible. The reality was largely the opposite. Building an atomic bomb was an incredible goal that required significant advances in our understanding of the underlying scientific principles. Getting to the moon was mainly a question of committing ourselves to spending a nontrivial chunk of our GDP on an undertaking that was hugely ambitious in terms of scale but which relied on technology that was already well-established by the beginning of the Sixties.

________________________________________________

The conquest of space by Willy Ley 1949
Page 48.

In general, however, the moon messenger [and unmanned test rocket designed to crash land on the moon – – MP] is close enough to present technological accomplishments so that its design and construction are possible without any major inventions. Its realization is essentially a question of hard work and money.

The manned moonship is a different story. The performance expected of it is, naturally, that it take off from the earth, go to the moon, land, takeoff from the moon, and return to earth. And that, considering known chemical fuels and customary design and construction methods, is beyond our present ability. But while the moon ship can make a round-trip is unattainable with chemical fuels, a moon ship which can land on the moon with a fuel supply insufficient for the return is a remote possibility. The point here is that one more attention of the step principle is possible three ships which landed might have enough fuel left among them for one to make the return trip.

This, of course, involves great risk, since the failure of one ship would doom them all. Probably the manned moon ship will have to be postponed until there is an orbital nation. Take off from the station, instead of from the ground, would require only an additional 2 mi./s, so that the total works out to about 7 mi./s, instead of the 12 mi./s mentioned on page 44.

Then, of course, there is the possibility of using atomic energy. If some 15 years ago, a skeptical audience had been polled as to which of the two "impossibilities" – – moon ship and large scale controlled-release of atomic energy – – they considered less fantastic, the poll would probably have been 100% in favor of the moon ship. As history turned out, atomic energy came first, and it is now permissible to speculate whether the one may not be the key to the other.

So far, unfortunately, we only know that elements like uranium, plutonium, etc., contain enough energy for the job. We also know that this energy is not completely accessible, that it can be released. He can't even be released in two ways, either fast in the form of a superexplosion, or slowly in a so-called "pile" where the energy appears mainly as he. But we don't know how to apply these phenomena to rocket propulsion. Obviously the fissionable matter should not form the exhaust; there should be an additional reactant, a substance which is thrown out: plain water, perhaps, which would appear as skiing, possibly even split up into its component atoms of hydrogen and oxygen, or perhaps peroxide.

The "how" is still to be discovered, but it will probably be based on the principle of using eight fissionable element's energy for the ejection of a relatively inert reactant. It may be that, when that problem has been solved, we will find a parallel to the problem of pumps in an ordinary liquid fuel rocket. When liquid fuel rockets were still small – – that was only about 17 years ago and I remember the vividly – – the fuels were forced into the rocket motor by pressurizing the whole fuel tank. But everybody knew then that this would not do for all time to come. The tank that had to stand the feeding pressure had to have strong walls. Consequently it was heavy. Consequently the mass ratio could not be I. The idea then was that the tank be only strong enough to hold the fuels, in the matter of the gasoline tank of a car or truck or an airplane, and that the feeding pressure should be furnished by a pop. Of course the pump had to weigh less than the saving in tank wall weight which they brought about. Obviously there was a minimum size and weight for a good home, and if that minimum weight was rather large, a rocket with pumps would have to be a big rocket.

It happened just that way. Efficient pumps were large and heavy and the rocket with pumps was the 46 foot the two. The "atomic motor" for rockets may also turn out to be large, the smallest really reliable and efficient model may be a compact little 7 ton unit. This would make for a large rocket – – but the size of a vehicle is no obstacle if you have the power to move it. Whatever the exhaust velocity, it will be high – – an expectation of 5 mi./s may be conservative. With such an exhaust velocity the mass ratio of the moon ship would be 11:1; with an exhaust velocity of 10 mi./s the mass ratio would drop .3:1!

The moon ship shown in the paintings of the second illustration section is based on the assumption of a mass ratio of this order of magnitude, which in turn is based on the assumption of an atomic rocket motor.

Naturally there would be some trouble with radioactivity in an atomic propelled rocket. But that is not quite as hard to handle as the radioactivity which would accompany atomic energy propulsion under different circumstances. A seagoing vessel propelled by time and energy could probably be built right now. It would operate by means of an atomic pile running at the center high enough to burden and water steam. The steam would drive a turbine, which would be coupled to the ships propeller. While all this mechanism would be reasonably small and light as ship engines go, it would have to be encased in many tons of concrete to shield the ships company against the radiation that would escape from the pile and from the water and the skiing the coolant. For a spaceship, no all-around shielding needed, only a single layer, separating the pilot's or crew's cabin in the nose from the rest of the ship. On the ground a ship which had grown "hot" through service would be placed inside a shielding structure, something like a massive concrete walls, open at the top. That would provide complete shielding or the public, but a shielding that the ship would not have to carry.
The problem that may be more difficult to handle is that of the radioactivity of the exhaust. A mood ship taking off with Lee behind a radioactive patch, caused by the ground/. Most likely that radioactivity would not last very long, but it would be a temporary danger spot. Obviously moon ship for some time to come will begin their journeys from desolate places. Of course they might take off by means of booster units producing nothing more dangerous in their exhaust them water vapor, carbon dioxide, and maybe a sulfurous smell.

Wednesday, November 7, 2018

Transportation cost overruns are nothing new

From Scientific American 1896








Tuesday, November 6, 2018

Monday, November 5, 2018

This nearly century and a quarter old discussion about rapid transit has a remarkably contemporary feel to it, starting with the phrase "rapid transit."

I always assumed it was a 20th Century term, but...



It's this paragraph, however, that struck me as particularly modern:











Friday, November 2, 2018

You should be concerned about the quality of the polls, but it's likely voter models that should worry you the most.

I've been meaning to do a good, substantial, well reasoned piece on fundamental misunderstandings about political polling. This is not that post. Things have been, let us say, busy of late and I don't have time to get this right, but I do need to get it written. I really want to get this one out in the first five days of November.

So here's the short version.

When the vast majority of journalists (even most data journalists) talk about polls being wrong, they tend to screw up the discussion on at least two levels. First because they do not grasp the distinction between data and model and second because they don't understand how either is likely to go kerplooie (okay, how would you spell it?).

The term "polls of registered voters" describes more or less raw data. A complete and detailed discussion would at this point mention weighting, stratification, and other topics but – – as previously mentioned – – this is not one of those discussions. For now, we will treat those numbers you see in the paper as summary statistics of the data.

Of course, lots of things can go wrong in the collecting. Sadly, most journalists are only aware of the least worrisome issue, sampling error. Far more troubling are inaccurate/dishonest responses and, even more importantly, nonrepresentative samples (a topic we have looked into at some depth earlier). For most reporters, "inside the margin of error" translates to "revealed word of God" and when this misunderstanding leads to disaster, they conclude that "the polls were wrong."

The term "likely voter" brings in an entirely different concept, one which is generally even less well understood by the people covering it because now we are talking not just about data, but about models. [Quick caveat: all of my experience with survey data and response models has been on the corporate side. I'm working under the assumption that the same basic approaches are being used here, but you should always consult your physician or political scientist before embarking on prognostications of your own.]

First off, it's worth noting that the very designation of "likely" is arbitrary. A model has been produced that attempts to predict the likelihood that a given individual will vote in an upcoming election, but the cut off between likely and unlikely is simply a number that the people in the field decided was reasonable. There's nothing scientific, let alone magical about it.

Far more important, particularly in the upcoming election, is the idea of range of data. Certain concepts somehow managed to be both painfully obvious and frequently forgotten. Perhaps the best example in statistics is that a model only describes the relationships found in the sample. When we try to extrapolate beyond the range of data, we can only hope that the relationships will continue to hold.

By their very nature, this is always a problem with predictive modeling, but it becomes a reason for skepticism bordering on panic when the variables you included in or perhaps more to the point, left out of your model start taking on values far in excess of anything you saw on the sample. 2018 appears to be a perfect example.

Will the relationships we've seen in the past hold? If not, will the shift favor the Democrats? The  Republicans? Or will the relationships break down in such a way that they cancel each other out? I have no intention of speculating. What I am saying is that we are currently so far out of the range of data on so many factors that I'm not sure it makes sense to talk about likely voters at all.

Thursday, November 1, 2018

Our regular repost on drinking from the wrong pipe

From Josh Marshall:

I managed to involve myself this weekend in a tiny eddy in the storm around the Pittsburgh synagogue massacre. As you can see below, early yesterday evening I happened upon this interview on Lou Dobbs’ Fox Business News show in which a guest, Chris Farrell, claimed the migrant caravan in southern Mexico was being funded and directed by the “Soros-occupied State Department.” This is, as I explained, straight out of The Protocols of the Elders of Zion, the foundational anti-Semitic tract, first circulated and perhaps authored by the Czarist secret police in the first years of the 20th century.

If you’re not familiar with this world, “ZOG” is a staple of white supremacist and neo-Nazi literature and websites. It stands for “Zionist Occupied Government” and is a shorthand for the belief that Jews secretly control the US government. Chris Farrell’s phrasing was no accident. All of this is straight out of the most rancid anti-Semitic propaganda. Rob Bowers, the shooter in the Pittsburgh massacre, appears to have been specifically inspired by this conspiracy theory. Indeed, Bowers had also reposted references to “ZOG” on his social media accounts.


All of the conspiracy theories around the caravan, particularly those involving George Soros and voter fraud, have a weird underwear gnomes quality to them. They make emotional sense for those deep in the conservative media bubble, but there's no way to make any kind of plausible argument for any of them.

It can be useful for the Republican Party if certain segments of the population believe these fantasies, even disseminate them as long as the discussion remains far enough on the recognized fringe to allow party leaders plausible deniability. It is not useful to have ranking politicians and influential conservative voices saying these things out loud on what are supposed to be respectable outlets.

Or as we said exactly two years ago...

Tuesday, November 1, 2016

In retrospect, it's surprising we don't use more sewage metaphors

A few stray thoughts on the proper flow of information (and misinformation) and a functional organization.

I know we've been through all of this stuff about Leo Strauss and the conservative movement before so I'm not going to drag this out into great detail except to reiterate that if you want to have a functional  institution that makes extensive use of internal misinformation, you have to make sure things move in the right direction.

With misinformation systems as with plumbing, when the flow starts going the wrong way, the results are seldom pretty. This has been a problem for the GOP for at least a few years now. A number of people in positions of authority, (particularly in the tea party wing) have bought into notions that were probably intended simply to keep the cannon-fodder happy. This may also partly explain the internal polling fiasco at the Romney campaign.

As always, though, it is Trump who takes things to a new level. We now have a Republican nominee who uses the fringier parts of the Twitter verse as briefings.

From Josh Marshall:


Here's what he said ...
Wikileaks also shows how John Podesta rigged the polls by oversampling democrats, a voter suppression technique. That's happening to me all the time. When the polls are even, when they leave them alone and do them properly, I'm leading. But you see these polls where they're polling democrats. How is Trump doing? Oh, he's down. They're polling democrats. The system is corrupt, rigged and broken. And we're going to change it. [ Cheers and applause ]
Thank you, thank you. In an e-mail podesta says he wants oversamples for our polling in order to maximize what we get out of our media polling. It's called voter suppression because people will say, oh, gee, Trump's down. Folks, we're winning. We're winning. We're winning. These thieves and crook, the immediate, yeah not all of it, not all of it, but much of it -- they're the most crooked -- they're almost as crooked as Hillary. They may even be more crooked than Hillary because without the media, she would be nothing.
Now this immediately this grabbed my attention because over the weekend I was flabbergasted to see this tweet being shared around the Trumposphere on Twitter.
I don't know who Taylor Egly is. But he has 250,000 followers - so he has a big megaphone on Twitter. This tweet and this new meme is a bracing example of just how many of the "scoops" from the Podesta emails are based on people simply not knowing what words mean.
Trump had already mentioned 'over-sampling' earlier. But here he's tying it specifically to the Podesta emails released by Wikileaks. This tweet above is unquestionably what he's referring to.
There are several levels of nonsense here. Let me try to run through them.
...

 More importantly, what Tom Matzzie is talking about is the campaign/DNC's own polls. Campaigns do extensive, very high quality polling to understand the state of the race and devise strategies for winning. These are not public polls. So they can't affect media polls and they can't have anything to do with voter suppression.

Now you may be asking, why would the Democrats skew their own internal polls? Well, they're not.
The biggest thing here is what the word 'oversampling' means. Both public and private pollsters will often over-sample a particular demographic group to get statistically significant data on that group.
...  You need to get an 'over-sample' to get solid numbers.

Whether it's public or private pollsters, the 'over-sample' is never included in the 'topline' number. So if you get 4 times the number of African-American voters as you got in a regular sample, those numbers don't all go into the mix for the total poll. They're segmented out. The whole thing basically amounts to zooming in on one group to find out more about them. To do so, to zoom in, you need to 'over-sample' their group as what amounts to a break-out portion of the poll.

What it all comes down to is that you're talking about a polling concept the Trumpers don't seem to understand (or are relying on supporters not understanding), about polls that are by definition secret (campaign polls aren't shared) and about an election eight years ago.

Wednesday, October 31, 2018

Halloween Greetings form College Humor

"The Internet Goes Trick-or-Treating"


Tuesday, October 30, 2018

We can't finish October without playing this at least once

The Danse Macabre (from the French language), also called the Dance of Death, is an artistic genre of allegory of the Late Middle Ages on the universality of death: no matter one's station in life, the Dance Macabre unites all.

The Danse Macabre consists of the dead or a personification of death summoning representatives from all walks of life to dance along to the grave, typically with a pope, emperor, king, child, and laborer. They were produced as mementos mori, to remind people of the fragility of their lives and how vain were the glories of earthly life. Its origins are postulated from illustrated sermon texts; the earliest recorded visual scheme was a now-lost mural at Holy Innocents' Cemetery in Paris dating from 1424 to 1425.



Monday, October 29, 2018

I think it's important to define subsidized journalism more broadly than just native advertising.

[This started out as a reply to Andrew Gelman's post "The Axios Turing test and the heat death of the journalistic universe," but when you break 500 words...]

First, there is the age-old problem of advertisers rewarding/punishing publications. If memory serves, Politico back in the Mike Allen days was notorious for questionable ethics along these lines.

Far more subtle and dangerous is the quid pro quo associated with access. While most news organizations have rules in place to prevent editorial influence from advertisers and generally avoid even the appearance of impropriety, giving favorable coverage to sources (often even to the extent of letting them set the narrative and distort the facts) is so widespread that many journalists don't even see the ethical problem. Politico and Axios both have bad reputations in this regard, but the worst offender may well be the New York Times.

This is both a carrot and a stick process. When Disney found itself the target of an LA Times exposé about its dealings with the city of Anaheim, they responded by publicly cutting off the paper's access to the studio's talent, even though that was an entirely different department of the paper dealing with entirely different Disney business lines. The good news is that, now that Disney has become by far the biggest and most powerful entertainment company in the world with the acquisition of Fox, I'm sure they will be much less inclined to abuse their position.

Particularly in fields such as entertainment, companies often go beyond merely providing journalists with the raw material and actually provide the stories themselves. Sometimes this is done by sending out press releases that can be repackaged as features with a minimum of work. Other times you have what can only be described as ghostwriting. Someone with the studio or one of the PR firms it employs will send a reporter an email about an upcoming project. The next day it will appear almost verbatim under the reporter's byline.

In addition to supplying money and content, companies can also provide an even more valuable service as promotion partners. That exclusive you just published about a new movie will get far more traffic because of the hundred million dollars the studio has spent on marketing the project, not to mention the effects of the SEO push and the social media blitz. If you write a story that the studios want to promote, you can literally see millions of PR dollars spent on getting you eyeballs.

While none of this is by any means new, Netflix and to a lesser extent the other streaming services have pushed things to a new level, spending billions on marketing and PR, even green lighting hundred million dollar documentaries and art-house projects for no other apparent reason than that they will generate lots of coverage and might snag a few awards. This unprecedented amount of money has distorted the narrative to such an extent that it is impossible to gauge the cultural impact or commercial viability of Netflix, but most of the journalists covering this story (virtually all of those on the East Coast) remain oblivious to this aspect, perhaps because ignorance of this particular detail makes their lives much easier.

Friday, October 26, 2018

More spooky stuff (In no way chosen as filler because I had a busy week)


Who better than Goldsmith and Hermann to send us off.
























Thursday, October 25, 2018

A Mercury Theatre Halloween

[repost]

The debut production of the Mercury Theatre of the Air, Dracula.




And, of course, the Mercury production of War of the Worlds.



While we're at it, here's a tour de force from Welles' favorite, Agnes Moorehead (don't let the corny intro turn you off) Sorry, Wrong Number.




Wednesday, October 24, 2018

We'll largely skip over the author's fixation on casual nudity.



When trying to understand attitudes toward and expectations for technology in previous eras, one needs to be especially cautious about relying on science fiction writers. Though they initially appear to be the richest source on the subject, numerous factors combine to make them unreliable. Most were attracted to the field by a fascination with the more dramatic aspects of the future and these tendencies are greatly heightened by working in a genre where the object was often to out geewhiz the next guy.

That said, it's a mistake to ignore them entirely, both because of the close relationship between the scientific and the science-fiction community, and because of the influence SF has played on the way we think about technology today either directly through books, film, television and indirectly through writers who alternated between science fact and science fiction (Asimov, Clarke, and to a degree, Willy Ley and Carl Sagan).

This mid century essay by Robert Heinlein on his predictions for the year 2000 is worth a look for a number of reasons. First, people did tend to take the man seriously in the postwar era. Though his standing has arguably declined somewhat at least relative to contemporaries like Asimov, at his peak, he was the best known and best respected hard science fiction writer among mainstream audiences. (Bradberry also had significant mainstream following, but even when writing about spaceships and aliens, his work tended to fall more in the category of fantasy).

Second, this essay is of particular value because the author not only makes a great number of detailed predictions (including a notable amount of time spent on the appeal of socially acceptable nudity), he also explicitly spells out the assumptions that underlie much of the period's attitudes toward the future. He even states his axioms and provides a handy graph of human advancement.




As a serious attempt at describing the rate of progress, this picture is fatally flawed. The year 1900 came at the end of a huge technological and scientific spike. Extending it back a couple of decades would have completely thrown off the curve. (Interestingly, you actually can justify an exponential curve describing progress in the 19th century.) Furthermore, it is difficult to argue a steady acceleration from the naughts to the teens, the teens to the 20s, and the 20s to the 30s.

This graph, however, is tremendously revealing when it comes to the ways people in the 1950s thought about progress. Like the end of the 19th century, the postwar era was a period when conditions lined up to cause a number of very steep S curves to cluster together. The result was a time of explosive, ubiquitous change. There was also, as mentioned before, a tendency to look at the two world wars and the interval between (particularly the Great Depression) as anomalous. It was natural for people in the postwar era to see themselves as living on an exponential slope that was on the verge of shooting past the comprehensible.

Tuesday, October 23, 2018

Cult of the CEO

This is Joseph

This is a revealing symptom of the cult of the CEO:
If you have a CEO this dead to rights on securities fraud, why let him continue as CEO? According to the SEC, Musk was indispensable. In a statement, SEC Chair Jay Clayton said “holding individuals accountable is important and an effective means of deterrence,” but that he must take the interests of investors into account, and “the skills and support of certain individuals may be important to the future success of a company.”
One of the most pernicious myths is that of the irreplaceable man.  We know that nobody is really irreplaceable, because in the end we are all replaced by the natural force of mortality.  But it is a terrible sign of a society when it sees the importance of a person to a business enterprise as an excuse for leniency for poor conduct.  The pressure to cheat to reach the top has to be high because the stakes are so incredibly meaningful in terms of wealth and status.  That suggests more scrutiny, not less. 

Irreplaceable men often beget disasters.  Great leaders, like Napoleon, often get cocky and make grave mistakes that end up costing a great deal despite the attributes that made them successful for a long time. 

I think that this line of thinking isn't ideal. 

Monday, October 22, 2018

It is always useful to go back and read the contemporary accounts.





One of the nails I've pounded flush to the board recently (apologies to long-suffering regular readers) is that much of the standard 21st-century narrative of technology consists of things that were at best sometimes true in the past and are almost entirely false now. The best example is probably the idea that the advances of old invariably came as a thief in the night with almost no one imagining the magnitude of their impact and what now seem obvious applications going undiscovered for years.



There are, of course, technological developments that caught people off guard or that moved in unexpected directions, but in most cases, if you go back and read early speculations about the potential of breakthrough technologies in the late 19th/early 20th centuries or the postwar era, you'll generally find that people had a pretty good sense of what was likely to come.

The same can be said for the dawn of the personal computing era




.






Friday, October 19, 2018

In case the aerospace allusions are getting a bit obscure, here's a week in video recommendation.




The Mouse on the Moon is an easy film to overlook. Between Peter Sellers' spectacular turn in the Mouse that Roared and the general tendency of the time to look at sequels as second-class cinematic citizens, particularly when none of the original stars made a return appearance), it is easy to think of the 1963 film as "the other one."

That's too bad, because the second film can easily hold its own. It's sharp and funny and like its predecessor. Both get 3 1/2 stars in the Leonard Maltin guide. What's more, it features the direction of a young Richard Lester just before he broke through with Hard Days Night.

From Wikipedia:

The Mouse on the Moon is a 1963 British comedy film, the sequel to The Mouse That Roared. It is an adaptation of the 1962 novel The Mouse on the Moon by Irish author Leonard Wibberley, and was directed by Richard Lester. In it, the people of the Duchy of Grand Fenwick, a microstate in Europe, attempt space flight using wine as a propellant. It satirises the space race, Cold War and politics.




Thursday, October 18, 2018


Continuing the visionary aerospace thread, there's almost a "mouse on the moon" quality to India's avatar.

No disrespect meant for India here. Quite the opposite. I think there's long been a tendency to underestimate the country and its extraordinary intellectual capital. Here's one of the projects I would definitely keep an eye on.

From Wikipedia:
:
The idea is to develop a spaceplane vehicle that can take off from conventional airfields. Its liquid air cycle engine would collect air in the atmosphere on the way up, liquefy it, separate oxygen and store it on board for subsequent flight beyond the atmosphere. The Avatar, a reusable launch vehicle, was first announced in May 1998 at the Aero India 98 exhibition held at Bangalore. 
Avatar seems to have, if you'll pardon the metaphor, stalled out (recent tests don't seem to involve any of the really cutting-edge stuff). It could be that the technology actually has hit a wall. That would hardly be surprising for something this ambitious. There's another possibility, however, that is both more encouraging and depressing at the same time, namely that it simply hasn't gotten the funding it needs. Depressing because that would mean we have unnecessarily delayed important advances. Encouraging because it suggests that we still might get this plane flying.

There's a lot of money floating around out there in the vanity aerospace industry and it would be nice to see it go to something ambitious and important. With all due respect to the recently departed, if Paul Allen had taken the money spent on 60 year old visions of space travel and poured it into something forward thinking, his greater legacy might've been what he did after Microsoft.

Wednesday, October 17, 2018

It also suggests that I chose the wrong major in college.

This long but well worth your time article approaches some familiar problems from a new and disturbing direction. We've talked a great deal about manipulation of data and research, the corrupting influence of big money on supposedly objective processes, and the tendency to equate overly complicated math with insight and profundity, but probably not nearly enough about how these things can undermine public policy decision-making.

Here are a few excerpts, but you should really read the whole thing.
These Professors Make More Than a Thousand Bucks an Hour… — ProPublica
Jesse Eisinger,Justin Elliott

If the government ends up approving the $85 billion AT&T-Time Warner merger, credit won’t necessarily belong to the executives, bankers, lawyers, and lobbyists pushing for the deal. More likely, it will be due to the professors.

A serial acquirer, AT&T must persuade the government to allow every major deal. Again and again, the company has relied on economists from America’s top universities to make its case before the Justice Department or the Federal Trade Commission. Moonlighting for a consulting firm named Compass Lexecon, they represented AT&T when it bought Centennial, DirecTV, and Leap Wireless; and when it tried unsuccessfully to absorb T-Mobile. And now AT&T and Time Warner have hired three top Compass Lexecon economists to counter criticism that the giant deal would harm consumers and concentrate too much media power in one company.

Today, “in front of the government, in many cases the most important advocate is the economist and lawyers come second,” said James Denvir, an antitrust lawyer at Boies, Schiller.

Economists who specialize in antitrust — affiliated with Chicago, Harvard, Princeton, the University of California, Berkeley, and other prestigious universities — reshaped their field through scholarly work showing that mergers create efficiencies of scale that benefit consumers. But they reap their most lucrative paydays by lending their academic authority to mergers their corporate clients propose. Corporate lawyers hire them from Compass Lexecon and half a dozen other firms to sway the government by documenting that a merger won’t be “anti-competitive”: in other words, that it won’t raise retail prices, stifle innovation, or restrict product offerings. Their optimistic forecasts, though, often turn out to be wrong, and the mergers they champion may be hurting the economy.

Some of the professors earn more than top partners at major law firms. Dennis Carlton, a self-effacing economist at the University of Chicago’s Booth School of Business and one of Compass Lexecon’s experts on the AT&T-Time Warner merger, charges at least $1,350 an hour. In his career, he has made about $100 million, including equity stakes and non-compete payments, ProPublica estimates. Carlton has written reports or testified in favor of dozens of mergers, including those between AT&T-SBC Communications and Comcast-Time Warner, and three airline deals: United-Continental, Southwest-Airtran, and American-US Airways.

American industry is more highly concentrated than at any time since the gilded age. Need a pharmacy? Americans have two main choices. A plane ticket? Four major airlines. They have four choices to buy cell phone service. Soon one company will sell more than a quarter of the quaffs of beer around the world.

Mergers peaked last year at $2 trillion in the U.S. The top 50 companies in a majority of American industries gained share between 1997 and 2012, and “competition may be decreasing in many economic sectors,” President Obama’s Council of Economic Advisers warned in April.

While the impact of this wave of mergers is much debated, prominent economists such as Lawrence Summers and Joseph Stiglitz suggest that it is one important reason why, even as corporate profits hit records, economic growth is slow, wages are stagnant, business formation is halting, and productivity is lagging. “Only the monopoly-power story can convincingly account” for high business profits and low corporate investment, Summers wrote earlier this year.

...


These complex mathematical formulations carry weight with the government because they purport to be objective. But a ProPublica examination of several marquee deals found that economists sometimes salt away inconvenient data in footnotes and suppress negative findings, stretching the standards of intellectual honesty to promote their clients’ interests.


 …

Recent research supports the classic view that large mergers, by reducing competition, hurt consumers. The 2008 merger between Miller and Coors spurred “an abrupt increase” in beer prices, an academic analysis found this year. In the most comprehensive review of the academic literature, Northeastern economist John Kwoka studied the effects of thousands of mergers. Prices on average increased by more than 4 percent. Prices rose on more than 60 percent of the products and those increases averaged almost 9 percent. “Enforcers clear too many harmful mergers,” American University’s Jonathan Baker, a Compass economist who has consulted for both corporations and the government, wrote in 2015.

Once a merger is approved, nobody studies whether the consultants’ predictions were on the mark. The Department of Justice and the Federal Trade Commission do not make available the reports that justify mergers, and those documents cannot be obtained through public records requests. Sometimes the companies file the expert reports with the courts, but judges usually agree to companies’ requests to seal the documents. After a merger is cleared, the government no longer has access to the companies’ proprietary data on their pricing.

The expert reports “are not public so only the government can check,” said Ashenfelter, the Princeton economist who has consulted for both government and private industry. “And the government no longer has the data so they can’t check.” How accurate are the experts? “The answer is no one knows and no one wants to find out.”

Tuesday, October 16, 2018

There's a lot of bold, visionary thinking coming out of aerospace, just not from the people who are supposed to be the bold visionaries.





Recently, we've been making the point that most of the "futuristic" and "revolutionary" proposals coming from the billionaire Messiah class (dominated by but not limited to the Silicon Valley variety) are usually postwar vintage and, even when you put aside those that probably won't work at all (see the Hyperloop), represent at best incremental advances. This is especially true in the vanity aerospace industry.

In making that point, I may have given the wrong impression about the aerospace industry as a whole. I'll try to post a couple more examples later. For now though, here is one technology that, if it proves viable and it is looking very promising, holds tremendous potential to revolutionize spaceflight.

From Wikipedia:

Like the RB545, the SABRE design is neither a conventional rocket engine nor jet engine, but a hybrid that uses air from the environment at low speeds/altitudes, and stored liquid oxygen (LOX) at higher altitude. The SABRE engine "relies on a heat exchanger capable of cooling incoming air to −150 °C (−238 °F), to provide oxygen for mixing with hydrogen and provide jet thrust during atmospheric flight before switching to tanked liquid oxygen when in space." 

And from the Guardian:


Spaceplanes are what engineers call single-stage-to-orbit (if you really want to geek out, just use the abbreviation: SSTO). They have long been a dream because they would be fully reusable, taking off and landing from a traditional runway.

By building reusable spaceplanes, the cost of reaching orbit could be reduced to a twentieth current levels. That makes spaceplanes a game changer both for taking astronauts into space and for deploying satellites and space probes.

If all goes to plan, the first test flights could happen in 2019, and Skylon – Reaction Engines' spaceplane – could be visiting the International Space Station by 2022. It will carry 15 tonnes of cargo on each trip. That's almost twice the amount of cargo that the European Space Agency's ATV vehicle can carry.
...

Rockets are cumbersome because not only must they carry fuel, they also need an oxidising agent to make it burn. This is usually oxygen, which is stored as a liquid in separate tanks. Spaceplanes do away with the need for carrying most of the oxidiser by using air from the atmosphere during the initial stages of their flight.

This is how a traditional jet engine works, and making a super-efficient version has been engineer Alan Bond's goal for decades. In 1989, he founded Reaction Engines and has painstakingly developed the Sabre engine, which stands for Synergetic Air-Breathing Rocket Engines.

In late 2012, tests managed by the European Space Agency showed that the key pieces of technology needed for Sabre worked. No one else has managed to successfully develop such a technology.
...

Spaceplanes should not to be confused with space tourism vehicles such as Virgin Galactic's Space Ship Two. The highest altitude this vehicle will reach is about 110km, giving passengers about six minutes of weightlessness as the craft plummets back to Earth before the controlled landing.

Although there is no fully agreed definition, space starts at around 100km in altitude. To have any hope of staying in orbit, you would have to reach twice that altitude. The International Space Station orbits at 340 kilometres, whereas the Hubble Space Telescope sits at 595 kilometres.