Out of the Storm News
The June 2 edition of the journal Pediatrics features a study of e-cigarette advertising on television from 2011 to 2013. The research, led by Jennifer Duke of RTI International in North Carolina, found that exposure among youth (12-17 years old) to television ads for e-cigs increased 256 percent over that period.
The author casts the industry as villain: “It appears that youth are being exposed to a sustained level of marketing about the benefits of e-cigarettes.” ABC News was less subtle, with a piece titled: “E-cigarette TV ads target kids.”
In fact, the results in Figure 1 of the study (left), which were not fully described by the authors, show that the advertisements’ effects were mainly on adults. The authors report exposure to e-cig ads in terms of target rating points (TRPs), a standard unit of measurement for the proportion of people exposed to an ad and its frequency.
Duke and colleagues report that exposure of 12-to-17-year-olds to e-cig advertising peaked in the second quarter 2013 at 347 TRPs. Young adults (18-to-24 years old), for whom purchase of e-cigarettes is legal almost everywhere, had peak exposure of 611 TRPs in 2013, indicating much higher exposure than youth.
The authors fail to note that older adults had far higher peak levels of exposure to e-cig ads. My table contains the estimated peak TRPs for each age group in Figure 1.Peak exposure to e-cigarette television advertising (TRPs) by age groups in Q2 2013 Age Group (years) Population (millions) Peak Exposure (TRPs) 12-17 24 347 18-24 30 611 55-34 41 611 35-54 86 820 55+ 77 850
This data shows that 234 million adults 18 years or older were the primary recipients of e-cigarette advertising.
Back in 2012, Kyle Bartlow and Mariah Gentry were juniors at the University of Washington’s Foster School of Business, looking to live out the American dream of entrepreneurship. Together, they formed JoeyBra LLC, a company dedicated to making and marketing specially designed women’s undergarments that include pockets on the wings (see the image above.) As the pair describes the item on their website:
Inspired by UW’s vibrant Greek system, JoeyBra was created for women who are constantly on the go and struggle to find a place to put their ID, keys, or phones. From their own personal experience, they know that women hate taking purses to dances, bars, or dance clubs. Leaving these items at home can pose a safety risk, but with JoeyBra women will never have to worry losing or damaging their valuables again.
Things got off to raring start for JoeyBra. A successful KickStarter campaign raised more than $10,000 in seed funding, which was supplemented by another $5,000 award when the company was named a finalist in the Foster School’s annual business plan competition.
Alas, like many fledgling entrepreneurs, Bartlow and Gentry soon discovered the road to riches was bound to have a few potholes. Theirs came in the personage of Mr. Charles Robinson, a British man who in 2001 was granted a patent (D448541) by USPTO for the ornamental design of a brassiere that, similarly, included pockets on the wings. An illustration of his design is below:
Robinson brought an infringement suit in Virginia against both JoeyBra LLC, and Bartlow and Gentry as individuals, seeking a preliminary injunction to bar them from marketing their wares.
Robinson’s basis for seeking an injunction was that it would cause him “irreparable harm,” even though he had never actually brought a product to market in the dozen years that he held the patent. In March 2013, U.S. District Court Judge Norman K. Moon denied the injunction, in part because Robinson could show no material harm, but also because, Moon wrote, “a brief patent search reveals that a pocketed bra is, in fact, not a ‘very novel element.’”
[T]hough the pocket on Defendants’ JoeyBra product is, like Plaintiff’s claimed design, on the side of the bra (rather than the cup), the size, orientation, and accessibility of that feature appear to be substantially different; as a consequence, and more significantly, the carrying capacity and overall functionality of the allegedly infringing product also appear substantially different. The pocket on Plaintiff’s design appears to be fit for a key, and after twelve years since receiving his patent, he does not have a product on the market. Defendants’ JoeyBra product, on the other hand, holds an iPhone and credit cards, among other items.
Moon also granted Bartlow and Gentry a motion to dismiss the claims against them personally, on jurisdictional grounds, though he would allow the case against JoeyBra to continue (three of the KickStarter funders lived in Virginia, which was enough to establish Robinson’s jurisdictional claim.)
Robinson filed a motion to reconsider, which Moon denied in May 2013. Following that ruling, there was no further action in the case until it was ultimately dismissed (but without prejudice) in March 2014 for “failure to prosecute.”
Having spent two crucial years of their young entrepreneurial career fending off this spurious claim, Bartlow and Gentry attempted to recover their legal fees. Specifically, under Title 35, Section 285 of the U.S. Code, “the court in exceptional cases may award reasonable attorney fees to the prevailing party.”
Alas, in a new order handed down June 3, Judge Moon broke the bad news: just because you won the case doesn’t mean you “prevailed.”
Moon noted that in only one case, Velez v. Portfolio Recovery Assoc., has a federal court determined that a party granted a motion to dismiss on jurisdictional meets the definition of a “prevailing party.”
What’s more, the Fourth Circuit has ruled that a party granted a preliminary injunction is not a “prevailing party,” because consideration of the merits of such cases are “necessarily abbreviated.” The Fourth Circuit has not ruled on whether denial of a preliminary injunction could render one a “prevailing party,” but the Tenth Circuit has ruled that it does not.
Of course, even if JoeyBra and its principals were declared “prevailing parties,” they’d face a whole other hurdle in establishing that their case met the definition of “exceptional.” In a unanimous decision in April in the case Octane Fitness LLC v. ICON Health & Fitness Inc., the U.S. Supreme Court ruled that the exceptionality test for award of attorney fees had set too high a bar, remanding the matter back to the Federal Circuit.
These are, of course, precisely the kinds of issues that were to be addressed by the patent reform bill that Senate Judiciary Committee Chairman Patrick Leahy, D-Vt., has tabled indefinitely. Perhaps one day Congress really will get back to work to resolve them.
Until then, JoeyBra isn’t taking any chances. They’ve applied for a patent of their own.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
R Street Executive Director Andrew Moylan joined David Williams of the Taxpayers Protection Alliance on TPA’s “Taxpayer Watch Podcast” to discuss the June 1 start of the Atlantic hurricane season and the work done by the SmarterSafer Coalition to bring budget watchdogs, insurance companies and environmental activists together for a common cause. You can check it out here or click the embedded player below.
This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
Residents of 11 Utah cities would be billed as much as $20 a month, as part of a plan to salvage the state’s once-heralded UTOPIA fiber optic network.
UTOPIA, short for the Utah Telecommunications Open Infrastructure Agency, was conceived in 2002 as a local government-managed alternative to commercial cable, telco and satellite broadband and has struggled ever since. It lost $19 million in fiscal year 2011. As of late 2012, the agency was $120 million in the red and had fewer than 10,000 customers
While there have been plenty of municipal broadband failures in the past, this may be the first time a government has actually considered reaching directly into consumer pockets to cover the cost of underperformance. Under the proposal, Macquarie Capital, an Australian investment group, would take over construction and operation of the financially troubled UTOPIA project in return for a share of the profits. The catch, however, is that households in the UTOPIA cities would be hit each month with a special “availability fee” of $18 to $20 to pay for the project’s completion, whether they opt to get service or not. This fee would be adjustable each year.
The terms themselves are measly. The “availability fee” would entitle consumers to only three megabytes per second of bandwidth (UTOPIA originally promised one gigabyte per second — 341 times as fast) and a monthly cap of 20 gigabytes, or enough for five or six high-definition Netflix movies. This compares to the 12 to 15 Mb/s available for $50 to $60 a month, with unlimited data, from most cable companies. AT&T’s U-Verse and Verizon’s FiOS offer faster connections.
In return for a 30-year public-private partnership, Macquarie would promise to complete UTOPIA’s build-out to 155,000 total residential and business addresses in 30 months. The network will continue to serve as a wholesale backbone for retail Internet service providers, and Macquarie would aggressively promote the network and extend it to any additional city that wants to accept its terms.
Five UTOPIA cities–Lindon, Murray, Layton, Tremonton and Centerville–have scheduled public meetings in the coming weeks to debate accepting the Macquarie proposal. Another three –Murray, Lindon and Orem–are polling residents to gauge sentiment on key elements of the plan, including its proposed $18 to $20 fee. Brigham City is studying the proposal and Midvale and West Valley City already have accepted it. Macquarie’s deadline is June 27.
Utah’s largest city, Salt Lake, in a prescient decision, opted out of UTOPIA participation.
It’s another unenviable position for cities that went the muni broadband route. The UTOPIA 11 are on the hook for the debt, and the best option means hitting up customers for service that’s inferior to what’s on the market now. At the same time, advocates of free-market solutions, and limiting government activity in the commercial sphere, warned that taxpayers were going to bear the brunt of any muni failures.
Even as this week’s news brings word of a new Google effort to launch satellites that will facilitate rural broadband, progressives still insist that government broadband is needed to respond to “market failure” in broadband provision. Really? It seems the only consistent failure in broadband have been municipal project after municipal project.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
Treating like cases alike is central to the U.S. legal system. Socially and morally, the advantages of juridical consistency are assumed to be foundational. For this reason, among the very first in a line of crucial terms that a first-year law student learns is “stare decicis“, the Latin phrase for “stand by things decided.”
A prime reason that courts “stand by things decided” is that, in cases coming before them, they typically encounter standard iterations of well-defined legal themes. It is a rare day when a truly novel legal or factual issue arises. In practice, stare decicis means that, when a court encounters a case with issues of fact and/or law implicated that it has encountered before, that court will defer to its existing patterns of thought in lieu of starting from scratch. As a matter of judicial economy, stare decisis is invaluable, but it serves other, perhaps more important, purposes: judicial predictability and private coordination.
Lines of precedent are relied upon by private parties as they plan for the future. Plainly, complying with the law is less of a chore when its application is discernible. Businesses, in particular, rely on the consistent application of the law, because they are among the most likely to be targeted for liability when precedent abruptly shifts or moves.
To draw attention to the problem of a moving legal target and other court-related maladies, the U.S. Chamber of Commerce releases an annual list of the nation’s “Judicial Hellholes“. These uncertain jurisdictions, found in red and blue states alike, may not be pure creations of political ideology. Instead, what connects them is that they exist within, and create expansive and amorphous zones of, civil liability. It’s bad enough when poor or ill-considered public policy choices of the electorate and their representatives are the cause. Deplorably, courts are also playing a significant role.
To describe the phenomenon of courts making public policy choices from the bench, commentators have popularized the phrase “judicial activism.” Unfortunately, hyperbole in its use and over-breadth in its application have left the phrase unmoored from easily discernable meaning.
Commendably, Colleen Pero, writing for the American Judicial Partnership, seeks to re-introduce meaning to the phrase by outlining indicia of judicial activism. She concludes that courts are being judicially activist when they deviate from an otherwise predictable outcome in favor of a particular policy outcome. Common examples include the use of non-mandatory authority as the basis of an opinion and the insertion of novel language into a statute to allow the law to support a meaning consistent with a desired outcome.
Courts typically have a degree of freedom to determine how they will interpret statutes. Some methodological flexibility is unavoidable. However, where methodological flexibility significantly reduces consistency and predictability, justice may be distorted. Differing interpretations of different facts may be legitimate, but inconsistent interpretations of similar bodies of law marks the very definition of judicial activism. Among all forms of judicial activism, methodological activism is particularly undesirable from a coordination perspective.
To counteract or defend against such judicial unpredictability, states should be led to embrace systems of methodological stare decicis. Such systems cement into law – for the benefit of courts and the public – interpretive road maps that provide a slew of benefits. Predictability is improved, since parties coming before the court understand what to prioritize in their briefs and what types of arguments the court would find compelling. Stability is improved, because parties can better evaluate what sorts of claims are worthy of pursuing in litigation. Efficiency is improved, because courts are not forced to select from a panoply of interpretive doctrines currently in circulation.
Most importantly, when methodological decisions are given a stare decisis effect, the complexion of the judiciary’s role changes. It shifts away from a law-making task for which it is ill-suited and toward greater predictability in the application of law.
One state, Oregon, conducted an experiment with methodological stare decicis. In PGE v. Bureau of Labor and Industries, the Oregon Supreme Court instructed its courts to adhere to a strict analytical hierarchy to help observers predict when, how and what to bring before Oregon courts. The system directed interpretive action to adhere to a three-step process. First, courts were instructed to take a look to the text of the statute and its surrounding context. Next, if the text contained unresolvable ambiguity, the court was to look to legislative history. Finally, in the event that understanding could not be gleaned from the first two tiers of scrutiny, the court was bound to look to maxims of statutory construction.
The result was that private parties in conflict, the Legislature and lower courts knew broadly what to expect, both in decisional reasoning and, ultimately, outcome. In an article on the topic, Abbe Gluck observed that from 1993 to 1998, the Oregon Supreme Court looked at 137 statutory-interpretation cases, reaching legislative history 33 times and substantive canons only 11 times. From 1999 to 2006, the court applied the statutory-interpretation framework 150 times, reaching legislative history nine times and never reaching the canons of construction. Clearly, a methodological preference for textualism was carrying the day in the period that followed PGE and its progeny.
Naturally, the content of any such interpretive roadmap is crucial. Some interpretive methodologies are preferable to others from a free-market perspective. Still, for the purposes of achieving the benefits of coordination and predictability, there is no need to come to any particular conclusion about the desirability of one system of statutory-interpretation over another, since methodological stare decisis would see the development of methodological precedent. Matters relating to a particular statute would be reliably interpreted according to an attendant methodological precedent.
At its core, methodological stare decisis is an institutional solution to the problem of extreme decisional variations and a reminder that all forms of legal precedent matter. Once an issue is decided – even if it is, from one’s own perspective, decided wrongly – the way that it is decided methodologically becomes ensconced in the body of law, to be referred to and relied upon by others moving forward.
This is not to say that stare decisis is immutable. It is not. Judges consistently revise, rework and reimagine the law as it relates to novel variations in its application. But the limiting function of stare decisis remains essential. At the very least, it serves to stymie rapid transformations that upset the public’s ability to rely upon the law. In essence, it is a procedural mechanism whereby the law expresses a preference for incrementalism.
Society as a whole benefits when risk, even in the form of judicial outcomes, can be assessed and accounted for.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
“Congress could have legislated this problem away many times … but obviously that hasn’t happened,” said Andrew Moylan, executive director of the R Street Institute, a conservative think tank that favors taxing carbon over “extraordinarily expensive” regulations…
…The R Street Institute is supportive. The group has been pitching a national “tax swap” that would create a fee on emissions while cutting tax rates on income, as a way to shape climate policy toward conservative principles.
“So basically a state-level equivalent of the [national] revenue-neutral carbon tax as a substitute for regulation,” Moylan said. “We think that’s the right sort of path forward.”
Kelly Ash is California director at the R Street Institute. She previously worked at the Personal Insurance Federation of California as the communications director and managed the political action committee. Before that, she spent five years working for the California State Assembly as a legislative director for various assembly members.
She was born and raised in San Diego, California.
Andrew Moylan of the R Street Institute, an independent think tank, said that contrary to supporters’ assertions, the MFA does not “level the playing field.”
Moylan said it would impose a “different and unequivocally harsher set of rules than exist for brick-and-mortar” transactions.
“Passage of the MFA would mean states could strong-arm remote sellers into complying with the more than 9,600 separate sales-tax rates that exist across the country, not to mention the 46 states with sales taxes that can issue their own unique set of edicts and definitions,” Moylan testified at a March Judiciary Committee hearing.
But, as Reihan Salam has written, the effects may also be more subtly damaging. As wages rise, businesses could simply seek to hire better educated and skilled employees, some of whom may well live outside the city limits, but suddenly find themselves happy to commute for a fatter paycheck. Seattle’s low-wage workers, meaning those who earn less than $15 an hour, are already more likely to have attended college than their counterparts in cities like Denver and San Francisco. As the pay floor rises, it seems reasonable to suspect that college-educated workers from around the region will take a growing share of jobs that might have once gone to high-school grads.
WASHINGTON (June 2, 2014) – New carbon emissions regulations proposed today by the U.S. Environmental Protection Agency are likely to prove expensive and damaging to a still-fragile economic recovery, according to the R Street Institute.
The regulations, which will be subject to a one-year review period, rely on the spectacularly ill-fitting regime of the Clean Air Act to impose a 30 percent reduction in carbon emissions from electricity generation by 2030. The EPA’s new scheme comes after years of unsuccessful legal and legislative challenges to the agency’s authority.
While climate change is a pressing environmental and economic issue deserving of a public policy response, these overly prescriptive regulations likely will impose enormous costs for relatively modest emissions reductions. R Street has instead urged officials to embrace the power of the free market by utilizing revenue-neutral carbon pricing as a complete substitute for command-and-control regulation.
“Carbon pricing could allow states to kill two birds with one stone,” said Andrew Moylan, executive director of R Street. “They could achieve mandated emissions reductions through a price signal instead of complicated regulation, while utilizing all resulting revenue to eliminate or reduce taxes that are damaging to the economy. This could get EPA off their backs while securing real pro-growth tax reform.”
Moylan also pointed out that Congress’ failure to legislate has helped lead to the imposition of this expensive plan, which the U.S. Chamber of Commerce estimates will cost $51 billion annually. This is simply the latest example of growth in executive power as Congress stands idly by, Moylan said.
“The fact that Congress has proven unable to legislate effectively in this area led directly to the unworkable regulations released today,” Moylan said. “Perhaps now that EPA’s economic threat is more real, Congress will reengage with its duty to guide policies of such enormous consequence.”
Perhaps the surest sign that the decade-long cupcake fad is really and truly over is that even cupcake-related lawsuits are wrapping up tidily.
Most recently, news has come down of a settlement in a design patent infringement suit brought in March 2013 by Kimber Cakeware LLC of Hilliard, Ohio against Rancho Cucamonga, Calif.-based Bradshaw International Inc. Bradshaw is a major homewares manufacturer that markets its devices through a variety of brand names, including Betty Crocker, Dawn, Mr. Clean, Black & Decker and, in this case, Good Cook.
The timeline laid out in the original complaint is somewhat tortured, but as Kimber tells it, the events went down something like this:
In 2009, Robert S. Reiser, who would later serve as one of the principals of Kimber, pitched his batter separator invention – the “Batter Daddy” – to a variety of firms, including Bradshaw. Bradshaw declined the pitch, on grounds that the then-current financial crisis had led them to hold off on developing new products.
In May 2010, Reiser founded Kimber, which would begin marketing an adapted version of the device called “Batter Babies” in December 2010.
In April 2011, Reiser submitted an application for a patent on the ornamental design of his batter separator. The illustration in the design patent application looked like this:
In March 2012, Bradshaw introduced at the International Housewares Show its own product, dubbed the “Sweet Creations by Good Cook cupcake divider.” That product looks like this:
In November 2012, Reiser’s application was approved as design patent D671376.
There is, of course, one major problem, which should be pretty obvious at first glance. Even in the exhibit Kimber included in its complaint, the design of the Good Cook cupcake divider just simply does not look anything like the Kimber batter divider:
Bradshaw felt the case was so weak that not only did they seek summary judgment to dismiss, but they argued in a filing that Kimber should be slapped with Rule 11 sanctions for bringing a frivolous lawsuit without any merit whatsoever.
Bradshaw conceded that, like Kimber’s design, theirs is a batter separator designed to fit inside a standard muffin tin, allowing two different types of batter to be cooked simultaneously. But since the only grounds Kimber would have for infringement is that the design so closely resembled their own as to confuse an ordinary observer, Bradshaw accused Kimber of committing the cardinal sin of design patent suits:
Kimber has wrongly asserted protection for a functional, utilitarian article through its design patent, which is the epitome of what cannot be claimed in a design patent.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
LUS Fiber, the municipal broadband system in Lafayette, La., last month received another warning from city auditors, an advisory that appears to have become an annual thing.
Although losses were anticipated during the initial five years of offering retail services to customers, management should carefully monitor the financial results of operations of the communications system. The projections calculated by operating and finance management should be compared to actual results on a regular basis and appropriate measures should be taken to minimize any significant negative variances. Additionally, management should continue to enhance its market strategy in order to increase its revenue base.
Lafayette’s auditors voiced similar concerns in their reports the last two years. In 2012, they punctuated it with a calculation that the $140-million system was costing the city $45,000 a day.
Now, after six years of operation, prospects aren’t much better. The city’s financial reports, provided by a source in Lafayette, show that for the fiscal year ended Oct. 31, 2013, LUS Fiber reported $23 million in operating revenues, compared to $36.7 million that was forecast in its feasibility study. The system incurred a $2.5 million operating loss for the year. According to the original plan, this was to be the point where the operation swung to a profit of $902,000.
The most staggering number, however, is LUS Fiber’s deficit, which stood at $47 million at the end of October, up from $37.1 million the year before.
LUS officials have been trying to put a smiling face on this by noting that the operation is cash-flow positive, which simply means that LUS Fiber is taking in more than it’s spending on a day-to-day basis, but does not factor in its enormous long-term debt liability. They’ve also tried some sales gimmicks, like doubling the amount of bandwidth capacity for an additional $5 a month. This deal goes for everyone except low-income customers, for whom provision of quality high-speed service was a major justification for LUS Fiber’s creation. They remain stuck with a 3 Mb/s connection, about a quarter to a fifth of the speed you now get from cable.
Officials also continue to assure Lafayette residents that profitability, like a Cubs pennant, is just one more year away, and that there is robust subscriber growth ahead. Yet it’s time to seriously question this proposition in light of broadband business trends. Cable customers have been turning to wireless and video on demand for video programming. Citing data from the ISI Group, an equity research firm, BusinessInsider.com reported that nearly 5 million cable TV subscribers cut the cord in the last five years, and that the number of cable TV-only subscribers remaining could sink below 40 million (see graph below).
True, in many cases you still may want a landline broadband connection to get Netflix or Amazon Prime, but the triple play business model—phone, cable, Internet—looks like it’s toast. Like the cable industry, muni broadband operations like LUS Fiber bet their entire long-term viability on triple play. Phone’s been gone for years. Cable TV is going. Commercial cable companies are scrambling to deal with this market shift, and muni operations are in the same boat. The difference is that, with private companies, the cost falls on shareowners. For municipalities, the cost is paid by taxpayers. In Lafayette, the meter stands at $47 million
What’s distressing is that today, when the turmoil in the service provider market is so measureable and visible, there are still cities and towns—Westminster, Md.; Princeton, Mass.; Vallejo, Calif., to name three I found here—think that municipal broadband is a sound idea.
For an in-depth look at the challenges municipal broadband faces, see my case study of LUS Fiber, downloadable here.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
From the Washington Examiner:
Steven Titch for R Street: The European Union’s highest court has ruled that Internet search engines must give serious consideration to users who request they remove links to any content or article that is personally or professionally unflattering, unfavorable or embarrassing.
This Sunday, June 1, marks the start of the 2014 Atlantic hurricane season. And though experts are predicting a relatively quiet one this year, our dutiful representatives in Congress have done their darnedest to ensure that it would take just one big storm to push a particularly troubled federal agency over the brink of bankruptcy.
The National Oceanic and Atmospheric Administration’s Climate Prediction Center recently unveiled its projection of a near-normal or below-normal hurricane season, with a 70 percent likelihood of eight to 13 named storms, three to six hurricanes and one or two major hurricanes.
Those projections more or less jibe with those released in April by forecasters Philip Klotzbach and William Gray of Colorado State University, who note the impact of a cooled Atlantic and at least a moderate-strength El Niño in their projection of nine named storms, three of which would become hurricanes and one of those would become a major hurricane. Klotzbach and Gray assign a 35 percent chance of a major hurricane making landfall somewhere on the U.S. coastline, compared to 52 percent in an average year.
Of course, it takes just one big storm to knock those living near the coast, as 126 million Americans do, for a major loop. It also would take just one big storm to render the National Flood Insurance Program – the federal agency that writes most flood coverage in the United States – unable to pay its claims.
In fact, the program is only able to pay any claims thanks to billions in loans from federal taxpayers. The NFIP currently owes the Treasury $24 billion, a tally mostly rung up during Hurricane Katrina in 2005 and Hurricane Sandy in 2012. The program hasn’t made a payment against its principal since 2010 and it’s been able to keep up with its interest payments only because interest rates have dropped dramatically in recent years. In 2008, the NFIP paid $730 million in interest to service its then-$18 billion of debt, compared to just $72 million in interest payments in 2011.
Current law allows the program to borrow up to $30.4 billion before it would need once again to go back to Congress, hat in hand. But the bottom line is that the premiums paid in — $3.8 billion last year, on $1.3 trillion of insured property – simply aren’t enough to keep up with claims. A major hurricane hitting any of the significantly populated areas along the Gulf and East coasts could easily cause another $10 billion or so in flood insurance claims, enough to bankrupt the program.
Congress had a plan to fix this mess. Years in the making, both houses ultimately passed – by wide margins and with strong bipartisan support – and President Barack Obama signed legislation in 2012 intended to set the NFIP on the path to solvency. The plan involved phasing out subsidies for the roughly 20 percent of policyholders who were getting them – more quickly, for vacation and business properties, less quickly for primary homes. It also involved updating and redrawing the program’s long-outdated maps, to ensure those properties most at risk were charged the rates they should be.
Alas, it didn’t take long for Congress to think better of its momentary rush into sane, responsible policymaking. The first complaints about rising rates from coastal policyholders started pouring in while the ink was still dry on the reform bill. Eventually the deafening chorus of complaint – emanating, as it did, from states like Louisiana, Arkansas, Georgia and North Carolina, which will prove crucial to control of the U.S. Senate this fall – was sufficiently to prompt action. Earlier this year, Congress set up re-breaking the program they had just attempted to fix, passing a law that, as the non-partisan Government Accountability Office put it, “will address affordability concerns, but may also reduce program revenues and weaken the financial soundness of the NFIP program.”
Entering into any storm season, one always hopes for the best and particularly, for no major loss of life or limb. But should the storm surge flood in and the flood program collapse, you’ll know where to point the finger of blame.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
Fifty-three tobacco research and policy experts from 15 countries today endorsed many of the tobacco harm reduction principles that I have advocated for 20 years. In a widely publicized open letter to Dr. Margaret Chan, director of the World Health Organization, they declared:
Tobacco harm reduction is part of the solution, not part of the problem. It could make a significant contribution to reducing the global burden of non-communicable diseases caused by smoking, and do so much faster than conventional strategies. If regulators treat low-risk nicotine products as traditional tobacco products and seek to reduce their use without recognising their potential as low-risk alternatives to smoking, they are improperly defining them as part of the problem
Just as I have done before, the experts warn that harsh regulation of e-cigarettes could have the unintended effect of protecting cigarettes:
On a precautionary basis, regulators should avoid support for measures that could have the perverse effect of prolonging cigarette consumption. Policies that are excessively restrictive or burdensome on lower-risk products can have the unintended consequence of protecting cigarettes from competition from less-hazardous alternatives, and cause harm as a result. Every policy related to low-risk, non-combustible nicotine products should be assessed for this risk.
The letter’s signatories also endorse a tax strategy that I have promoted for many years:
The tax regime for nicotine products should reflect risk and be organised to create incentives for users to switch from smoking to low-risk, harm-reduction products. Excessive taxation of low-risk products relative to combustible tobacco deters smokers from switching and will cause more smoking and harm than there otherwise would be.
The letter points to the enormous public health gains that are possible with tobacco harm reduction:
The potential for tobacco harm reduction products to reduce the burden of smoking-related disease is very large, and these products could be among the most significant health innovations of the 21st Century – perhaps saving hundreds of millions of lives.
It is encouraging to see such widespread international support for my long-held positions.
The U.S. Chamber of Commerce, a corporate advocacy group, has charged that regulations would cost $51 billion and 224,000 jobs. The National Mining Association ran advertisements predicting a near-doubling of electricity costs. Those claims were quickly rebutted, but they reflect a deeper fear that the EPA will try to micromanage the cuts rather than trusting companies to find their own solutions. That sort of command-and-control approach “is likely to be a costly disaster,” said Andrew Moylan, a senior fellow at R Street, a free market think tank…
…In addition to cap-and-trade, noted Moylan, states and power companies could also consider a carbon tax like British Columbia’s. Set at $30 Canadian dollars per metric ton, and ultimately reflected in the price of gasoline and fuel, it presently raises $1 billion in annual revenue, which is in turn used to fund business and income tax cuts.
We’ll be hosting a Google Hangout with the Mercatus Center at George Mason University, the Electronic Frontier Foundation and the Competitive Enterprise Institute to discuss copyright reform on Thursday, June 5, at 3 p.m. ET.
Copyright has long been a source of division among conservatives and libertarians. While some see creative works the same as any other property, and thus a natural right, others argue that copyright is different than traditional property, and that special interests have bloated the copyright system to the point where innovation is stifled. How much copyright is too much? At what point does cronyism trump innovation?
Panelists Tom W. Bell, author of “Intellectual Privilege,” Derek Khanna of R Street, Mitch Stoltz of EFF and Ryan Radia of CEI will debate these issues. We invite you to join the discussion by asking questions through the Hangout at j.mp/copyrighthangout or using the hashtag #iphangout on Twitter.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
There has been a great deal of heated discussion in North Carolina regarding auto insurance reform during the past several years. The insurance industry is divided and parties are trying to make their case to legislators in the N.C. General Assembly.
It can’t be denied that the average premium for drivers here is ranked as one of the lower in the nation and the market is not in a state of crisis.It also can’t be denied that our regulatory system is unique nationally and the rate bureau model is dated. This approach worked in the 1980s, before advances in technology, risk assessment and online resources. We now have usage-based insurance options that give the driver the ability to install a device that monitors driver habits in real time and allows discounts for driving carefully. Technology enables insurers to more accurately price risk on individual characteristics and reward less risky behavior. This is fairer than charging average rates based on large indiscriminate groupings.
We do not need to completely dismantle the current system, but it does seem reasonable to modify the system so that it makes it easier for insurers to operate and offer discount programs and enhancements commonly used in other states.
As it stands currently, the states with simpler, more modern systems are the first to receive new programs and North Carolina is left for last, or left out. Insurers could be given the opportunity to opt out of the current rate bureau system, provided that the insurance commissioner still has the option to review and approve new programs. This should lead to additional benefits for consumers, as insurers have more flexibility and incentives to compete.
I’ve taught insurance and personal finance courses for two decades – the last 12 at Appalachian State University. Each semester, I assign a project in which the students are required to obtain a quote for their auto insurance. I do not tell them where to get their quote or how to do it; they have to figure that out on their own.
The students are consistently able to complete this assignment and typically have the same general comments each semester. They are surprised how easy it is to get a quote online, over the phone or by visiting an agent. They also ask about commercials they have seen on TV and wonder why they couldn’t enroll in certain programs. “I have good grades, why doesn’t my insurer offer the good student discount in North Carolina?”
I believe the consumer is best served by allowing the commissioner to have authority to review and deny rate changes when appropriate. Standard forms with consistent policy language should continue and also be subject to the commissioner’s approval. There is concern that insurers may try to raise rates, and this could certainly happen for higher-risk drivers. However, if an insurer attempts to raise rates to an unfair level, they will quickly price themselves out of the market, as consumers shop around for the best deal.
I believe many consumers would be better served by revising our system and giving insurers the option to operate as they do in most other states. This would encourage insurers to market additional discount programs and unleash competitive products that would benefit consumers.Readmorehere: http://www.charlotteobserver.com/2014/05/27/4936249/nc-auto-insurance-price-controls.html#.U4dKgijEZOk#storylink=cpy
The annual World Series of Poker got underway in Las Vegas Tuesday. Poker enthusiasts are looking to see if legalization of online poker has an impact on the tournament series, which takes in 65 bracelet events over seven-and-a-half weeks, culminating with the $10,000 Main Event beginning July 5.
If there is a substantial bump in registrations – say, an increase of 1,000 entrants over the 6,352 who played in last year’s Main Event – it would be a strong indicator that online play has caused an uptick of interest in the game.
Nevada was the first state to legalize online gambling, in February 2013, and two of the three current sites got rolling early last summer. The Silver State limits online wagering to poker, but Delaware and New Jersey, the two states that followed shortly thereafter, permit Internet sites that offer the gamut of traditional casino house-banked games, such as blackjack, roulette, slots and video poker.
Overall, however, the first year of online gambling has fallen short of expectations. U.S. online poker revenues for the first 11 months of operation totaled $9.4 million, according to a recent Associated Press report. By comparison, the casinos in Nevada by themselves brought in $982 million in the month of March alone.
Gaming operators point to continued competition from off-shore providers, some of which still take deposits from players in the United States. There has also been the usual tech hiccups with some sites. In addition, Ultimate Poker, the first site to launch in Nevada, made a strategic decision to stick to very low-limit games at the outset —literally nickel-and-dime stakes—to keep professionals off the site. While this allowed newcomers the opportunity to play with less risk, it also suppressed early revenues.
Then there’s billionaire casino mogul Sheldon Adelson, principal owner of the Venetian and Palazzo hotels in Las Vegas, who says he’ll spend “whatever it takes” to stop online gambling. His principal goal is a federal law that would pre-empt state action, but he also aims to stop any state-level legalization.
Nonetheless, Morgan Stanley predicts that by 2020, legal online gambling in the U.S. will generate $8 billion a year. At least eight states have introduced legislation to legalize online poker or full-out house-banked gambling.
So despite the slow start, there is reason for optimism. Momentum for legalization is growing in Pennsylvania and California, although in the latter case, some tribal issues need to be worked out. However, the prevailing feeling in the Golden State is that online poker is an extension of the numerous brick-and-mortar poker rooms that already dot the state.
Here are five more reasons poker players should be patient, but optimistic, about the return of online play:
- The basic fact is that online poker is back.
After a meteoric rise, the last few years the industry has been dinged with shutdowns and (in the case of the Full Tilt Poker site) scandal. But legalization brings the ability to promote and advertise. The NBA’s Philadelphia 76ers and NHL’s New Jersey Devils have signed sponsorship deals with PartyPoker, which operates a site in New Jersey. This means PartyPoker signage in their playing arenas and consequent visibility on TV, as well as extensions into digital space. The pact will build mindshare about online poker in the key young male demographic. As awareness grows, so will online participation. In the meantime, Delaware and Nevada have agreed to a compact that allows residents of both states play on each other’s poker sites. New Jersey, guarding its large population, so far has stayed out. Nonetheless, compacts like these, which derive from multi-state lottery agreements, will help build player pools.
- The Feds made good on Full Tilt repayment.
Although the news didn’t get much play in the mainstream press, the government’s decision to return funds seized from online poker accounts was a huge development in the poker community. The money was seized as part of the Department of Justice’s investigation into Full Tilt Poker and few expected to see that money again. As a result of a complicated agreement (explained here), PokerStars was awarded Full Tilt’s assets under the stipulation that funds held in any accounts be returned to players.
- States need the money.
Call it cynical or realistic, but most states face enough financial pressure that any new source of tax revenue is attractive. Allowing online gambling through Pennsylvania’s casinos could bring in $307 million annually to the state, according to a study commissioned by the Pennsylvania Senate. Since gambling has the air of “vice,” higher percentages can be extracted than would be otherwise feasible with conventional sales taxes (not necessarily fair, but true all the same). Tobacco and alcohol serve as a model.
- Sheldon Adelson is hard to take seriously as a moral crusader.
While it’s one thing to give an ear to passionate critics who believe the social harms of gambling outweigh any fiscal benefits, it’s difficult to accept anti-gambling moralizing from a man who’s made billions operating casinos internationally. Adelson, 80, has made online gambling his personal bête noir, and seems bent on using a considerable part of his casino-generated wealth on winning political favor for a ban on Internet gambling.
Most observers believe there will be no federal legislative action here, despite Adelson’s lobbying. Adelson is in direct opposition to his fellow casino owners, represented by the American Gaming Association, which is pro-legalization. Such intra-industry conflicts favor the status quo, which in this case, tilts toward online.
- Poker is popular.
Finally, people like to play. Poker is recognized as a game of skill, in which satisfaction and profit can be derived from learning and improving on one’s play. Unlike lotteries, which rely on pure chance, or house-banked games, which give the casino a built-in mathematical edge, poker can be a winning proposition for those with the knowledge and self-control required to master the game.
Poker always was a social game, and online play served to introduce millions more to the experience. Its peak moment came when Chris Moneymaker won the 2003 World Series of Poker Main Event, after having played only online to that point. At that point, there was no turning back.
Right now, it’s enough to say that online poker has overcome a coordinated government effort to run it out of the country. Besides poker players, this simple statement should warm the hearts of anyone weary of paternalistic or moralistic governing. Yes, there can indeed be pushback when the government arbitrarily takes away a pastime that a substantial portion of the population enjoys.
Full Disclosure: Titch is an avid poker player and will play in his second $1,500 event in this year’s World Series of Poker.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.
Google has announced a truly “driverless” car that lacks a steering wheel, brakes or anything else that allows a driver to control its movements. This actually strikes me as a much bigger deal for the insurance industry than the optionally self-driving cars that have received so much attention and seem likely to come to market within the next few years. And, while the short-term insurance implications of the type of self-driving cars that are coming to market soon seem pretty modest, the new truly driverless models seem a lot more likely to have big effects.
Let’s start with what’s going to happen soon. Self-driving cars that require a “pilot” in a driver’s seat and have a steering wheel, accelerator and brakes are really just an evolutionary change in vehicles that progressively have become more automated over the last decade. As such, the changes in auto insurance are also going to be evolutionary. All major automakers already sell models with traction control, self-parking, collision avoidance and adaptive cruise control features. Together, these features make a car “half self-driving” already.
The “self-driving car” that will probably be on the market within a few years just adds a steering and navigation functions to this already widely deployed suite of features. Since drivers will still be able to control these cars, they will still need liability insurance. Some claims will become product liability claims—some types of accident claims already are—but most accidents would still result from human error of various kinds. There isn’t going to be a huge change. The insurance business will continue in much the same way for most insurers and most consumers.
There’s one exception: self-driving cars, even in the early generation, will be nigh-impossible for street criminals to steal. Any self-driving car is almost certainly going to be traceable via its GPS system and have a “kill switch.” Protection against theft isn’t a huge part of auto insurance premiums in most places but, in the long term, self-driving cars seem likely to more-or-less eliminate an entire category of crime. (Computer hackers might still steal cars from time-to-time but I’d suspect this is going to be about as difficult as raiding a bank account, which is hard.)
However, the kinds of fully self-driving cars that Google is now testing could represent a much bigger change. If a driver can’t manipulate a car in any way (except maybe to press a “stop” button) most crashes will probably result in product liability claims. Although such product-liability auto insurance could, in principle, be purchased as “master policies” by automakers, the mere fact that we have a longstanding cultural habit of buying auto insurance makes it quite possible that consumers will still buy polices. And, most likely, automakers as well as some insurance and consumer groups will argue for offering these policies on a no-fault basis. This, in turn, could lead to a real resurgence in no-fault coverage that’s fading elsewhere.
Second, under a full-self-driving model a significant fraction of people may well move towards a fractional ownership or “car-sharing subscription” service. Car sharing, of course, is already reasonably widespread in dense urban areas but remains a niche market because it’s not economically efficient to use car-sharing to commute to work, go on a road-trip or, really, do much of anything besides run a brief errand. By contrast, a service that lets you summon cars without drivers to wherever you are and have them take you where you want to go could replace automobiles for many people. One would think these services would probably bundle in some sort of policy in just the same way that existing car-sharing services do.
The near future of self-driving cars probably isn’t a big deal for insurers. The more distant future of fully autonomous cars, however, may well result in big, big changes.This work is licensed under a Creative Commons Attribution-NoDerivs 3.0 Unported License.