Skip to main content
SpectralShifts Blog 
Sunday, February 19 2012

The sports world is a great place to test and prove out management tactics and strategies.  Unlike in the commercial world where reasons for and measurement of product success are highly variable and of uncertain duration, sports leagues provide tightly controlled competitive conditions that enable consistent quantitative comparisons.  As a result, one can clearly see winning strategies and tactics based on season over season performance.  None is more remarkable than Sir Alex Ferguson’s record at Manchester United, probably the best record of any team/coach at the top leagues in any sport over the past 2+ decades.   Others have written about his management style here and here.

A great leader is one who gets the best from himself, his employees and customers regardless of the situation or context, be they sports teams, armies, corporations, academic institutions, societies, etc...  SAF's record is more remarkable in that it has improved the game of soccer at the same time.  Ultimately everyone is subservient and bound by the objective of the team, including himself.  SAF's style is best described as a disciplined command and control approach that leads to employee empowerment.   SAF has grown, bought, developed, and nurtured superstars and when they've outgrown the team, they are gone.

In his own words SAF says detail is important.  He is probably an acolyte of Edward Tufte, who says more information is better.  Through analytics SAF looks to gain the proverbial inch to give him the advantage over other teams.   In the process his staff has increased from 8 to 30 people over the past 2 decades.  In order to synthesize the information and implement strategy leading to individual decision making SAF has implemented an organizational matrix approach and not a management pyramid (or inverted one for that matter).  The result is that on game day ManU teams are capable of making decisions on the field and have the resolve to come from behind more often than not.

Furthermore, soccer records and trophies are won not just by outright wins, as in most other sports, but by a combination of wins, goal differentials and draws.  As well, Home and Away records are typically very important and a better than 50% winning percentage away from home means a lot.  In the table of SAF's points analysis we see that he achieved his remarkable record not just because he won, but because he managed to lose fewer than half his games away from home.  This record was achieved because he managed the goal differential ("for" over total goals in last column) significantly in his team's favor away from home. Anybody who knows the game of soccer knows the manager's strategy and tactics before and during the game typically has more of an impact on goal differential than luck.  The chart illustrates that ManU's goals for percentage mirrors the total winning points percentage.   For those who don't understand the subtleties of soccer's scoring, a 0-0 draw away from home can be the deciding factor in a trophy if a home win is managed.  Furthermore drawing 0-0 away from home to a big club is synonymous with a victory during the regular season.  (Of course an away goal in a cup competition is worth more than a home goal.)

So how did SAF manage this?  As a soccer purist, I love the style of play that is uptempo and attacking.  As a former defender I love when ManU’s right or left defenders become 3rd and 4th attackers up the wings.  But at the same time, striking the correct balance offensively and defensively is key.  By establishing a matrix approach SAF made it easier to move players around between and during games and to identify and increase their value relative to other players around them, so that a weakness in one area can be countered with strength in another.  Furthermore, the matrix approach empowers weak players and makes superstars out of "ordinary" players like Ryan Giggs and Paul Scholes; whom many would never have guessed to have such an impact on the team, or the game, as individuals.

In a matrix, employees have a 3D view of their decisions and the impact on the rest of the organization (resource constraints/availability).  From a tactical perspective (micro level) it enables forwards to play midfield and help on defense and defenders to develop the attack and even score.  A good example is a player like Valencia who used to just be a wing at another team, but has become a very dangerous midfielder and at times has started at right-back for an injury-riddled defense this season.  Strategically (at a macro level), the matrix defines key technology/product (supply) and customer/market (demand) parameters.  With a matrix approach micro and macro perspectives can be easily related.  For instance, SAF would never have let an individual player take over an entire organization’s public perception as Luis Suarez has at Liverpool this season.

Because of the matrix SAF’s strongest player is often the one others' consider weakest.  Rather than focusing on his superstars, he focuses on getting the superstars to work with the weakest players and vice versa.  Bringing this closer to home and more timely, the New York Knicks should implement a matrix approach both at the management/organizational and team levels.  Just look at the benefit the team gets from a player considered weak, Jeremy Lin, but who passes the ball effectively to very talented players versus superstars like Carmelo Anthony or Baron Davis who have a hard time being “matrix” players.  Every organization should develop its matrix and understand how employees play together as a team and how the organization is positioned in competitive industries.

Related Reading

Josh Linkner: Next Play, Mike Krzyzewski and Linkedin

John Doerr of KPCB used football example for OKRs for Google

No player is above the team according to SAF

Why soccer is so hard to predict

Posted by: Michael Elling AT 10:58 am   |  Permalink   |  0 Comments  |  Email
Sunday, February 12 2012

Last week we revisted our seminal analysis from 1996 of the 10 cent wireless minute plan (400 minutes for C$40) introduced by Microcell of Canada and came up with the investment theme titled “The 4Cs of Wireless”.  To generate sufficient ROI wireless needed to replace wireline as a preferred access method/device (PAD).  Wireless would have to satisfy minimal cost, coverage, capacity and clarity requirements to disrupt the voice market.  We found:

  • marginal cost of a wireless minute (all-in) was 1.5-3 cents
  • dual-mode devices (coverage) would lead to far greater penetration
  • software-driven and wideband protocols would win the capacity and price wars
  • CDMA had the best voice clarity (QoS); pre-dating Verizon’s “Can you hear me now” campaign by 6 years

In our model we concluded (and mathematically proved) that demand elasticity would drive consumption to 800 MOUs/month and average ARPUs to north of $70, from the low $40s.  It all happened within 2 short years; at least perceived by the market when wireless stocks were booming in 1998.  But in 1996, the pricing was viewed as the kiss of death for the wireless industry by our competitors on the Street.  BTW, Microcell, the innovator, was at a disadvantage based on our analysis, as the very reason they went to the aggressive pricing model to fill the digital pipes, namely lack of coverage due to a single-mode GSM phone, ended up being their downfall.  Coverage for a "mobility" product trumped price, as we see below.

What we didn’t realize at the time was that the 4Cs approach was broadly applicable to supply of communication services and applications in general.  In the following decade, we further realized the need for a similar checklist on the demand side to understand how the supply would be soaked up and developed the 4Us of Demand in the process.  We found that solutions and services progressed rapidly if they were:

  • easy to use interface
  • usable across an array of contexts
  • ubiquitous in terms of their access
  • universal in terms of appeal

Typically, most people refer only to user interface (UI or UX) or user experience (UE), but those aren't granular enough to accomodate the enormous range of demand at the margin.  Look at any successful product or service introduction over the past 30 years and they’ve scored high on all 4 demand elements.  The most profitable and self-sustaining products and solutions have been those that maximized perceived utility demand versus marginal cost.  Apple is the most recent example of this.

Putting the 4Cs and 4Us together in an iterative fashion is the best way to understand clearing of marginal supply and demand ex ante.  With rapid depreciation of supply (now in seconds, minutes and days) and infinitely diverse demand in digital networked ecosystems getting this process right is critical.

Back in the 1990s I used to say the difference between wireless and wired networks was like turning on a lightswitch in a dark room filled with people.  Reaction and interaction (demand) could be instantaneous for the wireless network.  So it was important to build out rapidly and load the systems quickly.  That made them generative and emergent, resulting in exponential demand growth.  (Importantly, this ubiquity resulted from interconnection mandated by regulations from the early 1980s and extended to new digital entrants (dual-mode) in the mid 1990s).  Conversely a wired network was like walking around with a flashlight and lighting discrete access points providing linear growth.

The growth in adoption we are witnessing today from applications like Pinterest, Facebook and Instagram (underscored in this blogpost from Fred Wilson) is like stadium lights compared with the candlelight of the 1990s.  What took 2 years is taking 2 months.  You’ll find the successful applications and technologies score high on the 4Cs and 4Us checklists before they turn the lights on and join the iOS and Android parties.

Related Reading:
Fred Wilson's 10 Golden Principles

Posted by: Michael Elling AT 11:37 am   |  Permalink   |  0 Comments  |  Email
Sunday, February 05 2012

You all know Monsieurs (MM.) Moore et Metcalfe.  But do you know Monsieur (M.) Zipf?  I made his acquaintance whilst researching infinite long tails.  Why does he matter, you inquire?  Because M. Zipf brings some respectability to Moore et Metcalfe, who can get a little out of control from time to time. 

Monsieur Moore is an aggressive chap who doubles his strength every 18 months or so and isn’t shy about it.  Monsieur Metcalfe has an insatiable appetite, and every bit he consumes increases his girth substantially.  Many people have made lots of money from MM Moore’s et Metcalfe’s antics over the past 30 years.  The first we refer to generally as the silicon or processing effect, the latter as the network effect.  Putting the two together should lead to declines of 50-60% in cost for like performance or throughput.  Heady, rather piggish stuff!

Monsieur Zipf, on the other hand, isn’t one for excess.  He follows a rather strict regimen; one that applies universally to almost everything around us; be it man-made or natural.  M. Zipf isn’t popular because he is rather unsocial.  He ensures that what one person has, the next chap only can have half as much, and the next chap half that, and so on.   It’s a decreasing, undemocratic principle.   Or is it?

Despite his unpopularity, lack of obvious charm and people’s general ignorance of him, M. Zipf’s stature is about to rise.  Why?  Because of the smartphone and everyone’s desire to always be on and connected; things MM Moore and Metcalfe wholeheartedly support.

M. Zipf is related to the family of power law distributions.  Over the past 20 years, technologists have applied his law to understanding network traffic.  In a time of plenty, like the past 20 years, M. Zipf’s not been that important.  But after 15 years of consolidation and relative underinvestment we are seeing demand outstrip supply and scarcity is looming.  M. Zipf can help deal with that scarcity.

Capacity will only get worse as LTE (4G) devices explode on the scene in 2012, not only because of improved coverage, better handsets, improved Android (ICS), but mostly because of the iconic iPhone 5 coming this summer!  Here’s the thing with 4G phones, they have bigger screens and they load stuff 5-10x faster.  So what took 30 seconds now takes 3-10 seconds to load and stuff will look 2-3x better!  People will get more and want more; much to MM Moore’s et Metcalfe’s great pleasure.

“Un moment!” cries M. Zipf.  “My users already skew access quite a bit and this will just make matters worse!  Today, 50% of capacity is used by 1% of my users.  The next 9% use 40% and the remaining 90% of users use just 10% of capacity.  With 4G the inequality can only get worse.  Indignez-vous!  Which is the latest French outcry for equality.”  It turns out Zipf's law is actually democratic in that each person consumes at their marginal not average rate.  The latter is socialism.

Few of us will see this distribution of usage as a problem short-term, except when we’re on overloaded cellsites and out of reach of a friendly WiFi hotspot.  The carriers will throw more capex at the problem and continue to price inefficiently and ineffectively.   The larger problem will become apparent within 2 years when the 90% become the 10% and the carriers tell Wall Street they need to invest another $50B after 2015 just after spending $53B between 2010-2014.

Most people aware of this problem say there is a solution.  More Spectrum = more Bandwidth to satisfy MM Moore et Metcalfe.  But they’ve never heard of M. Zipf nor understood fully how networks are used.  Our solution, extended as a courtesy by M. Zipf, is to “understand the customer” and work on “traffic offloading” at the margin.  Pricing strategies, some clever code, and marketing are the tools to implement a strategy that can minimize the capital outlays, and rapidly amortize investment and generate positive ROI.

We’ve been thinking about this since 1996 when we first introduced our 4Cs of Wireless (cost, coverage, capacity and clarity) analyzing, understanding and embracing 10 cent wireless pricing (introduced by French Canada's revolutionary MicroCell).  As a result we were 2-3 years ahead of everybody with respect to penetration, consumption and wireline substitution thinking and forecasts. Back in 1995 the best wireless prices were 50 cents per minute and just for buying a lot of local access.  Long-distance and roaming charges applied.  So a corporate executive who travelled a lot would regularly rack up $2000-3000 monthly phone bills.  The result was less than 10% penetration, 80 minutes of use per month, and ARPUs declining from $45 to $40 to $35 in analysts' models because the marginal customers being added to the network were using the devices infrequently and more often than not putting them into the glove compartment in case of emergencies.  Fewer than 3% of the population actually used the devices more than 1x day.

We used to poll taxi drivers continuously about wireless and found that their average perceived price of $0.75 per minute was simply too high to justify not having to pull over and use a payphone for $0.25.  So that was the magical inflection point in the elasticity curves.  When MicroCell introduced $0.10 late in the Spring of 1996 and we polled the same set of users, invariably we were just able to avoid an accident they got so excited.  So we reasoned and modeled that more than just taxi drivers would use wireless as a primary access device.  And use it a lot.  This wireless/wireline substitution would result in consumption of 700-800 minutes of use per month, penetration hitting 100% quickly and ARPUs, rather than declining, actually increasing to $70.  The forecast was unbelievably bullish.  And of course no one believed it in 1996, even though all those numbers were mostly reached within 5 years.   

But we also recognized that wireless was a two-edge sword with respect to localized capacity and throughput; taking into account the above 3 laws.  So we also created an optimal zone, or location-based, pricing and selling plan that increased ARPUs and effective yield and were vastly superior to all you can eat (AYCE) and eat what you want (EWYW) plans.  Unfortunately, carriers didn't understand or appreciate M. Zipf and within 2 years they were giving away night and weekend minutes for free, where they could have monetized them for 3-6 cents each.  Then some carriers responded by giving away long-distance (whose marginal cost was less than a minutes without access; but still could cost 2-3 cents).  Then AT&T responded with the One-rate plan, which destroyed roaming surcharges and led to one-rate everywhere; even if demand was different everywhere.

Here’s a snapshot of that analysis that is quite simple and consistent with Zipf's law and highly applicable today.  Unfortunately, where my approach would have kept effective yield at 8 cents or higher, the competitive carriers responded by going to all you can eat (AYCE) plans and the effective yield dropped to 4 cents by 2004.  Had intercarrier SMS not occurred in the 2003-4 timeframe, they would have all been sunk with those pricing models, as they were in the middle of massive 2G investment programs for the coming "wireless data explosion", which actually didn't happen until 3G and smartphones in the 2008-2009 timeframes.  It was still a voice and blackberry (texting and email) world in 2007 when the iPhone hit. With ubiquitous SMS and people's preference to text instead of leaving vmail, minutes dropped from 700 to 500, lowering carrier's costs, and they were able to generate incremental revenues on SMS pricing plans (called data) in the 2004-2007 timeframe.

All that said, the analysis and approach is even more useful today since extreme consumption of data will tend to occur disproportionately in the fixed mode (what many refer to as offload).  I let you come up with your own solutions.  À bientôt!  Oh, and look up Free in France to get an idea of where things are headed.  What is it about these French?  Must be something about Liberté, égalité, fraternité.

Related Reading:
It's All in Your Head by Robert Metcalfe

JD Power Surveys on Wireless Network Performance

The Law of the Few (the 20/80 rule) from the Tipping Point

Briscoe, Odlyzko get it wrong because of one dimensional thinking and mixing apples and oranges

See Larry Downes' Law of Disruption

Law of Bandwidth (but only in a period of monopoly)

See our own home cooked Law of Wireless Gravity

Keck's law of fiber capacity (may be coming to an end?

Kurzweil's 20 Laws of Telecosm (how many right and wrong also depends on timing)

Posted by: Michael Elling AT 11:38 am   |  Permalink   |  0 Comments  |  Email
Sunday, January 29 2012

Every institution, every industry, every company has or is undergoing the transformation from analog to digital.  Many are failing, superseded by new entrants.  No more so than in the content and media industries: music, retail, radio, newspaper and publishing.  But why, especially as they’ve invested in the tools and systems to go digital?  Their failure can be summed up by this simple quote, “Our retail stores are all about customer service, and (so and so) shares that commitment like no one else we’ve met,” said Apple’s CEO. “We are thrilled to have him join our team and bring his incredible retail experience to Apple.”

Think about what Apple’s CEO emphasized.  “Customer service.”  Not selling; and yet the stores accounted for $15B of product sold in 2011!  When you walk into an Apple store it is like no other retailing experience, precisely because Apple stood the retail model on its head.  Apple thought digital as it sold not just 4 or 5 products--yes that’s it—but rather 4-5 ecosystems that let the individual easily tailor their unique experience from beginning to end.

Analog does not scale.  Digital does.  Analog is manual.  Digital is automated.  Analog cannot easily be software defined and repurposed.  Digital can.  Analog is expensively two-way.  With Digital 2-way becomes ubiquitous and synchronous.  Analog is highly centralized.  Digital can be easily distributed.  All of this drives marginal cost down at all layers and boundary points meaning performance/price is constantly improving even as operator/vendor margins rise.

With Digital the long tail doesn’t just become infinite, but gives way to endless new tails.  The (analog) incumbent sees digital as disruptive, with per unit price declines and same-store revenues eroding.  They fail to see and benefit from relative cost declines and increased demand.  The latter invariably occurs due to a shift from "private" to public consumption, normal price elasticity, and "application" elasticity as the range of producers and consumers increases.  The result is overall revenue growth and margin expansion for every industry/market that has gone digital.

Digital also makes it easy for something that worked in one industry to be easily translated to another.   Bill Taylor of Fast Company recently wrote in HBR that keeping pace with rapid change in a digital world requires having the widest scope of vision, and implementing successful ideas from other fields.

The film and media industries are a case in point.  As this infographic illustrates Hollywood studios have resisted “thinking digital” for 80 years.  But there is much they could learn from the transformation of other information/content monopolies over the past 30 years.  This blog from Fred Wilson sums up the issues between the incumbents and new entrants well.  Hollywood would do well to listen and see what Apple did to the music industry and how it changed it fundamentally; because it is about to do the same to publishing and video.  If not Apple then others.

Another aspect of digital is the potential for user innovation.   Digital companies should constantly be looking for innovation at the edge.  This implies a focus on the “marginal” not average consumer.  Social media is developing tremendous tools and data architectures for this.  If companies don’t utilize these advances, those same users will develop new products and markets, as can be seen from the comments field of this blog on the financial services industry.

Digital is synonymous with flat which drives greater scale efficiency into markets.  Flat (horizontal) systems tend toward vertical completeness via ecosystems (the Apple or Android or WinTel approach).  Apple IS NOT vertically integrated.  It has pieced together and controls very effectively vertically complete solutions.  In contrast, vertically integrated monopolies ultimately fail because they don’t scale at every layer efficiently.  Thinking flat (horizontal) is the first step to digitization.

Related Reading:
Apple Answers the Question "Why?" Before "How?" And "What?"
US Govt to Textbook Publishers: You Will Go Digital!
This article confused vertical integration with vertical completeness
Walmart and Retailers Need to Rethink Strategy
Comparison of 3 top Fashion Retailers Web and App Strategies

Software Defined Networking translated to the real world

Posted by: Michael Elling AT 10:54 am   |  Permalink   |  0 Comments  |  Email
Sunday, January 22 2012

Data is just going nuts!  Big data, little data, smartphones, clouds, application ecosystems.  So why are Apple and Equinix two of only a few large cap companies in this area with stocks up over 35% over the past 12 months, while AT&T, Verizon, Google and Sprint are market performers or worse?   It has to do with pricing, revenues, margins and capex; all of which impact ROI.  The former’s ROI is going up while the latters’ are flat to declining.  And this is all due to the wildness of mobile data.

Data services have been revealing flaws and weaknesses in the carriers pricing models and networks for some time, but now the ante is being upped.  Smartphones now account for almost all new phones sold, and soon they will represent over 50% of every carriers base, likely ending this year over 66%.  That might look good except when we look at these statistics and facts:

  • 1% of wlx users use 50% of the bandwidth, while the top 10% account for 90%.  That means 9 out of 10 users account for only 10% of network consumption; clearly overpaying for what they get.
  • 4G smartphone displays (720x1280 pixels) allow video viewing that uses 300x more capacity than voice.
  • Streaming just 2 hours of music daily off a cloud service soaks up 3.5GB per month
  • Carriers still derive more than 2/3 of their revenues from voice.
  • Cellular wireless (just like WiFi) is shared. 

Putting this together you can see that on the one hand a very small percentage of users use the bulk of the network.  Voice pricing and revenues are way out of sync with corresponding data pricing and revenues; especially as OTT IP voice and other applications become pervasive. 

 

Furthermore, video, which is growing in popularity will end up using 90% of the capacity, crowding out everything else, unless carriers change pricing to reflect differences in both marginal users and marginal applications.  Marginal here = high volume/leading edge.

So how are carriers responding?  By raising data prices.  This started over a year ago as they started capping those “unlimited” data plans.  Now they are raising the prices and doing so in wild and wacky ways; ways we think that will come back to haunt them just like wild party photos on FB.  Here are just two of many examples:

  • This past week AT&T simplified its pricing and scored a marketing coup by offering more for more and lowering prices even as the media reported AT&T as “raising prices.”  They sell you a bigger block of data at a higher initial price and then charge the same rate for additional blocks which may or may not be used.  Got that?
  • On the other hand that might be better than Walmart’s new unlimited data plan which requires PhD level math skills to understand.  Let me try to explain as simply as possible.  Via T-Mobile they offer 5GB/month at 3G speed, thereafter (the unlimited part) they throttle to 2G speed.  But after March 16 the numbers will change to 250MB initially at 3G, then 2G speeds unlimited after that.  Beware the Ides of March’s consumer backlash!

Unless the carriers and their channels start coming up with realistic offload solutions, like France’s Free, and pricing to better match underlying consumption, they will continue to generate lower or negative ROI.  They need to get control of wild data.  Furthermore, if they do not, the markets and customers will.  With smartphones (like Apple's, who by the way drove WiFi as a feature knowing that AT&T's network was subpar) and cloudbased solutions (hosted by Equinix) it is becoming easier for companies like Republic Wireless to virtually bypass the expensive carrier plans using their very own networks.  AT&T, VZ, Sprint will continue to be market performers at best.

Related Reading/Viewing:
AT&T pricing relies heavily on breakage
Useful stats on data growth from MobileFuture
FoxNews report on AT&T Data Throttling
This article actually suggests "dissuading usage" as 1 of 4 solutions
Consumer reports article whereby data equivalent voice pricing = 18 cents, OTT on lowest plan = 6.6 cents, OTT on highest plan = 1 cent
New shared data plans set a new high standard

Posted by: Michael Elling AT 09:13 am   |  Permalink   |  0 Comments  |  Email
Sunday, January 15 2012

“There is no clear definition, but usually the Singularity is meant as a future time when societal, scientific and economic change is so fast we cannot even imagine what will happen from our present perspective, and when humanity will become posthumanity.”  Futurist Ray Kurzweil wrote the Singularity is near in 2005 and it sure felt that way at CES in Las Vegas in 2012. 

In any event, we didn’t come up with this association.  Matt Burns at TechCrunch did during one of the live streamed panel discussions from the show floor.  But given that we’ve been at the forefront of mobile devices and networks for 2 decades, we couldn’t have said it any better.  After all this time and all these promises, it really felt like all the devices were connected and this app-ecosystem-mass was beginning to move at a pace and direction that none of us could fully explain.

Vernor Vinge, a sci-fi writer, coined the expression back in the early 1990s.  When you look at the list of publications at the Singularity Institute one can’t help but think that the diaspora of different applications and devices self organizing out of CES 2012 isn’t the beginning of AI.  Will AI develop out of the search algorithms that tailor each individuals specific needs, and possibly out-thinking or out-guessing the individual in the process?

We probably won’t know the answer for at least a decade, but what we do know is that we are about to embark on a wave of growth, development and change that will make the past 30 year WinTel-Internet development look embryonic at best.  After all the Singularity is infinity in mathematical terms.  Hold onto your seats.

Posted by: Michael Elling AT 09:10 am   |  Permalink   |  0 Comments  |  Email
Sunday, January 08 2012

Yahoo executed on the portal concept better than any other company and one time was worth $113B.  Today it is worth $17.6B and is trading at 12.9x operating cash flow.  But its interest in Chinese and Japanese subsidiaries is $19B pretax or $13B aftertax.  Either way YHOO is absurdly cheap on $5B of revenue and $1.4B cash flow and 700 million users.  So new CEO Thompson has a chance to pull off the turnaround of the century for a company that is hard to define by many.

What would we do in his shoes?

First, we would go back to what made Yahoo so valuable in people’s eyes to begin with, namely an approach that used others’ content and turned it into gold via traffic streams and advertising revenue.  As a tools and application company they have been pretty poor at monetizing their innovations.  Ultimately, the chief reason people came to Yahoo was the breadth and easy accessibility of the information, making people shout “yahoo” in joy.  The brand is a great brand after all and one of the real internet originals which will fix itself.

So what is the “new” content today?  Applications.  Yahoo should evolve from content portal to "app-portal".  More specifically we are about to witness an explosion of cross application solutions and commerce that will develop entire new ecosystems of revenue opportunity.  And these will exist across multiple platforms and screens as one writer points out.  Facebook is trying to harness this opportunity with Opengraph, but already people might be concerned that it is just too much confusion and overkill and TMI.

Yahoo can not only remove that confusion but allow people to tailor their own environments.  HTML5 and connectivity across smartphone, TV, tablet and PC/ultrabook will pave the way.  Understanding and harnassing the big data required to do this is one of Thompson’s key attributes.  At the same time virtual currencies can be created to monetize this interactivity beyond mere advertising and that’s what Thompson brings from Paypal.

With over 70 different services/tools and 14,000 employees, Thompson will definitely have to restructure, reduce and refocus in order to have the resources to aggressively develop an app-portal future.

Posted by: Michael Elling AT 08:43 am   |  Permalink   |  0 Comments  |  Email
Thursday, January 05 2012

Counter-intuitive thinking often leads to success.  That’s why we practice and practice so that at a critical moment we are not governed by intuition (chance) or emotion (fear).  No better example of this than in skiing; an apt metaphor this time of year.  Few self-locomoted sports provide for such high risk-reward requiring mental, physical and emotional control.  To master skiing one has to master a) the fear of staying square (looking/pointing) downhill, b) keeping one’s center over (or keeping forward on) the skis, and c) keeping a majority of pressure on the downhill (or danger zone) ski/edge.  Master these 3 things and you will become a marvelous skier.  Unfortunately, all 3 run counter to our intuitions driven by fear and safety of the woods at the side of the trail, leaning back and climbing back up hill.  Overcoming any one is tough.

What got me thinking about all this was a Vint Cerf (one of the godfathers of the Internet) Op-Ed in the NYT this morning which a) references major internet access policy reports and decisions, b) mildly supports the notion of the Internet as a civil not human right, and c) trumpets the need for engineers to put in place controls that protect people’s civil (information) rights.  He is talking about policy and regulation from two perspectives, business/regulatory and technology/engineering, which is confusing.  In the process he weighs in, at a high level, on current debates over net neutrality, SOPA, universal service and access reform, from his positions at Google and IEEE and addresses the rights and governance from an emotional and intuitive sense.

Just as with skiing, let’s look at the issues critically, unemotionally and counter-intuitively.  We can’t do it all in this piece, so I will establish an outline and framework (just like the 3 main ways to master skiing) and we’ll use that as a basis in future pieces to expound on the above debates and understand corporate investment and strategy as 2012 unfolds.

First, everyone should agree that the value of networks goes up geometrically with each new participant.  It’s called Metcalfe’s law, or Metcalfe’s virtue.  Unfortunately people tend to focus on scale economies and cost of networks; rarely the value.  It is hard to quantify that value because most have a hard time understanding elasticity and projecting unknown demand.  Further few rarely distinguish marginal from average cost.  The intuitive thing for most to focus on is supply, because people fear the unknown (demand).

Second, everyone needs to realize that there is a fundamental problem with policy making in that (social) democrats tend to support and be supported by free market competitors, just as (conservative) republicans have a similar relationship with socialist monopolies.  Call it the telecom regulatory paradox.  This policy paradox is a function of small business vs big business, not either sides’ political dogma; so counter-intuitive and likely to remain that way.

Third, the internet was never open and free.  Web 1.0 resulted principally from a judicial action and a series of regulatory access frameworks/decisions in the mid to late 1980s that resulted in significant unintended consequences in terms of people's pricing perception.  Markets and technology adapted to and worked around inefficient regulations.  Policy makers did not create or herald the internet, wireless and broadband explosions of the past 25 years.  But in trying to adjust or adapt past regulation they are creating more, not less, inefficiency, no matter how well intentioned their precepts.  Accept it as the law of unintended consequences.  People feel more comfortable explaining results from intended actions than something unintended or unexplainable.

So, just like skiing, we’ve identified 3 principles of telecoms and information networks that are counter-intuitive or run contrary to accepted notions and beliefs.  When we discuss policy debates, such as net neutrality or SOPA, and corporate activity such as AT&T’s aborted merger with T-Mobile or Verizon’s spectrum and programming agreement with the cable companies, we will approach and explain them in the context of Metcalfe’s Virtue (demand vs supply), the Regulatory Paradox (vertical vs horizontal orientation; not big vs small), and  the law of unintended consequences (particularly what payment systems stimulate network investment).  Hopefully the various parties involved can utilize this approach to better understand all sides of the issue and come to more informed, balanced and productive decisions.

Vint supports the notion of a civil right (akin to universal service) for internet access.  This is misguided and unachievable via regulatory edict/taxation.  He also argues that there should be greater control over the network.  This is disingenuous in that he wants to throttle the open-ness that resulted in his godchild’s growth.  But consider his positions at Google and IEEE.  A “counter-intuitive” combination of competition, horizontal orientation and balanced payments is the best approach for an enjoyable and rewarding experience on the slopes of the internet and, who knows, ultimately and counterintuitively offering free access to all.  The regulators should be like the ski patrol to ensure the safety of all.   Ski school is now open.

Related reading:
A Perspective from Center for New American Security

Network Neutrality Squad (NNsquad) of which Cerf is a member

Sad State of Cyber-Politics from the Cato Institute

Bike racing also has a lot of counter-intuitive moments, like when your wheel locks with the rider in front.  Here's what to do!

Posted by: Michael Elling AT 01:23 pm   |  Permalink   |  0 Comments  |  Email
Thursday, December 29 2011

67 million Americans live in rural areas. The FCC says the benchmark broadband speed is at least 4 Mbps downstream and 1 Mbps upstream. Based on that definition 65% of Americans actually have broadband, but only 50% who live in rural markets do; or 35 million. The 50% is due largely because 19 million Americans (28%) who live in rural markets do not even have access to these speeds. Another way of looking at the numbers shows that 97% of non-rural Americans have access to these speeds versus 72% living in rural areas.  Rural Americans are at a significant disadvantage to other Americans when it comes to working from home, e-commerce or distance education.  Clearly 70% are buying if they have access to it.

Furthermore we would argue the FCC standard is no longer acceptable when it comes to basic or high-definition multimedia, video and file downloads.  These applications require 10+ Mbps downstream and 3+ Mbps upstream to make applications user friendly.  Without those speeds you get what we call the "world-wide-wait" in rural markets for most of today's high-bandwidth applications.  In the accompanying 2 figures we see a clear gap between the blue lines (urban) and green lines (rural) for both download and upload speeds.  The result is that only 7% of rural Americans use broadband service with 6+/1.5+ Mbps versus 22% nationwide today.

The problem in rural markets is lack of alternative and affordable service providers. In fact the NTIA estimates that 4% of Americans have no broadband provider to begin with, 12% only 1 service provider and 44% just 2 providers. Almost all rural subscribers fall into 1 of these 3 categories. Rural utilities, municipalities, businesses and consumers would benefit dramatically from alternative access providers as economic growth is directly tied to broadband penetration.

The accompanying chart shows how vital broadband is to regional economic growth.  If alternative access drives rural broadband adoption to levels similar to urban markets, then local economies will grow an additional 3% annually.  That's because new wireless technology and applications such as home energy management, video on demand, video conferencing and distance learning provide the economic justification for alternative, lower-cost, higher bandwidth solutions.

Related Reading

FCC Broadband Map

US 3G Wireless Coverage Map

The UK is Far Ahead of the US; Their deficient is our average

Rural Telcos Against FCC USF Reform

GA Tries to Reduce Subsidies, Again

 

Posted by: Michael Elling AT 08:09 am   |  Permalink   |  0 Comments  |  Email
Sunday, December 18 2011

 

(The web is dead, long live the apps)

 

Is the web dead?  According to George Colony, CEO of Forrester, at LeWeb (Paris, Dec 7-9) it is; and on top of that social is running out of time, and social is where the enterprise is headed.  A lot to digest at once, particularly when Google’s Schmidt makes a compelling case for a revolutionary smartphone future that is still in its very, very early stages; courtesy of an ice cream sandwich.

Ok, so let’s break all this down.  The Web, dead?  Yes Web 1.0 is officially dead, replaced by a mobile, app-driven future.  Social is saturated?  Yes, call it 1.0 and Social 2.0 will be utilitarian.  Time is money, knowledge is power.  Social is really knowledge and that’s where enterprises will take the real-time, always connected aspect of the smartphone ice cream sandwich applications that harness internal and external knowledge bases for rapid product development and customer support.  Utilitarian.  VIVA LA REVOLUTION!

Web 1.0 was a direct outgrowth of the breakup of AT&T; the US’ second revolution 30 years ago coinciding ironically with the bicentennial end of the 1st revolution.  The bandwidth bottleneck of the 1960s and 1970s (the telephone monopoly tyranny) that gave rise to Microsoft and Intel processing at the edge vs the core, began to reverse course in the late 1980s and early 1990s as a result of flat-rate data access and an unlimited universe of things to easily look for (aka web 1.0).  This flat-rate processing was a direct competitive response by the RBOCs to the competitive WAN (low-cost metered) threat.

As silicon scaled via Moore’s law (the WinTel sub-revolution) digital mobile became a low-cost, ubiquitous reality.  The same pricing concepts that laid the foundation for web 1.0 took hold in the wireless markets in the US in the late 1990s; courtesy of the software defined, high-capacity CDMA competitive approach (see pages 34 and 36) developed in the US.

The US is the MOST important market in wireless today and THE reason for its leadership in applications and smart cloud.  (Incidentally, it appears that most of LeWeb speakers were either American or from US companies.)  In the process the relationship between storage, processing and network has come full circle (as best described by Ben Horowitz).  The real question is, “will the network keep up?”  Or are we doomed to repeat the cycle of promise and dashed hopes we witnessed between 1998-2003?

The answer is, “maybe”; maybe the communications oligopolies will liken themselves to IBM in front of the approaching WinTel tsunami in 1987.  Will Verizon be that service provider that recognizes the importance of and embraces open-ness and horizontalization?  The 700 mhz auctions and recent spectrum acquisitions and agreements with the major cable companies might be a sign that they do.

But a bigger question is whether Verizon will adopt what I call a "balanced payment (or settlement) system" and move away from IP/ethernet’s "bill and keep" approach.  A balanced payment or settlement system for network interconnection simultaneously solves the issues of new service creation AND paves the way for the applications to directly drive and pay for network investment.  So unlike web 1.0 where communication networks were resistently pulled into a broadband present, maybe they can actually make money directly off the applications; instead of the bulk of the value accruing to Apple and Google.

Think of this as an “800” future on steroids or super advertising, where the majority of access is paid for by centralized buyers.  It’s a future where advertising, product marketing, technology, communications and corporate strategy converge.  This is the essence of what Colony and Schmidt are talking about.   Will Verizon CEO Seidenberg, or his rivals, recognize this?  That would indeed be revolutionary!

Related Reading:
February 2011 Prediction by Tellabs of Wireless Business Models Going Upside Down by 2013

InfoWeek Article on Looming Carrier Bandwidth Shortages

 

 

 

 

 

Posted by: Michael Elling AT 09:56 am   |  Permalink   |  0 Comments  |  Email

Email
Twitter
Facebook
Digg
LinkedIn
Delicious
StumbleUpon
Add to favorites

Information Velocity Partners, LLC
88 East Main Street, Suite 209
Mendham, NJ 07930
Phone: 973-222-0759
Email:
contact@ivpcapital.com

Mastodon

Design Your Own Website, Today!
iBuilt Design Software
Give it a try for Free