Recently released statistics by e-mail security and authentication service provider Valimail tell us that ~2% of e-mail communications worldwide are deemed “potentially malicious” because they’ve failed DMARC testing (domain-based message authentication, reporting and conformance) and also don’t originate from known, legitimate senders.
That’s a small percentage — seemingly trivial. But considering the volume of e-mail messages sent every day, it translates into nearly 6.4 billion e-mails sent every day that are “fake, faux and phony.”
Interestingly, the source of those fake e-mails is most often right here in the United States. Not Russia or Ukraine. Or Nigeria or Tajikistan.
In fact, no other country even comes close to the USA in the number of fraudulent e-mails.
The good news is that DMARC has made some pretty decent strides in recent times, with DMARC support now covering around 5 billion inboxes worldwide, up from less than 3 billion in 2015.
The federal government is the biggest user of DMARC, but nearly all U.S. tech companies and most Fortune 500 companies also participate.
Participation is one thing, but doing something about enforcement is another. At the moment, Valimail is finding that the enforcement failure rate is well above 70% — hardly an impressive track record.
The Valimail study findings came as the result of analyzing billions of e-mail message authentication requests, along with 3 million+ publicly accessible DMARC records. So, the findings are meaningful and provide good directional indications.
But what are the research implications? The findings underscore the degree to which name brands can be “hijacked” for nefarious purposes.
Additionally, there’s consumer fallout in that many people are increasingly skittish about opening any marketing-oriented e-mails at all, figuring that the risk of importing a virus outweighs any potential benefit from the marketing pitch.
That isn’t an over-abundance of caution, either, because 9 in 10 cyber attacks begin with a phishing e-mail.
It’s certainly enough to keep many people from opening the next e-mail that hits their inbox from a Penneys(?), DirecTV(?) or BestBuy(?).
How about you? Are you now sending those e-mails straight to the trash as a matter of course?
Video is supposed to be the “great equalizer”: evidence that doesn’t lie — particularly in the case of chronicling law enforcement events.
From New York City and Chicago to Baltimore, Charleston, SC and dozens of places in between, there have been a number of “high profile” police incidents in recent years where mobile video has made it possible to go beyond the sometimes-contradictory “he said/she said” statements coming from officers and citizens.
There’s no question that it’s resulted in some disciplinary or court outcomes that may well have turned out differently in times before.
In response, numerous police departments have responded in a way best described as, “If you can’t beat them, join them.” They’ve begun outfitting their law enforcement personnel with police body cams.
The idea is that having a “third party” digital witness on the scene will protect both the perpetrator and the officer when assessments need to be made about conflicting accounts of what actually happened.
This tidy solution seems to be running into a problem, however. Some security experts are calling into question the ability of body cameras to provide reliable evidence – and it isn’t because of substandard quality in the video footage being captured.
Recently, specialists at the security firm Nuix examined five major brands of security cameras … and have determined that all of them are vulnerable to hacking.
The body cam suppliers in question are CEESC, Digital Ally, Fire Cam, Patrol Eyes, and VIEVU. The cameras are described by Nuix as “full-feature computers walking around on your chest,” and as such, require the same degree of security mechanisms that any other digital device operating in security-critical areas would need to possess.
But here’s the catch: None of the body cameras evaluated featured digital signatures on the uploaded footage. This means that there would be no way to confirm whether any of the video evidence might have been tampered with.
In other words, a skilled technician with nefarious intent could download, edit and re-upload content – all while avoiding giving any sort of indication that it had been revised.
These hackers could be operating on the outside … or they could be rogue officers inside a law enforcement department.
Another flaw uncovered by Nuix is that malware can infect the cameras in the form of malicious computer code being disguised as software updates – updates that the cameras are programmed to accept without any additional verification.
Even worse, once a hacker successfully breached a camera device, he or she could easily gain access to the wider police network, thereby causing a problem that goes much further than a single camera or a single police officer.
Thankfully, Nuix is a “good guy” rather than a “bad actor” in its experimentation. The company is already working with several of the body cam manufacturers to remedy the problems uncovered by its evaluation, so as to improve the ability of the cameras to repel hacking attempts.
But the more fundamental issue that’s raised is this: What other types of security vulnerabilities are out there that haven’t been detected yet?
It doesn’t exactly reinforce our faith in technology to ensure fairer, more honest and more transparent law enforcement activities. If video footage can’t be considered verified proof that an event happened or didn’t happen, have we just returned to Square One again, with people pointing fingers in both directions but with even lower levels of trust?
Hopefully not. But with the polarized camps we have at the moment, with people only too eager to blame the motives of their adversaries, the picture doesn’t look particularly promising …
Looking back over the past eight years or so, something very interesting has been happening to the arrest rate statistics for people driving under the influence.
DUI arrests have been dropping – pretty steadily and inexorably.
The trend started in 2011, in which year an 8% decline in DUI arrests was experienced over the prior year. In 2012 the decline was 4.1% … in 2013, it was another 7.2%.
And arrest rates didn’t plateau after that, either. DUI arrests have continued to decline — even as police departments have continued to put plenty of cops on the beat for such purposes.
One of the most dramatic examples of the continued decline is in Miami-Dade County — the highest population county in the entire Southeastern U.S. The Miami-Dade police force made DUI arrests in 2017 that were 65% fewer than four years earlier.
Look around the country and you’ll see similar trends in places as diverse as San Antonio, TX, Phoenix, AZ, Portland, OR and Orange County, CA.
There are common thread, in what’s being seen across the nation:
DUI arrest levels are down in major metro areas — but not necessarily in exurban or rural areas.
DUI arrest levels have declined the nearly all of the metro areas where ride-sharing services are prominent.
This last point a significant factor to consider: The increasing popularity of ride sharing services has coincided with the drop in DUI arrests.
A 2017 University of Pennsylvania analysis found that in places where ride-hailing services were readily available, in most cases DUI arrests had declined upwards of 50% or more compared to just a few years earlier.
Ride-hailing services are particularly popular with younger adults, who like the smartphone apps that make them pretty effortless to use. They’re also popular with people who are looking for more affordable ways to get about town compared to what highly regulated taxi services choose to charge.
Plus, the “cool” factor of ride-sharing leaves old-fashioned taxi services pretty much in the dust.
The few exceptions of locations where DUI arrest declines haven’t been so pronounced are in places like Las Vegas and Reno, NV – tourist destinations that draw out-of-towners who frequently take public transportation, hail taxis, or simply walk rather than rent vehicles to get around town.
With the positive consequences of fewer DUI arrests – which also correlate to reductions in vehicular homicides and lower medical care costs due to fewer people injured in traffic accidents, as well as reductions in the cost of prosecuting and incarcerating the perpetrators – one might think that other urban areas would take notice and become more receptive to ride-sharing services than they have been up to now.
But where taxi services are well-entrenched and “wired” into the political fabric – a situation often encountered in older urban centers like Chicago, St. Louis, Philadelphia and Baltimore — the ancillary benefits of ride-sharing services don’t appear to hold much sway with city councils or city regulators – at least not yet.
One might suppose that overstretched urban police departments would welcome having to spend less time patrolling the streets for DUI drivers. And if that benefits police departments … well, the police also represents a politically important constituency, too.
It seems that some fresh thinking may be in order.
You’d think that with the continuing cascade of news about the exposure of personal information, people would be more skittish than ever about sharing their data.
But this isn’t the case … and we have a 2018 study from marketing data foundation firm Acxiom to prove it. The report, titled Data Privacy: What the Consumer Really Thinks, is the result of survey information gathered in late 2017 by Acxiom in conjunction with the Data & Marketing Association (formerly the Direct Marketing Association).
The research, which presents results from an online survey of nearly 2,100 Americans age 18 and older, found that nearly 45% of the respondents feel more comfortable with data exchange today than they have in the past.
Among millennial respondents, well over half feel more comfortable about data exchange today.
Indeed, the report concludes that most Americans are “data pragmatists”: people who are open about exchanging personal data with businesses if the benefits received in return for their personal information are clearly stated.
Nearly 60% of Americans fall into this category.
On top of that, another 20% of the survey respondents report that they’re completely unconcerned about the collection and usage of their personal data. Among younger consumers, it’s nearly one-third.
When comparing Americans’ attitudes to consumers in other countries, we seem to be a particularly carefree bunch. Our counterparts in France and Spain are much more wary of sharing their personal information.
Part of the difference in views may be related to feelings that Americans have about who is responsible for data security. In the United States, the largest portion of people (~43%) believe that 100% of the responsibility for data security lies with consumers themselves, versus only ~6% who believe that the responsibility resides solely with brands or the government. (The balance of people think that the responsibility is shared between all parties.)
To me, the bottom-line finding from the Acxiom/DMA study is that people have become so conditioned to receiving the benefits that come from data exchange, they’re pretty inured to the potential downsides. Probably, many can’t even fathom going back to the days of true data privacy.
Of course, no one wishes for their personal data to be used for nefarious purposes, but who is willing to forego the benefits (be it monetary, convenience or comfort) that come from companies and brands knowing their personal information and their personal preferences?
And how forgiving would these people be if their personal data were actually compromised? From Target to Macy’s, quite a few Americans have already had a taste of this, but what is it going to take for such “data pragmatism” to seem not so practical after all?
This past week, the European Union’s General Data Protection Regulation (GDPR) initiative kicked in. But what does it mean for businesses that operate in the EU region?
And what are the prospects for GDPR-like privacy coming to the USA anytime soon?
First off, let’s review what’s covered by the GDPR initiative. The GDPR includes the following rights for individuals:
The right to be informed
The right of access
The right to rectification
The right to be forgotten
The right to restrict processing
The right to data portability
The right to object
Rights in relation to automated decision making and profiling
The “right to be forgotten” means data subjects can request their information to be erased. The right to “data portability” is also a new factor. Data subjects now have the right to have data transferred to a third-party service provider in machine-readable format. However, this right arises only when personal data is provided and processed on the basis of consent, or when necessary to perform a contract.
Privacy impact assessments and “privacy by design” are now legally required in certain circumstances under GDPR, too. Businesses are obliged to carry out data protection impact assessments for new technologies. “Privacy by design” involves accounting for privacy risk when designing a new product or service, rather than treating it as an afterthought.
Implications for Marketers
A recent study investigated how much customer data will still be usable after GDPR provisions are implemented. Research was done involving more than 30 companies that have already gone through the process of making their data completely GDPR-compliant.
The sobering finding: Nearly 45% of EU audience data is being lost due to GDPR provisions. One of the biggest changes is that cookie IDs disappear, which is the basis behind so much programmatic and other data-driven advertising both in Europe and in the United States.
Doug Stevenson, CEO of Vibrant Media, the contextual advertising agency that conducted the study, had this to say about the implications:
“Publishers will need to rapidly fill their inventory with ‘pro-privacy’ solutions that do not require consent, such as contextual advertising, native [advertising] opportunities and non-personalized ads.”
New platforms are emerging to help publishers manage customer consent for “privacy by design,” but the situation is sure to become more challenging in the ensuing months and years as compliance tracking the regulatory authorities ramps up.
It appears that some companies are being a little less proactive than is advisable. A recent study by compliance consulting firm CompliancePoint shows that a large contingent of companies, simply put, aren’t ready for GDPR.
As for why they aren’t, nearly half report that they’re taking a “wait and see” attitude to determine what sorts of enforcement actions ensue against scofflaws. Some marketers admit that their companies aren’t ready due to their own lack of understanding of GDPR issues, while quite a few others claim simply that they’re unconcerned.
I suspect we’re going to get a much better understanding of the implications of GDPR over the coming year or so. It’ll be good to check back on the status of implementation and enforcement measure by this time next year.
Perhaps it’s the rash of daily reports about data breaches. Or the one-too-many compromises of protection of people’s passwords.
Whatever the cause, it appears that Americans are becoming increasingly interested in the use of biometrics to verify personal identity or to enable payments.
And the credit card industry has taken notice. Biometrics – the descriptive term for body measurements and calculations – is becoming more prevalent as a means to authenticate identity and enable proper access and control of accounts.
A recent survey of ~1,000 American adult consumers, conducted in Fall 2017 by AYTM Marketing Research for VISA, revealed that two-thirds of the respondents are now familiar with biometrics.
What’s more, for those who understand what biometrics entails, more than 85% of the survey’s respondents expressed interest in their use for identity authentication.
About half of the respondents think that adopting biometrics would be more secure than using PIN numbers or passwords. Even more significantly, ~70% think that biometrics would make authentication faster and easier – whether it be done via voice recognition or by fingerprint recognition.
Interestingly, the view that biometrics are “easier” than traditional methods appears to be the case despite the fact that fewer than one-third of the survey respondents use unique passwords for each of their accounts.
As a person who does use unique passwords for my various accounts – and who has the usual “challenges” managing so many different ones – I would have thought that people who use only a few passwords might find traditional methods of authentication relatively easy to manage. Despite this, the “new world” of biometrics seems like a good bet for many of these people.
That stated, it’s also true that people are understandably skittish about ID theft in general. To illustrate, about half of the respondents in the AYTM survey expressed concerns about the risk of a security breach of biometric data – in other words, that the very biometric information used to authenticate a person could be nabbed by others who could use it the data for nefarious purposes.
And lastly, a goodly percentage of “Doubting Thomases” question whether biometric authentication will work properly – or even if it does work, whether it might require multiple attempts to do so.
In other words, it may end up being “déjà vu all over again” with this topic …
For an executive summary of the AYTM research findings, click or tap here.
I don’t know about you, but I’ve never forgotten the late afternoon of August 14, 2003 when problems with the North American power grid meant that people in eight states stretching from New England to Detroit suddenly found themselves without power.
Fortunately, my company’s Maryland offices were situated about 100 miles beyond the southernmost extent of the blackout. But it was quite alarming to watch the power outage spread across a map of the Northeastern and Great Lakes States (plus Ontario) in real-time, like some sort of creeping blob from a science fiction film.
“Essential services remained in operation in some … areas. In others, backup generation systems failed. Telephone networks generally remained operational, but the increased demand triggered by the blackout left many circuits overloaded. Water systems in several cities lost pressure, forcing boil-water advisories to be put into effect. Cellular service was interrupted as mobile networks were overloaded with the increase in volume of calls; major cellular providers continued to operate on standby generator power. Television and radio stations remained on the air with the help of backup generators — although some stations were knocked off the air for periods ranging from several hours to the length of the entire blackout.”
Another (happier) thing I remember from this 15-year-old incident is that rather than causing confusion or bedlam, the massive power outage brought out the best in people. This anecdote from the blackout was typical: Manhattanites opening their homes to workers who couldn’t get to their own residences for the evening.
For most of the 50 million+ Americans and Canadians affected by the blackout, power was restored after about six hours. But for some, it would take as long as two days for power restoration.
Upon investigation of the incident, it was discovered that high temperatures and humidity across the region had increased energy demand as people turned on air conditioning units and fans. This caused power lines to sag as higher currents heated the lines. The precipitating cause of the blackout was a software glitch in the alarm system in a control room of FirstEnergy Corporation, causing operators to be unaware of the need to redistribute the power load after overloaded transmission lines had drooped into foliage.
In other words, what should have been, at worst, a manageable localized blackout cascaded rapidly into a collapse of the entire electric grid across multiple states and regions.
But at least the incident was borne out of human error, not nefarious motives.
That 2003 experience should make anyone hearing last week’s testimony on Capitol Hill about the risks faced by the U.S. power grid think long and hard about what could happen in the not-so-distant future.
The bottom-line on the testimony presented in the hearings is that malicious cyberattacks are becoming more sophisticated – and hence more capable of causing damage to American infrastructure. The Federal Energy Regulatory Commission (FERC) is cautioning that hackers are increasingly threatening U.S. utilities ranging from power plants to water processing systems.
Similar warnings come from the Department of Homeland Security, which reports that hackers have been attacking the U.S. electric grid, power plants, transportation facilities and even targets in commercial sectors.
The Energy Department goes even further, reporting in 2017 that the United States electrical power grid is in “imminent danger” from a cyber-attack. To underscore this threat, the Department contends that more than 100,000 cyber-attacks are being mounted every day.
With so many attacks of this kind happening on so many fronts, one can’t help but think that it’s only a matter of time before we face a “catastrophic event” that’s even more consequential than the one that affected the power grid in 2003.
Even more chilling, if it’s borne out of intentional sabotage – as seems quite likely based on recent testimony – it’s pretty doubtful that remedial action could be taken as quickly or as effectively as what would be done in response to an accidental incident likr the one that happened in 2003.
Put yourself in the saboteurs’ shoes: If your aim is to bring U.S. infrastructure to its knees, why plan for a one-off event? You’d definitely want to build in ways to cause cascading problems – not to mention planting additional “land-mines” to frustrate attempts to bring systems back online.
Contemplating all the implications is more than sobering — it’s actually quite frightening. What are your thoughts on the matter? Please share them with other readers.
Recently I read a column in MediaPost by contributing writer Ted McConnell (The Block Chain [sic]Gang, 3/8/18) that seemed pretty wide of the mark in terms of some of its (negative) pronouncements about blockchain technology.
Nelson cited four key areas where he disagreed with the positions of the column’s author – particularly in the fallacy of conflating blockchain technology – the system of keeping records which are impossible to falsify – with cryptocurrencies (systems for trading assets).
Interested in learning more about this larger issue, I asked Nelson to expand on this aspect. Presented below is what he reported back to me. His perspectives are interesting enough, I think, to share with others in the world of business.
Hyperbole All Over Again?
Most folks in the technology world – and many business professionals outside of it – are familiar with the “hype cycle.” It’s a branded, graphic representation of the maturity, adoption and social application of technologies from Gartner, Inc., a leading IT research outfit.
According to Gartner, the hype cycle progresses through five phases, starting with the Technology Trigger:
“A potential technology breakthrough kicks things off. Early proof-of-concept stories and media interest trigger significant publicity. Often no usable products exist and commercial viability is unproven.”
Next comes the Peak of Inflated Expectations, which implies that everyone is jumping on the bandwagon. But Gartner is a bit less sanguine:
“Early publicity produces a number of success stories — often accompanied by scores of failures. Some companies take action; most don’t.”
There follows a precipitous plunge into the Trough of Disillusionment. Gartner says:
“Interest wanes as experiments and implementations fail to deliver. Producers of the technology shake out or fail. Investment continues only if the surviving providers improve their products to the satisfaction of early adopters.”
If the hype cycle is to be believed, the Trough of Disillusionment cannot be avoided; but from a technology provider’s perspective it seems a dreadful place to be.
With nowhere to go but up, emerging technologies begin to climb the Slope of Enlightenment. Gartner explains:
“More instances of how the technology can benefit the enterprise start to crystallize and become more widely understood. Second- and third-generation products appear from technology providers. More enterprises fund pilots; conservative companies remain cautious.”
Finally, we ascend to the Plateau of Productivity. Gartner portrays a state of “nirvana” in which:
“Mainstream adoption starts to take off. Criteria for assessing provider viability are more clearly defined. The technology’s broad market applicability and relevance are clearly paying off. If the technology has more than a niche market, then it will continue to grow.”
Gartner publishes dozens of these “hype cycles,” refreshing them every July or August. They mark the progression of specific technologies along the curve – as well as predicting the number of years to mainstream adoption.
Below is an infographic I’ve created which plots Gartner’s view for cloud computing overlaid by a plot of global public cloud computing revenues during that time, as reported by Forrester Research, another prominent industry observer.
This infographic provides several interesting insights. For starters, cloud computing first appeared on Gartner’s radar in 2008. In hindsight that seems a little late — especially considering that Salesforce.com launched its first product all the way back in 2000. Amazon Web Services (AWS) first launched in 2002 and re-launched in 2006.
Perhaps Gartner was paying more attention to Microsoft, which announced Azure in 2008 but didn’t release it until 2010. Microsoft, Amazon, IBM and Salesforce are the top four providers today, holding ~57% of the public cloud computing market between them.
Cloud computing hit the peak of Gartner’s hype cycle just one year later, in 2009, but it lingered at or near the Peak of Inflated Expectations for three full years. All the while, Gartner was predicting mainstream adoption within 2 to 5 years. Indeed, they have made the same prediction for ten years in a row (although I would argue that cloud computing has already gained mainstream market adoption).
It also turns out that the Trough of Disillusionment isn’t quite the valley of despair that Gartner would have us believe. In fact, global public cloud revenues grew from $26 billion during 2011 to $114 billion during 2016 — roughly half the 64% compound annual growth rate (CAGR) during the Peak of Inflated Expectations, but a respectable 35% CAGR nonetheless.
Indeed, there’s no evidence here of waning market interest – nor did an industry shakeout occur. With the exception of IBM, the leading producers in 2008 remain the leading producers today.
All in all, it seems the hype curve significantly lagged the revenue curve during the plunge into the Trough of Disillusionment.
… Which brings us to blockchain technology. Just last month (February 2018), Gartner Analyst John Lovelock stated, “We are in the first phase of blockchain use, what I call the ‘irrational exuberance’ phase.” The chart below illustrates shows how Gartner sees the “lay of the land” currently:
This suggests that blockchain is at the Peak of Inflated Expectations, though it appeared ready to jump off a cliff into the Trough of Disillusionment in Gartner’s 2017 report for emerging technologies. (It wasn’t anywhere on Gartner’s radar before 2016.)
Notably, cryptocurrencies first appeared in 2014, just past the peak of the cycle, even though the first Bitcoin network was established in 2009. Blockchain is the distributed ledger which underpins cryptocurrencies and was invented at the same time.
It’s also interesting that Gartner doesn’t foresee mainstream adoption of blockchain for another 5 to 10 years. Lovelock reiterated that point in February, reporting that “Gartner does not expect large returns on blockchain until 2025.”
All the same, blockchain seems to be progressing through the curve at three times the pace of cloud computing. If cloud computing’s history is any guide, blockchain provider revenues are poised to outperform the hype curve and multiply into some truly serious money during the next five years.
It’s reasonable to think that blockchain has been passing through a phase of “irrational exuberance” lately, but it’s equally reasonable to believe that industry experts are overly cautious and a bit late to the table.
________________________
So that’s one specialist’s view. What are your thoughts on where we are with blockchain technology? Please share your perspectives with other readers here.
The Internet of Things is making major headway in consumer product categories — but it turns out it’s bringing its share of headaches along for the ride.
It shouldn’t be particularly surprising that security could be a potential issue around IoT, of course. But recent evaluations point to the incidence being more significant than first thought.
That’s the conclusion of research conducted by management consulting firm Altman Vilandrie & Company. Its findings are based on a survey of ~400 IT decision-makers working at companies that have purchased some form of IoT security solutions.
According to the Vilandrie survey, approximately half of the respondents reported that they have experienced at least one IoT-related security intrusion or breach within the past two years. The companies included in the research range across some 19 industry segments, so the issue of security doesn’t appear to be confined to one or two sectors alone.
What’s more, smaller firms experienced higher relative “pain” caused by a security breach. In the Vilandrie survey, companies with fewer than $5 million in annual revenues reported an average loss of $255,000 associated with IoT security breaches.
While that’s substantially lower in dollar amount to the average loss reported by large companies, the loss for small business as a percentage of total revenues is much greater.
More findings from the Altman Vilandrie research study can be accessed here.
It’s common knowledge by now that the data breach at credit reporting company Equifax earlier this year affected more than 140 million Americans. I don’t know about you personally, but in my immediate family, it’s running about 40% of us who have been impacted.
And as it turns out, the breach occurred because one of the biggest companies in the world — an enterprise that’s charged with collecting, holding and securing the sensitive personal and financial data of hundreds of millions of people — was woefully ill-prepared to protect any of it.
How ill-prepared? The more you dig around, the worse it appears.
Since my brother, Nelson Nones, works every day with data and systems security issues in his dealings with large multinational companies the world over, I asked him for his thoughts and perspectives on the Equifax situation.
What he reported back to me is a cautionary tale for anyone in business today – whether you’re working in a big or small company. Nelson’s comments are presented below:
Background … and What Happened
According to Wikipedia, “Equifax Inc. is a consumer credit reporting agency. Equifax collects and aggregates information on over 800 million individual consumers and more than 88 million businesses worldwide.”
Founded in 1899, Equifax is one of the largest credit risk assessment companies in the world. Last year it reported having more than 9,500 employees, turnover of $3.1 billion, and a net income of $488.1 million.
On September 8, 2017, Equifax announced a data breach potentially impacting 143 million U.S. consumers, plus anywhere from 400,000 to 44 million British residents. The breach was a theft carried out by unknown cyber-criminals between mid-May 2017 until July 29, 2017, which is when Equifax first discovered it.
It took another 4 days — until August 2, 2017 — for Equifax to engage a cybersecurity firm to investigate the breach.
Equifax has since confirmed that the cyber-criminals exploited a vulnerability of Apache Struts, which is an open-source model-view-controller (MVC) framework for developing web applications in the Java programming language.
The specific vulnerability, CVE-2017-5638, was disclosed by Apache in March 2017, but Equifax had not applied the patch for this vulnerability before the attack began in mid-May 2017.
The workaround recommended by Apache back in March consists of a mere 27 lines of code to implement a Servlet filter which would validate Content-Type and throw away requests with suspicious values not matching multipart/form-data. Without this workaround or the patch, it was possible to perform Remote Code Execution through a REST API using malicious Content-Type values.
Subsequently, on September 12, 2017, it was reported that a company “online portal designed to let Equifax employees in Argentina manage credit report disputes from consumers in that country was wide open, protected [sic] by perhaps the most easy-to-guess password combination ever: ‘admin/admin’ … anyone authenticated with the ‘admin/admin’ username and password could … add, modify or delete user accounts on the system.”
Existing user passwords were masked, but:
“… all one needed to do in order to view [a] password was to right-click on the employee’s profile page and select ‘view source’. A review of those accounts shows all employee passwords were the same as each user’s username. Worse still, each employee’s username appears to be nothing more than their last name, or a combination of their first initial and last name. In other words, if you knew an Equifax Argentina employee’s last name, you also could work out their password for this credit dispute portal quite easily.”
The reporter who broke this story contacted Equifax and was referred to their attorneys, who later confirmed that the Argentine portal “was disabled and that Equifax is investigating how this may have happened.”
The Immediate Impact on Equifax’s Business
In the wake of these revelations, Equifax shares fell sharply: 15% on September 8, 2017, reducing market capitalization (shareholder value) by $3.97 billion in a single trading day.
Over the next 5 trading days, shares fell another 24%, reducing shareholder value by another $5.4 billion.
What this means is that the cost of the breach, measured in shareholder value lost by the close of business on September 15, 2017 (6 business days), was $9.37 billion – which is equivalent to the entire economic output of the country of Norway over a similar time span.
This also works out to losses of $347 million per line of code that Equifax could have avoided had it deployed the Apache Struts workaround back in March 2017.
The company’s Chief Information Officer and Chief Security Officer also “retired” on September 15, 2017.
Multiple lawsuits have been filed against Equifax. The largest is seeking $70 billion in damages sustained by affected consumers. This is more than ten times the company’s assets in 2016, and nearly three times the company’s market capitalization just before the breach was announced.
The Long-Term Impact on Equifax’s Brand
This is yet to be determined … but it’s more than likely the company will never fully recover its reputation. (Just ask Target Corporation about this.)
Takeaway Points for Other Companies
If something like this could happen at Equifax — where securely keeping the private information of consumers is the lifeblood of the business — one can only imagine the thousands of organizations and millions of web applications out there which are just as vulnerable (if not as vital), and which could possibly destroy the entire enterprise if compromised.
At most of the companies I’ve worked with over the past decade, web application development and support takes a back seat in terms of budgets and oversight compared to so-called “core” systems like SAP ERP. That’s because the footprint of each web application is typically small compared to “core” systems.
Of necessity, due to budget and staffing constraints at the Corporate IT level, business units have haphazardly built out and deployed a proliferation of web applications — often “on the cheap” — to address specific and sundry tactical business needs.
“Kid’s Day” at Equifax’s Argentine offices. Were the kids in command there, one is tempted to wonder …
I strongly suspect the Equifax portal for managing credit report disputes in Argentina — surely a backwater business unit within the greater Equifax organization — was one of those.
If I were a CIO or Chief Security Officer right now, I’d either have my head in the sand, or I’d be facing a choice. I could start identifying and combing through the dozens or hundreds of web applications currently running in my enterprise (each likely to be architecturally and operationally different from the others) to find and patch all the vulnerabilities. Or I could throw them all out, replacing them with a highly secure and centrally-maintainable web application platform — several of which have been developed, field-tested, and are readily available for use.
__________________________
So, there you have it from someone who’s “in the arena” of risk management every day. To all the CEOs, CIOs and CROs out there, here’s your wakeup call: Equifax is the tip of the spear. It’s no longer a question of “if,” but “when” your company is going to be attacked.
And when that attack happens, what’s the likelihood you’ll be able to repel it?
… Or maybe it’ll be the perfect excuse to make an unforeseen “early retirement decision” and call it a day.
__________________________
Update (9/25/17): And just like clockwork, another major corporation ‘fesses up to a major data breach — Deloitte — equally problematic for its customers.