Source: http://ansonalex.com/infographics/big-data-analytics-profitability-statistics-infographic/
Friday, 26 April 2013
Thursday, 25 April 2013
Wednesday, 24 April 2013
Cordray defends consumer bureau big-data effort - Business - The Boston Globe
By Carter Dougherty
| BLOOMBERG NEWS APRIL 24, 2013
WASHINGTON — The Consumer Financial Protection Bureau’s data-collection program does not invade individuals’ privacy, is a vital regulatory tool, and mimics techniques in wide use by the private sector, the agency’s director said Tuesday.
‘‘The big banks know more about you than you know about yourself,’’ director Richard Cordray told the Senate Banking Committee. At a hearing, he faced criticism about the initiative.
The bureau is collecting data to improve its rule-writing and supervisory work, sometimes through purchasing it from credit bureaus like Experian. It is also requesting large amounts of information from major banks in areas such as checking account overdrafts and credit cards.
Senator Mike Johanns, Republican of Nebraska, highlighted potential violations of privacy.‘‘
‘‘To many people, this is going to sound downright creepy,’’ Johanns said. ‘‘People are going to be bothered by the fact that there’s this federal agency that’s collecting data on the behavior of people like you and me and everybody else who’s paying off a mortgage, who’s paying credit card bills every month.’’
Cordray said the bureau does not monitor people and guards data carefully. Consumers, though, do sometimes share personally identifiable information, as when they file complaints.
‘‘The notion that we’re tracking individual consumers or invading their privacy is quite wrong,’’ Cordray said.
Cordray emphasized the bureau needs this information to do the work, such as cost-benefit analysis, that Congress has ordered. He also said the private sector has fully embraced data analytics, often labeled ‘‘Big Data.’’
Republican Senator Michael Crapo of Idaho complained that he ‘‘specifically’’ asked the agency about its data collection last month, but learned more about the subject from an April 17 article published by Bloomberg News.
‘‘The responses I received downplayed the nature and extent of the issue,’’ Crapo said.
Crapo said that private-sector data collection is fundamentally different from the bureau’s activities. ‘‘Nobody in the private sector has the power that the federal government has,’’ Crapo said. ‘‘The power of the government is behind this data collection.’’
Cordray said the Federal Reserve sometimes purchases data for its own work on consumer finances. The goal is to establish a starting point for where the market stands, he said. ‘‘There are times where as we examine institutions, we have to begin by getting a baseline of data to calibrate what is going on,’’ he said.
In its first 21 months, the bureau has secured $425 million in relief for 6 million consumers wronged by financial service providers, Cordray said.
‘‘We also imposed penalties on the companies to deter such activity in the future,’’ Cordray said.
The bureau, created by the 2010 Dodd-Frank Act, has taken enforcement actions in areas including credit-card add-on products such as credit monitoring, and mortgage insurance.
Tuesday, 23 April 2013
With Big Data, Context is a Big Issue | Innovation Insights | Wired.com
- ALISSA LORENTZ |
- 04.23.13
In the war on noise, contextual applications serve as crucial ammunition.
The Age of Content
With all the hype around Big Data, we’ve become extremely proficient at collecting data – be it from enterprise systems, documents, social interactions, or e-mail and collaboration services. The expanding smorgasbord of data collection points are turning increasingly portable and personal, including mobile phones and wearable sensors, resulting in a data mining gold rush that will soon have companies and organizations accruing Yottabytes (10^24) of data.
To put things into perspective, 1 Exabyte (10^18) of data is created on the internet daily, amounting to roughly the equivalent of data in 250 million DVDs. Humankind produces in two days the same amount of data it took from the dawn of civilization until 2003 to generate, and as the Internet of Things become a reality and more physical objects become connected to the internet, we will enter the Brontobyte (10^27) Era.
So it’s all dandy that we’re getting better and better at sucking up every trace of potential information out there, but what do we do with these mountains of data? Move over Age of Content, enter the Age of Context.
The Age of Context
Big Data has limited value if not paired with its younger and more intelligent sibling, Context. When looking at unstructured data, for instance, we may encounter the number “31” and have no idea what that number means, whether it is the number of days in the month, the amount of dollars a stock increased over the past week, or the number of items sold today. Naked number “31” could mean anything, without the layers of context that explain who stated the data, what type of data is it, when and where it was stated, what else was going on in the world when this data was stated, and so forth. Clearly, data and knowledge are not the same thing.
Take the example of a company that has invested heavily in business intelligence (BI) software that organizes internal data. In an increasingly connected world, this company has not leveraged its data to its potential. Why not? Because the company’s internal data is isolated from the rest of the data universe including news, social media, blogs, and other relevant sources. What if you could join the dots between all these data sources and surface hidden connections?
Using an application that, unlike the average BI product, pulls in information from the entire data universe would help the company answer questions like “Why did our sales plummet last month?” as opposed to just “What happened to our sales figures last month?”
The company could further question, “Did our slip in sales have anything to do with the recent elections and uncertainty?” and “How can we make sure our sales do not slip next time there is a shift in the political landscape?” Identifying potential causality is key in spotting patterns that enable prediction.
For organizations and businesses to survive today, they have to contextualize their data. Just as a doctor diagnosing a patient with diabetes based on body temperature alone is incorrect, so is making business decisions derived from data out of context. A doctor needs to know about the patient’s age, lifestyle, diet, weight, family history, and more in order to make a probable and guarded diagnosis and prognosis. Contextualization is crucial in transforming senseless data into real information – information that can be used as actionable insights that enable intelligent corporate decision-making.
At the end of the day, our overworked minds want to be spoon-fed insights. We want key signals and tightly packaged summaries of relevant, intriguing information that will arm us with knowledge and augment our intelligence.
But how do we extract real intelligence from data?
Extract Real Intelligence From Data
Besides cross-referencing internal data with a plethora of other sources, we need algorithms to boil off the noise and extract the signals, or real human meaning, from the data. What do we mean? Let’s say you have a million tweets from New York City on the eve of the U.S. elections and rather than read them all, you want to know quickly how people are feeling based on these tweets.
To do this you must apply complex algorithms derived from machine learning, computational linguistics, and natural language processing to harvest the key words and corresponding emotions from the tweets. Then you get the key signals: Are people feeling anxious? Hopeful? Confident? Fearful? This is precisely what we do here at Augify: Paving the way for the future of understanding by surfacing signals from the noise in our cloud-based, algorithm-packed product. We wrap the signals in a dashboard of slickly designed, color-coded gauges and visualizations that enhance your understanding of key insights.
Getting Contextual
So it’s great that applications out there can gather and analyze data, detect human-based meaning from it, and visualize it all, but any application is limiting itself if it is only useful once you open the application and enter a query. We wanted to go beyond this. That’s why we decided to get contextual on you and increase our technology’s reach. We go beyond contextual data by developing contextual applications – we wanted to put data to good use by applying it to real-life situations in our daily lives.
The Don of ubiquitous computing, Mark Weiser, stated in 1991 that “the most profound technologies are those that disappear. They weave themselves into the fabric of our everyday life until they are indistinguishable from it,” much like the telephone or electricity. They have seeped into our surroundings, playing an integral role in our everyday lives.
Modern day examples include adaptive technologies such as Google Now, which tracks your online behavior and uses this data to predict the information that you will need, such as local traffic or weather updates. Similarly, “learning thermostat” Nest self-adjusts your home’s temperature based on your activity, saving energy usage and bills. Pervasive technologies mean they are everywhere and nowhere at the same time – an invisible layer listening to your actions in order to be more helpful.
Along the same vein, we believe in pervasive Augmented Intelligence, meaning that you can reap the benefits of Augify’s signal-detecting technology anywhere you go in the connected world. So even when you are working outside of the Augify application screen, be it writing an e-mail or reading a website, Augify’s algorithms will work silently and invisibly in the background, culling key insights from the text you are reading at that very moment and feeding them to you in a pop-up window.
Pervasive Augmented Intelligence
Imagine being able to save time reading long text by getting a synopsis of the key themes, topics, emotions, people, and entities in a news article? By seeing the emotions, sentiments, intentions, and credibility in someone’s Twitter feed, and by spotting the key influencers in their network? What if, as you typed an e-mail to your boss about a new enterprise software, you received pop-up recommendations of credible reference articles related to said software?
In other words, what if you could detect the important signals in any web content in real-time? With Augify’s contextual application, this is truly possible, enabling you to augment your intelligence anywhere your online activity takes you.
Augment Your Intelligence Anywhere Your Online Activity Takes You
The Age of Context demands that contextual data be applied to everyday situations in useful ways. How do we make use of this data? Since we’ve gotten good at collecting data, now it’s all about putting it into context and making sense out of it – mining for the nuggets of insights that answer the “So What?” question. Data is meaningless and even cumbersome without context – the key holistic and interpretive lens through which data is filtered and turned into real information.
Surprise: You’re using ‘Big Data’- MSN Money
Big Data is the popular name for the field of data analytics, the process of sifting through massive amounts of information with large numbers of variables in split seconds to produce results that can be used make evidence-based decisions.
Think statistics blogger Nate Silver accurately predicting Obama's win in the 2012 presidential election, or IBM's Watson computer trouncing two human "Jeopardy" champs -- that's Big Data.
The bits of data generated by websites, social networks and smartphones apps are creating a "nuclear explosion" of information, as a recent Wired report called it. Combined with the power of relatively low-cost cloud-based software and digital storage, it's contributing to a global mountain of data that in 2013 is expected to reach an unprecedented 1.2 zettabytes, according to a report from Tableau Software. That's equal to 1 billion terabytes.
Companies and other organization are jumping into big data with abandon. It's being used by everyone from health insurance carriers to analyze bills for fraud to police departments to analyze past crime patterns to better predictwhere future thefts or robbers are most likely to occur.
The average Joe and Jane are benefiting because retailers and other companies are using big data to offer more personalized services. People have always put a high value on one-on-one interactions, and the experience that Big Data offers "simulates a parallel world online that's as personalized and trustworthy as offline," says Vikas Sabnani, chief data scientist with Glassdoor.com, the jobs and careers website. "Big data is the only way we can get there."
Big data, no price tag
Many big data-based services for shoppers, job seekers and other consumers are free.
Job hunters pay nothing to use Glassdoor to search through millions of job openings, see salary information for specific jobs and read reviews of what it's like to work at myriad companies.
Companies post the job openings. The rest of the information comes from reviews or data on salaries, bonuses and commissions that job hunters share with the site. Glassdoor compiles and analyzes all that data to provide job hunters with as many specifics as possible when they search on the site. If someone who's interested in software engineer jobs does a search on the term and then clicks on a job at Google, Glassdoor takes that as feedback that they want to work in the tech industry or at a large company like Google, and uses the information to show ratings and reviews of companies they might be interested in.
Glassdoor also lets job seekers link to their Facebook accounts to see if any of their friends or connections work at any of the companies. All of it is possible because Big Data doesn't take as much computing power or money as it would have only a short time ago, Sabnani says. "The cloud makes it much cheaper for us to process and scale as opposed to hosting it on our own servers."
Amazon is a long-time proponent of Big Data, which runs the recommendations engine on its online bookstore. Browse through the bookstore and you'll see recommendations pop up for novels or other books you might be interested in. The company analyzes stored data on books you've previously looked up on the site, and compares that with data from customers who've either looked up or bought the same books. Based on the information, the website shows you books that those customers have looked at or purchased that might appeal to you.
Not all big-data services are free. A recommendation service on Netflix is only available to people who pay the $7.99 monthly subscription. If they do, they can opt to use the online entertainment company's Taste Profile tool to rate films and TV shows with up to five stars. The company then compiles that data with billions of ratings from its other 36 million members to suggest other shows or films a subscriber might want to watch. The more shows or movies you watch and rate, the better the suggestions get. "It's the network effect," says Phil Simon, author of the just-published book, "Too Big to Ignore: The Business Case for Big Data."
Mind-boggling amounts of data
Companies such as Google, Twitter and Facebook are built on Big Data, Simon says. Google uses it in multiple information products, from its ubiquitous search engine to Google News Alerts, a news finding service that can be set up to monitor specific subjects and have updates sent to an email address on a real-time, daily or weekly basis.
Facebook uses Big Data to analyze what users "like" or share, and then uses that to sell ad space on a person's timeline or newsfeed. Simon, a long-time Rush devotee, likes the rock band's Facebook page and takes other actions that show he's a fan. If a music promoter was trying to fill seats at a Rush concert in Simon's hometown of Henderson, Nev., it wouldn't surprise him to see an ad for tickets on his Facebook page. "The amount of data Facebook has on people is mind boggling," he says.
Twitter’s Trends feature is also built on Big Data. When someone on Twitter writes a status update with a hashtag-labeled keyword signifying a topic the update is related to, and agrees to share their location, the company combines the information with other users' updates, hashtags and locations and shares what topics are trending in specific locations or throughout the service.
Big Data’s Usability Problem - Bill Wise - Voices - AllThingsD
Bill Wise, CEO, Mediaocean | APRIL 23, 2013
Sen. Lindsay Graham just told Fox News that the reason the FBI never realized that Boston Marathon bombing suspect Tamerlan Tsarnaev went to Russia in 2011 is that “when he got on the Aeroflot plane, they misspelled his name, so it never went into the system that he actually went to Russia.” Meanwhile, the Reinhart-Rogoff paper that has been a catalyst for government austerity policies worldwide since 2010 has, in fact, accidentally left out several countries’ worth of critical data in Excel.
As one blogger sums up scathingly: “One of the core empirical points providing the intellectual foundation for the global move to austerity in the early 2010s was based on someone accidentally not updating a row formula in Excel.”
Taken together, these factors offer a critical lesson here about the power and limits of Big Data today. In both scenarios, data management tools (i.e., the FBI’s systems and Excel) were undone by fairly simple errors: In one situation, a misspelling; in another, a failure to code a spreadsheet properly. And in both scenarios, the results were dire — an awful tragedy, and a potentially misdirected government economic policy in the midst of a recession.
As someone who spends day and night thinking through data management and workflow, these two stories lead me to three observations:
- As a society, we’re hugely reliant on data management platforms for our most critical information.
- Our core data platforms often aren’t set up to handle human error, from basic coding flaws to spelling mistakes.
- The wealth of data in our data tools can mask that human error. Consider: TheReinhart-Rogoff study examined “new data on forty-four countries spanning about two hundred years” with “over 3,700 annual observations covering a wide range of political systems, institutions, exchange rate arrangements, and historic circumstances.”
In such a wide sea of data, a few lines of code can be very easy to overlook, even if they have strong ramifications for analysis.
There are lots of things to take away from these three points, but I’ll just focus on one: The promise of Big Data is that it can make everyday processes — from critical analyses to mundane tasks — work smarter through data intelligence. Ultimately, all that data management translates into an economy and society that lets machines handle the minutiae as humans think through the larger picture.
To a large extent, that vision is already here. But at the same time, more human/data interaction means a lot more room for error (and inefficiency) around increasingly critical data sets — which, as we’ve seen, can have very serious results. Which means that, if we want to make the reality of Big Data match the dream, we need to spend serious time around providing usability that guides human users in the best way to engage with the data, and automation that takes human interaction (and human error) out of the picture for a lot of the basic calculations and tasks — and for some of the complicated ones, too.
If Big Data can’t fit hand-in-glove with usability and workflow, a lot of the promise of big data will be empty data crunching. That’s not just a problem for getting where we want to be in the evolution of computing. It’s a situation that can lead to bad data management — which translates into bad economics and, sometimes, far worse.
What big data means for marketing decision making | Media Network | guardian.co.uk
David Lloyd | Tuesday 23 April 2013
"Not everything that counts can be counted, and not everything that can be counted counts." This quote, often attributed to Albert Einstein, originated before the concept of big data. Yet now, more than ever, we should reflect on what it means in the context of how our marketing decisions are made.
The first part of the quote reveals an inconvenient truth of analytics; as much as we strive for a world where everything is measurable and all decisions can be based on real-time, infallible data, this is not always possible. Take "social media ROI" for instance; a term that pervades marketing industry commentary. Often, the data to attribute sales revenue confidently simply isn't there and no amount of white papers and positive thinking can fill the gap.
In such cases, it's important to understand the options quickly along with the associated cost, commitment and accuracy. Be prepared not to do it at all. Yes, really. That may seem controversial and may be something that doesn't fit with cultures that only approve projects based on hard commercial benefits, but it is better to focus efforts on what you can do than stretch data beyond credibility.
Instead, focus on more readily available measures that indicate performance against objectives. Have I got good reach across my target audience compared with other channels? Does my target audience even use this social platform? How do they engage with my content compared with the content from competitors? You may find that some of these measures, once properly benchmarked and tracked, along with a good dose of marketing nous, will lead to a well-informed social media strategy.
The second part of the quote provides a counter-perspective yet is equally important. We are told that big data is all around us; that we must collect it or be hopelessly behind the curve and lose huge amounts of revenue. Almost every day a new vendor, seminar or white paper springs up promising the solution to our big data problems. If anything, big data (the campaign) is an unintentional mirror of big data (the real stuff) itself; it creates more noise, more pitfalls and requires more precision to navigate. The amount of data available does not necessarily correlate with its value – not everything that can be counted counts.
Yet the solution to such concerns is similar to the solution when you don't have enough data. Align the data you need to your strategic objectives, assess the options and make the decision accordingly. This could be big data, small data or somewhere in-between data. These principles hold true now, as they did when customer relationship management (CRM) was marketing's next big thing, and before the quote itself was conceived.
We live in an environment where certain parties obfuscate some of these truths, whether they do that to sell technology or simply amplify the hype to appear to be leading the conversation.
I'm aware that I'm focusing on the fallibility of data; we can't take for granted that data can drive every decision, we must use also our judgment. Well, I'm prepared to be honest about this as, despite the limitations of data, far more decisions are made by instinct alone when data could have improved those decisions. This gap is a huge opportunity and to close it, we just need the foresight to collect the right data and ensure it is used correctly by people with the right skills.
If you're a business focusing on the big data question and how it might revolutionise your marketing, I advise you to take stock of your objectives. Is there a clear connection between your strategy and the data? Are you only planning to collect the data you really need (regardless of size)? Do you have a clear roadmap and the expertise to steer the course? If you can't provide a confident yes to these questions, make sure you rectify that before spending considerable time and money on solutions. Otherwise you could find yourself another wreck stranded on the rocks.
Monday, 22 April 2013
Phil Simon: Boston, Privacy, and Big Data
Phil Simon
The events last week were nothing short of horrible, although Bostonians and Americans together rejoiced on Friday as the second bombing suspect was captured. I for one shed a tear Friday night while watching citizens cheering police officers and government officials. You just don't see that every day.
Without question, the Brothers Tsarnaeva would not have been identified and apprehended so quickly-if at all-were it not for the scores of videos and pictures people willingly provided to authorities. You'll get no argument from me that we should have used every means at our disposal to apprehend these dangerous individuals.
Catching the bad guys is great, but in an ideal world federal, state, and local governments wouldalways prevent attacks like these from happening in the first place. Unfortunately, that's just not a likely scenario, although Big Data and powerful analytic tools should help us reduce the likelihood of (successful) future terrorist attacks.
In Too Big to Ignore, I address the far-reaching privacy and security implications of Big Data. Yes, good old-fashioned police work still matters, as events of last week proved. At the same time, though, government agencies and officials have at their disposal much better technologies and tools than even ten years ago. They could not realistically mine billions of call records, search emails, penetrate networks with geolocation data, etc. Today, these limitations cease to exist. Authorities can effectively use the data generated by suspected and real terrorists and criminals as a weapon. And here's where privacy implications start to matter.
The sticky ethical conundrum involves drawing the line. Should Internet service providers, cell phone carriers, and email services give government agencies user data? Should Amazon let the FBI and CIA in on who is buying what books? Should Google flag certain videos and report stats (along with individually identifiable information) to authorities?
I don't have the answers to these questions. According to one school of thought, no one has a right to know which videos people have watched, emails we've sent, and calls we've made. On the other end of the spectrum, if the government can use this information to prevent another Boston-like tragedy and you have nothing to hide, then isn't a little sacrifice of personal privacy in order?
In an era of Big Data, events like the Boston Marathon bombings make answering this question more important than ever.
Unlock the creativity of your security team with big data | Media Network | Guardian Professional
Mark Seward | Monday 22 April 2013
When we talk about data security, there's an increasingly popular adage that there are two types of companies: those that have been breached and those that don't know they've been breached.
The reason that the latter type of company exists is because attacks on enterprise infrastructure have become both more persistent and more sophisticated. In other words, unless the company has bullet-proof security processes and technology in place, it's almost inevitable that some form of attack will have penetrated the enterprise perimeter, whether the company realises it or not.
Once an attacker or piece of malware is inside the network, it can often lurk unseen among the mass of data that enterprise systems generate and trying to locate it, even if you're aware that an attack has taken place is extremely difficult.
That's why the new frontier of enterprise security is big data and statistical analysis specifically in machine data. Every interaction with a 'machine' – whether it's a website, mobile device, application server, corporate network, sensor or electronic tag, and whether it's automatically generated or a manual transaction – leaves a trail and a record.
In this new world of security, chief security officers (CSOs) and IT teams have to unlearn their over-reliance on traditional data protection technologies such as anti-virus software, firewalls and security information and event-management (SIEM) systems. Instead, the tools which might be most useful in the future include a statistics textbook, a subscription to Psychology Today, and the security professional's own brain.
Over the past few years, three major issues have come to dominate the current enterprise security landscape. First, the enterprise is now under pretty much constant attack, with common sense dictating that this increases the likelihood of a breach taking place. The attacker only has to be right once, while the IT team has to be right all the time. This being the case, relying on an attacker to alert you to their presence by tripping a specific rule that you've set in a SIEM clearly isn't a sufficient or reliable strategy.
Second, attackers understand that the pressure on IT teams has increased enormously as a result and with much of their attention and resources dedicated to protecting the perimeter, companies don't tend to do a great job at monitoring what's happening inside the network. As such, if they make it past the barricades, attackers can become very difficult to spot and can act with impunity.
Finally, and perhaps most worrying for the security industry as a whole, this constant bombardment has turned security into a reactive, administrative role, where team members are just responding to systems alerts rather than thinking more laterally about threats. Security is an exciting industry to be in, but too often, both seasoned professionals and new entrants aren't being challenged to be creative and come up with new ways to defend the enterprise – they're just doing what they are told to do by the tools they use.
So what's the solution to these problems? It's about having much greater oversight of everything that's happening inside the enterprise and developing operational intelligence. It's also about giving security professionals the right tools to quickly analyse and sift through enterprise data sets that include data generated through normal interaction with IT systems – right down to the machine data – in order to identify unusual patterns and abnormal behaviour in that data which might indicate that an attack is taking place.
It's here that statistics, psychology and old-fashioned brainpower really come into their own.
Big data analysis technologies exist that can help to identify possible anomalies, but it still requires human insight and intelligence to interpret what they might mean.
For example, there are giveaways that malware might be trying to compromise your network available in the machine data logs – these include URL strings that might be two to three times longer than normal, indicating the possible presence of command and control instructions attempting to launch a web protocol attack, or it might be the 'tell' of a network access password being entered 10 times faster than it's possible for a human to type.
In addition to these web protocol attacks, there are other examples of data logs that track human interactions with IT systems and facilities that might raise the security teams' suspicions. Why is a user repeatedly trying to access a file they don't have permission to view, or why has their ID card been used to enter the office when they're meant to be on holiday in the Bahamas?
Achieving this level of operational intelligence not only opens up new possibilities for how companies defend themselves against the myriad security threats that they face, but also re-engages the interest and creativity of the IT teams entrusted with this vital task.
The days of rules-based security engines looking for known threats are drawing to a close, as they're simply not built to handle the volume and sophistication of attacks today. To truly understand the nature of the threats they face, companies need to move beyond traditional approaches to security and delve deeper into the machine data being generated every second of every day.
Mark Seward is senior director of security and compliance at Splunk
Friday, 19 April 2013
James Grundvig: Making Big Data Small (and Accessible): Interview With SiSense CEO Amit Bendov
Two weeks after the GigaOm Conference on Structure Data in New York, Gerstein Fisher held its annual Real Talk series. The subject of the investment management firm's lecture followed GigaOm's footsteps, as if by design.
"Knowledge Redefined: How Technology is Changing the Way We Learn, What We Know, and How We Invest," centered on the lecture of David Weinberger, a Harvard researcher and bestselling author.
After Weinberger showed a pyramid with four levels of intelligence -- data, information, knowledge, and wisdom at the top -- to the packed Times Center auditorium, he went on a fast-paced tear explaining the eccentricities, broken paradigms, and counter-intuitive insights into three of the four levels.
He punted on the fourth giving up trying to explain, "What is wisdom?"
Like the CIA and its CTO Gus Hunt at the GigaOm Conference, Mr. Weinberger claimed that we should be collecting everything with respect to data. "We are living in a time of inclusion, not curating data. Today, we filter on the way out."
He went on to give an example that even data of reality TV shows on celebrities should be gathered, as what might be thought of nonsense one day, became a Harvard course on celebrities and their influence on fashion and culture last semester.
"Order doesn't scale in the big data era," he said. "Rows and columns don't scale. Agreements don't scale." Weinberger went through a few, often funny, examples of how humans always agree to disagree. "But messy data does scale. It gives meaning."
In mashing up all kinds of data, from polar opposite categories to data sets of different variety, velocity, and volume, the Harvard researcher drove home his point that we should no longer be excluding (censoring) data, but collecting all of it. As Mr. Hunt said at the GigaOm Conference, the CIA will collect and keep all its data storing it in a $600 million Amazon cloud to enable the intelligence agency "to connect the dots" in some unforeseen way for future events.
So how do we make sense of all the terabytes (TB) and petabytes of data that different companies, institutions, and industries are collecting and storing in the cloud today?
A Beacon into Making 'Sense' of Big Data
At the same GigaOm Conference, I came across the answer: SiSense.
In meeting Bruno Aziza, vice president of marketing at SiSense, I found an energetic man excited about the future of data. He believed that his firm, which just closed a $10 million Series B investment with Battery Ventures, has engineered the road out of the dark void of messy data.
Prism™, SiSense's flagship business intelligence (BI) software, is a technology that reinforces both the company's mantra -- "You Don't Have to be a Data Scientist" to mine big data -- and the 10x10x10 challenge for analytics.
Last month, SiSense partnered with AOL to make sense of a decade worth of stored big data in the CrunchBase ecosystem of venture capital and technology startups for free.
In test-driving a demo of CrunchBase through Prism, Mr. Aziza showed the average investment by round for various startups, from mobile to cloud computing. He pulled up different tables, joined them together, then ran the analytics without typing a single key.
"It's very simple to use," he said. "Call up pivot tables, index them. Ask if you want to count it. Auto-create and aggregate the data. With other solutions, when users make changes along the way accumulating evermore data, things breakdown. But that's when SiSense gets smarter."
Once users have the data, they want it in a simple, visual way. "They can share it, export it, publish it, and schedule distribution," Aziza said. "Using CrunchBase, people can mine fifty tables joined from different sources and different systems.
Not only is the BI software useful, but also its user-friendly as all this non-coder had to do was use a laptop glidepad to pull up data and join the tables.
SiSense's 10x10x10 challenge: "10 TB of data in 10 seconds on a $10K machine."
Without the heavy investment in legacy systems, SiSense has made the world of messy data accessible and affordable to the small-to-medium sizes businesses.
Interview with CEO Amit Bendov
In a follow up interview with SiSense CEO Amit Bendov, he made the point that his firm has solved the "Tower of Babel" problem of BI and data analytics of different software programs that don't speak to one another, conflicting pivot tables, and big untenable data. What seemed impossible -- at least for a reasonable price -- to "pull" key data from vast oceans and warehouses of unrelated data in 2012 appears to have turned the corner this year.
"Launching Prism last October, we set out to make data analytics simple with what feels like a monster of complicated models that cost millions to crunch and do it at a fraction of the cost," Bendov said. Probably in a fraction of the time, too.
SiSense's co-founders, Chief Product Officer Elad Israeli and CTO Eldad Farkash, both veterans of Israel's technology hub, started developing the BI software many years ago.
To put that database analytics model into commercial use, Amit Bendov came on board to accelerate growth of the company: "A couple of years ago our growth soared from 300% to 520% last year. We have been very aggressive in achieving our goals and currently we are one month ahead of our latest milestone," he said.
"We are in high demand to keep support people in New York City, where we will soon open a regional office. Our client-base there is growing tremendously as we offer unmatched power, speed, and simplicity compared to IBM, HP, and Oracle. We do a different job than they do. They're very complicated for non-technical people, who can't handle big data or computer storage. Our product is revolutionary," Bendov said.
"Who are your competitors?" I asked.
Without hesitation, Amit Bendov replied, "They are the lower end solutions, from Tableau (which just announced an IPO) to other ediscovery tools and software like Qliktech. We are cheaper to use. But the critical difference, the more users and larger portions of data in a database, the smarter Prism works."
Machine learning isn't new, but for big data that's hard to manage and impossible to pull key analytics easily and at a low cost, this could be an inflection point.
"Data connectivity and visualizing everything comes in a package, to scale. When I did a demo for a non-profit in Washington, DC, they were impressed that a non-techie CEO could demo the product and show them how easy it is to use," he said.
"We are thrilled to add Battery Ventures as an investor. We are experiencing exponential sales growth and incredible buzz in the market -- now it's time to add oil to the fire. Our biggest challenge right now is growing the sales force and the support teams quickly enough to keep up with the demand," said Bendov in a press release.
That isn't the only challenge for SiSense. Getting the word out that their software can add value to information from unthreaded, messy forms of data, and for businesses to make intelligent, actionable decisions may be just as big.
If it all works out, big data may not be as big and imposing as we once thought. That's hard to imagine as each day we send, share, and collect a trillion emails, tweets, millions of images and photographs, and other skynormous troves of data.
Source: www.huffingtonpost.com/james-grundvig/making-big-data-small-and_b_3099551.html
"Knowledge Redefined: How Technology is Changing the Way We Learn, What We Know, and How We Invest," centered on the lecture of David Weinberger, a Harvard researcher and bestselling author.
After Weinberger showed a pyramid with four levels of intelligence -- data, information, knowledge, and wisdom at the top -- to the packed Times Center auditorium, he went on a fast-paced tear explaining the eccentricities, broken paradigms, and counter-intuitive insights into three of the four levels.
He punted on the fourth giving up trying to explain, "What is wisdom?"
Like the CIA and its CTO Gus Hunt at the GigaOm Conference, Mr. Weinberger claimed that we should be collecting everything with respect to data. "We are living in a time of inclusion, not curating data. Today, we filter on the way out."
He went on to give an example that even data of reality TV shows on celebrities should be gathered, as what might be thought of nonsense one day, became a Harvard course on celebrities and their influence on fashion and culture last semester.
"Order doesn't scale in the big data era," he said. "Rows and columns don't scale. Agreements don't scale." Weinberger went through a few, often funny, examples of how humans always agree to disagree. "But messy data does scale. It gives meaning."
In mashing up all kinds of data, from polar opposite categories to data sets of different variety, velocity, and volume, the Harvard researcher drove home his point that we should no longer be excluding (censoring) data, but collecting all of it. As Mr. Hunt said at the GigaOm Conference, the CIA will collect and keep all its data storing it in a $600 million Amazon cloud to enable the intelligence agency "to connect the dots" in some unforeseen way for future events.
So how do we make sense of all the terabytes (TB) and petabytes of data that different companies, institutions, and industries are collecting and storing in the cloud today?
A Beacon into Making 'Sense' of Big Data
At the same GigaOm Conference, I came across the answer: SiSense.
In meeting Bruno Aziza, vice president of marketing at SiSense, I found an energetic man excited about the future of data. He believed that his firm, which just closed a $10 million Series B investment with Battery Ventures, has engineered the road out of the dark void of messy data.
Prism™, SiSense's flagship business intelligence (BI) software, is a technology that reinforces both the company's mantra -- "You Don't Have to be a Data Scientist" to mine big data -- and the 10x10x10 challenge for analytics.
Last month, SiSense partnered with AOL to make sense of a decade worth of stored big data in the CrunchBase ecosystem of venture capital and technology startups for free.
In test-driving a demo of CrunchBase through Prism, Mr. Aziza showed the average investment by round for various startups, from mobile to cloud computing. He pulled up different tables, joined them together, then ran the analytics without typing a single key.
"It's very simple to use," he said. "Call up pivot tables, index them. Ask if you want to count it. Auto-create and aggregate the data. With other solutions, when users make changes along the way accumulating evermore data, things breakdown. But that's when SiSense gets smarter."
Once users have the data, they want it in a simple, visual way. "They can share it, export it, publish it, and schedule distribution," Aziza said. "Using CrunchBase, people can mine fifty tables joined from different sources and different systems.
Not only is the BI software useful, but also its user-friendly as all this non-coder had to do was use a laptop glidepad to pull up data and join the tables.
SiSense's 10x10x10 challenge: "10 TB of data in 10 seconds on a $10K machine."
Without the heavy investment in legacy systems, SiSense has made the world of messy data accessible and affordable to the small-to-medium sizes businesses.
Interview with CEO Amit Bendov
In a follow up interview with SiSense CEO Amit Bendov, he made the point that his firm has solved the "Tower of Babel" problem of BI and data analytics of different software programs that don't speak to one another, conflicting pivot tables, and big untenable data. What seemed impossible -- at least for a reasonable price -- to "pull" key data from vast oceans and warehouses of unrelated data in 2012 appears to have turned the corner this year.
"Launching Prism last October, we set out to make data analytics simple with what feels like a monster of complicated models that cost millions to crunch and do it at a fraction of the cost," Bendov said. Probably in a fraction of the time, too.
SiSense's co-founders, Chief Product Officer Elad Israeli and CTO Eldad Farkash, both veterans of Israel's technology hub, started developing the BI software many years ago.
To put that database analytics model into commercial use, Amit Bendov came on board to accelerate growth of the company: "A couple of years ago our growth soared from 300% to 520% last year. We have been very aggressive in achieving our goals and currently we are one month ahead of our latest milestone," he said.
"We are in high demand to keep support people in New York City, where we will soon open a regional office. Our client-base there is growing tremendously as we offer unmatched power, speed, and simplicity compared to IBM, HP, and Oracle. We do a different job than they do. They're very complicated for non-technical people, who can't handle big data or computer storage. Our product is revolutionary," Bendov said.
"Who are your competitors?" I asked.
Without hesitation, Amit Bendov replied, "They are the lower end solutions, from Tableau (which just announced an IPO) to other ediscovery tools and software like Qliktech. We are cheaper to use. But the critical difference, the more users and larger portions of data in a database, the smarter Prism works."
Machine learning isn't new, but for big data that's hard to manage and impossible to pull key analytics easily and at a low cost, this could be an inflection point.
"Data connectivity and visualizing everything comes in a package, to scale. When I did a demo for a non-profit in Washington, DC, they were impressed that a non-techie CEO could demo the product and show them how easy it is to use," he said.
"We are thrilled to add Battery Ventures as an investor. We are experiencing exponential sales growth and incredible buzz in the market -- now it's time to add oil to the fire. Our biggest challenge right now is growing the sales force and the support teams quickly enough to keep up with the demand," said Bendov in a press release.
That isn't the only challenge for SiSense. Getting the word out that their software can add value to information from unthreaded, messy forms of data, and for businesses to make intelligent, actionable decisions may be just as big.
If it all works out, big data may not be as big and imposing as we once thought. That's hard to imagine as each day we send, share, and collect a trillion emails, tweets, millions of images and photographs, and other skynormous troves of data.
Source: www.huffingtonpost.com/james-grundvig/making-big-data-small-and_b_3099551.html
Will Big Data Mark the End of the Market Research Industry? | SmartData Collective
Mark van Rijmenam | April 19, 2013
Big data is disrupting the world, society and all industries. One industry that will definitely notice these waves of change is the market research industry. The new big data startups, that are founded everywhere around the world, can be viewed as new, but improved, market research agencies. Consequently, the market research industry as we know it today will cease to exist.
Customers interacting with organisations provide valuable feedback for companies. With big data this feedback, whether provided via the website, via the call centre or by simply using the product, can be turned into market information for product improvements. With big data it is possible to turn every conversation recorded by a call centre into text, which could then be data mined for insights. This would provide very useful content about what and how consumers feel about the brand and the products.
Even more, consumer panels, panels of business managers or focus groups will not be necessary anymore in the world of big data. In an inter-connected world, where products have sensors and are connected to the internet, companies will know in real-time how people use their product, when they use it, for how long they use it and when things go wrong. Whenever a product needs to be improved or a new product needs to be developed, organisations can simply look at the real-time sensor data pouring into the organisation and understand the needs and wishes of their customers. If they want to understand how people think of their products, the company or what the sentiment of the brand is, they can simply connect to a big data startup and real-time data starts appearing on their screen, analysed and visualized, to be understood by everyone in the organisation. A market research organisation that requires a few weeks to create and perform a
questionnaire and another few weeks to analyse the results, is likely to be by-passed in this fast world.
In a blog post by Brigid Kilcoin it is noted though that market research agencies can actually thrive due to big data because market research organisations focus on the ‘why’. She writes that it is only in understanding the context of why something is happening, that an enterprise can drive meaningful change within its organization. According to Kiloin, market research organisations are there to provide an answer to this ‘why’ question. In a world of big data, however, understanding of the context is done automatically by the algorithms. Members of an organisation can use the analyses and visualizations made by the big data tools to under the context of why something is happening.
In a big data society, researchers are consequently not needed anymore to tell a compelling story using visualizations. There are great tools available on the market already that can make these visualizations automatically and deliver them in real-time to the decision-maker.
Gigi DeVault, from about.com, notes that “all theories of human behavior can be tossed out since data analysts no longer need to know why people do what they do. Human behavior can be tracked and measured with “‘unprecedented fidelity.”” Market research organisations will therefore have to adapt to the new rules of the game or they will be by-passed by big data startups who can do the work for market research agencies at a fraction of the costs and in real-time.
It will require market researchers to understand the power of Big Data and to become comfortable with the tools and techniques of Big Data. If they want to survive, they should start hiring big data scientists to be able to deliver the same kind of services as big data startups around the world already offer.
So yes, big data will mark the end of the market research industry as we know it. Unless, of course, they are able to adapt and pivot on time. Otherwise, the big data scientist will replace the market researcher and this will happen faster than one might expect.
Source: http://smartdatacollective.com/bigdatastartups/118931/will-big-data-mark-end-market-research-industry
Big data is disrupting the world, society and all industries. One industry that will definitely notice these waves of change is the market research industry. The new big data startups, that are founded everywhere around the world, can be viewed as new, but improved, market research agencies. Consequently, the market research industry as we know it today will cease to exist.
Customers interacting with organisations provide valuable feedback for companies. With big data this feedback, whether provided via the website, via the call centre or by simply using the product, can be turned into market information for product improvements. With big data it is possible to turn every conversation recorded by a call centre into text, which could then be data mined for insights. This would provide very useful content about what and how consumers feel about the brand and the products.
Even more, consumer panels, panels of business managers or focus groups will not be necessary anymore in the world of big data. In an inter-connected world, where products have sensors and are connected to the internet, companies will know in real-time how people use their product, when they use it, for how long they use it and when things go wrong. Whenever a product needs to be improved or a new product needs to be developed, organisations can simply look at the real-time sensor data pouring into the organisation and understand the needs and wishes of their customers. If they want to understand how people think of their products, the company or what the sentiment of the brand is, they can simply connect to a big data startup and real-time data starts appearing on their screen, analysed and visualized, to be understood by everyone in the organisation. A market research organisation that requires a few weeks to create and perform a
questionnaire and another few weeks to analyse the results, is likely to be by-passed in this fast world.
In a blog post by Brigid Kilcoin it is noted though that market research agencies can actually thrive due to big data because market research organisations focus on the ‘why’. She writes that it is only in understanding the context of why something is happening, that an enterprise can drive meaningful change within its organization. According to Kiloin, market research organisations are there to provide an answer to this ‘why’ question. In a world of big data, however, understanding of the context is done automatically by the algorithms. Members of an organisation can use the analyses and visualizations made by the big data tools to under the context of why something is happening.
In a big data society, researchers are consequently not needed anymore to tell a compelling story using visualizations. There are great tools available on the market already that can make these visualizations automatically and deliver them in real-time to the decision-maker.
Gigi DeVault, from about.com, notes that “all theories of human behavior can be tossed out since data analysts no longer need to know why people do what they do. Human behavior can be tracked and measured with “‘unprecedented fidelity.”” Market research organisations will therefore have to adapt to the new rules of the game or they will be by-passed by big data startups who can do the work for market research agencies at a fraction of the costs and in real-time.
It will require market researchers to understand the power of Big Data and to become comfortable with the tools and techniques of Big Data. If they want to survive, they should start hiring big data scientists to be able to deliver the same kind of services as big data startups around the world already offer.
So yes, big data will mark the end of the market research industry as we know it. Unless, of course, they are able to adapt and pivot on time. Otherwise, the big data scientist will replace the market researcher and this will happen faster than one might expect.
Source: http://smartdatacollective.com/bigdatastartups/118931/will-big-data-mark-end-market-research-industry
Intel CIO Kim Stevenson on big data, OpenStack, women in IT | ZDNet
By Larry Dignan for Between the Lines |
Intel CIO Kim Stevenson, who has been at the helm for a little more than a year, said OpenStack is the most useful cloud architecture for avoiding lock-in, outlined how the chip giant is using big data techniques, and talked capacity planning for her company.
Stevenson, formerly the vice president of global operations and services at Intel, has a dual role at the company. First, she's in charge of the IT operations for Intel globally and has the usual CIO headaches. And second, Stevenson is the guinea-pig-in-chief for Intel's products.
Intel's IT group is typically the "first implementer" for new hardware and software, said Stevenson. In other words, Intel eats its own dog food and gives product people honest feedback on what needs to happen. Stevenson quipped that the conversations from an IT customer perspective don't always go well.
Here are a few key highlights from a breakfast conversation in New York:
Big data. One area where Intel is the first implementer is on its Hadoop distribution. Like other large enterprises, Stevenson's group is looking toward big data and analytics to boost revenue, improve time to market and cut costs. Here are Stevenson's biggest takeaways on using big data internally.
Here's how the cloud fits in with Intel's key areas: Office applications, silicon design and manufacturing.
Beyond design and manufacturing, though, Stevenson requires that an application request has to prove why it needs physical resources. In other words, applications need to be virtualized or they won't get capacity.
Capacity planning. Stevenson said the goal of Intel is to have 15 days of capacity on hand. This 15 day rule means that if the typical usage occurs within Intel with the usual new requests, the company will be out of compute in 15 days. For Stevenson, compute capacity is her inventory. The 15-day capacity cushion is based on memory mostly since that's the first to go when resources are being taxed. Compute and storage are also watched. Stevenson acknowledges that the 15-day capacity best practice may not work for all companies and could even be shorter. "Some companies may have to convince the CFO that having 15 days of excess capacity sitting around is a good idea," said Stevenson. "We didn't have to convince our CFO of that."
Playing the cloud provider game. Stevenson said Intel has gone with OpenStack, a cloud architecture that's gaining momentum. Why? Intel uses multiple cloud providers working under a master services agreement and wants to hop between them to maximize performance and costs. "We use several cloud providers mostly in the U.S.," she said. "We've had no issues with SLAs, but we're only bursting. We're careful with what workloads we use in the public cloud." OpenStack enables Intel to use multiple providers and avoid lock-in, said Stevenson.
About those cloud brokers. I asked Stevenson what she thought of cloud brokers---companies that would manage providers to maximize savings. "Cloud brokers are transitory. That dog doesn't hunt as an independent business model," she said. Ultimately, the ability to hop between cloud providers will be built into an infrastructure like OpenStack and automated.
Women in IT. Finally, Stevenson talked a bit about women in IT and was asked about Sheryl Sandberg's Lean In: Women, Work, and the Will to Lead, which drew some fire over what the Facebook operating chief had to say. Stevenson said the book was largely on target and added that Sandberg had good advice for anyone in business not just women. Her main takeaway on women in IT is that they have to understand the communication differences in a male dominated industry. In other words, know how to navigate the subcultures such as sports chatter and international differences. Also use good context at the start of a business problem to navigate functional and relational communication styles.
Source: http://www.zdnet.com/intel-cio-kim-stevenson-on-big-data-openstack-women-in-it-7000014221/
Intel CIO Kim Stevenson, who has been at the helm for a little more than a year, said OpenStack is the most useful cloud architecture for avoiding lock-in, outlined how the chip giant is using big data techniques, and talked capacity planning for her company.
Stevenson, formerly the vice president of global operations and services at Intel, has a dual role at the company. First, she's in charge of the IT operations for Intel globally and has the usual CIO headaches. And second, Stevenson is the guinea-pig-in-chief for Intel's products.
Intel's IT group is typically the "first implementer" for new hardware and software, said Stevenson. In other words, Intel eats its own dog food and gives product people honest feedback on what needs to happen. Stevenson quipped that the conversations from an IT customer perspective don't always go well.
Here are a few key highlights from a breakfast conversation in New York:
Big data. One area where Intel is the first implementer is on its Hadoop distribution. Like other large enterprises, Stevenson's group is looking toward big data and analytics to boost revenue, improve time to market and cut costs. Here are Stevenson's biggest takeaways on using big data internally.
- Talent: "Whether developing or implementing big data, talent is in short supply," said Stevenson. She added that Intel is working university partnerships and programs to get talent, but it takes time to turn interns into big data visionaries. Intel is also training employees, but noted that the transition to big data can be difficult. "It's not easy for a database person who thinks in rows and columns to learn Hadoop," said Stevenson. "It's also difficult for business people because they usually want clean data."
- Time to market is the big data win: Intel's biggest goal with big data and analytics "is to reduce the product cycle time," said Stevenson. After all, Intel is entering the smartphone market and the product cycles are twice as fast as PCs. To speed up product development time, Intel looked at the post silicon validation process. This process revolves around making sure that all of the transistors---all 1 billion or so---are functioning properly. The data out of a silicon wafer equates to 750 terabytes. Intel deployed a machine learning model that isolates testing and will cut 25 percent of a five quarter product cycle period.
- Manufacturing: Big data is also used to cut costs in manufacturing by analyzing silicon wafers and how they are cut. The goal is to have less waste and more processors for better yields. Hadoop and Intel's analytics stack -- which consists of IBM's Netezza, Hadoop and proprietary algorithms -- logs information, finds errors and best practices and passes the data along to the next tool. If each step improves, there are better good die yields and lower unit costs, explained Stevenson.
- Marketing fraud: Big data is also used to detect fraud and errors in the chip maker's Intel Inside marketing. In a nutshell, Intel partners market Intel Inside and get reimbursed by the company on their marketing spend. Stevenson said that Intel used its machine learning engine to look at fraudulent claims, mistakes and incomplete reimbursement requests. Intel also looked to cut down on manual post audits. The return was $7 million in savings in a quarter and the expulsion of bad guys. Reimbursement time also improved.
Here's how the cloud fits in with Intel's key areas: Office applications, silicon design and manufacturing.
- Typical enterprise applications: According to Stevenson, 75 percent of Intel's typical office systems such as email and ERP are in the cloud in a private-public hybrid model. "In 2012, we were bursting to other Intel regions and now we burst out to other providers," she said. One public cloud use case is Will.I.Am. The artist can offer a free download on Intel.com and get a ton of Web traffic. That use case is perfect for the public cloud.
- Design: "Silicon design will never go out to the cloud. That's our core IP," said Stevenson. She added that no cloud service level agreement or chargeback would ever compensate for Intel's intellectual property being leaked. Instead, product design runs on a high performance computing grid that's internal.
- Manufacturing: Manufacturing is another area that won't be put into the cloud. The information is housed in small data centers near the manufacturing site and later aggregated.
Beyond design and manufacturing, though, Stevenson requires that an application request has to prove why it needs physical resources. In other words, applications need to be virtualized or they won't get capacity.
Capacity planning. Stevenson said the goal of Intel is to have 15 days of capacity on hand. This 15 day rule means that if the typical usage occurs within Intel with the usual new requests, the company will be out of compute in 15 days. For Stevenson, compute capacity is her inventory. The 15-day capacity cushion is based on memory mostly since that's the first to go when resources are being taxed. Compute and storage are also watched. Stevenson acknowledges that the 15-day capacity best practice may not work for all companies and could even be shorter. "Some companies may have to convince the CFO that having 15 days of excess capacity sitting around is a good idea," said Stevenson. "We didn't have to convince our CFO of that."
Playing the cloud provider game. Stevenson said Intel has gone with OpenStack, a cloud architecture that's gaining momentum. Why? Intel uses multiple cloud providers working under a master services agreement and wants to hop between them to maximize performance and costs. "We use several cloud providers mostly in the U.S.," she said. "We've had no issues with SLAs, but we're only bursting. We're careful with what workloads we use in the public cloud." OpenStack enables Intel to use multiple providers and avoid lock-in, said Stevenson.
About those cloud brokers. I asked Stevenson what she thought of cloud brokers---companies that would manage providers to maximize savings. "Cloud brokers are transitory. That dog doesn't hunt as an independent business model," she said. Ultimately, the ability to hop between cloud providers will be built into an infrastructure like OpenStack and automated.
Women in IT. Finally, Stevenson talked a bit about women in IT and was asked about Sheryl Sandberg's Lean In: Women, Work, and the Will to Lead, which drew some fire over what the Facebook operating chief had to say. Stevenson said the book was largely on target and added that Sandberg had good advice for anyone in business not just women. Her main takeaway on women in IT is that they have to understand the communication differences in a male dominated industry. In other words, know how to navigate the subcultures such as sports chatter and international differences. Also use good context at the start of a business problem to navigate functional and relational communication styles.
Source: http://www.zdnet.com/intel-cio-kim-stevenson-on-big-data-openstack-women-in-it-7000014221/
Subscribe to:
Posts (Atom)