<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>The Speculist &#187; Big Data</title>
	<atom:link href="https://blog.speculist.com/tag/big-data/feed" rel="self" type="application/rss+xml" />
	<link>https://blog.speculist.com</link>
	<description>Live to see it.</description>
	<lastBuildDate>Thu, 25 Jul 2019 23:07:25 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.6.1</generator>
		<item>
		<title>Where the Possibilities Are</title>
		<link>https://blog.speculist.com/big-data/where-the-possibilities-are.html</link>
		<comments>https://blog.speculist.com/big-data/where-the-possibilities-are.html#comments</comments>
		<pubDate>Wed, 28 Jan 2015 22:36:37 +0000</pubDate>
		<dc:creator>Phil Bowermaster</dc:creator>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Book of the Possible]]></category>
		<category><![CDATA[Datafication]]></category>

		<guid isPermaLink="false">https://blog.speculist.com/?p=4955</guid>
		<description><![CDATA[Where does the value of big data truly present itself, in the data itself or in the algorithms we use to make sense of it? Bill Franks of Teradata comes down sharply  on the side of the data: &#8230;I’m convinced that new information will beat new algorithms and new metrics based on existing information almost [&#8230;]]]></description>
				<content:encoded><![CDATA[<p><img class=" wp-image-4956 alignright" alt="??????????????????" src="https://blog.speculist.com/wp-content/uploads/2015/01/Possibilities.jpg" width="390" height="260" />Where does the value of big data truly present itself, in the data itself or in the algorithms we use to make sense of it? <a href="http://www.forbes.com/sites/teradata/2015/01/06/a-man-his-dog-and-their-microchips/">Bill Franks</a> of Teradata comes down sharply  on the side of the data:</p>
<p style="padding-left: 30px;">&#8230;I’m convinced that <em>new information will beat new algorithms and new metrics based on existing information almost every time.</em> Indeed, new information can be so powerful that, once it is found, analytics professionals should stop worrying about improving existing models with existing data and focus instead on incorporating and testing that new information.</p>
<p>By &#8220;new information,&#8221; he means information that didn&#8217;t exist before or that we now have to a level of depth never before possible. Sensor data in Internet of Things environments can represent either of these kinds of data. For example, we may have always used temperature data in performing some calculation, but back in the day we used a daily average. Now we have sensors providing temperature data every few minutes (or seconds.) That&#8217;s data to a greater depth. For data that we didn&#8217;t have before, Bill cites sensors on cars that track wear and tear as the vehicle is driven. Previously, vehicle repair occurred in a primarily reactive way. Now we can begin to anticipate repairs before they are needed.</p>
<p>Somehow this reminds me of a talk that Eliezer Yudkowsky gave at the <a href="http://archive.today/I6akm">Singularity Summit</a> back in 2007. He said:</p>
<p style="padding-left: 30px;">In the intelligence explosion the key threshold is criticality of recursive self-improvement. It&#8217;s not enough to have an AI that improves itself a little. It has to be able to improve itself enough to significantly increase its ability to make further self-improvements, which sounds to me like a software issue, not a hardware issue. So there is a question of, Can you predict that threshold using Moore&#8217;s Law at all?</p>
<p style="padding-left: 30px;">Geordie Rose of D-Wave Systems recently was kind enough to provide us with a startling illustration of software progress versus hardware progress. Suppose you want to factor a 75-digit number. Would you rather have a 2007 supercomputer, IBM&#8217;s Blue Gene/L, running an algorithm from 1977, or a 1977 computer, an Apple II, running a 2007 algorithm? And Geordie Rose calculated that Blue Gene/L with 1977&#8242;s algorithm would take ten years, and an Apple II with 2007&#8242;s algorithm would take three years.</p>
<p>There is a progression here, albeit a counter-intuitive one. We might be inclined to think that hardware adds more value than &#8220;mere&#8221; software and that software is inherently more valuable than &#8220;mere&#8221; (or the term we like to throw around a lot is &#8220;raw&#8221;) data. The opposite turns out to be the truth. The data itself is where the value is. Hardware and software only help us to focus on the potentialities, the possibilities, that it already contains.</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.speculist.com/big-data/where-the-possibilities-are.html/feed</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
		<item>
		<title>Datafication in Three Easy Steps</title>
		<link>https://blog.speculist.com/scenarios/datafication-in-three-easy-steps.html</link>
		<comments>https://blog.speculist.com/scenarios/datafication-in-three-easy-steps.html#comments</comments>
		<pubDate>Sat, 20 Sep 2014 04:48:25 +0000</pubDate>
		<dc:creator>Phil Bowermaster</dc:creator>
				<category><![CDATA[Acceleration]]></category>
		<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Datafication]]></category>
		<category><![CDATA[Ephemeralization]]></category>
		<category><![CDATA[Scenarios]]></category>
		<category><![CDATA[Movies]]></category>

		<guid isPermaLink="false">https://blog.speculist.com/?p=4899</guid>
		<description><![CDATA[The relentless wave of change that is transforming our world from being one made primarily out of stuff to one made primarily out of data has a name. It&#8217;s called datafication. Over the past few decades, we have witnessed the datafication of business, of society, and of everyday life. There appear to be three major [&#8230;]]]></description>
				<content:encoded><![CDATA[<p>The relentless wave of change that is transforming our world from being one made primarily out of stuff to <a href="https://blog.speculist.com/big-data/bigger-than-we-realize.html">one made primarily out of data</a> has a name. It&#8217;s called <a href="http://en.wikipedia.org/wiki/Datafication">datafication</a>.</p>
<p>Over the past few decades, we have witnessed the datafication of business, of society, and of everyday life. There appear to be three major phases of datafication. In the first phase, an activity or process becomes increasingly reliant on data. In the second, data begins to transform the activity or process by taking a central role in its execution. In the third phase, the activity is moved entirely into the data substrate.</p>
<p><img class="wp-image-4904 alignright" alt="film" src="https://blog.speculist.com/wp-content/uploads/2014/11/film.jpg" width="199" height="331" /></p>
<p>Take the movie business. Putting artistic considerations aside, the success of any film has always been a measurement of how much revenue it generates. Originally, this was a pretty straightforward matter of counting box office receipts. (Today, what with many and varied distribution channels and considerations such as licensing and merchandising that often come into play, the math for calculating success is considerably more complex.) The film industry entered the first phase of datafication relatively early on, as studios began trying develop formulas for repeat box office success. have witnessed the datafication of business, of society, and of everyday life. There appear to be three major phases of datafication. In the first phase, an activity or process becomes increasingly reliant on data. In the second, data begins to transform the activity or process by taking a central role in its execution. In the third phase, the activity is moved entirely into the data substrate.</p>
<p>The data points were, at first, relatively few and far between: geographic differences in box office; one star&#8217;s draw vs. another; westerns vs. romances vs. war movies vs. musicals; summer releases vs. Christmas releases. Over time, the analysis evolved in terms of sophistication until the industry reached the second phase of datafication. This is how we  came to live in an age of scripts written for a target adolescent male audience and re-edits and even rewrites following test screenings. The data began to drive the process.</p>
<p>But data wasn&#8217;t done with the movies yet. The film industry is moving rapidly into the third phase of datafication. Once upon a time, filmmakers made <em>films.</em> Long strips of celluloid with images on them. We&#8217;ve all heard of efforts to preserve decaying movies from the early part of the last century. Film was a chemical and mechanical process resulting in a physical artifact. But not today. The product of the film-making process is now essentially a data artifact. Movies are consumed over digital networks on TVs, laptops, and smartphones. And, in fact, they can now be <em>made</em> <a href="http://www.thewrap.com/movies/article/director-movie-made-smartphones-puts-all-his-chips-oscar-34080/">entirely on smartphones</a>.  Short messages, tweets, motions pictures&#8230;it&#8217;s all the same. It&#8217;s all data.</p>
<p>The big data revolution is ultimately about this kind of transformation in all sectors of all industries. The movie and music businesses are obvious examples of industries that have made it at least part of the the way to phase three. But then so is the telecommunications industry. Shipping and logistics have become as much about data as they are about moving stuff around. Even manufacturing is moving in that direction &#8212; and will continue to do so as digital fabrication and 3D printing become increasingly mainstream.</p>
<p>Right now the world as whole is really just beginning to move from phase 1 to phase 2. Data is beginning to influence and direct the world in ways never before considered. And we are still in the very early days.</p>
<p>&nbsp;</p>
]]></content:encoded>
			<wfw:commentRss>https://blog.speculist.com/scenarios/datafication-in-three-easy-steps.html/feed</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
	</channel>
</rss>
