Are we getting lost in this huge sea of humongous data? Is storage capability of our network start to become so much larger than ability of our brain to digest that data?? The tipping point might be that soon nobody will know what they have stored and where.
Of course the author will plug the advantages of flash memory, since his company uses it.
But the Invisible Hand stills plays a role. We are in the early stages of shifting to SSDs in PCs and the like, because old fashioned hard drives are simply a lot *cheaper*. I don't see the industry as a whole shifting till costs drop by an order of magnitude. It's the reason why an awful lot of backup is to tape: the cheapest cost per megabyte of storage.
Computing could become big data oriented if it were all based on big look up tables. But accessing memory to carry this out would have to be much more parallel than existing multi-core threads to be appreciably fast. Probably on the order of 1000x at least.
Not sure that few Tbs of storage I have helps me. I am at the point where I am not sure where everything is located. And don't have the time to sort thru 100 Gbs+ storage in family pictures not to mention other stuff. Help is needed. Start-up opps?
I love the fact that I have 2 TB of at home storage! I also really like having alternate remote site backups! It is a wonderful time to be on a machine with so many options for storage/performance and capacity..
Valid point! Most storage area networks and storage services will have 4X to 6X redundancy and my point above is this can get way out of hand with hardware limitations in the coming days. Solidstate drives have to improve their reliability drastically from where they are now.
At some point we'll have to stop hoarding data and do some old-fashioned spring cleaning. I see a horrible practice of multiple instances of the exact same files here on the corporate servers basically because people are in herently lazy and nobody forces the issue. We could easily contain this data explosion if people could just discipline themselves to keep only what they need. I dare say, 98% of the data is probably worthless (repeated, out-of-date, superseded, obselete, etc.).
One thing abundantly clear to every one is the explosion of data in the last five years. If projects based on IoT forge ahead and we end up living with 50billion nodes networked by 2050, a storage environment where round trip latencies for accessing and fetching data have to scale exponentially and that is possible only with solid state drives. The question is how can be make the nonvolatile storage faster and more reliable, over and above what is afforded by redundancy today.
What are the engineering and design challenges in creating successful IoT devices? These devices are usually small, resource-constrained electronics designed to sense, collect, send, and/or interpret data. Some of the devices need to be smart enough to act upon data in real time, 24/7. Are the design challenges the same as with embedded systems, but with a little developer- and IT-skills added in? What do engineers need to know? Rick Merritt talks with two experts about the tools and best options for designing IoT devices in 2016. Specifically the guests will discuss sensors, security, and lessons from IoT deployments.