My understanding is that in standard big-little operation the operating system sees just four cores.
As load increases applications move up through the dynamic voltae and clock frequency scaling (DVFS) on the A7 until it is close to maximum then they are moved on to the low-end of the DVFS for for the companion A15 where the scaling continues to maximum voltage and clock frequency. As the processing load reduces the reverse happens.
There is not much latency in switching from A7 to A15 but it is still important to avoid a hysteresis effect where the software repearedly bounces from A7 to A15 and back.
You bet they can! Think real-time image/data processing and data collection tasks. I have applications that consume all of my dual-core processing right now and could use more. The issue, now, is power consumption (how long can I run on all four cores going at full speed) and memory bandwidth (can my portable device's memory feed all those processors fast enough to keep them running at full speed).
Thanks for information for Nvidia Core.
I am surprised that my comments cause so much discussion. I will agree that nothing prevent all 8 cores running at the same time although it is not likely choice. A7 performance per MHz is only half of A15 and is running at 60% of clock frequency. Therefore, each A7 core will only deliver 1/3 A15 performance. So it doesn't make lots of sense to allow A7 join A15 cluster in most cases. Therefore, its performance is only equivalent to 5.3 A15 Core performance. So we should call Samsung 5.3 Core in the future. :-)
Well, that has to be the main application target, isnt't it? If Samsung wanted to have just more power, they would have not bothered with the big-little implementation. They would have used eight full-speed, full-power cores, all active at the same time and with the usual clock scaling. This seems to be a sensible compromise, two distinct, selectable computational levels, each optimized for different operational conditions, plus the option to combine both to have higher processing power when really needed.
You're right, but I doubt all cores will be used except for very specific applications meant to take advantage of it - like any app that could take advantage of heterogeneous computing.
The rest 99.9% of them will only use 4 cores at most.
What are the engineering and design challenges in creating successful IoT devices? These devices are usually small, resource-constrained electronics designed to sense, collect, send, and/or interpret data. Some of the devices need to be smart enough to act upon data in real time, 24/7. Specifically the guests will discuss sensors, security, and lessons from IoT deployments.