With the maturing of 3D-IC processing and design tool capabilities, it will be interesting to see if more companies begin migrating the analog portions to separate die/process and then integrating them vertically.
New (ie smaller geometries) have fueled digital innovation by allowing us to pack more and more digital transistors into the same silicon area. Analog "in general" doesn't scale quite the same so going to smaller geometries doesn't make the circuit smaller and less expensive. These smaller geometry nodes also have thinner oxides and lower operating voltages so it is often more difficult to design high performance analog at these smaller nodes. That's why you hear a lot of folks saying "More than Moore" these days because a lot of really cool analog and mixed signal designs are being done at the older 0.35u and 0.18u nodes.
I would take exception to the statement that 70% of today's chips are mixed-signal. IMHO, the true number is 100%. Even (or especially) the most gargantuan digital SoC at least has a PLL or two, some DDR I/Os, and other cells that it would be a stretch to refer to as "digital."
Yes, the toolsets are very different and there is no question that the bulk of EDA innovation over the last couple decades has been on the digital side.
The really interesting part, and a big divergence of methodologies, comes at the chip top level where all the blocks get integrated.
In a Big D/Little A chip, the integration is done in a "digital" P&R tool and the analog designers just deliver their blocks (cells), however large or small -- data converters, PLLs, I/O buffers, voltage regulators or whatever. In a Big A/Little D chip, it's the other way around -- the "digital" blocks are treated as cells, like any analog cell, and the whole thing gets integrated into a full chip in an "analog" P&R tool.
To add another level of complexity and weirdness, none of the "analog" blocks are truly analog -- there are always some digital standard cells in there too. This creates interesting problems for incorporating that loose logic into scan chains and making the whole thing ATPG-tool-friendly.
The analog designers don't know the digital tools and the digital designers don't know the analog tools, so there need to be at least a few people who could rightly be called AMS designers -- those who can play in both sandboxes, to try to cobble the whole mess together into something that is thoroughly verified and tapeout-worthy.
Many companies tried to automate the analog design flow (both schematic and layout) and failed. Things like small mismatch in the design can severely affect the performance of the analog designs. Moreover many companies are still using old-reliable analog processes and are reluctant to move to the new processes.
What are the engineering and design challenges in creating successful IoT devices? These devices are usually small, resource-constrained electronics designed to sense, collect, send, and/or interpret data. Some of the devices need to be smart enough to act upon data in real time, 24/7. Are the design challenges the same as with embedded systems, but with a little developer- and IT-skills added in? What do engineers need to know? Rick Merritt talks with two experts about the tools and best options for designing IoT devices in 2016. Specifically the guests will discuss sensors, security, and lessons from IoT deployments.