ICH9/10 have 6 PCI-E v1.1 lanes for shared onbard device bus' and low bandwidth PCIE v1.1 slots, ie 1x or 4x.
You should also note when you overclock the system on the Host Bus side, you'll improve data going out (read burst) due to additional Host Bus bandwidth and cycles but on the other hand you are also creating a queue on the data coming in because Host Bus can push more data faster than DMI bus bandwidth is there. Raising PCIE frequency helps stabilize the spikes in IO reads, and improves IO write bandwidth because DMI clock is generated by clock generator PCIE dividers. Raising PCIE frequency beyond design can introduce signal integrity degradation from unshielded external EMI or cross chatter from selecting a similar reference frequency to a vendors own unique reference frequency, and then all hell breaks loose on your system
X38/X48/X58 have 36 PCIE v2.0 lanes. 32 lanes for PCIE slot allocation. 4 lanes for DMI (direct media interface, which is a dedicated two way point to point link on PCIE v2.0 4x bus) interconnect host link between PCH (x58) / MCH (x38/48) and IOH (ICHx).
P35/P45/etc have 20 PCIE v2.0 lanes. 16 for PCIE slot, 4 lanes for DMI.
ICH has always been on DMI, which behaves similarly to PCI while at the same time behaving like PCIE with VC (virtual channels) layer for QoS and transaction scheduling / interleaving for all IOH root bridge outgoing transactions with TAD (target address decode) destination beyond IOH scope or incoming bus transactions from MCH with TAD lying inside IOH.
Bookmarks