The number of primes dictates the amount of memory used, but its not fine grained as the Linpack configuration file since you are limited to only 4 choices. For a quick stability tool, SuperPrime fits the bill.
Select the max number of primes that doesn't cause paging and use it to "check stability".
Asus P5K-E/WiFi-AP (1305 Bios)
Intel E8400 @ 4.33 GHz) [1.416V Load/1.424 Idle]
4x2GB DDR2 (5-5-5-15-2T PL 8) [2.0V]
GeForce GTS 250 1024MB
HT Omega Claro Halo (2xOPA2107 -> LME49860)
Take a look at Commit Charge in the Performance tab of the Task Manager and you will see how much memory is being used. 144M is definitely too much for a 1GB machine, really meant for a 2GB machine unless you want paging to occur.
Asus P5K-E/WiFi-AP (1305 Bios)
Intel E8400 @ 4.33 GHz) [1.416V Load/1.424 Idle]
4x2GB DDR2 (5-5-5-15-2T PL 8) [2.0V]
GeForce GTS 250 1024MB
HT Omega Claro Halo (2xOPA2107 -> LME49860)
I tried running the linpak64.exe directly w/o relying on the .bat file and i noticed the CPU usage is much more stable at 100%
could be an incorrect string on the batch file ?
Im still inputting the same values of
1
2600
2600
4
Well running the .exe gives me a real time monitor of the progress, i think ill stick with the .exe
I would just like to say this is the best stress test ever. I was experiencing occasional freezes in UT3, where the game would just lock up and I'd have to reset the system. Prime95 and OCCT were 100% stable and error free for hours on end, and so was ATItool artifact scanning. I thought I'd give this a try, and not only does it get the CPU close to 6c hotter than prime95 (using small FTT), but it identified errors with it set to only 5 repetitions! Increased vcore some more and now it's stable as well as UT3!
Asus Rampage II Gene | Core i7 920 | 6*2GB Mushkin 998729 | BFG GTX280 OCX | Auzentech X-Fi Forte | Corsair VX550
—Life is too short to be bound by the moral, ethical and legal constraints imposed on us by modern day society.
It seems like this program (linpak 64) killed my PCI Network card.
After 16 minutes of testing my hottest cores were at 73 C, compared to 62-63 with p95 with 4 threads.
After 17 minutes i heard wierd sounds from the computer and my network stopped working. I am now running on an other network card.
CPU frequency: 3.294 GHz
Number of CPUs: 4
Number of threads: 4
Parameters are set to:
Number of tests : 1
Number of equations to solve (problem size) : 14500
Leading dimension of array : 14500
Number of trials to run : 128
Data alignment value (in Kbytes) : 4
Maximum memory requested that can be used = 1682294096, at the size = 14500
============= Timing linear equation system solver =================
Size LDA Align. Time(s) GFlops Residual Residual(norm)
14500 14500 4 53.596 37.9286 2.404240e-010 4.050385e-002
14500 14500 4 53.534 37.9726 2.404240e-010 4.050385e-002
14500 14500 4 53.785 37.7957 2.404240e-010 4.050385e-002
14500 14500 4 53.857 37.7453 2.404240e-010 4.050385e-002
14500 14500 4 54.063 37.6010 2.404240e-010 4.050385e-002
14500 14500 4 54.111 37.5681 2.404240e-010 4.050385e-002
14500 14500 4 54.152 37.5397 2.404240e-010 4.050385e-002
14500 14500 4 54.289 37.4449 2.404240e-010 4.050385e-002
14500 14500 4 54.406 37.3641 2.404240e-010 4.050385e-002
14500 14500 4 54.493 37.3044 2.404240e-010 4.050385e-002
14500 14500 4 54.626 37.2141 2.404240e-010 4.050385e-002
14500 14500 4 54.742 37.1350 2.404240e-010 4.050385e-002
14500 14500 4 54.790 37.1021 2.404240e-010 4.050385e-002
14500 14500 4 55.007 36.9557 2.404240e-010 4.050385e-002
14500 14500 4 56.042 36.2734 2.404240e-010 4.050385e-002
Seems like my cpu is Linpak 64 stable but my NIC was not =(
Anyone got an idea on why it burned my NIC?
I've hit over 80c sporadically in Linpack 64bit, provided the main CPU temp doesnt ever get that high you should be fine. Highest I've seen CPU is at 60c or so with Linpack, but most cores are between 70-80c. Once memory/cpu/fsb clocks are all aligned and in phase Linpack temps go through the roof! It holds at 100% cpu for over 40sec on 15000x15008 dimensions! Most I let Linpack run CPU temp is 65 constantly before I kill the process. All depends on the chip.
@graysky, you mentioned you didn't quite get the differences between core 0/1 and 2/3 in linpack. The differences in load are due to out of skew CPU/DRAM clock signals on either side of the MCH. If data clock signal from Channel A arrives at MCH before Channel B, then it is passed to which even particular die is using the smallest time slice. CPU clock signals are still based on same AGTL+ ref voltage though even for each die, so while 1 die is handling the data which arrived from first memory channel, the other die is sitting there idle until it arrives, how ever many clocks late. So thats the reason for the unstable CPU usage. Mine runs nearly at 100% for every test, it usually takes a few seconds for the PLL to stabilize the clock signals, but the bouncing is the PLL not able to keep them in phase for the duration of the test.
Last edited by mikeyakame; 07-13-2008 at 05:47 AM.
DFI LT-X48-T2R UT CDC24 Bios | Q9550 E0 | G.Skill DDR2-1066 PK 2x2GB |
Geforce GTX 280 729/1566/2698 | Corsair HX1000 | Stacker 832 | Dell 3008WFP
Damn nice! My kentsfield peaks at 42GFlops average at 1940mhz FSB! Thats impressive for a notebook CPU!
Batch file I use to launch Linpack from inside Windows Powershell cause I got sick of doing cd and typing the whole input line. Modify to your needs, I launch it from C:\Users\MyUser
< lp.bat >
-----------------------------------------------------------------------------------------
@ECHO OFF
cd AppData\Roaming\Intel.Linpack.Benchmark
.\linpack_xeon64.exe .\lininput_xeon64
-----------------------------------------------------------------------------------------
< lininput_xeon64 >
-----------------------------------------------------------------------------------------
Intel(R) LINPACK data file
User Specified Datafile
1 # number of tests
14000 # problem sizes ( x = problem_sz; 8 * x^2 = mem_usage; )
14008 # leading dimensions ( x = problem_sz, y = lead_dim; y > x && y % 8 == 0 && y % 16 == 8; )
5 # times to run a test
4 # alignment values ( in kBytes )
-----------------------------------------------------------------------------------------
Last edited by mikeyakame; 07-22-2008 at 05:35 AM.
DFI LT-X48-T2R UT CDC24 Bios | Q9550 E0 | G.Skill DDR2-1066 PK 2x2GB |
Geforce GTX 280 729/1566/2698 | Corsair HX1000 | Stacker 832 | Dell 3008WFP
NIce looking app.
The problem with most software testing is it is coded by man and i can guarantee there will be mistake somewhere that can cause error, so why bother? Until some one can code 100% with no errors and no all bugs on all known hardware then we will have true stability, but that will never happen. A combination of Man made product and man coding is always going to give issues.
Mine: i5 2500k @ 4.7gig - CM 212+ - GTX570 - Giga P67A-UD4 - 2 x 4gig Ripjaws 1600Mhz - OCZ vertex 2 120g - Corsair TX750 - Antec 300
Wifies : Asrock 880GMH - Phenom II B50@3.5 gig - Gskill 2 x 2gig DDR3 1333 - HIS 5770 - 74g Raptor - CM TX3 - Vantec 620w PSU
Kids: Asus M4N68 - Athlon x2 3800+ - 2 x 2gig Kingston - 8800GTS - CM TX3 - 74gig raptor - TT 550w
May I ask why don't you use LinX or IntelBurnTest?
If it ain't broke... fix it until it is.
Bookmarks