Advanced search

Message boards : News : WU: NOELIA_KLEBEs

Author Message
Profile Stoneageman
Avatar
Send message
Joined: 25 May 09
Posts: 224
Credit: 34,057,224,498
RAC: 190
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32423 - Posted: 28 Aug 2013 | 21:49:21 UTC

So far I've had
one long wu error after 13s

one short wu complete OK
one long wu stall, ran for 2hr on a 660 but little progress, so got the bullet

Be on your guard!


5pot
Send message
Joined: 8 Mar 12
Posts: 411
Credit: 2,083,882,218
RAC: 0
Level
Phe
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32430 - Posted: 28 Aug 2013 | 23:52:33 UTC

Encountering issues with Noelia WUs:

http://www.gpugrid.net/result.php?resultid=7218125

http://www.gpugrid.net/result.php?resultid=7218124

http://www.gpugrid.net/result.php?resultid=7218111

All had the same error, swanMemset failed

Stefan
Project administrator
Project developer
Project tester
Project scientist
Send message
Joined: 5 Mar 13
Posts: 348
Credit: 0
RAC: 0
Level

Scientific publications
wat
Message 32438 - Posted: 29 Aug 2013 | 8:18:06 UTC
Last modified: 29 Aug 2013 | 8:24:32 UTC

Since I saw a few error posts popping out about Noelia's new WU's and there was no official thread... I make this thread to collect them all. Once they all come to the office I will inform them.

flashawk
Send message
Joined: 18 Jun 12
Posts: 297
Credit: 3,572,627,986
RAC: 0
Level
Arg
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwat
Message 32443 - Posted: 29 Aug 2013 | 10:00:33 UTC

I've had 5 NOELIA's fail in the past 24 hours.

noelia
Send message
Joined: 5 Jul 12
Posts: 35
Credit: 393,375
RAC: 0
Level

Scientific publications
wat
Message 32444 - Posted: 29 Aug 2013 | 10:16:26 UTC

Hi, for some reason some of you are having problems with this WUs in the new application and we've moved them to the beta queue to have a proper look. I've also just sent 50 WU under the name KLEBEbeta with a much simpler configuration file. These simulations are really important, and fixing this bug will also help for future similar projects in drug discovery. Please report any problems you might have on groups KLEBEs and KLEBEbeta.

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32446 - Posted: 29 Aug 2013 | 11:34:38 UTC - in response to Message 32444.
Last modified: 29 Aug 2013 | 11:57:33 UTC

Noelia,

Thanks for moving to beta to try to fix the problems!
I'm noticing a "new" problem with the NOELIA_KLEBEbeta tasks, though.

Essentially, they get assigned to a GPU, and then when they try to "get going", BOINC shows that they run for about 15-40 seconds, and then the task resets back to the beginning (with Elapsed back to 0 seconds), and it retries.
It just keeps retrying until failure.
Additionally, if the user closes BOINC, the acemd.800-55.exe process for that task does not close properly (it still remains in the Task Manager's process list, even though all other related BOINC processes have exited normally).

Also, looking at stderr.txt for one of the tasks that I aborted (http://www.gpugrid.net/result.php?resultid=7221709), said the following lines that might give a hint as to what's happening:
swanMemset failed
Can't acquire lockfile - exiting
FILE_LOCK::unlock(): close failed.: No error

I have not seen this behavior before today, so I think there is at least 1 new bug here.

This happens both on my GTX 660 Ti, as well as my GTX 460, in Windows 8.1 Preview x64.

The current task exhibiting this behavior is:
109nx4-NOELIA_KLEBEbeta-0-3-RND0846_0

I hope this information helps you to track it down to correct the problem(s) quickly, as right now my GPU is spinning in circles and doing no work. Are you able to reproduce the problem in your testing?

If there's anything else you might need, please let us know.

Thanks,
Jacob

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32447 - Posted: 29 Aug 2013 | 11:59:05 UTC

I've probably fixed the fault. There'll be an updated acemdbeta app very soon.

MJH

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32449 - Posted: 29 Aug 2013 | 12:19:56 UTC - in response to Message 32447.

801 is now live.

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32450 - Posted: 29 Aug 2013 | 12:29:45 UTC
Last modified: 29 Aug 2013 | 12:35:27 UTC

Lol What a "luck" ^^

http://www.gpugrid.net/workunit.php?wuid=4729004

One of my maschine failed this, and i got exactly this wu to my next machine where it stucks O.o saw it now after 2 hours. Puh earliy enough before weekend ^^
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32451 - Posted: 29 Aug 2013 | 12:30:16 UTC - in response to Message 32449.

Thanks for the prompt response. I got an 801 KLEBEbeta, and it's at least getting off the ground now.

I hope that you can see that it is very difficult for us to know if the problem is in the task set, or if the problem is in the application.

Will continue to monitor...

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32452 - Posted: 29 Aug 2013 | 12:36:38 UTC
Last modified: 29 Aug 2013 | 12:36:56 UTC

Would anyone with a cc 1.3 card - Geforce GTX 200 series - please try some of the current acemdbeta v801 Noelia-KLEBE WUs and report back here?

MJH

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32453 - Posted: 29 Aug 2013 | 12:37:07 UTC - in response to Message 32451.
Last modified: 29 Aug 2013 | 12:37:40 UTC

I tested suspending one of these KLEBEbeta tasks, and it caused a driver reset. So, the problem still persists.

Can you please look into it more closely? The issue has to deal with how the KLEBE tasks are exiting - it seems they are not releasing the GPU in a timely fashion, as compared to every other GPU task I run (across all my GPU projects).

Maybe compare the exit logic of a KLEBE task, versus the exit logic of other GPUGrid task types?

Operator
Send message
Joined: 15 May 11
Posts: 108
Credit: 297,176,099
RAC: 0
Level
Asn
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32454 - Posted: 29 Aug 2013 | 12:42:04 UTC

On my Titan box I've gotten two of these NOELIAs.

Both exhibited the same behavior.

"8/29/2013 7:36:55 AM | GPUGRID | Task 063px38-NOELIA_KLEBEs-1-3-RND3786_0 exited with zero status but no 'finished' file"

Over and over again without making much progress.

So I pulled the trigger to kill them.

So it's back to babysitting to make sure I only get NATHAN longs for the time being.

That NOELIA sure has a reputation! ;-}

Operator

____________

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32455 - Posted: 29 Aug 2013 | 12:43:38 UTC - in response to Message 32453.

Jacob,

The current beta addresses the "swanMemset failed" and "access violation" errors.
The suspend problem I have not yet investigated. (Is it with 'suspend to memory' or 'suspend and exit'? )

MJH

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32456 - Posted: 29 Aug 2013 | 12:49:56 UTC - in response to Message 32455.
Last modified: 29 Aug 2013 | 13:00:27 UTC

Jacob,

The current beta addresses the "swanMemset failed" and "access violation" errors.
The suspend problem I have not yet investigated. (Is it with 'suspend to memory' or 'suspend and exit'? )

MJH


There's a whole thread about it, where I posted in as much detail as I could about the problem, on 4/4/2013 (4+ months ago), here:
http://www.gpugrid.net/forum_thread.php?id=3333

It happens whenever a NOELIA task (especially KLEBE) is suspended for any reason, including:
- BOINC set to Snooze
- BOINC set to Snooze GPU
- BOINC set to Suspend
- BOINC set to Suspend GPU
- BOINC set to Suspend due to exclusive app running
- BOINC set to Suspend GPU due to exclusive GPU app running
- GPUGrid project set to Suspend
- NOELIA KLEBE task set to Suspend
- BOINC exited with "Stop running tasks" checked

Something in the KLEBE exit logic has been causing driver resets and watchdog timeouts, for several months, for many of your Windows users. I sure hope you guys can work together to get a handle on it!

Note: I do use the "Leave application in memory when suspended" setting, but so far as I know, that is irrelevant to GPU tasks. When a GPU task is suspended, BOINC has to remove it from memory, regardless of that user setting. It treats GPU tasks differently because there's no PageFile backing the GPU RAM.

Thanks for looking into this. It's my biggest problem across all of my 20 BOINC projects.

TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32457 - Posted: 29 Aug 2013 | 13:16:01 UTC

I have two NOELIA KLEEBEbeata on my 770 and they start, then when 0.021% complete, no more progress but they keep running, 2h57m16s elapsed and 0h0m0s remaining. This was in app 8.00 and I have now aborted these WU's and try the new 8.01 app.

I have now 1 with the 8.01 (cuda55) app (NOELIA KLEEBEbeta) and it is running normal. Twelve hour to finish, progress runs up, elapsed time runs up, and remaining runs down. Win7 x64, BOINC 7.0.64, driver 326.80
____________
Greetings from TJ

Carlos Augusto Engel
Send message
Joined: 5 Jun 09
Posts: 38
Credit: 2,880,758,878
RAC: 0
Level
Phe
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32459 - Posted: 29 Aug 2013 | 14:07:08 UTC - in response to Message 32457.

I have two NOELIA KLEEBEbeata on my 770 and they start, then when 0.021% complete, no more progress but they keep running, 2h57m16s elapsed and 0h0m0s remaining. This was in app 8.00 and I have now aborted these WU's and try the new 8.01 app.




Same thing here , but after 30 minutes i stop it.

http://www.gpugrid.net/result.php?resultid=7221521


____________

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32464 - Posted: 29 Aug 2013 | 14:51:21 UTC - in response to Message 32452.
Last modified: 29 Aug 2013 | 15:02:22 UTC

Would anyone with a cc 1.3 card - Geforce GTX 200 series - please try some of the current acemdbeta v801 Noelia-KLEBE WUs and report back here?

MJH


Ok i started one with 8.01. But this can take some time even on my 670mhz 285gtx..I normaly dont run gpugrid on this anymore. It will need about 33hours. The short run 8.00 was ok on this card. I dont think anybody still uses a powerhungry 200series on long runs O.o
____________
DSKAG Austria Research Team: http://www.research.dskag.at



TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32467 - Posted: 29 Aug 2013 | 15:12:49 UTC

With the new 8.01 app they run normal!

I noticed the following:
On the 770 I have this one: 063px79-NOELIA_KLEEBEbeta2-0-3-RND678_0
MEM use: 1003MB
Clock: 1097MHz (however I have set the clock to 1060MHz!)
GPU load: 87%
Temp: 65°C
7.5% done in 40 minutes

On the 660 I have this one: 109nx37-NOELIA_KLEEBEbeta-0-3-RND0283_0
MEM use: 779MB
Clock: 1045MHZ (as I set it)
GPU load: ~88%
Temp: 67°C
5.6% done in 40 minutes

I now these are not the same WU's and the GPU's are not the same as well. But it is strange that the WU can manage to get the clock higher, or this must have been the result of the faulty WU that I aborted and not reboot afterwards.

Due to the difference in memory load it can also be that cards with only 1MB can not do these WU's as before. That may result in some comments ;-)
____________
Greetings from TJ

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32468 - Posted: 29 Aug 2013 | 15:16:55 UTC - in response to Message 32464.

Would anyone with a cc 1.3 card - Geforce GTX 200 series - please try some of the current acemdbeta v801 Noelia-KLEBE WUs and report back here?

MJH


Ok i started one with 8.01. But this can take some time even on my 670mhz 285gtx..I normaly dont run gpugrid on this anymore. It will need about 33hours. The short run 8.00 was ok on this card. I dont think anybody still uses a powerhungry 200series on long runs O.o


Oh and when somebody started one too on 200series, plz tell me, got my energybill today, so i would love to stop it the next hours when not needed :p
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32470 - Posted: 29 Aug 2013 | 15:22:38 UTC - in response to Message 32464.
Last modified: 29 Aug 2013 | 15:46:47 UTC

063px6-NOELIA_KLEBEbeta-0-3-RND7897_0 Workunit stuck at 0.021% (8.00 app though).

While 'running' the stderr.txt file already reported swanMemset failed, but the WU didn't terminate:

# GPU [GeForce GTX 660] Platform [Windows] Rev [3170M] VERSION [55]
# SWAN Device 1 :
# Name : GeForce GTX 660
# ECC : Disabled
# Global mem : 2048MB
# Capability : 3.0
# PCI ID : 0000:02:00.0
# Device clock : 1032MHz
# Memory clock : 3004MHz
# Memory width : 192bit
# Driver version : r325_00
swanMemset failed

Suspended the WU. When I resumed it, 5min later, I got the error message,
"Display driver nvlddmkm stopped responding and has successfully recovered".

When I checked my Windows logs I saw,
"A request to disable the Desktop Window Manager was made by process (4)" - listed 2sec before the driver crash/restart entry. The driver log entry was made after the driver restarted rather than when the failure was triggered.

The WU again continued 'running' without progressing. I aborted it but now the stderr has nothing of any use,

    Stderr output

    <core_client_version>7.0.64</core_client_version>
    <![CDATA[
    <message>
    aborted by user
    </message>
    ]]>


____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32471 - Posted: 29 Aug 2013 | 15:25:21 UTC
Last modified: 29 Aug 2013 | 15:25:35 UTC

I have noticed that, using the 8.01 app on a NOELIA_KLEBEbeta task, on my GTX 660 Ti, the process does not utilize a full CPU core (like other GPUGrid tasks normally do for that GPU). It's like SWAN_SYNC is not set correctly. Though I'm still getting good (85-91%) GPU utilization for the task.

Is this behavior new? Also, is it expected?

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32472 - Posted: 29 Aug 2013 | 15:32:51 UTC - in response to Message 32468.


Oh and when somebody started one too on 200series, plz tell me, got my energybill today, so i would love to stop it the next hours when not needed :p


If it is still running then that's plenty long enough to demonstrate that all is well, thanks. You can kill it off.

Matt

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32473 - Posted: 29 Aug 2013 | 15:33:51 UTC - in response to Message 32471.


I have noticed that, using the 8.01 app on a NOELIA_KLEBEbeta task, on my GTX 660 Ti, the process does not utilize a full CPU core (like other GPUGrid tasks normally do for that GPU). It's like SWAN_SYNC is not set correctly. Though I'm still getting good (85-91%) GPU utilization for the task.


It should have exactly the same load profile as 8.00 did.

MJH

Profile dskagcommunity
Avatar
Send message
Joined: 28 Apr 11
Posts: 456
Credit: 817,865,789
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32475 - Posted: 29 Aug 2013 | 15:50:04 UTC - in response to Message 32472.


Oh and when somebody started one too on 200series, plz tell me, got my energybill today, so i would love to stop it the next hours when not needed :p


If it is still running then that's plenty long enough to demonstrate that all is well, thanks. You can kill it off.

Matt


Ok it ran one hour, was at 3,3%, 95% gpu load, used 515MB VRAM, cpu was busy working on LHC and still computed normal. Thx Aborted it. ^^
____________
DSKAG Austria Research Team: http://www.research.dskag.at



Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32476 - Posted: 29 Aug 2013 | 16:00:08 UTC

Have revved the beta app to 8.02. This might also fix the driver-hang-on-suspend problem.

MJH

flashawk
Send message
Joined: 18 Jun 12
Posts: 297
Credit: 3,572,627,986
RAC: 0
Level
Arg
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwat
Message 32481 - Posted: 29 Aug 2013 | 17:30:49 UTC - in response to Message 32471.

I have noticed that, using the 8.01 app on a NOELIA_KLEBEbeta task, on my GTX 660 Ti, the process does not utilize a full CPU core (like other GPUGrid tasks normally do for that GPU). It's like SWAN_SYNC is not set correctly. Though I'm still getting good (85-91%) GPU utilization for the task.

Is this behavior new? Also, is it expected?


I thought NOELIA's never used a full CPU core, that's the way it's always been. We've talked about it before in different threads.

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32482 - Posted: 29 Aug 2013 | 17:36:48 UTC - in response to Message 32481.

That's fine if that's the case. I don't know, which is why I asked. I'm "used" to seeing tasks on my "Kepler" (GTX 660 Ti) taking a full CPU core (via SWAN_SYNC) automatically. Maybe NOELIA tasks work differently.

Profile nenym
Send message
Joined: 31 Mar 09
Posts: 137
Credit: 1,308,230,581
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32484 - Posted: 29 Aug 2013 | 18:19:04 UTC
Last modified: 29 Aug 2013 | 19:11:45 UTC

ID 156955: i7-3770K (no HT, 4 CPU cores) GTX560Ti, W7 64bit driver 328.80.
NOELIA_KLEBEbeta 8.02 CUDA55 application
GPU [GeForce GTX 560 Ti] Platform [Windows] Rev [3182M] VERSION [55]
# SWAN Device 0 :
# Name : GeForce GTX 560 Ti
# ECC : Disabled
# Global mem : 1024MB
# Capability : 2.1
# PCI ID : 0000:01:00.0
# Device clock : 1720MHz
# Memory clock : 2100MHz
# Memory width : 256bit
# Driver version : r325_00
GPU load 91 - 94%, process priority tamed to realtime, GPU load is 87 - 91% if not tamed
CPU load 20% of one core (5% in all)

active_task>
<project_master_url>http://www.gpugrid.net/</project_master_url>
<result_name>063px30-NOELIA_KLEBEbeta2-0-3-RND2325_0</result_name>
<checkpoint_cpu_time>564.224400</checkpoint_cpu_time>
<checkpoint_elapsed_time>2994.810513</checkpoint_elapsed_time>
<fraction_done>0.048235</fraction_done>
</active_task>

Seems to run OK so far.

Concurrently running: 4x CPU Asteroids SSE3, 1x GPU Einstein BPRS on Intel HD4000
Note: previous 6.18 CUDA 4.2 application could run on 875 MHz core clock. Factory OC of the GPU is 900 MHz.

flashawk
Send message
Joined: 18 Jun 12
Posts: 297
Credit: 3,572,627,986
RAC: 0
Level
Arg
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwat
Message 32487 - Posted: 29 Aug 2013 | 18:29:41 UTC - in response to Message 32482.

That's fine if that's the case. I don't know, which is why I asked. I'm "used" to seeing tasks on my "Kepler" (GTX 660 Ti) taking a full CPU core (via SWAN_SYNC) automatically. Maybe NOELIA tasks work differently.


I understand, I know your a very busy man, I thought I saw you debugging apps for other projects in some different forums some place else. I don't know how you manage to keep track of them all.

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32488 - Posted: 29 Aug 2013 | 18:35:58 UTC - in response to Message 32487.

:) Yeah, thanks. I've helped Einstein fix a bug, MindModeling fix a bug, GPUGrid fix a couple things, Test4Theory fix a bug, Rosetta fix their app, SETI fix a GPU estimate problem, got nVidia to fix a monitor-sleep issue, and more. And I also do alpha/beta testing of the actual BOINC software, and have worked directly with the BOINC devs.

Regarding this particular case, I believe I was aware of some tasks "not using a full CPU core on my Kepler card", but I did not know it was NOELIA ones. I'll try to keep that in mind.

Thanks again.

Profile Retvari Zoltan
Avatar
Send message
Joined: 20 Jan 09
Posts: 2343
Credit: 16,201,255,749
RAC: 6,169
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32503 - Posted: 29 Aug 2013 | 20:05:10 UTC - in response to Message 32482.

That's fine if that's the case. I don't know, which is why I asked. I'm "used" to seeing tasks on my "Kepler" (GTX 660 Ti) taking a full CPU core (via SWAN_SYNC) automatically. Maybe NOELIA tasks work differently.

I still consider this different CPU load as a malfunction.
However, with this low CPU load the GPU load is still above 95%, so we can turn this question the way around: is it sure that the other tasks need a full CPU thread to feed a Kepler GPU?

Profile Retvari Zoltan
Avatar
Send message
Joined: 20 Jan 09
Posts: 2343
Credit: 16,201,255,749
RAC: 6,169
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32504 - Posted: 29 Aug 2013 | 20:09:14 UTC - in response to Message 32476.

Have revved the beta app to 8.02. This might also fix the driver-hang-on-suspend problem.

MJH

I think you could promote this 8.02 to the production queue at once, as it is proved to be better than the 8.00.

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32505 - Posted: 29 Aug 2013 | 20:14:33 UTC - in response to Message 32504.

It's there now.

MJH

Jim1348
Send message
Joined: 28 Jul 12
Posts: 819
Credit: 1,591,285,971
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32507 - Posted: 29 Aug 2013 | 21:08:18 UTC - in response to Message 32503.
Last modified: 29 Aug 2013 | 21:32:37 UTC

That's fine if that's the case. I don't know, which is why I asked. I'm "used" to seeing tasks on my "Kepler" (GTX 660 Ti) taking a full CPU core (via SWAN_SYNC) automatically. Maybe NOELIA tasks work differently.

I still consider this different CPU load as a malfunction.
However, with this low CPU load the GPU load is still above 95%, so we can turn this question the way around: is it sure that the other tasks need a full CPU thread to feed a Kepler GPU?

On the Folding forum, there have been extended discussions of Nvidia CPU core usage under CUDA. It contrasts to the case of AMD cards running OpenCL, which typically require only a few percent of a CPU core.

As I recall, Nvidia provides the option to the developers to reserve a full CPU core when running under CUDA using spin states, which I don't understand anyway. If the application developers want to ensure that they have enough CPU support, they can reserve it, even though typically not all of it is actually in use

So maybe the other tasks don't really require a full core, except that it may be useful to reserve it for stability or performance or whatever.

EDIT: To further complicate matters, Nvidia cards running OpenCL always require a full CPU core; there is no option not to.

5pot
Send message
Joined: 8 Mar 12
Posts: 411
Credit: 2,083,882,218
RAC: 0
Level
Phe
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32520 - Posted: 30 Aug 2013 | 4:35:07 UTC

8.02 beta tasks seem to work ok on 780s, but now all other tasks fail

Richard Haselgrove
Send message
Joined: 11 Jul 09
Posts: 1576
Credit: 5,605,686,851
RAC: 8,689,502
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32525 - Posted: 30 Aug 2013 | 8:13:54 UTC - in response to Message 32507.

I still consider this different CPU load as a malfunction.
However, with this low CPU load the GPU load is still above 95%, so we can turn this question the way around: is it sure that the other tasks need a full CPU thread to feed a Kepler GPU?

On the Folding forum, there have been extended discussions of Nvidia CPU core usage under CUDA. It contrasts to the case of AMD cards running OpenCL, which typically require only a few percent of a CPU core.

As I recall, Nvidia provides the option to the developers to reserve a full CPU core when running under CUDA using spin states, which I don't understand anyway. If the application developers want to ensure that they have enough CPU support, they can reserve it, even though typically not all of it is actually in use

So maybe the other tasks don't really require a full core, except that it may be useful to reserve it for stability or performance or whatever.

EDIT: To further complicate matters, Nvidia cards running OpenCL always require a full CPU core; there is no option not to.

Watching two different third-party developers working on SETI (one specialising in CUDA, the other in OpenCL), we get the opposite outcome: OpenCL on ATI is inefficient unless a spare CPU core is available, but CUDA on Nvidia requires very little CPU.

I'm not a developer myself (at least, not at the level these guys program), but from the peanut gallery it looks as if CPU usage is very much down to the skill of the developer, and how well they know their platform and tools.

But I'm interested by the OpenCL on Nvidia point. That does seem to be a common observation - I wonder if it has necessarily to be so? Or maybe Mvidia didn't port some of their synch technology from CUDA to the OpenCL toolchain yet?

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32526 - Posted: 30 Aug 2013 | 8:19:17 UTC - in response to Message 32520.
Last modified: 30 Aug 2013 | 8:27:09 UTC

8.02 app running the Noelia Beta WU's (one on the CUDA4.2 and the other on the 5.5 app).

When I use snooze the driver still restarts. I have the driver timeout set to 20sec, and it takes 20seconds for the driver to crash/restart.

When I suspended the WU's individually they didn't cause a driver restart.

However when I suspended both at the same time the driver restarted, again after 20sec. (These situation driver restarts/or lack of restarts are repeatable).

I noted that the 5.5 WU kept running (progressing) for about 4seconds after I suspended it.

But I'm interested by the OpenCL on Nvidia point. That does seem to be a common observation - I wonder if it has necessarily to be so? Or maybe Mvidia didn't port some of their synch technology from CUDA to the OpenCL toolchain yet?

The GK104 cards are supposed to be OpenCL 1.2 but the drivers are only OpenCL1.1, which means the toolkit can't be 1.2.
AMD/ATI supports OpenCL1.2, Intel supports OpenCL1.2, NVidia says it's GPU's are OpenCL1.2 but their drivers prevent the cards from being used for OpenCL1.2.
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32531 - Posted: 30 Aug 2013 | 8:55:38 UTC - in response to Message 32526.

On my Linux systems I have the STABLE Repository drivers (304.88), supposedly only CUDA 5.0.
However I'm presently running a CUDA 5.5 NOELIA Beta WU (12h in 3 to go).
I thought CUDA 5.5 would only be used if the system had the correct drivers?
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32535 - Posted: 30 Aug 2013 | 10:33:23 UTC - in response to Message 32531.
Last modified: 30 Aug 2013 | 10:33:34 UTC


I thought CUDA 5.5 would only be used if the system had the correct drivers?


The intent is that you'll get 55 only if the driver revision is >= 315.15
Alas, the scheduler has a will of its own.

MJH

Jim1348
Send message
Joined: 28 Jul 12
Posts: 819
Credit: 1,591,285,971
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32537 - Posted: 30 Aug 2013 | 13:25:39 UTC - in response to Message 32525.
Last modified: 30 Aug 2013 | 13:27:35 UTC

Watching two different third-party developers working on SETI (one specialising in CUDA, the other in OpenCL), we get the opposite outcome: OpenCL on ATI is inefficient unless a spare CPU core is available, but CUDA on Nvidia requires very little CPU.

I'm not a developer myself (at least, not at the level these guys program), but from the peanut gallery it looks as if CPU usage is very much down to the skill of the developer, and how well they know their platform and tools.

That is quite true from my own experience also (as a user only), but I think we are talking about two different things. Neither ATI on OpenCL nor Nvidia on CUDA require a CPU core unless the project developer requires it. And usually CUDA can be made more efficient with CPU usage. Certainly that is the case with Folding with their separate OpenCL core_16 (for AMD cards only) and CUDA core_15 versions (obviously for Nvidia cards only); the CUDA one is much better (less than 1 percent verses maybe 20 percent or more).

But I'm interested by the OpenCL on Nvidia point. That does seem to be a common observation - I wonder if it has necessarily to be so? Or maybe Mvidia didn't port some of their synch technology from CUDA to the OpenCL toolchain yet?

All I know is that on Folding with their newest OpenCL core_17, which runs on both AMD and Nvidia, the situation is reversed. It requires only 1 or 2 percent on AMD cards (e.g., my HD 7870 on an i7-3770), whereas on an Nvidia card it reserves a full core (e.g., on my GTX 660 Ti). The question has been asked on the Folding forum as to whether that is necessary, and the answer is that Nvidia has not implemented the option in OpenCL to use less than a full core. Apparently they could if they wanted to, but maybe for performance reasons (so the speculation goes) they want their cards to perform the best they can, so they just grab the whole core. It helps solve the problem you mentioned above, where users don't always know to leave a core free I suppose.

Richard Haselgrove
Send message
Joined: 11 Jul 09
Posts: 1576
Credit: 5,605,686,851
RAC: 8,689,502
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32539 - Posted: 30 Aug 2013 | 15:30:19 UTC - in response to Message 32537.

But I'm interested by the OpenCL on Nvidia point. That does seem to be a common observation - I wonder if it has necessarily to be so? Or maybe Mvidia didn't port some of their synch technology from CUDA to the OpenCL toolchain yet?

All I know is that on Folding with their newest OpenCL core_17, which runs on both AMD and Nvidia, the situation is reversed. It requires only 1 or 2 percent on AMD cards (e.g., my HD 7870 on an i7-3770), whereas on an Nvidia card it reserves a full core (e.g., on my GTX 660 Ti). The question has been asked on the Folding forum as to whether that is necessary, and the answer is that Nvidia has not implemented the option in OpenCL to use less than a full core. Apparently they could if they wanted to, but maybe for performance reasons (so the speculation goes) they want their cards to perform the best they can, so they just grab the whole core. It helps solve the problem you mentioned above, where users don't always know to leave a core free I suppose.

That was my suspicion too. In trying to pass messages between the two developers - apparently the new CUDA way is to use 'callback' rather than 'spin' synch - I was invited to refer to the NVidia toolkit documentation to find examples for the OpenCL implementation. I couldn't find any.

If there are any unbiased developer observers of this thread, it would be useful to hear if there is any factual basis for our observations - and for the rumour I've heard that NVidia might pull away from OpenCL support entirely. That would be a shame, if true - both NVidia and ATI (as it was then) were founder members of the Khronos Group in January 2000. It would be a pity if competition drove out collaboration, and we returned to the days of two incompatible native-code development environments.

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32540 - Posted: 30 Aug 2013 | 15:49:38 UTC - in response to Message 32539.
Last modified: 30 Aug 2013 | 15:50:09 UTC

Perhaps you might create a new thread devoted toward finding the OpenCL/CUDA information.
This thread is for "WU: NOELIA_KLEBEs" :)

TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32541 - Posted: 30 Aug 2013 | 15:58:22 UTC - in response to Message 32540.
Last modified: 30 Aug 2013 | 15:59:09 UTC

Perhaps you might create a new thread devoted toward finding the OpenCL/CUDA information.
This thread is for "WU: NOELIA_KLEBEs" :)

And NOELIA_KLEBEbeta's which run fine by the way on my 660 and 770 with 8.02!
Noelia and MJH did a good job with this.
____________
Greetings from TJ

Profile The King's Own
Avatar
Send message
Joined: 25 Apr 12
Posts: 32
Credit: 945,543,997
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwat
Message 32544 - Posted: 30 Aug 2013 | 17:08:01 UTC

http://www.gpugrid.net/result.php?resultid=7221215

Would progress to 0.21% and then sit while elapsed time increased. Switched from 660Ti to 580 with same result. Aboirted after 3 hrs 38 min on 580.
____________

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32545 - Posted: 30 Aug 2013 | 17:17:35 UTC - in response to Message 32544.

King's Own. 8.00 is deprecated - your problem is fixed in the current release.

Profile The King's Own
Avatar
Send message
Joined: 25 Apr 12
Posts: 32
Credit: 945,543,997
RAC: 0
Level
Glu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwat
Message 32546 - Posted: 30 Aug 2013 | 17:31:23 UTC - in response to Message 32545.

Thank you.

However;

i. I would have to be convinced that this my problem?

ii. Why are deprecated WU being dispatched? That certainly is not a problem caused by me.

____________

Profile Ascholten
Send message
Joined: 21 Dec 10
Posts: 7
Credit: 78,122,357
RAC: 0
Level
Thr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwat
Message 32551 - Posted: 30 Aug 2013 | 21:01:37 UTC

IM getting zero output file errors every few minutes and it slams the fan on my card and resets the task.

It is also showing my video cards in like slots 7 and 8. I believe the should be slot 0 and 1 or is that moot?

This has been going on a few days, I reset the project and aborted a few tasks thinking they were the problem to find it's ongoing.

I see this is an issue that is known?? Any etr?

Thank you
Aaron

ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32553 - Posted: 30 Aug 2013 | 21:04:22 UTC - in response to Message 32525.

Richard wrote:
Or maybe Mvidia didn't port some of their synch technology from CUDA to the OpenCL toolchain yet?

That's what I suppose as well, without being a GPU developer. Over a year ago nVidias performance at POEM OpenCL was horrible, but they only used ~50% of one core. A driver update doubled performance but since then they're using a full CPu core.

To me it seems like "just use a full core" was a quick fix. And now they don't want to push OpenCL any further than they have to and just stick with this solution.

MrS
____________
Scanning for our furry friends since Jan 2002

TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32560 - Posted: 31 Aug 2013 | 5:17:44 UTC

How can you all see that a full core is used with one GPUGRID WU?
____________
Greetings from TJ

Profile Retvari Zoltan
Avatar
Send message
Joined: 20 Jan 09
Posts: 2343
Credit: 16,201,255,749
RAC: 6,169
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32562 - Posted: 31 Aug 2013 | 8:31:57 UTC - in response to Message 32560.
Last modified: 31 Aug 2013 | 8:46:15 UTC

How can you all see that a full core is used with one GPUGRID WU?

It can be checked in the Windows Task Manager: look for the acemd.80x-55.exe (or acemd.80x-42.exe) on the "Processes" tab. If its CPU usage is 1-2%, then it's not using a full core, otherwise the CPU usage is 100/the number of your CPU's threads (12-13% on a 8-threaded CPU, 8% on a 12-threaded CPU). You can check the past workunits' CPU usage at your hosts' task list: if the "CPU time" (almost) equals the "run time", then the task used a full core, if the "CPU time" is significantly less than the "run time", then it didn't use a full core.

Profile skgiven
Volunteer moderator
Volunteer tester
Avatar
Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32564 - Posted: 31 Aug 2013 | 12:18:12 UTC - in response to Message 32562.

Just note that the NOELIA_KLEBE WU's don't use a full CPU core/thread - Never have.
My Boinc scheduler has them at 0.595 CPU's, but actual use is less than that (2 or 3% of the entire CPU, which means <=0.25 CPU threads).
____________
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help

Jacob Klein
Send message
Joined: 11 Oct 08
Posts: 1127
Credit: 1,901,927,545
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32568 - Posted: 31 Aug 2013 | 14:39:01 UTC - in response to Message 32564.

Thanks for the info, skgiven.

I have overridden the CPU requirements, via app_config.xml. Because I have 2 GPUs that do GPUGrid (1 Fermi, 1 Kepler), I had set cpu_usage to 0.5 for all GPUGrid app types, so that when both cards are working on GPUGrid, BOINC reserves 1 total CPU core for them, keeping the CPU slightly above saturation. I've since changed my logic a bit so as to slightly undersaturate the CPU; I accomplished that by changing cpu_usage to 1.0 for all GPUGrid app types, so a logical CPU core is reserved for each, which I think is what you guys always recommended anyway.

Long story short, I used Process Explorer to confirm that NOELIA_KLEBE units strangely do not use a full CPU on my Kepler card, whereas it seems to me that every other GPUGrid task does use a full CPU on my Kepler card. It matters to me since they are "mixed in" with other tasks in the "long" app, and my cpu_usage setting now applies to some tasks that won't use a full core. In a perfect world, and if I were an admin, I might consider placing "strange types" like this in a separate app queue, maybe.

Thank you very much for confirming this is "normal" for NOELIA_KLEBE on Kepler.
Jacob.

TJ
Send message
Joined: 26 Jun 09
Posts: 815
Credit: 1,470,385,294
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32571 - Posted: 31 Aug 2013 | 15:26:26 UTC - in response to Message 32562.

How can you all see that a full core is used with one GPUGRID WU?

It can be checked in the Windows Task Manager: look for the acemd.80x-55.exe (or acemd.80x-42.exe) on the "Processes" tab. If its CPU usage is 1-2%, then it's not using a full core, otherwise the CPU usage is 100/the number of your CPU's threads (12-13% on a 8-threaded CPU, 8% on a 12-threaded CPU). You can check the past workunits' CPU usage at your hosts' task list: if the "CPU time" (almost) equals the "run time", then the task used a full core, if the "CPU time" is significantly less than the "run time", then it didn't use a full core.

Thanks Zoltan,
This is what I thought and in this way I look at task manager. With Noelia WU, the one we have now and in the past use 1-3%. Rosetta is using 13% per core.
I have also seen Nathans not using less then 13% and Santi's that use not 13% all the time. It was fluctuating from 2% steady to 11% for seconds and then back to 2% again. But I am not watching task manager a lot.
____________
Greetings from TJ

Bedrich Hajek
Send message
Joined: 28 Mar 09
Posts: 467
Credit: 8,194,571,966
RAC: 10,457,610
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32575 - Posted: 31 Aug 2013 | 20:09:27 UTC

I had a NOEL_KLEBEbeta WU error out because of this:

8/31/2013 3:59:44 PM | GPUGRID | Aborting task 063px53-NOELIA_KLEBEbeta2-2-3-RND7138_0: exceeded elapsed time limit 4172.44 (250000000.00G/2062.63G)
8/31/2013 3:59:47 PM | GPUGRID | Computation for task 063px53-NOELIA_KLEBEbeta2-2-3-RND7138_0 finished


Here is the link:


http://www.gpugrid.net/result.php?resultid=7230639

The unit was good to that point.


Bedrich Hajek
Send message
Joined: 28 Mar 09
Posts: 467
Credit: 8,194,571,966
RAC: 10,457,610
Level
Tyr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 32576 - Posted: 31 Aug 2013 | 23:25:51 UTC - in response to Message 32575.

I had another NOEL_KLEBEbeta WU error out.


http://www.gpugrid.net/result.php?resultid=7231327


8/31/2013 7:16:37 PM | GPUGRID | Aborting task 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0: exceeded elapsed time limit 3857.75 (250000000.00G/64804.69G)
8/31/2013 7:16:41 PM | GPUGRID | Computation for task 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0 finished
8/31/2013 7:16:41 PM | GPUGRID | Output file 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0_1 for task 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0 absent
8/31/2013 7:16:41 PM | GPUGRID | Output file 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0_2 for task 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0 absent
8/31/2013 7:16:41 PM | GPUGRID | Output file 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0_3 for task 063px63-NOELIA_KLEBEbeta2-2-3-RND4579_0 absent



nanoprobe
Send message
Joined: 26 Feb 12
Posts: 184
Credit: 222,376,233
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwat
Message 32587 - Posted: 1 Sep 2013 | 12:14:42 UTC

I updated to the 326.98 driver last night. The 1 NOEL_KLEBEbeta WU I received failed with 197 (0xc5) EXIT_TIME_LIMIT_EXCEEDED error message.
http://www.gpugrid.net/result.php?resultid=7232695

The 20 or so MJHARVEY_TEST14 betas I received all finished and validated.

nanoprobe
Send message
Joined: 26 Feb 12
Posts: 184
Credit: 222,376,233
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwat
Message 32598 - Posted: 1 Sep 2013 | 16:42:00 UTC - in response to Message 32587.

I updated to the 326.98 driver last night. The 1 NOEL_KLEBEbeta WU I received failed with 197 (0xc5) EXIT_TIME_LIMIT_EXCEEDED error message.
http://www.gpugrid.net/result.php?resultid=7232695

The 20 or so MJHARVEY_TEST14 betas I received all finished and validated.


Spoke too soon. One of the MJHARVEY_TEST betas failed with the same 197 (0xc5) EXIT_TIME_LIMIT_EXCEEDED error message.

http://www.gpugrid.net/result.php?resultid=7233613

Profile MJH
Project administrator
Project developer
Project scientist
Send message
Joined: 12 Nov 07
Posts: 696
Credit: 27,266,655
RAC: 0
Level
Val
Scientific publications
watwat
Message 32678 - Posted: 4 Sep 2013 | 10:26:16 UTC

Just a note: there are also NOELA_KLEBE WUs on the acemdbeta queue. Somewhat confusingly, those are test WUs for the beta app and aren't part of this batch. If you have problems, please check the application that was used and report it over on thread about the beta application if appropriate:

http://www.gpugrid.net/forum_thread.php?id=3465

Thanks!

MJH

Post to thread

Message boards : News : WU: NOELIA_KLEBEs

//