[email protected] - Failed Server Upload p2411_Ribo_aspartic

Associate
Joined
3 Jul 2004
Posts
866
Location
Helena, Montana
I have a 600 point WU that's been trying to send for a few days now, but the server says full accept. How long does it take for them to clear the backlog on these servers. It's not a problem on my connection end, as I've processed a couple other uints since. Here's the relevant log portions on the WU:

[08:49:24] + Processing work unit
[08:49:24] Core required: FahCore_78.exe
[08:49:24] Core found.
[08:49:24] - Autosending finished units...
[08:49:24] Trying to send all finished work units
[08:49:24] + No unsent completed units remaining.
[08:49:24] - Autosend completed
[08:49:24] Working on Unit 06 [October 2 08:49:24]
[08:49:24] + Working ...


[08:49:24] *------------------------------*
[08:49:24] [email protected] Gromacs Core
[08:49:24] Version 1.90 (March 8, 2006)
[08:49:24]
[08:49:24] Preparing to commence simulation
[08:49:24] - Assembly optimizations manually forced on.
[08:49:24] - Not checking prior termination.
[08:49:51] - Expanded 3064521 -> 16583281 (decompressed 541.1 percent)
[08:49:54]
[08:49:54] Project: 2411 (Run 5, Clone 3, Gen 12)
[08:49:54]
[08:49:56] Assembly optimizations on if available.
[08:49:56] Entering M.D.
[08:50:19] (Starting from checkpoint)
[08:50:19] Protein: p2411_Ribo_aspartic


02:10:57] Completed 250000 out of 250000 steps (100)
[02:10:57] Writing final coordinates.
[02:11:01] Past main M.D. loop
[02:12:02]
[02:12:02] Finished Work Unit:
[02:12:02] - Reading up to 2112072 from "work/wudata_06.arc": Read 2112072
[02:12:02] - Reading up to 113640 from "work/wudata_06.xtc": Read 113640
[02:12:02] goefile size: 0
[02:12:02] logfile size: 189634
[02:12:02] Leaving Run
[02:12:03] - Writing 2931110 bytes of core data to disk...
[02:12:04] ... Done.
[02:12:04] - Shutting down core
[02:12:04]
[02:12:04] [email protected] Core Shutdown: FINISHED_UNIT
[02:12:07] CoreStatus = 64 (100)
[02:12:07] Unit 6 finished with 93 percent of time to deadline remaining.
[02:12:07] Updated performance fraction: 0.953525
[02:12:08] Sending work to server


[02:12:08] + Attempting to send results
[02:12:08] - Reading file work/wuresults_06.dat from core
[02:12:08] (Read 2931110 bytes from disk)
[02:12:08] Connecting to http://171.64.122.142:8080/
[02:12:09] - Couldn't send HTTP request to server
[02:12:09] + Could not connect to Work Server (results)
[02:12:09] (171.64.122.142:8080)
[02:12:09] - Error: Could not transmit unit 06 (completed October 3) to work server.
[02:12:09] - 1 failed uploads of this unit.
[02:12:09] Keeping unit 06 in queue.
[02:12:09] Trying to send all finished work units


[09:09:18] + Attempting to send results
[09:09:18] - Reading file work/wuresults_06.dat from core
[09:09:18] (Read 2931110 bytes from disk)
[09:09:18] Connecting to http://171.64.122.142:8080/
[09:09:19] - Couldn't send HTTP request to server
[09:09:19] + Could not connect to Work Server (results)
[09:09:19] (171.64.122.142:8080)
[09:09:19] - Error: Could not transmit unit 06 (completed October 3) to work server.
[09:09:19] - 23 failed uploads of this unit.
.
.
 
Soldato
Joined
16 Dec 2005
Posts
14,443
Location
Manchester
Stop the client and restart it in -configonly mode. Now ask for Deadlineless WUs. Restart client again and when your current WU finishes it should force a server change and your WU may very well get sent off.

I did this a while ago and it worked.
 
Associate
Joined
3 Jul 2004
Posts
866
Location
Helena, Montana
SiriusB said:
Stop the client and restart it in -configonly mode. Now ask for Deadlineless WUs. Restart client again and when your current WU finishes it should force a server change and your WU may very well get sent off.

I did this a while ago and it worked.

I reconfigured it, but failed again on that attempt. Will give it a day or so and see how things go.
.
 
Associate
Joined
3 Jul 2004
Posts
866
Location
Helena, Montana
When I pulled off the port 8080 it repsonded with OK ... will take the laptop home and try it from my network there. My be some sort of port issue here, not sure.
.
 
Top