This repository has been archived by the owner on Sep 23, 2020. It is now read-only.
/
changelog.html
2327 lines (2226 loc) · 77.5 KB
/
changelog.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
m4_include(/mcs/m4/worksp.lib.m4)
_NIMBUS_HEADER(Changelog)
_NIMBUS_HEADER2(n,n,n,n,y,n,n)
_NIMBUS_LEFT2_COLUMN
_NIMBUS_LEFT2_ABOUT_SIDEBAR(n,n,n,y)
_NIMBUS_LEFT2_COLUMN_END
_NIMBUS_CENTER2_COLUMN
<h2>Changelog</h2>
<p>For cloud client changes, see
<a href="http://github.com/nimbusproject/nimbus/raw/master/cloud-client/nimbus-cloud-client-src/CHANGES.txt">here.</a>
</p>
<a href="2.3"> </a>
<h3>2.3</h3>
<i>Summary</i>
<ul>
<li>
<p>
Support for the EC2 Query API.
</p>
</li>
<li>
<p>
Introduction of administrative web portal interface. Supports
securely distributing user credentials.
</p>
</li>
<li>
<p>
Refactored workspace-control and integrated with
<a href="http://libvirt.org/">libvirt</a>. Includes initial
support for the
<a href="http://www.linux-kvm.org">KVM</a> hypervisor.
</p>
</li>
<li>
<p>
Assorted bug fixes and minor enhancements.
</p>
</li>
</ul>
<i>Services</i>
<ul>
<li>
<p>
Support for the EC2 Query API. Tested with the Python
<a href="http://code.google.com/p/boto/">boto</a> client but should
work with others. The service does not run in the standard Globus container,
it spawns a separate Jetty process. While installed by default, it requires
configuration before it can be used.
</p>
</li>
<li>
<p>
EC2 SOAP API support has been upgraded to version 2009-08-15. This means
ec2-api-tools clients must be upgraded to
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools-1.3-42584.zip">
this version</a>. Early work has been done to support multiple versions
concurrently, but this functionality is not yet available.
</p>
</li>
<li>
<p>
The new Nimbus web portal is based on Django and is a standalone
component with (in this version) no ties to the other Nimbus services. This
component's current sole functionality to facilitate securely
providing users with their X509 and query credentials. It will be expanded in future releases to include more functionality for both users and administrators.
</p>
</li>
<li>
<p>
The Context Broker has been refactored and merged into the main Nimbus
source tree. It is installed by default but is not enabled because it
needs configuration.
</p>
</li>
<li>
<p>
The Nimbus Derby configuration now supports network access, though it
is disabled by default. At install, passwords are generated and stored
in <i>var/derby.properties</i>. The Nimbus service still uses the
embedded interface. See
<a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6516">Bug 6516</a>
for details.
</p>
</li>
<li>
<p>Assorted bugfixes:</p>
<ul>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6631">
Bug 6631 - inconsistency with resource reservations
</a></li>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6633">
Bug 6633 - backwards compatibility with context broker is broken
</a></li>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6735">
Bug 6735 - blankspace regression
</a></li>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6856">
Bug 6856 - user data is broken
</a></li>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6869">
Bug 6869 - blankspace creation regression
</a></li>
</ul>
</li>
</ul>
<i>Control Agents</i>
<ul>
<li>
<p>
The workspace-control component has been significantly refactored.
It has been moved from <i>backend/</i> to <i>control/</i> in the
source tree.
Direct command-line invocation of Xen operations has been replaced with
calls to the excellent <a href="http://libvirt.org/">libvirt</a>
library. This opens the door to easier integration with several other
hypervisors, starting with KVM.
</p>
</li>
<li>
<p>
Initial <a href="http://kvm.qumranet.com/kvmwiki">KVM</a> support is
provided.
</p>
</li>
<li>
<p>Assorted bugfixes:</p>
<ul>
<li><a href="http://bugzilla.mcs.anl.gov/globus/show_bug.cgi?id=6868">
Bug 6868 - in non-cloud configurations, .gz support is broken
</a></li>
</ul>
</li>
</ul>
<i>Cloud client</i>
<ul>
<li>
<p>
The current cloud client as of this release is cloud-client-014. This service
release should also work with cloud clients 011 through 013.
</p>
<p>
For cloud client changes, see
<a href="http://github.com/nimbusproject/nimbus/raw/master/cloud-client/nimbus-cloud-client-src/CHANGES.txt">here.</a>
</li>
</ul>
<a name="TP2.2"> </a>
<h3>TP2.2</h3>
<i>Summary</i>
<ul>
<li>
<p>
Introduction of the metadata server which mimics the EC2 HTTP
query based metadata server.
</p>
</li>
<li>
<p>
Introduction of a standalone context broker, see the downloads
page. This runs by itself so that you can use just the context
broker to contextualize virtual clusters on EC2. No Nimbus cluster
is necessary.
</p>
</li>
<li>
<p>
Bug fixes, see below for specifics.
</p>
</li>
</ul>
<i>Services</i>
<ul>
<li>
<p>
Added a metadata server which responds to VMs HTTP queries, using
the same path names as the EC2 metadata server. The URL for this
is obtained by looking at <i>/var/nimbus-metadata-server-url</i>
on the VM, which is an optional VM customization that can be made.
See <i>"etc/nimbus/workspace-service/metadata.conf"</i> for the
details.
</p>
<p>
It responds based on source IP address so there is an
assumption that the immediately local network is non-spoofable.
</p>
<p>
The metadata server is disabled by default.
</p>
</li>
<li>
<p>
Introduction of a standalone context broker, see the downloads
page. This runs by itself so that you can use just the context
broker to contextualize virtual clusters on EC2. No Nimbus cluster
is necessary.
</p>
</li>
<li>
<p>
Added user-data support to EC2 remote interfaces.
</p>
</li>
<li>
<p>
Added user-data support to the WSRF operations, but namespaces did
not change. This maintains client forward compatibility. If the
user data element is missing, that is not an issue for the service.
</p>
</li>
<li>
<p>
Added getGlobalAll to the RM API, see enhancement request
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6556">6556</a>
</p>
</li>
<li>
<p>
Added <i>MetadataServer</i> module and user-data to <i>VM</i> to
the RM API.
</p>
</li>
<li>
<p>
Added user-data support to EC2 remote interfaces.
</p>
</li>
<li>
<p>
Fixed these EC2 interface bugs:
<a href="http://bugzilla.globus.org/globus/show_bug.cgi?id=6530">wrong
instance ID is returned</a> and
<a href="http://bugzilla.globus.org/globus/show_bug.cgi?id=6537">describe
instances fails with parameter</a>.
</p>
</li>
<li>
<p>
Fixed misc bugs
<a href="http://bugzilla.globus.org/globus/show_bug.cgi?id=6546">6546</a>
and <a href="http://bugzilla.globus.org/globus/show_bug.cgi?id=6545">6545
(pilot plugin initialization failure)</a>.
</p>
</li>
</ul>
<i>Cloud client</i>
<ul>
<li>
<p>
Current cloud client as of this release is cloud-client-011. This
supports contextualization using the new standalone context broker.
</p>
</li>
<li>
<p>
A lone invocation of "--status" (which prints all your currently
running instances) will now print the associated cloud handle
of each workspace.
</p>
</li>
<li>
<p>
Java 1.5 (Java 5) is now a requirement
</p>
</li>
<li>
<p>
The TP2.2 service side is backwards compatible with the "old
style" contextualization but this cloud client only supports
the new one. <i>You can only use this against Nimbus TP2.1
installations if you are not using contextualization</i>.
</p>
</li>
<li>
<p>
Support for contextualizing easily with EC2 resources. See the
output of "--extrahelp" for the new "--ec2script" option. Sample
EC2 cluster.xml file is @ "samples/ec2basecluster.xml"
</p>
<p>
This will take care of the context broker interactions for you and
give you a suggested set of EC2 commands to run (including files
for metadata) for the virtual cluster to contextualize while running
on EC2.
</p>
</li>
<li>
<p>
Fixed bug in the "lib/this-globus-environment.sh" script, the
X509_CERT_DIR variable was being set incorrectly
</p>
</li>
</ul>
<i>Context agent</i>
<ul>
<li>
<p>
A new version of the context agent is necessary to contextualize a
virtual cluster with Nimbus TP2.2's metadata server and the new
context broker.
</p>
</li>
</ul>
<a name="TP2.1"> </a>
<h3>TP2.1</h3>
<i>Summary</i>
<ul>
<li>
<p>
Introduction of an auto-configuration program which
guides you through many of the initial configuration steps and
run several validity tests.
</p>
</li>
<li>
<p>
Introduction of the Nimbus AutoContainer program which
allows you to set up a Globus Java web services environment from
scratch (including security) in less than a minute.
</p>
</li>
<li>
<p>
Introduction of the <i>cloud-admin</i> program which allows you
to very easily manage new users in a cloud configuration.
</p>
</li>
<li>
<p>
No protocol changes to WSRF based messaging. Previous
clients such as cloud-client-010 are compatible.
</p>
</li>
<li>
<p>
Protocol update to match the current Amazon EC2 deployment,
see below for details.
</p>
</li>
<li>
<p>
New workspace-control configurations options to support more
kinds of deployments, see below for details.
</p>
</li>
<li>
<p>
New service requirement: Java JDK5+ (<i>aka</i> Java 1.5+)
</p>
</li>
<li>
<p>
Updated documentation.
Added an <a href="plugins/index.html">extensibility guide</a>
and <a href="admin/upgrading.html">upgrade guide</a>.
</p>
</li>
<li>
<p>
Bug fixes, see below for specifics.
</p>
</li>
</ul>
<i>Services</i>
<ul>
<li>
<p>
Introduction of an auto-configuration program which will
guide you through many of the initial configuration steps and
run several validity tests.
</p>
<p>
See <a href="admin/quickstart.html#part-IIb">this section</a> of
the administrator quickstart for more information.
</p>
</li>
<li>
<p>
Introduction of the Nimbus AutoContainer program which will
allow you to set up a Globus Java web services environment from
scratch (including security) in less than a minute.
</p>
<p>
It requires a separate download.
See <a href="admin/quickstart.html#auto-container">this section</a>
of the administrator quickstart for more information.
</p>
</li>
<li>
<p>
Introduction of the "cloud-admin" program which will allow you
to very easily manage new users in a
<a href="doc/cloud.html">cloud</a> configuration.
</p>
<p>
It is installed at the same time as the auto-configuration program,
installed as
<i>$GLOBUS_LOCATION/share/nimbus-autoconfig/cloud-admin.sh</i>,
see <a href="doc/cloud.html#cloud-admin">this section</a>
of the cloud guide for more information
</p>
</li>
<li>
<p>
Protocol update to match the current Amazon EC2 deployment:
</p>
<p>
Nimbus TP2.1 supports the <i>2008-05-05</i> WSDL
(used by
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools-1.3-24159.zip">this
EC2 client</a>) as opposed to Nimbus TP2.0 which supported
the <i>2008-02-01</i> WSDL
(used by
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools-1.3-19403.zip">this
EC2 client</a>).
</p>
</li>
<li>
<p>
New service requirement: Java JDK5+ (<i>aka</i> Java 1.5+)
</p>
</li>
<li>
<p>
Resolved bug
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6390">6390</a>:
"notifications script is not sh compliant"
</p>
<p>
The notification scripts now directly use the intended "bash" shell.
</p>
</li>
<li>
<p>
Resolved bug
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6474">6474</a>:
"destruction callbacks were not registered"
</p>
<p>
An internal problem was fixed which made the logs wrong as well
as causing problems for the client at destroy time. In particular,
a VM would be destroyed but the remote client would not hear the
last notification of the event causing it to hang.
</p>
</li>
<li>
<p>
Resolved bug
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6397">6397</a>:
"reservation ID mapping verification wrong for single-VM reservations"
</p>
<p>
The EC2 reservation emulation is now working correctly with single
VMs.
</p>
</li>
<li>
<p>
Resolved bug
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6475">6475</a>:
"repository + scp propagation"
</p>
<p>
The EC2 messaging system now works with setups that use
SCP propagation, there is a new relevant
configuration in the <i>elastic.conf</i> file.
</p>
</li>
<li>
<p>
Resolved miscellaneous/cosmetic bugs
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6393">6393</a>,
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6394">6394</a>,
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6396">6396</a>,
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6398">6398</a>, and
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6416">6416</a>.
</p>
</li>
</ul>
<i>Reference clients</i>
<ul>
<li>
<p>
Cloud and reference clients did not change. Current cloud client
as of this release is cloud-client-010.
</p>
</li>
<li>
<p>
You will need to update any EC2 client you use with Nimbus:
</p>
<p>
Nimbus TP2.1 supports the <i>2008-05-05</i> WSDL
(used by
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools-1.3-24159.zip">this
EC2 client</a>) as opposed to Nimbus TP2.0 which supported
the <i>2008-02-01</i> WSDL
(used by
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools-1.3-19403.zip">this
EC2 client</a>).
</p>
</li>
</ul>
<i>Control agents</i>
<ul>
<li>
<p>
Added a new option to create VMs with "tap:aio" instead of using
the "file" method (these are Xen terms for methods of mounting
the disks). The "tap:aio" method is often
used in Xen 3.2 setups and is now possible to use via
workspace-control. See the new <i>worksp.conf.sample</i>.
</p>
</li>
<li>
<p>
Resolved enhancement request
<a href="http://bugzilla.globus.org/bugzilla/show_bug.cgi?id=6326">6326</a>:
"use matching initrd with kernel"
</p>
<p>
This allows you to configure workspace-control to take the kernel
filename it is launching a VM with and search for a matching
initrd based on suffix rules you set up. This allows you to
easily use many of the Xen guest kernels that are created with popular
Linux distributions.
</p>
</li>
</ul>
<a name="TP2.0"> </a>
<h3>TP2.0</h3>
<i>Summary</i>
<ul>
<li>
<p>
Introduction of the FAQ which explains
many things you may already know, but it also includes new descriptions
of the component system now more clearly articulated in the Nimbus
TP2.0 release.
</p>
</li>
<li>
<p>
Introduction of the Java
RM API which is
a bridge between protocols and resource management implementations.
The resource managers can remain protocol/framework/security agnostic
(they can be "pure Java") and various protocol implementations
can be implemented independently (and even simultaneously). Runtime
orchestration of implementation choices is directed by industry
standard
<a href="http://www.springframework.org/">Spring</a> dependency
injection.
</p>
</li>
<li>
<p>
Introduction of an alternative remote protocol implementation based
on Amazon <a href="http://aws.amazon.com/ec2">EC2</a>'s WSDL interface
description. It is only a partial implementation (see below).
It can be used simultaneously alongside the WSRF based protocols.
</p>
</li>
<li>
<p>
More friendly configuration mechanism for administrators including
area-specific ".conf" files instead of any XML and the addition of
some helper scripts.
</p>
</li>
<li>
<p>
No protocol changes (only an additional remote protocol). Previous
clients such as cloud-client-009 are compatible.
</p>
</li>
</ul>
<i>Services</i>
<ul>
<li>
<p>
Introduction of the Java
RM API which is
a bridge between protocols and resource management work. The
resource managers below can remain protocol/framework agnostic
(they can be "pure Java") and various protocol implementations
can be implemented independently. Runtime directions of choices
is directed by <a href="http://www.springframework.org/">Spring</a>
dependency injection.
</p>
</li>
<li>
<p>
Introduction of an alternative remote protocol implementation based
on Amazon EC2's WSDL interface description
(namespace <i>http://ec2.amazonaws.com/doc/2008-02-01/</i>)
</p>
<p>
It can be used simultaneously alongside the previous remote
interfaces. If the EC2 protocol layer does not recognize instance
identifiers being reported by the underling resource manager
(for example when gathering "describe-instances" results), it
will create new, unique instance and reservation IDs on the fly for
them.
</p>
<p>
It is only a partial protocol implementation, the operations behind
these EC2 commandline clients are currently provided:
</p>
<ul>
<li>
<p>
ec2-describe-images - See what images in your personal cloud
directory you can run.
</p>
</li>
<li>
<p>
ec2-run-instances - Run images that are in your personal cloud
directory.
</p>
</li>
<li>
<p>
ec2-describe-instances - Report on currently running instances.
</p>
</li>
<li>
<p>
ec2-terminate-instances - Destroy currently running instances.
</p>
</li>
<li>
<p>
ec2-reboot-instances - Reboot currently running instances.
</p>
</li>
<li>
<p>
ec2-add-keypair [*] - Add personal SSH public key that can be
installed for root SSH logins
</p>
</li>
<li>
<p>
ec2-delete-keypair - Delete keypair mapping.
</p>
</li>
</ul>
<p>
[*] - One of two add-keypair implementations can be chosen by
the administrator.
</p>
<ul>
<li>
<p>
One is the normal implementation where the
server-side generates a private and public key (using
<a href="http://www.jcraft.com/jsch/">jsch</a>) and delivers
the private key to you.
</p>
</li>
<li>
<p>
The other (configured by default) is a break from the
regular semantics. It allows the keypair "name" you
send in the request to be the name AND the public key value.
This means there is never a private key server-side and
also that you can keys you aready have on your system.
</p>
</li>
</ul>
</li>
<li>
<p>
More friendly configuration mechanism for administrators including
area-specific ".conf" files (instead of XML) and the addition of
some helper scripts.
</p>
<p>
If you are familiar with a previous Nimbus versions (VWS), these
".conf" files hold anything found in the old "jndi-config.xml" file
which you don't need to look at anymore.
The files hold name=value pairs with surrounding comments. They
are organized by area: accounting.conf, global-policies.conf,
logging.conf, pilot.conf, network.conf, ssh.conf, vmm.conf.
</p>
</li>
<li>
<p>
Service configurations are now in "etc/nimbus/workspace-service" and
"etc/nimbus/elastic". Advanced configurations (which you should
not need to alter normally are now in
"etc/nimbus/workspace-service/other" and "etc/nimbus/elastic/other".
</p>
</li>
<li>
<p>
New persistence management wrapper scripts are in "share/nimbus"
and the persistence directory has moved to "var/nimbus"
</p>
</li>
<li>
<p>
Support for site-to-site file management (staging) was removed.
</p>
</li>
<li>
<p>
Developers: Significant directory reworkings (and subsequent build
file changes) to organize modules more coherently, allowing for
easier module independence.
</p>
<p>
Build system now clearly separates anything to do with the target
deployment (only one target deployment at the moment, GT4.0.x).
</p>
</li>
<li>
<p>
New Java dependencies:
</p>
<ul>
<li>
<a href="http://www.springframework.org/">Spring</a> - just the
core dependency injection library. The
RM API
depends on Spring import statements but no other module has any
direct coupling to it.
</li>
<li>
<a href="http://cglib.sourceforge.net/">cglib</a> - used
"invisibly" alongside Spring to provide some limited code
generation when convenient.
</li>
<li>
<a href="http://ehcache.sourceforge.net/">ehcache</a> - used
for in-memory object caching.
</li>
<li>
<a href="http://jug.safehaus.org/">jug</a> - used for UUID
generation instead of needing an axis dependency.
</li>
<li>
<a href="http://www.jcraft.com/jsch/">jsch</a> - used for
SSH keypair generation if necessary (see [*] in the EC2
section).
</li>
</ul>
</li>
</ul>
<i>Reference clients</i>
<ul>
<li>
<p>
The clients have stayed the same (on purpose, to reduce too much
changing) except for some library package name changes.
</p>
</li>
<li>
<p>
When using a cloud running the EC2 front end implementation, you
can download this
<a href="http://s3.amazonaws.com/ec2-downloads/ec2-api-tools.zip">EC2
client</a> from Amazon or try a number of different client that are
<a href="http://www.google.com/search?hl=en&q=ec2%20client">out there</a>.
</p>
</li>
</ul>
<i>Control agents</i>
<ul>
<li>
<p>
Workspace-control has stayed the same (on purpose, to reduce too
much changing).
</p>
</li>
</ul>
<i>Workspace pilot system</i>
<ul>
<li>
<p>
No changes except that the server side configuration location
has moved from the "jndi-config.xml" file to "pilot.conf"
</p>
</li>
</ul>
<a name="TP1.3.3"> </a>
<h3>TP1.3.3.1</h3>
<i>Summary</i>
<ul>
<li>
<p>
Introduction of support for contextualization with virtual
clusters. See the <a href="clouds/">clouds page</a> and the new
<a href="clouds/clusters.html">one-click clusters</a> page to see
the various new features in action.
</p>
</li>
<li>
<p>
New ensemble service report operation allows efficient queries
about a large number of workspaces.
</p>
</li>
<li>
<p>
Support for storing images at the repository in gzip format and
retrieving them from the repository in gzip format. This can
save a lot of time in cluster situations.
</p>
</li>
<li>
<p>
Support for pegging the number of vcpus clients receive.
</p>
</li>
<li>
<p>
Various client enhancements including internal organization,
cleaner output, and new commandline options. Embedded security
tools (like grid-proxy-init) work more out of the box now.
</p>
</li>
<li>
<p>
No configuration migrations are necessary for moving to this
version from TP1.3.2. Some configuration additions will be
necessary if you'd like to take advantage of features.
</p>
</li>
<li>
<p>
There was a WSDL update: additions, changes and new namespaces.
The base namespace for workspace schemas is now
<i>http://www.globus.org/2008/06/workspace/</i>
</p>
</li>
<li>
<p>
Some bug fixes.
</p>
</li>
</ul>
<i>Services</i>
<ul>
<li>
<p>
Integration with context broker.
</p>
</li>
<li>
<p>
New ensemble service report operation allows efficient queries
about a large number of workspaces. Can retrieve status and
error messages about entire ensemble at once.
</p>
</li>
<li>
<p>
Fixed scheduler backout to correctly handle situation where
ensemble wasn't launched yet but ensemble-destroy was invoked.
</p>
</li>
<li>
<p>
Fixed bug where IP address updates were not passing through cache
layer to DB correctly causing a possible inconsistency if container
restarted in certain circumstances. <b>NOTE:</b> <i>this bugfix
was not present in TP1.3.3 but is present in TP1.3.3.1</i>.
</p>
</li>
<li>
<p>
Various internal changes (see CVS log)
</p>
</li>
<li>
<p>
No configuration changes are necessary for moving to this version
from TP1.3.2. But to enable the context broker, you need to
configure paths to a credential for it in the jndi-config file
and make sure the WSDD file lists the context broker as in the
source file "deploy-server.wsdd" (which becomes server-config.wsdd)
</p>
</li>
</ul>
<i>Reference clients</i>
<ul>
<li>
<p>
Added cloud-client cluster and contextualization support. Includes
new "--cluster" flag (see cloud-client CHANGES.txt for full changes
there).
</p>
<p>
See the <a href="clouds/">clouds page</a> and the new
<a href="clouds/clusters.html">clusters</a> page.
</p>
</li>
<li>
<p>
The regular commandline client has new flags for ensemble
and context broker support. See "-h" output.
</p>
</li>
</ul>
<i>Workspace-control</i>
<ul>
<li>
<p>
Support for gzip via filename-sense. See cloud
<a href="clouds/clusters.html#compression">notes</a> on image
compression/decompression. This can save a lot of time in cluster
launch situations since the gzip/gunzip happens on the VMMs
simultaneously, cutting transfer times (where there is contention)
considerably.
</p>
</li>
<li>
<p>
Local-locked the control of dhcpd start and stop: now works for
situations where multiple workspaces are deployed on a VMM
simultaneously (such as one VM per core and launching as part of
a cluster). The DHCP adjustment was being excercised
simultaneously, revealing the race.
</p>
</li>
<li>
<p>
There is no need to change the workspace-control configuration
file from a TP1.3.2 compatible one. There is a new configuration
if you want to use it, though. The "[behavior] --> num_cpu_per_vm"
configuration allows you to peg the number of vcpus that are
assigned to every workspace.
</p>
<p>
You can choose to not upgrade workspace-control at all if you don't
want the features listed here.
</p>
</li>
</ul>
<i>Workspace pilot program</i>
<ul>
<li>
<p>
No changes.
</p>
</li>
</ul>
<br />
<a name="TP1.3.2"> </a>
<h3>TP1.3.2</h3>