Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 1 addition & 3 deletions cookbooks/aws-parallelcluster-awsbatch/recipes/install.rb
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,7 @@
# limitations under the License.

return if platform?('redhat')
return if aws_region.start_with?("us-iso")

include_recipe "::awsbatch_virtualenv"

Expand All @@ -28,9 +29,6 @@

# Check whether install a custom aws-parallelcluster-awsbatch-cli package or the standard one
# Install awsbatch cli into awsbatch virtual env
if aws_region.start_with?("us-iso") && !node['cluster']['custom_awsbatchcli_package'].empty?
node.default['cluster']['custom_awsbatchcli_package'] = "#{node['cluster']['artifacts_s3_url']}/dependencies/awsbatch/aws-parallelcluster.tgz"
end
Comment on lines -31 to -33
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is only removing the code enough? In the below code, it will try to install from pip and will fail, right

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

From what I know we dont even support Batch scheduler in ADC

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

it skips the whole install part for aws batch in ADC due to this line: return if aws_region.start_with?("us-iso")

if !node['cluster']['custom_awsbatchcli_package'].nil? && !node['cluster']['custom_awsbatchcli_package'].empty?
# Install custom aws-parallelcluster package
bash "install aws-parallelcluster-awsbatch-cli" do
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,9 @@
end

if aws_region.start_with?("us-iso") && !is_custom_node?
node.default['cluster']['custom_node_package'] = "#{node['cluster']['artifacts_s3_url']}/dependencies/node/aws-parallelcluster-node.tgz"
node_package = "aws-parallelcluster-node-#{node['cluster']['parallelcluster-node-version']}.tgz"

node.default['cluster']['custom_node_package'] = "#{node['cluster']['s3_url']}/parallelcluster/#{node['cluster']['parallelcluster-node-version']}/node/#{node_package}"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Are we making changes in the S3 location to upload default Node package in this location. As I dont see node package in s3 for other regions?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

when we release in adc, we will also have to release a node and cookbook package so that they can use those for build image. I think we have a similar path where we have been publish the cookbook in s3 in commercial.

end

if is_custom_node?
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,4 +5,5 @@

# URL for ParallelCluster Artifacts stored in public S3 buckets
# ['cluster']['region'] will need to be defined by image_dna.json during AMI build.
default['cluster']['artifacts_s3_url'] = "https://#{node['cluster']['region']}-aws-parallelcluster.s3.#{node['cluster']['region']}.#{node['cluster']['aws_domain']}/archives"
default['cluster']['s3_url'] = "https://#{node['cluster']['region']}-aws-parallelcluster.s3.#{node['cluster']['region']}.#{node['cluster']['aws_domain']}"
default['cluster']['artifacts_s3_url'] = "#{node['cluster']['s3_url']}/archives"