Skip to content

Commit

Permalink
pre-commit fixes
Browse files Browse the repository at this point in the history
  • Loading branch information
askulkarni2 committed May 12, 2024
1 parent 8eebcaa commit 7a0e65f
Show file tree
Hide file tree
Showing 14 changed files with 21 additions and 24 deletions.
1 change: 1 addition & 0 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,3 +37,4 @@ repos:
rev: v2.2.6
hooks:
- id: codespell
args: ["--skip=*.excalidraw"]
4 changes: 0 additions & 4 deletions ai-ml/trainium-inferentia/main.tf
Original file line number Diff line number Diff line change
Expand Up @@ -47,10 +47,6 @@ provider "kubectl" {
}
}

data "aws_eks_cluster_auth" "this" {
name = module.eks.cluster_name
}

data "aws_ecrpublic_authorization_token" "token" {
provider = aws.ecr
}
Expand Down
2 changes: 1 addition & 1 deletion analytics/terraform/superset-on-eks/addons.tf
Original file line number Diff line number Diff line change
Expand Up @@ -109,7 +109,7 @@ module "eks_data_addons" {
superset_helm_config = {
values = [templatefile("${path.module}/helm-values/superset-values.yaml", {})]
}
depends_on = [module.eks_blueprints_addons]
depends_on = [module.eks_blueprints_addons]

}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -50,10 +50,10 @@ supersetWorker:
requests:
cpu: 200m
memory: 400Mi

persistence:
enabled: true


postgresql:
## Set to false if bringing your own PostgreSQL.
Expand Down Expand Up @@ -92,4 +92,4 @@ redis:
## Access mode:
accessModes:
- ReadWriteOnce
runAsUser: 1000
runAsUser: 1000
2 changes: 1 addition & 1 deletion gen-ai/inference/gradio-ui/gradio-app-mistral.py
Original file line number Diff line number Diff line change
Expand Up @@ -62,4 +62,4 @@ def filter_harmful_content(text):
)

# Launch the ChatInterface
chat_interface.launch()
chat_interface.launch(server_name="0.0.0.0")
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ def __init__(self):
n_positions=4096
)
self.neuron_model.to_neuron()

# Initialize tokenizer for the model
self.tokenizer = AutoTokenizer.from_pretrained(model_id)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -169,4 +169,4 @@ spec:
service:
name: mistral
port:
number: 8000
number: 8000
Original file line number Diff line number Diff line change
Expand Up @@ -161,4 +161,4 @@ spec:
service:
name: stablediffusion
port:
number: 8000
number: 8000
4 changes: 2 additions & 2 deletions website/docs/blueprints/distributed-databases/clickhouse.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
sidebar_position: 4
sidebar_label: ClickHouse
---
# ClickHouse on EKS
# ClickHouse on EKS
[ClickHouse](https://clickhouse.com/) is a high-performance, column-oriented SQL database management system (DBMS) for online analytical processing (OLAP) that is open sourced under the Apache 2.0 license.


Expand All @@ -17,4 +17,4 @@ OLAP is software technology you can use to analyze business data from different
* SQL Support: ClickHouse supports a subset of SQL, making it familiar and easy to use for developers and analysts who are already familiar with SQL-based databases.
* Integrated Data Formats: ClickHouse supports various data formats, including CSV, JSON, Apache Avro, and Apache Parquet, making it flexible for ingesting and querying different types of data.

**To deploy Clickhouse on EKS**, we recommend this [Clickhouse on EKS blueprint](https://github.com/Altinity/terraform-aws-eks-clickhouse) from Altinity, an AWS Partner who maintains the [Clickouse Kubernetes Operator](https://github.com/Altinity/clickhouse-operator). If you have any issues with the blueprint or operator, please create an issue on the corresponding Altinity GitHub repository.
**To deploy Clickhouse on EKS**, we recommend this [Clickhouse on EKS blueprint](https://github.com/Altinity/terraform-aws-eks-clickhouse) from Altinity, an AWS Partner who maintains the [Clickouse Kubernetes Operator](https://github.com/Altinity/clickhouse-operator). If you have any issues with the blueprint or operator, please create an issue on the corresponding Altinity GitHub repository.
6 changes: 3 additions & 3 deletions website/docs/gen-ai/excalidraw/llama3.svg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
4 changes: 2 additions & 2 deletions website/docs/gen-ai/inference/Llama2-inf2.md
Original file line number Diff line number Diff line change
Expand Up @@ -235,7 +235,7 @@ Discover how to create a user-friendly chat interface using [Gradio](https://www

Let's move forward with setting up the Gradio app as a Docker container running on localhost. This setup will enable interaction with the Stable Diffusion XL model, which is deployed using RayServe.

### Build the Gradio app docker container
### Build the Gradio app docker container

First, lets build the docker container for the client app.

Expand Down Expand Up @@ -297,4 +297,4 @@ This script will cleanup the environment using `-target` option to ensure all th
```bash
cd ../../../ai-ml/trainium-inferentia/
./cleanup.sh
```
```
2 changes: 1 addition & 1 deletion website/docs/gen-ai/inference/StableDiffusion-inf2.md
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ Discover how to create a user-friendly chat interface using [Gradio](https://www

Let's move forward with setting up the Gradio app as a Docker container running on localhost. This setup will enable interaction with the Stable Diffusion XL model, which is deployed using RayServe.

### Build the Gradio app docker container
### Build the Gradio app docker container

First, lets build the docker container for the client app.

Expand Down
6 changes: 3 additions & 3 deletions website/docs/gen-ai/inference/llama3-inf2.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ In this tutorial, you will not only learn how to harness the power of Llama-3, b

### What is Llama-3-8B Instruct?

Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Further, in developing these models, we took great care to optimize helpfulness and safety.
Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Further, in developing these models, we took great care to optimize helpfulness and safety.

More information on Llama3 sizes and model architecture can be found [here](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct).

Expand All @@ -36,7 +36,7 @@ One of the key challenges in deploying large language models (`LLMs`) like Llama

In contrast, `Trn1/Inf2` instances, such as `trn1.32xlarge`, `trn1n.32xlarge`, `inf2.24xlarge` and `inf2.48xlarge`, are purpose built for high-performance deep learning (DL) training and inference of generative AI models, including LLMs. They offer both scalability and availability, ensuring that you can deploy and scale your `Llama-3` models as needed, without resource bottlenecks or delays.

**Cost Optimization**
**Cost Optimization**

Running LLMs on traditional GPU instances can be cost-prohibitive, especially given the scarcity of GPUs and their competitive pricing. **Trn1/Inf2** instances provide a cost-effective alternative. By offering dedicated hardware optimized for AI and machine learning tasks, Trn1/Inf2 instances allow you to achieve top-notch performance at a fraction of the cost. This cost optimization enables you to allocate your budget efficiently, making LLM deployment accessible and sustainable.

Expand Down Expand Up @@ -234,7 +234,7 @@ Discover how to create a user-friendly chat interface using [Gradio](https://www

Let's move forward with setting up the Gradio app as a Docker container running on localhost. This setup will enable interaction with the Stable Diffusion XL model, which is deployed using RayServe.

### Build the Gradio app docker container
### Build the Gradio app docker container

First, lets build the docker container for the client app.

Expand Down
2 changes: 1 addition & 1 deletion website/docs/gen-ai/inference/stablediffusion-gpus.md
Original file line number Diff line number Diff line change
Expand Up @@ -152,7 +152,7 @@ Discover how to create a user-friendly chat interface using [Gradio](https://www

Let's move forward with setting up the Gradio app as a Docker container running on localhost. This setup will enable interaction with the Stable Diffusion XL model, which is deployed using RayServe.

### Build the Gradio app docker container
### Build the Gradio app docker container

First, lets build the docker container for the client app.

Expand Down

0 comments on commit 7a0e65f

Please sign in to comment.