class Value:
def __init__(self,data):
self.data=data
def __repr__(self):
return f"a={self.data}"
**def __add__(self,other):
out=Value(self.data-other.data)**
return out
def __mul__(self,other):
out=Value(self.data*other.data)
return out
a=Value(10.0)
b=Value(3.0)
c=a.__add__(b)
d=a.__mul__(b)
**a+b**
Certainly! Here's a concise summary of why _s
(or variables prefixed with a single underscore) is commonly used, without the actual code implementation:
- Signaling Internal Use:
_s
signals that the variable is intended for internal use within a module or class, distinguishing it from the public interface. - Encapsulation: Encourages encapsulation by suggesting that
_s
should be accessed or modified through defined methods or properties, rather than directly. - Preventing Name Clashes: Reduces the likelihood of naming conflicts with external libraries or modules.
- Readability and Documentation: Enhances code readability and documentation by clearly indicating the intended usage of variables.
- Consistency: Aligns with industry best practices and conventions, promoting consistency across Python codebases.
CNN Done For unsupervised learning Auto encoders are used in Convolution neural networks
restnet
cross-GPU parallelization
Concept | Complete |
---|---|
weights-decay | |
relu | |
residuals | |
dropout | ✅ |
batch-norm | |
layer-norm | |
gelu | |
adam | |
early-stopping |
include stopping the training as soon as performance on a validation set starts to get worse, introducing weight penalties of various kinds such as L1 and L2 regularization and soft weight sharing (Nowlan and Hinton, 1992).
A dropout network typically takes 2-3 times longer to train than a standard neural network of the same ar- chitecture.
Tokenization is the process of translating string of texts into sequences of tokens and vice-versa.
Ollama 2 We trained on 2 trillion tokens of data
gpt 4(cl100k base) has less tokens compared to gpt2
ord('A') encodings are the way by which we can take the unicode text and store as binary data.
tiktoken
Why the LLM break if I ask it about "SolidGoldMagikarp"? Tokenization.
regex and re
It is taking the space and the character as one token
gcp vm ssh lengthy process compared to aws
Why can't LLM spell words? Tokenization. Why can't LLM do super simple string processing tasks like reversing a string? Tokenization. Why is LLM worse at non-English languages (e.g. Japanese)? Tokenization. Why is LLM bad at simple arithmetic? Tokenization. Why did GPT-2 have more than necessary trouble coding in Python? Tokenization. Why did my LLM abruptly halt when it sees the string "<|endoftext|>"? Tokenization. What is this weird warning I get about a "trailing whitespace"? Tokenization. Why the LLM break if I ask it about "SolidGoldMagikarp"? Tokenization. Why should I prefer to use YAML over JSON with LLMs? Tokenization. Why is LLM not actually end-to-end language modeling? Tokenization. What is the real root of suffering? Tokenization.
Installing LLMs on the command line APPLICATIONS
Install this tool using pip:
pip install llm
Or using pipx:
pipx install llm
Or using Homebrew (see warning note):
brew install llm
llm models default
gpt-4o-mini
llm install llm-gemini
llm keys set gemini Enter key: https://aistudio.google.com/app/apikey
llm models default gemini
llm -m palm "hi"
(base) @NShravanReddy ➜ /workspaces/DeepLearning (main) $ llm -m palm "hi how are you?"
I am doing well, thank you for asking! How are you today?
llm logs -c to view the logs
llm logs path
(base) @NShravanReddy ➜ /workspaces/DeepLearning (main) $ llm logs path /home/codespace/.config/io.datasette.llm/logs.db
pip install datasette
llm install llm-cmd It is not like cmd
llm install llama-cpp-python
llm llama-cpp download-model
https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/resolve/main/llama-2-7b-chat.ggmlv3.q8_0.bin
--alias llama2-chat --alias l2c --llama2-chat
llm install llm-ollama
Installing Ollama
curl -fsSL https://ollama.com/install.sh | sh
Terminal 1 : ollama serve
Terminal 2L: ollama run phi3
cat README.md | llm -s 'create code snippert for readme' >test1.md
docker pull ollama/ollama
docker run -it
--rm
-v ollama:/root/.ollama
-p 11434:11434
--name ollama
ollama/ollama
docker exec -it ollama bash
apt-get install curl https://github.com/NShravanReddy/DeepLearning?tab=readme-ov-file#installation
mkdir -p ~/miniconda3
wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh -O ~/miniconda3/miniconda.sh bash ~/miniconda3/miniconda.sh -b -u -p ~/miniconda3 rm -rf ~/miniconda3/miniconda.sh
~/miniconda3/bin/conda init bash ~/miniconda3/bin/conda init zsh
curl -fsSL https://ollama.com/install.sh | sh
ollama start ollama pull llama3.1 ollama run llama3.1
ollama pull nomic-embed-text
git clone https://github.com/ItzCrazyKns/Perplexica.git
cd Perplexica
mv sample.config.toml config.toml (rename)
docker compose up -d
now a window pop up to open port 4000 docker stop $(docker ps -q)
pip install inspect-ai pip install google-generativeai
export GOOGLE_API_KEY= getkeyfrom
nano theory_of_mind.py
from inspect_ai import Task, eval, task
from inspect_ai.dataset import example_dataset
from inspect_ai.scorer import model_graded_fact
from inspect_ai.solver import (
chain_of_thought, generate, self_critique
)
@task def theory_of_mind(): return Task( dataset=example_dataset("theory_of_mind"), plan=[ chain_of_thought(), generate(), self_critique() ], scorer=model_graded_fact() )
inspect eval theory_of_mind.py --model google/gemini-1.0-pro
Damm success. but it took more than 25 minutes
Experiment tracking is the process of keeping track of all the relevant information from an ML experiment which includes:
source code Environment Data Model Hyperparameters Metrics ...
why experiment tracking important?
1.Reproducibility 2.Organizatio 3.Optimization
Like tracking in spreadsheets has
- Error prone 2.No standard format 3.Visiblity & Collaboration
Platform for the machine learning lifecycle Four modules 1.Tracking 2.Models 3.Model Registry 4.Projects
1.Parameters 2.Metrics 3.Metadata 4.Artifacts 5.Models It also logs extra info 1.Source code 2.Version of the code 3.Start and endtime 4.Author
pip install mlflow git remote -v
Techniques 1.Few shot prompting
Examples will be given like english to telugu translation then the actual task will be given at end of prompt
- Chain-of-thought reasoning
Fine-tuning when you’ve already deployed LLMs in prod
- Unless you can't hit your quality target 2.Latency target 3.Cost target
1.Choosing a Base Model