Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Circcash should promote AI safety instead of just AI progress. #13

Open
jvanname opened this issue Apr 17, 2023 · 0 comments
Open

Circcash should promote AI safety instead of just AI progress. #13

jvanname opened this issue Apr 17, 2023 · 0 comments

Comments

@jvanname
Copy link
Owner

jvanname commented Apr 17, 2023

Circcash's mining algorithm is designed to accelerate the development of reversible computing hardware. The capabilities of AI systems with reversible computing hardware will vastly outperform any possible AI systems without such reversible computing hardware. AI technologies are inherently risky especially if these technologies include energy efficient reversible hardware. We believe that the benefits of reversible computing technologies outweigh the risks, but Circcash is committed to mitigating these risks. Circcash is in a good position to mitigate the risks posed by AI systems since AI systems that have been developed to investigate Circcash's mining algorithm seem to be inherently safer than other AI systems for several reasons. Circcash is committed to developing AI systems that are more safe, understandable, interpretable, predictable, and mathematical, but are just as powerful as other AI systems. Circcash is also committed to open communication about its AI research and making as much of its research open source as possible. Circcash is also non-alarmist at the moment in the sense that we do not believe that AI progress should be halted as we do not see any clear threat from AI. We simply want to perform our basic responsibility at addressing any risks that may arise from reversible computing technologies.

Comparing efficiency of current computation to reversible computation

Our current computing hardware is very inefficient compared to the future reversible computing hardware. The paper Mechanical Computing Systems Using Only Links and Rotary Joints concludes that it may be feasible (but probably exceedingly difficult) to make necessarily reversible computing hardware that computes "10^21 FLOPS in a sugar cube using 1 watt of power with a 100 MHz clock (10 ns)." To put this into comparison, an NVIDA GeForce RTX 4090 uses 450 W of power while it produces 10^14 FLOPS of computation. Training the 175 billion parameter version of GPT 3 used 3*10^23 FLOPS which would translate to 100 Joules of energy using future reversible computing hardware. I am not an expert in nano-technology nor am I an engineer, so I cannot comment too much on the feasibility of using mechanical molecular computing systems for computation. But there are plenty of other ideas for reversible computing hardware that are more efficient than the hardware we have today.

Some principles of AI safety

  1. Low entropy random variables: The entropy of a discrete random variable $\mathcal{X}$ on the set $A$ is the value
    $H(\mathcal{X})=-\sum_{a\in A}P(\mathcal{X}=a)\cdot\log(P(\mathcal{X}=a)).$ Suppose that $D$ is our data set. Then let $\mathcal{X}_D$ denote the machine learning model trained with the data set $D$. Then a lower value of $H(\mathcal{X}_D)$ should be preferred over a higher value of $H(\mathcal{X})$. The reason for this is that if $H(\mathcal{X}_D)$ is higher, then the trained model $\mathcal{X}_D$ will contain much random information that is not part of the training data $D$. This random information will make the machine learning model less understandable and interpretable.

  2. Smoothness: The function $D\mapsto\mathcal{X}_D$ should be smooth.

  3. Strong convexity around optima: If $A$ is a non-zero positive semidefinite matrix, then let $A(r)=r=\lambda_n/\lambda_1$ where $\lambda_1\geq\dots\geq\lambda_n$ are the eigenvalues of $A$. Suppose now that $F_D:X\rightarrow\mathbb{R}$ is the loss function for the training set $D$ and that $X$ is given a Riemannian metric so that there is a notion of a Hessian for the function $F_D$. Let $x_0$ be the value such that $F_D(x_0)$ is locally maximized. Then the ratio $A(H(F_D)(x_0))$ of the smallest to the largest eigenvalues of the Hessian $H(F_D)(x_0)$ should be large. A large value of $H(F_D)(x_0)$ means that the local maximum $x_0$ is more robust to changes in the training data $D$ and other changes in the loss function $F_D$. It is preferrable if most of the eigenvalues of $H(F_D)(x_0)$ are close together.

  4. Large basin of attraction. Suppose that $U$ is an open subset of Euclidean space. Let $F_D$ be a loss function. Suppose that $F_D(x_0)$ is our loss function. Let $r$ be the maximum value such that if $y\in B_r(x_0)$, then
    $\langle\nabla F_d(y),y-x_0\rangle\leq 0$. Then the value $r$ should be large in order to ensure that the gradient descent process converges to the same local minimum $F_D(x_0)$ around $x_0$.

Spectral and spectral-like algorithms

Suppose that $A_1,\dots,A_r\in M_n(\mathbb{C})$. Define a partial function $F_{A_1,\dots,A_r;d}:M_d(\mathbb{C})\rightarrow[0,\infty)$ by setting $F_{A_1,\dots,A_r;d}(X_1,\dots,X_r)=\frac{\rho(A_1\otimes\overline{X_1}+\dots+A_r\otimes\overline{X_r})}{\rho(X_1\otimes\overline{X_1}+\dots+X_r\otimes\overline{X_r})^{1/2}}$. Then we say that $(X_1,\dots,X_r)\in M_d(\mathbb{C})$ is an $L_{2,d}$-spectral radius dimensionality reduction LSRDR of $A_1,\dots,A_r$ if $(X_1,\dots,X_r)\mapsto F_{A_1,\dots,A_r;d}(X_1,\dots,X_r)$ is locally maximized.

If $(X_1,\dots,X_r)$ is an LSRDR trained from $(A_1,\dots,A_r)$, then $(X_1,\dots,X_r)$ tends to have low entropy or even zero entropy. Algorithms that are similar to LSRDRs tend to have low entropy as well. LSRDRs behave mathematically since they are related to quantum channels in quantum information theory. Since LSRDRs have low entropy and behave mathematically, one should consider LSRDRs safer than machine learning models with higher entropy.

LSRDRs and similar constructions seem to be capable of performing machine learning tasks including that of constructing matrix-valued word embeddings, graph embeddings, evaluating the security of block ciphers, but LSRDRs do not seem to be capable of replacing neural networks. More research is needed in order to explore the full potential of LSRDRs.

Matrix-valued natural language processing

In natural language processing, tokens are typically represented as vectors. This may be problematic since vectors are useful for representing a single meaning of a token but vectors are not very well suited for handling tokens which may have multiple meanings. It is better for the tokens to be represented as a sort of superposition of vectors to represent the variety of meanings that a token could have. In other words, it is better for tokens to be represented by matrices than simply as vectors.

Suppose that $A$ is our set of tokens and $f:A\rightarrow M_d(K)$ is a matrix-valued word embedding. In order to save space, the matrix $f(a)$ may be of low rank or have many columns or rows consisting entirely of zeros. Here, the matrix $f(a)f(a)^*$ will represent the superposition of all meanings of the token $a$ while each individual meaning of the token $a$ will be represented as a vector.

Suppose that $a_1\dots a_r\in A^\ast$ is a sequence of tokens. $R(a_1,\dots,a_r)$ will be a sequence of vectors $(v_1,\dots,v_r)\in (K^d)^r$ where each vector $v_j$ represents the meaning of the token $a_j$ in the text $a_1\dots a_r$. We shall call the vector $v_j$ an instantiation of the token $a_j$. Here, the word embedding $f$ together with the function $R$ will be trained so that $f(a)f(a)^\ast=E(vv^\ast)$ where $v$ is a randomly selected instantiation of the token $a$. For further processing, the machine learning model will not use any information about $a_1\dots a_r$ other than the sequence of vectors $R(a_1,\dots,a_r)$. If there is an issue with the natural language model, then one can more easily determine whether the issue is with the tokenization of text, the word embedding $f$, the instantiation $R(a_1,\dots,a_r)$ of the sequence $(f(a_1),\dots,f(a_r))$ of matrices, or some other part of the natural language model. Since vectors are simpler than matrices, the object $R(a_1,\dots,a_r)$ will take less memory than the sequence of matrices $(f(a_1),\dots,f(a_r))$. This will simplify the natural language model; a simplified natural language model will be more interpretable, understandable, energy efficient, and safe than an unsimplified natural language model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant