Skip to content

Commit

Permalink
Merge pull request #115 from jackschedel/2.1.0
Browse files Browse the repository at this point in the history
2.1.0b major bug fixes
  • Loading branch information
jackschedel committed Feb 12, 2024
2 parents b563ce1 + b9d1ba9 commit cfd6f60
Show file tree
Hide file tree
Showing 5 changed files with 23 additions and 12 deletions.
13 changes: 13 additions & 0 deletions src/App.tsx
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,8 @@ function App() {
const apiAuth = useStore((state) => state.apiAuth);
const apiKey = useStore((state) => state.apiKey);
const setApiKey = useStore((state) => state.setApiKey);
const modelDefs = useStore((state) => state.modelDefs);
const setModelDefs = useStore((state) => state.setModelDefs);

const initialiseNewChat = useInitialiseNewChat();
const addChat = useAddChat();
Expand All @@ -41,6 +43,17 @@ function App() {
setApiKey('');
}

// migration from broken 2.1.0 release
if (
modelDefs[1].model === 'gpt-4-turbo-preview' &&
modelDefs[1].model_max_tokens == 128000
) {
modelDefs[1].model_max_tokens = 4096;
modelDefs[1].model_max_context = 128000;

setModelDefs(modelDefs);
}

const handleGenerate = () => {
if (useStore.getState().generating) return;
const updatedChats: ChatInterface[] = JSON.parse(
Expand Down
13 changes: 6 additions & 7 deletions src/components/ChatConfigMenu/ChatConfigMenu.tsx
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,6 @@ const ChatConfigPopup = ({
config.model_selection
);
const [_maxTokens, _setMaxTokens] = useState<number>(config.max_tokens);
const [_maxToken, _setMaxToken] = useState<number>(config.max_tokens);
const [_temperature, _setTemperature] = useState<number>(config.temperature);
const [_topP, _setTopP] = useState<number>(config.top_p);
const [_presencePenalty, _setPresencePenalty] = useState<number>(
Expand All @@ -65,7 +64,7 @@ const ChatConfigPopup = ({
const handleSave = () => {
setDefaultChatConfig({
model_selection: _modelSelection,
max_tokens: _maxToken,
max_tokens: _maxTokens,
temperature: _temperature,
top_p: _topP,
presence_penalty: _presencePenalty,
Expand All @@ -76,7 +75,7 @@ const ChatConfigPopup = ({
};

const handleReset = () => {
_setMaxToken(_defaultChatConfig.max_tokens);
_setMaxTokens(_defaultChatConfig.max_tokens);
_setTemperature(_defaultChatConfig.temperature);
_setTopP(_defaultChatConfig.top_p);
_setPresencePenalty(_defaultChatConfig.presence_penalty);
Expand All @@ -99,15 +98,15 @@ const ChatConfigPopup = ({
_setModel={_setModelSelection}
showHidden={true}
/>
<DefaultSystemChat
_systemMessage={_systemMessage}
_setSystemMessage={_setSystemMessage}
/>
<MaxTokenSlider
_maxToken={_maxTokens}
_setMaxToken={_setMaxTokens}
_model={_modelSelection}
/>
<DefaultSystemChat
_systemMessage={_systemMessage}
_setSystemMessage={_setSystemMessage}
/>
<TemperatureSlider
_temperature={_temperature}
_setTemperature={_setTemperature}
Expand Down
1 change: 0 additions & 1 deletion src/components/ConfigMenu/ConfigMenu.tsx
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,6 @@ const ConfigMenu = ({
title={t('configuration') as string}
setIsModalOpen={setIsModalOpen}
handleConfirm={handleConfirm}
handleClickBackdrop={handleConfirm}
>
<div className='p-6 border-b bg-neutral-base'>
<label className='block text-sm font-medium text-custom-white pb-2'>
Expand Down
2 changes: 1 addition & 1 deletion src/hooks/useSubmit.ts
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ const useSubmit = () => {
const messages = limitMessageTokens(
chats[currentChatIndex].messages,
modelDef.model,
modelDef.model_max_tokens,
modelDef.model_max_context,
config.max_tokens
);
if (messages.length === 0) throw new Error('Message exceeds max token!');
Expand Down
6 changes: 3 additions & 3 deletions src/store/auth-slice.ts
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ export const createAuthSlice: StoreSlice<AuthSlice> = (set) => ({
name: 'gpt-3',
model: 'gpt-3.5-turbo',
endpoint: 0,
model_max_context: 4096,
model_max_context: 16385,
model_max_tokens: 4096,
prompt_cost_1000: 0.0005,
completion_cost_1000: 0.0005,
Expand All @@ -32,8 +32,8 @@ export const createAuthSlice: StoreSlice<AuthSlice> = (set) => ({
name: 'gpt-4',
model: 'gpt-4-turbo-preview',
endpoint: 0,
model_max_context: 4096,
model_max_tokens: 128000,
model_max_context: 128000,
model_max_tokens: 4096,
prompt_cost_1000: 0.01,
completion_cost_1000: 0.03,
swap_visible: true,
Expand Down

0 comments on commit cfd6f60

Please sign in to comment.