Skip to content

Commit

Permalink
docs: Fix broken imports in documentation (#19655)
Browse files Browse the repository at this point in the history
Found via script in #19611
  • Loading branch information
ccurme committed Mar 27, 2024
1 parent 0019d8a commit 4e9b358
Show file tree
Hide file tree
Showing 15 changed files with 20 additions and 17 deletions.
2 changes: 1 addition & 1 deletion docs/docs/guides/model_laboratory.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -167,7 +167,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain.chains import SelfAskWithSearchChain\n",
"from langchain.agents.self_ask_with_search.base import SelfAskWithSearchChain\n",
"from langchain_community.utilities import SerpAPIWrapper\n",
"\n",
"open_ai_llm = OpenAI(temperature=0)\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@
"outputs": [],
"source": [
"from langchain_community.document_loaders import ApifyDatasetLoader\n",
"from langchain_community.document_loaders.base import Document"
"from langchain_core.documents import Document"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/document_loaders/quip.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain_community.document_loaders import QuipLoader\n",
"from langchain_community.document_loaders.quip import QuipLoader\n",
"\n",
"loader = QuipLoader(\n",
" api_url=\"https://platform.quip.com\", access_token=\"change_me\", request_timeout=60\n",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/llms/llm_caching.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -1378,7 +1378,7 @@
},
"outputs": [],
"source": [
"from langchain.cache import AzureCosmosDBSemanticCache\n",
"from langchain_community.cache import AzureCosmosDBSemanticCache\n",
"from langchain_community.vectorstores.azure_cosmos_db import (\n",
" CosmosDBSimilarityType,\n",
" CosmosDBVectorSearchType,\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -154,7 +154,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain_community.vectorstores import XataVectorStore\n",
"from langchain_community.vectorstores.xata import XataVectorStore\n",
"from langchain_openai import OpenAIEmbeddings\n",
"\n",
"embeddings = OpenAIEmbeddings()\n",
Expand Down
3 changes: 2 additions & 1 deletion docs/docs/integrations/retrievers/cohere-reranker.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -334,7 +334,8 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain.retrievers import CohereRerank, ContextualCompressionRetriever\n",
"from langchain.retrievers.contextual_compression import ContextualCompressionRetriever\n",
"from langchain_cohere import CohereRerank\n",
"from langchain_community.llms import Cohere\n",
"\n",
"llm = Cohere(temperature=0)\n",
Expand Down
4 changes: 2 additions & 2 deletions docs/docs/integrations/retrievers/jaguar.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,7 @@
"outputs": [],
"source": [
"from langchain_community.document_loaders import TextLoader\n",
"from langchain_community.vectorstores import Jaguar\n",
"from langchain_community.vectorstores.jaguar import Jaguar\n",
"from langchain_openai import OpenAIEmbeddings\n",
"from langchain_text_splitters import CharacterTextSplitter\n",
"\n",
Expand Down Expand Up @@ -147,7 +147,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain_community.vectorstores import Jaguar\n",
"from langchain_community.vectorstores.jaguar import Jaguar\n",
"from langchain_openai import OpenAIEmbeddings\n",
"\n",
"# Instantiate a Jaguar vector store object\n",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/retrievers/llmlingua.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -340,7 +340,7 @@
],
"source": [
"from langchain.retrievers import ContextualCompressionRetriever\n",
"from langchain_community.retrievers import LLMLinguaCompressor\n",
"from langchain_community.document_compressors import LLMLinguaCompressor\n",
"from langchain_openai import ChatOpenAI\n",
"\n",
"llm = ChatOpenAI(temperature=0)\n",
Expand Down
3 changes: 2 additions & 1 deletion docs/docs/integrations/toolkits/airbyte_structured_qa.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -28,8 +28,9 @@
"import os\n",
"\n",
"import pandas as pd\n",
"from langchain.agents import AgentType, create_pandas_dataframe_agent\n",
"from langchain.agents import AgentType\n",
"from langchain_community.document_loaders.airbyte import AirbyteStripeLoader\n",
"from langchain_experimental.agents import create_pandas_dataframe_agent\n",
"from langchain_openai import ChatOpenAI\n",
"\n",
"stream_name = \"customers\"\n",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/toolkits/spark.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -334,7 +334,7 @@
"source": [
"import os\n",
"\n",
"from langchain.agents import create_spark_dataframe_agent\n",
"from langchain_experimental.agents import create_spark_dataframe_agent\n",
"from langchain_openai import OpenAI\n",
"\n",
"os.environ[\"OPENAI_API_KEY\"] = \"...input your openai api key here...\"\n",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/toolkits/xorbits.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -380,7 +380,7 @@
],
"source": [
"import xorbits.numpy as np\n",
"from langchain.agents import create_xorbits_agent\n",
"from langchain_experimental.agents.agent_toolkits import create_xorbits_agent\n",
"from langchain_openai import OpenAI\n",
"\n",
"arr = np.array([1, 2, 3, 4, 5, 6])\n",
Expand Down
4 changes: 2 additions & 2 deletions docs/docs/integrations/tools/google_drive.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -99,8 +99,8 @@
},
"outputs": [],
"source": [
"from langchain_community.tools.google_drive.tool import GoogleDriveSearchTool\n",
"from langchain_community.utilities.google_drive import GoogleDriveAPIWrapper\n",
"from langchain_googldrive.tools.google_drive.tool import GoogleDriveSearchTool\n",
"from langchain_googledrive.utilities.google_drive import GoogleDriveAPIWrapper\n",
"\n",
"# By default, search only in the filename.\n",
"tool = GoogleDriveSearchTool(\n",
Expand Down
2 changes: 1 addition & 1 deletion docs/docs/integrations/vectorstores/vikingdb.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@
"outputs": [],
"source": [
"from langchain.document_loaders import TextLoader\n",
"from langchain.vectorstores.vikingdb import VikingDB, VikingDBConfig\n",
"from langchain_community.vectorstores.vikingdb import VikingDB, VikingDBConfig\n",
"from langchain_openai import OpenAIEmbeddings\n",
"from langchain_text_splitters import RecursiveCharacterTextSplitter"
]
Expand Down
3 changes: 2 additions & 1 deletion docs/docs/modules/callbacks/async_callbacks.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,8 @@
"from typing import Any, Dict, List\n",
"\n",
"from langchain.callbacks.base import AsyncCallbackHandler, BaseCallbackHandler\n",
"from langchain_core.messages import HumanMessage, LLMResult\n",
"from langchain_core.messages import HumanMessage\n",
"from langchain_core.outputs import LLMResult\n",
"from langchain_openai import ChatOpenAI\n",
"\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
"outputs": [],
"source": [
"import nest_asyncio\n",
"from langchain.chains.graph_qa import GremlinQAChain\n",
"from langchain.chains.graph_qa.gremlin import GremlinQAChain\n",
"from langchain.schema import Document\n",
"from langchain_community.graphs import GremlinGraph\n",
"from langchain_community.graphs.graph_document import GraphDocument, Node, Relationship\n",
Expand Down

0 comments on commit 4e9b358

Please sign in to comment.