diff --git a/README.md b/README.md index 81c5b84..58e2806 100644 --- a/README.md +++ b/README.md @@ -6,7 +6,7 @@ A RAG LLM co-pilot for browsing the web, powered by local LLMs. ![Screenshot of Lumos](./screenshots/lumos_screenshot_5.png) -This Chrome extension is powered by [Ollama](https://ollama.ai/). Inference is done on your local machine without any _external_ server support. However, due to security constraints in the Chrome extension platform, the app does rely on _local_ server support to run the LLM. This app is inspired by the [Chrome extension example](https://github.com/mlc-ai/web-llm/tree/main/examples/chrome-extension) provided by the [Web LLM project](https://webllm.mlc.ai/) and the [local LLM examples](https://js.langchain.com/docs/use_cases/question_answering/local_retrieval_qa) provided by [LangChain](https://github.com/langchain-ai/langchainjs). +This Chrome extension is powered by [Ollama](https://ollama.ai/). Inference is done on your local machine without any _remote_ server support. However, due to security constraints in the Chrome extension platform, the app does rely on _local_ server support to run the LLM. This app is inspired by the [Chrome extension example](https://github.com/mlc-ai/web-llm/tree/main/examples/chrome-extension) provided by the [Web LLM project](https://webllm.mlc.ai/) and the [local LLM examples](https://js.langchain.com/docs/use_cases/question_answering/local_retrieval_qa) provided by [LangChain](https://github.com/langchain-ai/langchainjs). - [Ollama (Home)](https://ollama.ai/) - [Ollama (GitHub)](https://github.com/jmorganca/ollama) @@ -78,7 +78,7 @@ See the section about [running tests](https://facebook.github.io/create-react-ap ### `npm run lint` -Runs `eslint` and `prettier` on `src` files. +Runs `eslint` and `prettier` on `src` and `__tests__` files. ### `npm run build` @@ -256,3 +256,4 @@ Lumos invokes [Tools](https://js.langchain.com/docs/modules/agents/tools/) autom - [Let’s Normalize Online, In-Memory RAG! (Part 3)](https://medium.com/@andrewnguonly/lets-normalize-online-in-memory-rag-88e8169e9806) - [Supercharging If-Statements With Prompt Classification Using Ollama and LangChain (Part 4)](https://blog.langchain.dev/supercharging-if-statements-with-prompt-classification-using-ollama-and-langchain/) - [Bolstering LangChain’s MemoryVectorStore With Keyword Search (Part 5)](https://medium.com/@andrewnguonly/bolstering-langchains-memoryvectorstore-with-keyword-search-3d35cc278dc1) +- [A Guide to Gotchas with LangChain Document Loaders in a Chrome Extension (Part 6)](https://medium.com/@andrewnguonly/a-guide-to-gotchas-with-langchain-document-loaders-in-a-chrome-extension-6228369f79f6) diff --git a/src/components/ChatBar.tsx b/src/components/ChatBar.tsx index ad57622..2b34677 100644 --- a/src/components/ChatBar.tsx +++ b/src/components/ChatBar.tsx @@ -13,6 +13,7 @@ import InfoIcon from "@mui/icons-material/Info"; import PlaylistRemoveIcon from "@mui/icons-material/PlaylistRemove"; import RefreshIcon from "@mui/icons-material/Refresh"; import SaveAltIcon from "@mui/icons-material/SaveAlt"; +import SettingsIcon from "@mui/icons-material/Settings"; import { Alert, Box, @@ -724,6 +725,13 @@ const ChatBar: React.FC = () => { +