diff --git a/README.md b/README.md
index 81c5b84..58e2806 100644
--- a/README.md
+++ b/README.md
@@ -6,7 +6,7 @@ A RAG LLM co-pilot for browsing the web, powered by local LLMs.
![Screenshot of Lumos](./screenshots/lumos_screenshot_5.png)
-This Chrome extension is powered by [Ollama](https://ollama.ai/). Inference is done on your local machine without any _external_ server support. However, due to security constraints in the Chrome extension platform, the app does rely on _local_ server support to run the LLM. This app is inspired by the [Chrome extension example](https://github.com/mlc-ai/web-llm/tree/main/examples/chrome-extension) provided by the [Web LLM project](https://webllm.mlc.ai/) and the [local LLM examples](https://js.langchain.com/docs/use_cases/question_answering/local_retrieval_qa) provided by [LangChain](https://github.com/langchain-ai/langchainjs).
+This Chrome extension is powered by [Ollama](https://ollama.ai/). Inference is done on your local machine without any _remote_ server support. However, due to security constraints in the Chrome extension platform, the app does rely on _local_ server support to run the LLM. This app is inspired by the [Chrome extension example](https://github.com/mlc-ai/web-llm/tree/main/examples/chrome-extension) provided by the [Web LLM project](https://webllm.mlc.ai/) and the [local LLM examples](https://js.langchain.com/docs/use_cases/question_answering/local_retrieval_qa) provided by [LangChain](https://github.com/langchain-ai/langchainjs).
- [Ollama (Home)](https://ollama.ai/)
- [Ollama (GitHub)](https://github.com/jmorganca/ollama)
@@ -78,7 +78,7 @@ See the section about [running tests](https://facebook.github.io/create-react-ap
### `npm run lint`
-Runs `eslint` and `prettier` on `src` files.
+Runs `eslint` and `prettier` on `src` and `__tests__` files.
### `npm run build`
@@ -256,3 +256,4 @@ Lumos invokes [Tools](https://js.langchain.com/docs/modules/agents/tools/) autom
- [Let’s Normalize Online, In-Memory RAG! (Part 3)](https://medium.com/@andrewnguonly/lets-normalize-online-in-memory-rag-88e8169e9806)
- [Supercharging If-Statements With Prompt Classification Using Ollama and LangChain (Part 4)](https://blog.langchain.dev/supercharging-if-statements-with-prompt-classification-using-ollama-and-langchain/)
- [Bolstering LangChain’s MemoryVectorStore With Keyword Search (Part 5)](https://medium.com/@andrewnguonly/bolstering-langchains-memoryvectorstore-with-keyword-search-3d35cc278dc1)
+- [A Guide to Gotchas with LangChain Document Loaders in a Chrome Extension (Part 6)](https://medium.com/@andrewnguonly/a-guide-to-gotchas-with-langchain-document-loaders-in-a-chrome-extension-6228369f79f6)
diff --git a/src/components/ChatBar.tsx b/src/components/ChatBar.tsx
index ad57622..2b34677 100644
--- a/src/components/ChatBar.tsx
+++ b/src/components/ChatBar.tsx
@@ -13,6 +13,7 @@ import InfoIcon from "@mui/icons-material/Info";
import PlaylistRemoveIcon from "@mui/icons-material/PlaylistRemove";
import RefreshIcon from "@mui/icons-material/Refresh";
import SaveAltIcon from "@mui/icons-material/SaveAlt";
+import SettingsIcon from "@mui/icons-material/Settings";
import {
Alert,
Box,
@@ -724,6 +725,13 @@ const ChatBar: React.FC = () => {
+