Pre-configured AI server // How? #706
Replies: 14 comments
-
Ah, I see! To use the environment variables from the OS, you need to pass the environment variables to the container manually. So if you're running
But if you are using Let me know if it helps. |
Beta Was this translation helpful? Give feedback.
-
I think that is what I have been asking, in regard to mounting the volumes. that way I can create and edit the .env. I have full control of the machine. |
Beta Was this translation helpful? Give feedback.
-
Ah ok, then I need to clarify it in the Readme. Please try this:
Edit the
This will build and run the docker container, already taking all the files in And to update it to the latest version, stop the container, and run:
P.S. I know you know the commands/sequence already, but I'm just writing in detail so I can reference this message if other issues are open about this later. |
Beta Was this translation helpful? Give feedback.
-
I don't use those commands for my containers. I'm stuck with a GUI for building and maintaining containers (unRaid OS). Are there any flags to turn off the browser or openai interface? Also, point of confusion, the AI Processing Location has default options of What option leads to what DEFAULT_INFERENCE_TYPE ? Consider naming them the same thing and graying out options that are not available. |
Beta Was this translation helpful? Give feedback.
-
Ah yes, about the confusion, it's because I've recently renamed those options, but I forgot to update the But here's how they map:
No, as this would go against the main purpose of the app, which was to run models directly from the browser (the OpenAI-Compatible API was only added recently). Could you tell me more about why you think there should be this option? But going back to the topic, if unRaid OS runs Please let me know if any of the solutions worked. |
Beta Was this translation helpful? Give feedback.
-
No, as this would go against the main purpose of the app, which was to run models directly from the browser (the OpenAI-Compatible API was only added recently). Could you tell me more about why you think there should be this option? I'm not using the app with browser based LLMs. Ill use it with my ollama server for private search. I view it is a very lite weight webui for llms. |
Beta Was this translation helpful? Give feedback.
-
Ah, so it's to avoid loading the libraries that run the LLMs in the browser, making the app lighter, right? No need to worry, those libraries are only loaded if you have the "In the browser (Private)" AI Processing Location selected. It was all modularized to use only what users needed. By the way, this discussion gave me great insights, so I thought about simplifying the server configuration, which is becoming complicated with all those environment variables: |
Beta Was this translation helpful? Give feedback.
-
Back to the original question, were you able to make it display the custom option to use the internal API? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Still not working. I'm stuck. |
Beta Was this translation helpful? Give feedback.
-
I’m out of ideas on how to address this issue. In the meantime, as you search for a solution to get the Internal API working, can you use the "Remote Server (API)" option to connect to Ollama directly from the browser? |
Beta Was this translation helpful? Give feedback.
-
I can, and thats what I currently have set up. Works great. |
Beta Was this translation helpful? Give feedback.
-
That's good! Then, I'm converting this issue into a discussion to keep it open in case more information related to this arises. |
Beta Was this translation helpful? Give feedback.
-
Sounds good. I still think this problem has to do with read/write access on my system. Any chance I can get configuration settings so I can mount a volume? |
Beta Was this translation helpful? Give feedback.
-
Bug description
On your demo site, you have a pre-configured AI server.
How did you set this up?
Steps to reproduce
Here are my current docker settings:
Expected behavior
When I load my site https://search.robocanvas.io it should auto load with my pre-configured ollama server AND the user should not have the option to change the configuration.
Device info
No response
Additional context
No response
Beta Was this translation helpful? Give feedback.
All reactions