|
|
- ```
- _ _ _
- | |__ __ _ ___ __ _| |__ _ _ _ _| | ____ _
- | '_ \ / _` / __|/ _` | '_ \| | | | | | | |/ / _` |
- | |_) | (_| \__ \ (_| | |_) | |_| | |_| | < (_| |
- |_.__/ \__,_|___/\__,_|_.__/ \__,_|\__,_|_|\_\__,_|
-
- ```
-
- This whole code is licensed under the License of Redistribution 1.0.
- You can find a copy of this licence on this instance, under
-
- https://code.basabuuka.org/alpcentaur/license-of-redistribution
-
-
-
- The whole prototype is nested in a tensorflow/tensorflow:2.2.0-gpu docker
- container.
- Install graphic card drivers according to your hardware and your OS.
- To make the tensorflow docker container work.
-
- Also get Ollama running in a docker container, sharing the same network protoRustNet.
- That it is reachable from the ollama-cli-server under http://ollama:11434/api/generate.
-
- I run my ollama container seperately together with open web ui.
- Like that, I can administrate the models over the web ui, and then use them by changing the code
- in the fastapi_server.py file of the ollama-cli-server container.
-
- After having set up the ollama container and the gpu docker drivers, just start the whole project in the
- ```
- compose
- ```
- directory with
- ```
- docker compose up -d
- ```
-
- The deb-rust-interface will be running on port
-
-
- For instructions how to setup a webserver as reverse proxy, you can contact basabuuka.
- My nginx configuration for the basabuuka prototype is the following:
-
- ```
- upstream protointerface {
- server 127.0.0.1:1030;
- }
-
- server {
-
- server_name example.org;
- # access logs consume resources
- access_log /var/log/nginx/example.org.access.log;
-
- location / {
- proxy_pass http://protointerface;
- }
-
- # set high timeouts for using 14b model on 200 euro gpu
- proxy_connect_timeout 300;
- proxy_send_timeout 300;
- proxy_read_timeout 300;
- send_timeout 300;
- keepalive_timeout 300;
-
-
- }
-
- ```
|