[go: nahoru, domu]

Skip to content

Commit

Permalink
fix: Bare metal (bionic-gpt#409)
Browse files Browse the repository at this point in the history
  • Loading branch information
9876691 committed Mar 19, 2024
1 parent a5b04d8 commit e4424e3
Show file tree
Hide file tree
Showing 18 changed files with 92 additions and 676 deletions.
24 changes: 23 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -93,8 +93,30 @@

See the [open issues](https://github.com/purton-tech/bionicgpt/issues) for a full list of proposed features (and known issues).

# Enterprise
For companies that need better security, user management and professional support

[Talk to the founders](https://calendly.com/bionicgpt)

This covers:
-**Help with integrations**
-**Feature Prioritization**
-**Custom Integrations**
-**LTS (Long Term Support) Versions**
-**Professional Support**
-**Custom SLAs**
-**Secure access with Single Sign-On**
-**Continuous Batching**
-**Data Pipelines**

# Support / talk with founders

- [Schedule a Chat 👋](https://calendly.com/bionicgpt)
- [Connect on Linked in 💭](https://www.linkedin.com/in/kulbinderdio/)
- Our emails ✉️ ian@bionic-gpt.com / dio@bionic-gpt.com

<!-- Try it out -->
## Try it out with `docker compose`
## Try it out

follow [our guide](https://bionic-gpt.com/docs/running-locally/introduction/) to running Bionic-GPT on your local machine.

Expand Down
4 changes: 2 additions & 2 deletions crates/ui-pages/team_members/team_role.rs
Original file line number Diff line number Diff line change
Expand Up @@ -10,14 +10,14 @@ pub fn Role(cx: Scope, role: Role) -> Element {
Label {
class: "mr-2",
label_role: LabelRole::Highlight,
"SystemAdministrator"
"System Administrator"
}
)),
Role::TeamManager => cx.render(rsx!(
Label {
class: "mr-2",
label_role: LabelRole::Neutral,
"Team Manger"
"Team Manager"
}
)),
Role::Collaborator => cx.render(rsx!(
Expand Down
1 change: 1 addition & 0 deletions install.sh
Original file line number Diff line number Diff line change
Expand Up @@ -32,6 +32,7 @@ reset_k3s() {
sudo /usr/local/bin/k3s-uninstall.sh
curl -sfL https://get.k3s.io | sh -
sudo chmod 444 /etc/rancher/k3s/k3s.yaml
mkdir -p ~/.kube
cp /etc/rancher/k3s/k3s.yaml ~/.kube/config
sed -i "s,127.0.0.1,$1,g" ~/.kube/config
}
Expand Down
4 changes: 4 additions & 0 deletions website/content/.spelling
Original file line number Diff line number Diff line change
Expand Up @@ -183,6 +183,10 @@ Airbyte
localhost
RabbitMQ
MQ
Ollama
systemd
K3s
IoT
- docs/rag/generating-embeddings.md
80
Instruction-Finetuned
Expand Down
82 changes: 35 additions & 47 deletions website/content/docs/enterprise-edition/adding-models.md
Original file line number Diff line number Diff line change
@@ -1,54 +1,42 @@
+++
title = "Adding LLM Models"
title = "Connecting to Ollama"
weight = 95
sort_by = "weight"
+++

To add a model to your cluster you can create a [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/) and a [Service](https://kubernetes.io/docs/concepts/services-networking/service/).

## Example Adding Mixtral 8x7B (Work in progress)

This section is a work in progress but deploying a new model will look something like this.

```yml
apiVersion: v1
kind: Service
metadata:
name: mixtral-8x7b
namespace: bionic-gpt
spec:
selector:
app: mixtral-8x7b
ports:
- protocol: TCP
port: 8000
targetPort: 8000

---

apiVersion: apps/v1
kind: Deployment
metadata:
name: mixtral-8x7b-deployment
namespace: bionic-gpt
spec:
replicas: 1
selector:
matchLabels:
app: mixtral-8x7b
template:
metadata:
labels:
app: mixtral-8x7b
spec:
containers:
- name: bionic-gpt-operator
image: ghcr.io/huggingface/text-generation-inference:sha-0eabc83
args:
- --model-id
- mistralai/Mixtral-8x7B-Instruct-v0.1
- --quantize
- gptq
We need to get Ollama to listen on `0.0.0.0`.

Edit the systemd service by calling `sudo vi /etc/systemd/system/ollama.service`. This will open an editor.

For each environment variable, add a line Environment under section [Service]:

```
[Service]
Environment="OLLAMA_HOST=0.0.0.0"
```

Save and exit.

Reload systemd and restart Ollama:

```sh
systemctl daemon-reload
systemctl restart ollama
```

You can run the following to view the logs

```
journalctl -u ollama
```

## Test Ollama

Get you host with `hostname` then curl using that host.

The model and inference engine must run in the same container.
```sh
curl http://pop-os:11434/api/generate -d '{
"model": "phi",
"prompt":"Why is the sky blue?"
}'
```
33 changes: 0 additions & 33 deletions website/content/docs/enterprise-edition/install-bionic-operator.md

This file was deleted.

Loading

0 comments on commit e4424e3

Please sign in to comment.