Skip to content

Commit cfdc15a

Browse files
committed
Normalize all line endings to LF
1 parent cbc15b3 commit cfdc15a

39 files changed

+2187
-2185
lines changed

.github/workflows/canary.yml

+41-41
Original file line numberDiff line numberDiff line change
@@ -1,41 +1,41 @@
1-
name: Build and Commit to Canary
2-
3-
on:
4-
push:
5-
branches:
6-
- main
7-
8-
jobs:
9-
build:
10-
runs-on: ubuntu-latest
11-
12-
permissions:
13-
contents: write # Grant write permission to the contents scope
14-
15-
steps:
16-
- name: Checkout code
17-
uses: actions/checkout@v4
18-
19-
- name: Set up Node.js
20-
uses: actions/setup-node@v4
21-
with:
22-
node-version: '20.x'
23-
24-
- name: Install frontend dependencies
25-
working-directory: frontend
26-
run: npm ci
27-
28-
- name: Build frontend
29-
working-directory: frontend
30-
run: npm run build
31-
32-
- name: Commit and push to canary branch
33-
run: |
34-
git config --global user.name 'github-actions[bot]'
35-
git config --global user.email 'github-actions[bot]@users.noreply.github.com'
36-
git checkout -b canary
37-
git add -f frontend/dist
38-
git commit -m "Update canary with latest build from main"
39-
git push origin canary --force
40-
env:
41-
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
1+
name: Build and Commit to Canary
2+
3+
on:
4+
push:
5+
branches:
6+
- main
7+
8+
jobs:
9+
build:
10+
runs-on: ubuntu-latest
11+
12+
permissions:
13+
contents: write # Grant write permission to the contents scope
14+
15+
steps:
16+
- name: Checkout code
17+
uses: actions/checkout@v4
18+
19+
- name: Set up Node.js
20+
uses: actions/setup-node@v4
21+
with:
22+
node-version: '20.x'
23+
24+
- name: Install frontend dependencies
25+
working-directory: frontend
26+
run: npm ci
27+
28+
- name: Build frontend
29+
working-directory: frontend
30+
run: npm run build
31+
32+
- name: Commit and push to canary branch
33+
run: |
34+
git config --global user.name 'github-actions[bot]'
35+
git config --global user.email 'github-actions[bot]@users.noreply.github.com'
36+
git checkout -b canary
37+
git add -f frontend/dist
38+
git commit -m "Update canary with latest build from main"
39+
git push origin canary --force
40+
env:
41+
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}

.vscode/tasks.json.sample

+75-75
Original file line numberDiff line numberDiff line change
@@ -1,75 +1,75 @@
1-
{
2-
"version": "2.0.0",
3-
"tasks": [
4-
{
5-
"label": "Start Frontend Server",
6-
"type": "shell",
7-
"command": "npm run dev",
8-
"isBackground": true,
9-
"presentation": {
10-
"reveal": "always",
11-
"panel": "new",
12-
"group": "frontend"
13-
},
14-
"options": {
15-
"cwd": "${workspaceFolder}/frontend"
16-
},
17-
"problemMatcher": {
18-
"owner": "custom",
19-
"pattern": [
20-
{
21-
"regexp": ".",
22-
"file": 1,
23-
"location": 2,
24-
"message": 3
25-
}
26-
],
27-
"background": {
28-
"activeOnStart": true,
29-
"beginsPattern": ".",
30-
"endsPattern": "."
31-
}
32-
}
33-
},
34-
{
35-
"label": "Start Backend Server",
36-
"type": "shell",
37-
"command": "./.vscode/tasks/start_backend.sh",
38-
"windows": {
39-
"command": "pwsh .vscode/tasks/start_backend.ps1",
40-
},
41-
"presentation": {
42-
"reveal": "always",
43-
"panel": "new",
44-
"group": "backend"
45-
},
46-
"problemMatcher": {
47-
"owner": "custom",
48-
"pattern": [
49-
{
50-
"regexp": ".",
51-
"file": 1,
52-
"location": 2,
53-
"message": 3
54-
}
55-
],
56-
"background": {
57-
"activeOnStart": true,
58-
"beginsPattern": ".",
59-
"endsPattern": "."
60-
}
61-
}
62-
},
63-
{
64-
"label": "Run Both Servers",
65-
"dependsOn": [
66-
"Start Frontend Server",
67-
"Start Backend Server"
68-
],
69-
"presentation": {
70-
"reveal": "always",
71-
"panel": "shared"
72-
}
73-
}
74-
]
75-
}
1+
{
2+
"version": "2.0.0",
3+
"tasks": [
4+
{
5+
"label": "Start Frontend Server",
6+
"type": "shell",
7+
"command": "npm run dev",
8+
"isBackground": true,
9+
"presentation": {
10+
"reveal": "always",
11+
"panel": "new",
12+
"group": "frontend"
13+
},
14+
"options": {
15+
"cwd": "${workspaceFolder}/frontend"
16+
},
17+
"problemMatcher": {
18+
"owner": "custom",
19+
"pattern": [
20+
{
21+
"regexp": ".",
22+
"file": 1,
23+
"location": 2,
24+
"message": 3
25+
}
26+
],
27+
"background": {
28+
"activeOnStart": true,
29+
"beginsPattern": ".",
30+
"endsPattern": "."
31+
}
32+
}
33+
},
34+
{
35+
"label": "Start Backend Server",
36+
"type": "shell",
37+
"command": "./.vscode/tasks/start_backend.sh",
38+
"windows": {
39+
"command": "pwsh .vscode/tasks/start_backend.ps1",
40+
},
41+
"presentation": {
42+
"reveal": "always",
43+
"panel": "new",
44+
"group": "backend"
45+
},
46+
"problemMatcher": {
47+
"owner": "custom",
48+
"pattern": [
49+
{
50+
"regexp": ".",
51+
"file": 1,
52+
"location": 2,
53+
"message": 3
54+
}
55+
],
56+
"background": {
57+
"activeOnStart": true,
58+
"beginsPattern": ".",
59+
"endsPattern": "."
60+
}
61+
}
62+
},
63+
{
64+
"label": "Run Both Servers",
65+
"dependsOn": [
66+
"Start Frontend Server",
67+
"Start Backend Server"
68+
],
69+
"presentation": {
70+
"reveal": "always",
71+
"panel": "shared"
72+
}
73+
}
74+
]
75+
}

.vscode/tasks/start_backend.ps1

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,2 @@
1-
& .\.venv\Scripts\Activate.ps1
2-
python __main__.py
1+
& .\.venv\Scripts\Activate.ps1
2+
python __main__.py

__main__.py

+22-22
Original file line numberDiff line numberDiff line change
@@ -1,22 +1,22 @@
1-
import sys
2-
from pathlib import Path
3-
4-
# Ensure the parent directory is in sys.path so relative imports work.
5-
base_dir = Path(__file__).parent
6-
if base_dir not in sys.path:
7-
sys.path.append(str(base_dir))
8-
from backend.env import check_env
9-
from backend.paths import backend_dir, venv_dir
10-
11-
if __name__ == "__main__":
12-
# Check if the environment is set up and activated before importing dependencies
13-
check_env()
14-
from starlette.staticfiles import StaticFiles
15-
16-
# Create the app
17-
from app import create_app
18-
app = create_app()
19-
20-
# Run the app
21-
import uvicorn
22-
uvicorn.run("app:create_app", host="localhost", port=3080, factory=True, workers=1, reload=True, reload_dirs=[backend_dir], reload_excludes=[venv_dir])
1+
import sys
2+
from pathlib import Path
3+
4+
# Ensure the parent directory is in sys.path so relative imports work.
5+
base_dir = Path(__file__).parent
6+
if base_dir not in sys.path:
7+
sys.path.append(str(base_dir))
8+
from backend.env import check_env
9+
from backend.paths import backend_dir, venv_dir
10+
11+
if __name__ == "__main__":
12+
# Check if the environment is set up and activated before importing dependencies
13+
check_env()
14+
from starlette.staticfiles import StaticFiles
15+
16+
# Create the app
17+
from app import create_app
18+
app = create_app()
19+
20+
# Run the app
21+
import uvicorn
22+
uvicorn.run("app:create_app", host="localhost", port=3080, factory=True, workers=1, reload=True, reload_dirs=[backend_dir], reload_excludes=[venv_dir])

abilities/.gitignore

+5-5
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
1-
# Ignore pAI-OS installed files
2-
installing
3-
installed
4-
upgrading
5-
uninstalling
1+
# Ignore pAI-OS installed files
2+
installing
3+
installed
4+
upgrading
5+
uninstalling
+35-35
Original file line numberDiff line numberDiff line change
@@ -1,35 +1,35 @@
1-
{
2-
"id": "llama_cpp",
3-
"name": "llama.cpp",
4-
"versions": {
5-
"package": "0.1.0",
6-
"product": "3091"
7-
},
8-
"description": "LLM inference in C/C++",
9-
"repository": {
10-
"type": "git",
11-
"url": "https://github.com/ggerganov/llama.cpp.git"
12-
},
13-
"author": {
14-
"name": "ggml authors",
15-
"url": "https://github.com/ggerganov/llama.cpp"
16-
},
17-
"maintainer": {
18-
"name": "Sam Johnston",
19-
"email": "[email protected]",
20-
"url": "http://samjohnston.org"
21-
},
22-
"language": "c++",
23-
"license": "MIT",
24-
"dependencies": [
25-
{
26-
"type": "python",
27-
"id": "fetch_llama_cpp",
28-
"name": "fetch_llama_cpp",
29-
"description": "Fetches the latest and best version of llama.cpp for your system.",
30-
"versions": {
31-
"required": ">=0.1.0"
32-
}
33-
}
34-
]
35-
}
1+
{
2+
"id": "llama_cpp",
3+
"name": "llama.cpp",
4+
"versions": {
5+
"package": "0.1.0",
6+
"product": "3091"
7+
},
8+
"description": "LLM inference in C/C++",
9+
"repository": {
10+
"type": "git",
11+
"url": "https://github.com/ggerganov/llama.cpp.git"
12+
},
13+
"author": {
14+
"name": "ggml authors",
15+
"url": "https://github.com/ggerganov/llama.cpp"
16+
},
17+
"maintainer": {
18+
"name": "Sam Johnston",
19+
"email": "[email protected]",
20+
"url": "http://samjohnston.org"
21+
},
22+
"language": "c++",
23+
"license": "MIT",
24+
"dependencies": [
25+
{
26+
"type": "python",
27+
"id": "fetch_llama_cpp",
28+
"name": "fetch_llama_cpp",
29+
"description": "Fetches the latest and best version of llama.cpp for your system.",
30+
"versions": {
31+
"required": ">=0.1.0"
32+
}
33+
}
34+
]
35+
}

0 commit comments

Comments
 (0)