Added docker setup
This commit is contained in:
parent
1ac28f4cb8
commit
7a7a857fdc
7 changed files with 4010 additions and 3668 deletions
20
cognee-frontend/Dockerfile
Normal file
20
cognee-frontend/Dockerfile
Normal file
|
|
@ -0,0 +1,20 @@
|
|||
# Use an official Node.js runtime as a parent image
|
||||
FROM node:18.17.0
|
||||
|
||||
# Set the working directory to /app
|
||||
WORKDIR /app
|
||||
|
||||
# Copy package.json and package-lock.json to the working directory
|
||||
COPY package*.json ./
|
||||
|
||||
# Install any needed packages specified in package.json
|
||||
RUN npm install
|
||||
|
||||
# Copy the rest of the application code to the working directory
|
||||
COPY . .
|
||||
|
||||
# Make port 3000 available to the world outside this container
|
||||
EXPOSE 3000
|
||||
|
||||
# Run npm install and then npm run dev
|
||||
CMD ["sh", "-c", "npm install && npm run dev"]
|
||||
7627
cognee-frontend/package-lock.json
generated
7627
cognee-frontend/package-lock.json
generated
File diff suppressed because it is too large
Load diff
|
|
@ -37,7 +37,7 @@ export default function Home() {
|
|||
}, [refreshDatasets]);
|
||||
|
||||
const openDatasetData = (dataset: { id: string }) => {
|
||||
fetch(`http://localhost:8000/datasets/${dataset.id}/data`)
|
||||
fetch(`http://0.0.0.0:8000/datasets/${dataset.id}/data`)
|
||||
.then((response) => response.json())
|
||||
.then(setDatasetData)
|
||||
.then(() => setSelectedDataset(dataset.id));
|
||||
|
|
@ -56,7 +56,7 @@ export default function Home() {
|
|||
const file = files[0];
|
||||
formData.append('data', file, file.name);
|
||||
|
||||
return fetch('http://localhost:8000/add', {
|
||||
return fetch('http://0.0.0.0:8000/add', {
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
})
|
||||
|
|
@ -88,7 +88,7 @@ export default function Home() {
|
|||
const cognifyDataset = useCallback((dataset: { id: string }) => {
|
||||
showNotification(`Cognification started for dataset "${dataset.id}".`, 5000);
|
||||
|
||||
return fetch('http://localhost:8000/cognify', {
|
||||
return fetch('http://0.0.0.0:8000/cognify', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
|
|
@ -120,7 +120,7 @@ export default function Home() {
|
|||
}, [cognifyDataset, disableCognifyRun, enableCognifyRun]);
|
||||
|
||||
const deleteDataset = useCallback((dataset: { id: string }) => {
|
||||
fetch(`http://localhost:8000/datasets/${dataset.id}`, {
|
||||
fetch(`http://0.0.0.0:8000/datasets/${dataset.id}`, {
|
||||
method: 'DELETE',
|
||||
})
|
||||
.then(() => {
|
||||
|
|
@ -146,7 +146,7 @@ export default function Home() {
|
|||
}, [showExplorationWindow]);
|
||||
|
||||
const exploreDataset = useCallback((dataset: { id: string }) => {
|
||||
fetch(`http://localhost:8000/datasets/${dataset.id}/graph`)
|
||||
fetch(`http://0.0.0.0:8000/datasets/${dataset.id}/graph`)
|
||||
.then((response) => response.text())
|
||||
.then((text) => text.replace('"', ''))
|
||||
.then((graphUrl: string) => {
|
||||
|
|
|
|||
|
|
@ -37,7 +37,7 @@ export default function DataView({ datasetId, data, onClose, onDataAdd }: DataVi
|
|||
const showRawData = useCallback((dataItem: Data) => {
|
||||
setSelectedData(dataItem);
|
||||
|
||||
fetch(`http://localhost:8000/datasets/${datasetId}/data/${dataItem.id}/raw`)
|
||||
fetch(`http://0.0.0.0:8000/datasets/${datasetId}/data/${dataItem.id}/raw`)
|
||||
.then((response) => response.arrayBuffer())
|
||||
.then(setRawData);
|
||||
}, [datasetId]);
|
||||
|
|
|
|||
|
|
@ -14,7 +14,7 @@ function useDatasets() {
|
|||
const statusTimeout = useRef<any>(null);
|
||||
|
||||
const fetchDatasetStatuses = useCallback((datasets: Dataset[]) => {
|
||||
fetch(`http://localhost:8000/datasets/status?dataset=${datasets.map(d => d.id).join('&dataset=')}`)
|
||||
fetch(`http://0.0.0.0:8000/datasets/status?dataset=${datasets.map(d => d.id).join('&dataset=')}`)
|
||||
.then((response) => response.json())
|
||||
.then((statuses) => setDatasets(
|
||||
(datasets) => (
|
||||
|
|
@ -65,7 +65,7 @@ function useDatasets() {
|
|||
}, []);
|
||||
|
||||
const fetchDatasets = useCallback(() => {
|
||||
fetch('http://localhost:8000/datasets')
|
||||
fetch('http://0.0.0.0:8000/datasets')
|
||||
.then((response) => response.json())
|
||||
.then((datasets) => datasets.map((dataset: string) => ({ id: dataset, name: dataset })))
|
||||
.then((datasets) => {
|
||||
|
|
|
|||
|
|
@ -37,11 +37,24 @@ services:
|
|||
depends_on:
|
||||
- postgres
|
||||
- neo4j
|
||||
- frontend
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
cpus: "4.0"
|
||||
memory: 8GB
|
||||
frontend:
|
||||
container_name: frontend
|
||||
build:
|
||||
context: ./cognee-frontend
|
||||
dockerfile: Dockerfile
|
||||
volumes:
|
||||
- "./cognee-frontend:/app"
|
||||
ports:
|
||||
- "3000:3000"
|
||||
networks:
|
||||
- cognee_backend
|
||||
|
||||
|
||||
postgres:
|
||||
image: postgres
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue