Skip to content

Commit

Permalink
fix builds
Browse files Browse the repository at this point in the history
  • Loading branch information
louis030195 committed Jan 22, 2025
1 parent fba156b commit 487aa90
Show file tree
Hide file tree
Showing 5 changed files with 10 additions and 131 deletions.
28 changes: 2 additions & 26 deletions pipes/data-table/src/lib/hooks/use-settings.tsx
Original file line number Diff line number Diff line change
@@ -1,34 +1,10 @@
"use client";

import { useState, useEffect } from "react";
import type { Settings } from "@screenpipe/browser";
import { getDefaultSettings, type Settings } from "@screenpipe/browser";

export function useSettings() {
const defaultSettings: Settings = {
analyticsEnabled: true,
openaiApiKey: "",
deepgramApiKey: "",
aiModel: "gpt-4",
aiUrl: "https://api.openai.com/v1",
customPrompt: "",
port: 3030,
dataDir: "default",
disableAudio: false,
ignoredWindows: [],
includedWindows: [],
aiProviderType: "openai",
embeddedLLM: {
enabled: false,
model: "llama3.2:1b-instruct-q4_K_M",
port: 11438,
},
enableFrameCache: true,
enableUiMonitoring: false,
aiMaxContextChars: 128000,
user: {
token: "",
},
};
const defaultSettings = getDefaultSettings();

const [settings, setSettings] = useState<Settings>(defaultSettings);
const [loading, setLoading] = useState(false);
Expand Down
28 changes: 2 additions & 26 deletions pipes/identify-speakers/src/lib/hooks/use-settings.tsx
Original file line number Diff line number Diff line change
@@ -1,34 +1,10 @@
"use client";

import { useState, useEffect } from "react";
import type { Settings } from "@screenpipe/browser";
import { getDefaultSettings, type Settings } from "@screenpipe/browser";

export function useSettings() {
const defaultSettings: Settings = {
analyticsEnabled: true,
openaiApiKey: "",
deepgramApiKey: "",
aiModel: "gpt-4",
aiUrl: "https://api.openai.com/v1",
customPrompt: "",
port: 3030,
dataDir: "default",
disableAudio: false,
ignoredWindows: [],
includedWindows: [],
aiProviderType: "openai",
embeddedLLM: {
enabled: false,
model: "llama3.2:1b-instruct-q4_K_M",
port: 11438,
},
enableFrameCache: true,
enableUiMonitoring: false,
aiMaxContextChars: 128000,
user: {
token: "",
},
};
const defaultSettings = getDefaultSettings();

const [settings, setSettings] = useState<Settings>(defaultSettings);
const [loading, setLoading] = useState(false);
Expand Down
28 changes: 2 additions & 26 deletions pipes/memories/src/lib/hooks/use-settings.tsx
Original file line number Diff line number Diff line change
@@ -1,34 +1,10 @@
"use client";

import { useState, useEffect } from "react";
import type { Settings } from "@screenpipe/browser";
import { getDefaultSettings, type Settings } from "@screenpipe/browser";

export function useSettings() {
const defaultSettings: Settings = {
openaiApiKey: "",
deepgramApiKey: "",
aiModel: "gpt-4",
aiUrl: "https://api.openai.com/v1",
customPrompt: "",
port: 3030,
dataDir: "default",
disableAudio: false,
ignoredWindows: [],
includedWindows: [],
aiProviderType: "openai",
embeddedLLM: {
enabled: false,
model: "llama3.2:1b-instruct-q4_K_M",
port: 11438,
},
enableFrameCache: true,
enableUiMonitoring: false,
aiMaxContextChars: 128000,
user: {
token: "",
},
analyticsEnabled: true,
};
const defaultSettings = getDefaultSettings();

const [settings, setSettings] = useState<Settings | null>(null);
const [loading, setLoading] = useState(true);
Expand Down
28 changes: 2 additions & 26 deletions pipes/reddit-auto-posts/lib/hooks/use-settings.ts
Original file line number Diff line number Diff line change
@@ -1,33 +1,10 @@
"use client";

import { useState, useEffect } from "react";
import type { Settings } from "@screenpipe/browser";
import { getDefaultSettings, type Settings } from "@screenpipe/browser";

export function useSettings() {
const defaultSettings: Settings = {
openaiApiKey: "",
deepgramApiKey: "",
aiModel: "gpt-4",
aiUrl: "https://api.openai.com/v1",
customPrompt: "",
port: 3030,
dataDir: "default",
disableAudio: false,
ignoredWindows: [],
includedWindows: [],
aiProviderType: "openai",
embeddedLLM: {
enabled: false,
model: "llama3.2:1b-instruct-q4_K_M",
port: 11438,
},
enableFrameCache: true,
enableUiMonitoring: false,
aiMaxContextChars: 128000,
user: {
token: "",
},
};
const defaultSettings = getDefaultSettings();

const [settings, setSettings] = useState<Settings | null>(null);
const [loading, setLoading] = useState(true);
Expand Down Expand Up @@ -204,4 +181,3 @@ export function useSettings() {
resetSettings,
};
}

29 changes: 2 additions & 27 deletions pipes/search/src/lib/hooks/use-settings.tsx
Original file line number Diff line number Diff line change
@@ -1,35 +1,10 @@
"use client";

import { useState, useEffect } from "react";
import type { Settings } from "@screenpipe/browser";
import { getDefaultSettings, type Settings } from "@screenpipe/browser";

export function useSettings() {
const defaultSettings: Settings = {
analyticsEnabled: true,
openaiApiKey: "",
deepgramApiKey: "",
aiModel: "gpt-4",
aiUrl: "https://api.openai.com/v1",
customPrompt: "",
port: 3030,
dataDir: "default",
disableAudio: false,
ignoredWindows: [],
includedWindows: [],
aiProviderType: "openai",
embeddedLLM: {
enabled: false,
model: "llama3.2:1b-instruct-q4_K_M",
port: 11438,
},
enableFrameCache: true,
enableUiMonitoring: false,
aiMaxContextChars: 128000,
user: {
token: "",
},
};

const defaultSettings = getDefaultSettings();
const [settings, setSettings] = useState<Settings>(defaultSettings);
const [loading, setLoading] = useState(false);
const [error, setError] = useState<Error | null>(null);
Expand Down

0 comments on commit 487aa90

Please sign in to comment.