- useToast.ts: global reactive toast singleton for cross-component toasts - App.vue: sticky demo mode banner + global toast slot - router: bypass wizard gate entirely in demo mode (pre-seeded data) - ApplyWorkspace, CompanyResearchModal: guard generate() in demo mode - fineTune store: guard submitJob() in demo mode - ui_switcher.py: remove Vue→Streamlit fallback in demo mode (now handled natively) All LLM-triggering actions show a toast and no-op in demo mode. Backend already blocks inference via DEMO_MODE env; Vue layer adds UX signal. Closes #46
89 lines
2.6 KiB
TypeScript
89 lines
2.6 KiB
TypeScript
import { ref } from 'vue'
|
|
import { defineStore } from 'pinia'
|
|
import { useApiFetch } from '../../composables/useApi'
|
|
import { useAppConfigStore } from '../appConfig'
|
|
import { showToast } from '../../composables/useToast'
|
|
|
|
export interface TrainingPair {
|
|
index: number
|
|
instruction: string
|
|
source_file: string
|
|
}
|
|
|
|
export const useFineTuneStore = defineStore('settings/fineTune', () => {
|
|
const step = ref(1)
|
|
const inFlightJob = ref(false)
|
|
const jobStatus = ref<string>('idle')
|
|
const pairsCount = ref(0)
|
|
const quotaRemaining = ref<number | null>(null)
|
|
const uploading = ref(false)
|
|
const loading = ref(false)
|
|
const pairs = ref<TrainingPair[]>([])
|
|
const pairsLoading = ref(false)
|
|
let _pollTimer: ReturnType<typeof setInterval> | null = null
|
|
|
|
function resetStep() { step.value = 1 }
|
|
|
|
async function loadStatus() {
|
|
const { data } = await useApiFetch<{ status: string; pairs_count: number; quota_remaining?: number }>('/api/settings/fine-tune/status')
|
|
if (!data) return
|
|
jobStatus.value = data.status
|
|
pairsCount.value = data.pairs_count ?? 0
|
|
quotaRemaining.value = data.quota_remaining ?? null
|
|
inFlightJob.value = ['queued', 'running'].includes(data.status)
|
|
}
|
|
|
|
function startPolling() {
|
|
loadStatus()
|
|
_pollTimer = setInterval(loadStatus, 2000)
|
|
}
|
|
|
|
function stopPolling() {
|
|
if (_pollTimer !== null) { clearInterval(_pollTimer); _pollTimer = null }
|
|
}
|
|
|
|
async function submitJob() {
|
|
if (useAppConfigStore().isDemo) { showToast('AI features are disabled in demo mode'); return }
|
|
const { data, error } = await useApiFetch<{ job_id: string }>('/api/settings/fine-tune/submit', { method: 'POST' })
|
|
if (!error && data) { inFlightJob.value = true; jobStatus.value = 'queued' }
|
|
}
|
|
|
|
async function loadPairs() {
|
|
pairsLoading.value = true
|
|
const { data } = await useApiFetch<{ pairs: TrainingPair[]; total: number }>('/api/settings/fine-tune/pairs')
|
|
pairsLoading.value = false
|
|
if (data) {
|
|
pairs.value = data.pairs
|
|
pairsCount.value = data.total
|
|
}
|
|
}
|
|
|
|
async function deletePair(index: number) {
|
|
const { data } = await useApiFetch<{ ok: boolean; remaining: number }>(
|
|
`/api/settings/fine-tune/pairs/${index}`, { method: 'DELETE' }
|
|
)
|
|
if (data?.ok) {
|
|
pairs.value = pairs.value.filter(p => p.index !== index).map((p, i) => ({ ...p, index: i }))
|
|
pairsCount.value = data.remaining
|
|
}
|
|
}
|
|
|
|
return {
|
|
step,
|
|
inFlightJob,
|
|
jobStatus,
|
|
pairsCount,
|
|
quotaRemaining,
|
|
uploading,
|
|
loading,
|
|
pairs,
|
|
pairsLoading,
|
|
resetStep,
|
|
loadStatus,
|
|
startPolling,
|
|
stopPolling,
|
|
submitJob,
|
|
loadPairs,
|
|
deletePair,
|
|
}
|
|
})
|