diff --git a/.gitignore b/.gitignore index 04916c9..9185a5d 100644 --- a/.gitignore +++ b/.gitignore @@ -13,6 +13,9 @@ dist-ssr *-dist *.local +# Exception: Include dist folder for workspace packages +!packages/*/dist + # Editor directories and files .vscode/* !.vscode/extensions.json diff --git a/packages/spark-tools/dist/hooks/index.d.ts b/packages/spark-tools/dist/hooks/index.d.ts new file mode 100644 index 0000000..3ee64e4 --- /dev/null +++ b/packages/spark-tools/dist/hooks/index.d.ts @@ -0,0 +1,12 @@ +/** + * useKV Hook - Persistent key-value storage with localStorage and window.spark.kv integration + * + * This hook provides persistent state management that syncs with localStorage + * and integrates with the Spark KV storage system if available. + * + * @param key - Storage key + * @param defaultValue - Default value if key doesn't exist + * @returns Tuple of [value, setValue, deleteValue] + */ +export declare function useKV(key: string, defaultValue: T): [T, (value: T | ((prev: T) => T)) => void, () => void]; +//# sourceMappingURL=index.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/hooks/index.d.ts.map b/packages/spark-tools/dist/hooks/index.d.ts.map new file mode 100644 index 0000000..221370c --- /dev/null +++ b/packages/spark-tools/dist/hooks/index.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/hooks/index.ts"],"names":[],"mappings":"AAEA;;;;;;;;;GASG;AACH,wBAAgB,KAAK,CAAC,CAAC,EACrB,GAAG,EAAE,MAAM,EACX,YAAY,EAAE,CAAC,GACd,CAAC,CAAC,EAAE,CAAC,KAAK,EAAE,CAAC,GAAG,CAAC,CAAC,IAAI,EAAE,CAAC,KAAK,CAAC,CAAC,KAAK,IAAI,EAAE,MAAM,IAAI,CAAC,CA8ExD"} \ No newline at end of file diff --git a/packages/spark-tools/dist/index.d.ts b/packages/spark-tools/dist/index.d.ts new file mode 100644 index 0000000..4d76510 --- /dev/null +++ b/packages/spark-tools/dist/index.d.ts @@ -0,0 +1,8 @@ +/** + * @github/spark - Main Hooks Entry Point + * + * This is the entry point for the hooks exports from the Spark package. + */ +export { useKV } from './hooks/index'; +export { sparkRuntime } from './lib/spark-runtime'; +//# sourceMappingURL=index.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/index.d.ts.map b/packages/spark-tools/dist/index.d.ts.map new file mode 100644 index 0000000..57dab23 --- /dev/null +++ b/packages/spark-tools/dist/index.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,EAAE,KAAK,EAAE,MAAM,eAAe,CAAA;AACrC,OAAO,EAAE,YAAY,EAAE,MAAM,qBAAqB,CAAA"} \ No newline at end of file diff --git a/packages/spark-tools/dist/index.js b/packages/spark-tools/dist/index.js new file mode 100644 index 0000000..ee56d88 --- /dev/null +++ b/packages/spark-tools/dist/index.js @@ -0,0 +1,2 @@ +import{useState as r,useCallback as e,useEffect as o}from"react";export{s as sparkRuntime}from"./spark-runtime-wNXbhm34.js";function t(t,n){const[s,a]=r(()=>{try{if("undefined"!=typeof window&&window.spark?.kv){const r=window.spark.kv.get(t);if(void 0!==r)return r}const r=localStorage.getItem(t);return r?JSON.parse(r):n}catch(r){return console.error("Error reading from storage:",r),n}}),i=e(r=>{try{a(e=>{const o="function"==typeof r?r(e):r;return localStorage.setItem(t,JSON.stringify(o)),"undefined"!=typeof window&&window.spark?.kv&&window.spark.kv.set(t,o),o})}catch(r){console.error("Error writing to storage:",r)}},[t]),w=e(()=>{try{localStorage.removeItem(t),"undefined"!=typeof window&&window.spark?.kv&&window.spark.kv.delete(t),a(n)}catch(r){console.error("Error deleting from storage:",r)}},[t,n]);return o(()=>{const r=r=>{if(r.key===t&&null!==r.newValue)try{a(JSON.parse(r.newValue))}catch(r){console.error("Error parsing storage event:",r)}};return window.addEventListener("storage",r),()=>window.removeEventListener("storage",r)},[t]),[s,i,w]}export{t as useKV}; +//# sourceMappingURL=index.js.map diff --git a/packages/spark-tools/dist/index.js.map b/packages/spark-tools/dist/index.js.map new file mode 100644 index 0000000..f5ddd95 --- /dev/null +++ b/packages/spark-tools/dist/index.js.map @@ -0,0 +1 @@ +{"version":3,"file":"index.js","sources":["../src/hooks/index.ts"],"sourcesContent":["import { useState, useEffect, useCallback } from 'react'\n\n/**\n * useKV Hook - Persistent key-value storage with localStorage and window.spark.kv integration\n * \n * This hook provides persistent state management that syncs with localStorage\n * and integrates with the Spark KV storage system if available.\n * \n * @param key - Storage key\n * @param defaultValue - Default value if key doesn't exist\n * @returns Tuple of [value, setValue, deleteValue]\n */\nexport function useKV(\n key: string,\n defaultValue: T\n): [T, (value: T | ((prev: T) => T)) => void, () => void] {\n // Initialize state from localStorage or default value\n const [value, setValueInternal] = useState(() => {\n try {\n // Try to get from window.spark.kv first\n if (typeof window !== 'undefined' && window.spark?.kv) {\n const sparkValue = window.spark.kv.get(key)\n if (sparkValue !== undefined) {\n return sparkValue as T\n }\n }\n\n // Fallback to localStorage\n const item = localStorage.getItem(key)\n return item ? JSON.parse(item) : defaultValue\n } catch (error) {\n console.error('Error reading from storage:', error)\n return defaultValue\n }\n })\n\n // Set value and sync to storage\n const setValue = useCallback(\n (newValue: T | ((prev: T) => T)) => {\n try {\n setValueInternal((prevValue) => {\n const valueToStore =\n typeof newValue === 'function'\n ? (newValue as (prev: T) => T)(prevValue)\n : newValue\n\n // Store in localStorage\n localStorage.setItem(key, JSON.stringify(valueToStore))\n\n // Store in window.spark.kv if available\n if (typeof window !== 'undefined' && window.spark?.kv) {\n window.spark.kv.set(key, valueToStore)\n }\n\n return valueToStore\n })\n } catch (error) {\n console.error('Error writing to storage:', error)\n }\n },\n [key]\n )\n\n // Delete value from storage\n const deleteValue = useCallback(() => {\n try {\n localStorage.removeItem(key)\n if (typeof window !== 'undefined' && window.spark?.kv) {\n window.spark.kv.delete(key)\n }\n setValueInternal(defaultValue)\n } catch (error) {\n console.error('Error deleting from storage:', error)\n }\n }, [key, defaultValue])\n\n // Sync with localStorage changes from other tabs\n useEffect(() => {\n const handleStorageChange = (e: StorageEvent) => {\n if (e.key === key && e.newValue !== null) {\n try {\n setValueInternal(JSON.parse(e.newValue))\n } catch (error) {\n console.error('Error parsing storage event:', error)\n }\n }\n }\n\n window.addEventListener('storage', handleStorageChange)\n return () => window.removeEventListener('storage', handleStorageChange)\n }, [key])\n\n return [value, setValue, deleteValue]\n}\n"],"names":["useKV","key","defaultValue","value","setValueInternal","useState","window","spark","kv","sparkValue","get","undefined","item","localStorage","getItem","JSON","parse","error","console","setValue","useCallback","newValue","prevValue","valueToStore","setItem","stringify","set","deleteValue","removeItem","delete","useEffect","handleStorageChange","e","addEventListener","removeEventListener"],"mappings":"4HAYM,SAAUA,EACdC,EACAC,GAGA,MAAOC,EAAOC,GAAoBC,EAAY,KAC5C,IAEE,GAAsB,oBAAXC,QAA0BA,OAAOC,OAAOC,GAAI,CACrD,MAAMC,EAAaH,OAAOC,MAAMC,GAAGE,IAAIT,GACvC,QAAmBU,IAAfF,EACF,OAAOA,CAEX,CAGA,MAAMG,EAAOC,aAAaC,QAAQb,GAClC,OAAOW,EAAOG,KAAKC,MAAMJ,GAAQV,CACnC,CAAE,MAAOe,GAEP,OADAC,QAAQD,MAAM,8BAA+BA,GACtCf,CACT,IAIIiB,EAAWC,EACdC,IACC,IACEjB,EAAkBkB,IAChB,MAAMC,EACgB,mBAAbF,EACFA,EAA4BC,GAC7BD,EAUN,OAPAR,aAAaW,QAAQvB,EAAKc,KAAKU,UAAUF,IAGnB,oBAAXjB,QAA0BA,OAAOC,OAAOC,IACjDF,OAAOC,MAAMC,GAAGkB,IAAIzB,EAAKsB,GAGpBA,GAEX,CAAE,MAAON,GACPC,QAAQD,MAAM,4BAA6BA,EAC7C,GAEF,CAAChB,IAIG0B,EAAcP,EAAY,KAC9B,IACEP,aAAae,WAAW3B,GACF,oBAAXK,QAA0BA,OAAOC,OAAOC,IACjDF,OAAOC,MAAMC,GAAGqB,OAAO5B,GAEzBG,EAAiBF,EACnB,CAAE,MAAOe,GACPC,QAAQD,MAAM,+BAAgCA,EAChD,GACC,CAAChB,EAAKC,IAkBT,OAfA4B,EAAU,KACR,MAAMC,EAAuBC,IAC3B,GAAIA,EAAE/B,MAAQA,GAAsB,OAAf+B,EAAEX,SACrB,IACEjB,EAAiBW,KAAKC,MAAMgB,EAAEX,UAChC,CAAE,MAAOJ,GACPC,QAAQD,MAAM,+BAAgCA,EAChD,GAKJ,OADAX,OAAO2B,iBAAiB,UAAWF,GAC5B,IAAMzB,OAAO4B,oBAAoB,UAAWH,IAClD,CAAC9B,IAEG,CAACE,EAAOgB,EAAUQ,EAC3B"} \ No newline at end of file diff --git a/packages/spark-tools/dist/lib/spark-runtime.d.ts b/packages/spark-tools/dist/lib/spark-runtime.d.ts new file mode 100644 index 0000000..3eff122 --- /dev/null +++ b/packages/spark-tools/dist/lib/spark-runtime.d.ts @@ -0,0 +1,34 @@ +/** + * Spark Runtime - Core runtime services for Spark applications + * + * This module provides mock implementations of Spark services including: + * - KV storage (key-value store) + * - LLM service (language model integration) + * - User authentication + */ +export declare const sparkRuntime: { + kv: { + get: (key: string) => T | undefined; + set: (key: string, value: any) => void; + delete: (key: string) => void; + clear: () => void; + keys: () => string[]; + }; + llm: { + (prompt: string, model?: string, jsonMode?: boolean): Promise; + chat(messages: any[]): Promise<{ + role: string; + content: string; + }>; + complete(prompt: string): Promise; + }; + user: { + getCurrentUser: () => { + id: string; + name: string; + email: string; + }; + isAuthenticated: () => boolean; + }; +}; +//# sourceMappingURL=spark-runtime.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/lib/spark-runtime.d.ts.map b/packages/spark-tools/dist/lib/spark-runtime.d.ts.map new file mode 100644 index 0000000..bb13e5e --- /dev/null +++ b/packages/spark-tools/dist/lib/spark-runtime.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"spark-runtime.d.ts","sourceRoot":"","sources":["../../src/lib/spark-runtime.ts"],"names":[],"mappings":"AAAA;;;;;;;GAOG;AAwBH,eAAO,MAAM,YAAY;;cAEf,CAAC,aAAa,MAAM,KAAG,CAAC,GAAG,SAAS;mBAa/B,MAAM,SAAS,GAAG;sBAQf,MAAM;;;;;iBAzCW,MAAM,UAAU,MAAM,aAAa,OAAO,GAAG,OAAO,CAAC,GAAG,CAAC;uBAKxD,GAAG,EAAE;;;;yBAQH,MAAM;;;;;;;;;;CA4D3C,CAAA"} \ No newline at end of file diff --git a/packages/spark-tools/dist/lib/spark.d.ts b/packages/spark-tools/dist/lib/spark.d.ts new file mode 100644 index 0000000..5a2ff20 --- /dev/null +++ b/packages/spark-tools/dist/lib/spark.d.ts @@ -0,0 +1,14 @@ +/** + * Spark Initialization Module + * + * This module initializes the Spark runtime and makes it available globally + * via window.spark. It should be imported early in the application lifecycle. + */ +import { sparkRuntime } from './spark-runtime'; +declare global { + interface Window { + spark: typeof sparkRuntime; + } +} +export default sparkRuntime; +//# sourceMappingURL=spark.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/lib/spark.d.ts.map b/packages/spark-tools/dist/lib/spark.d.ts.map new file mode 100644 index 0000000..8c72cbb --- /dev/null +++ b/packages/spark-tools/dist/lib/spark.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"spark.d.ts","sourceRoot":"","sources":["../../src/lib/spark.ts"],"names":[],"mappings":"AAAA;;;;;GAKG;AAEH,OAAO,EAAE,YAAY,EAAE,MAAM,iBAAiB,CAAA;AAG9C,OAAO,CAAC,MAAM,CAAC;IACb,UAAU,MAAM;QACd,KAAK,EAAE,OAAO,YAAY,CAAA;KAC3B;CACF;AAOD,eAAe,YAAY,CAAA"} \ No newline at end of file diff --git a/packages/spark-tools/dist/spark-runtime-wNXbhm34.js b/packages/spark-tools/dist/spark-runtime-wNXbhm34.js new file mode 100644 index 0000000..2a37e36 --- /dev/null +++ b/packages/spark-tools/dist/spark-runtime-wNXbhm34.js @@ -0,0 +1,2 @@ +const e=new Map,o=async(e,o,r)=>(console.log("Mock LLM called with prompt:",e,"model:",o,"jsonMode:",r),"This is a mock response from the Spark LLM service.");o.chat=async e=>(console.log("Mock LLM chat called with messages:",e),{role:"assistant",content:"This is a mock response from the Spark LLM service."}),o.complete=async e=>(console.log("Mock LLM complete called with prompt:",e),"This is a mock completion from the Spark LLM service.");const r={kv:{get:o=>{try{const r=e.get(o);if(void 0!==r)return r;const t=localStorage.getItem(o);return t?JSON.parse(t):void 0}catch(e){return void console.error("Error getting KV value:",e)}},set:(o,r)=>{try{e.set(o,r),localStorage.setItem(o,JSON.stringify(r))}catch(e){console.error("Error setting KV value:",e)}},delete:o=>{try{e.delete(o),localStorage.removeItem(o)}catch(e){console.error("Error deleting KV value:",e)}},clear:()=>{try{const o=Array.from(e.keys());e.clear(),o.forEach(e=>localStorage.removeItem(e))}catch(e){console.error("Error clearing KV storage:",e)}},keys:()=>Array.from(e.keys())},llm:o,user:{getCurrentUser:()=>({id:"mock-user-id",name:"Mock User",email:"mock@example.com"}),isAuthenticated:()=>!0}};export{r as s}; +//# sourceMappingURL=spark-runtime-wNXbhm34.js.map diff --git a/packages/spark-tools/dist/spark-runtime-wNXbhm34.js.map b/packages/spark-tools/dist/spark-runtime-wNXbhm34.js.map new file mode 100644 index 0000000..030ca20 --- /dev/null +++ b/packages/spark-tools/dist/spark-runtime-wNXbhm34.js.map @@ -0,0 +1 @@ +{"version":3,"file":"spark-runtime-wNXbhm34.js","sources":["../src/lib/spark-runtime.ts"],"sourcesContent":["/**\n * Spark Runtime - Core runtime services for Spark applications\n * \n * This module provides mock implementations of Spark services including:\n * - KV storage (key-value store)\n * - LLM service (language model integration)\n * - User authentication\n */\n\n// Mock KV Storage\nconst kvStorage = new Map()\n\n// Create llm function with additional properties\nconst llmFunction = async (prompt: string, model?: string, jsonMode?: boolean): Promise => {\n console.log('Mock LLM called with prompt:', prompt, 'model:', model, 'jsonMode:', jsonMode)\n return 'This is a mock response from the Spark LLM service.'\n}\n\nllmFunction.chat = async (messages: any[]) => {\n console.log('Mock LLM chat called with messages:', messages)\n return {\n role: 'assistant',\n content: 'This is a mock response from the Spark LLM service.'\n }\n}\n\nllmFunction.complete = async (prompt: string) => {\n console.log('Mock LLM complete called with prompt:', prompt)\n return 'This is a mock completion from the Spark LLM service.'\n}\n\nexport const sparkRuntime = {\n kv: {\n get: (key: string): T | undefined => {\n try {\n const value = kvStorage.get(key)\n if (value !== undefined) {\n return value as T\n }\n const stored = localStorage.getItem(key)\n return stored ? JSON.parse(stored) : undefined\n } catch (error) {\n console.error('Error getting KV value:', error)\n return undefined\n }\n },\n set: (key: string, value: any) => {\n try {\n kvStorage.set(key, value)\n localStorage.setItem(key, JSON.stringify(value))\n } catch (error) {\n console.error('Error setting KV value:', error)\n }\n },\n delete: (key: string) => {\n try {\n kvStorage.delete(key)\n localStorage.removeItem(key)\n } catch (error) {\n console.error('Error deleting KV value:', error)\n }\n },\n clear: () => {\n try {\n // Get keys before clearing\n const keysToRemove = Array.from(kvStorage.keys())\n kvStorage.clear()\n // Clear corresponding keys from localStorage\n keysToRemove.forEach(key => localStorage.removeItem(key))\n } catch (error) {\n console.error('Error clearing KV storage:', error)\n }\n },\n keys: () => Array.from(kvStorage.keys())\n },\n \n llm: llmFunction,\n \n user: {\n getCurrentUser: () => ({\n id: 'mock-user-id',\n name: 'Mock User',\n email: 'mock@example.com'\n }),\n isAuthenticated: () => true\n }\n}\n"],"names":["kvStorage","Map","llmFunction","async","prompt","model","jsonMode","console","log","chat","messages","role","content","complete","sparkRuntime","kv","get","key","value","undefined","stored","localStorage","getItem","JSON","parse","error","set","setItem","stringify","delete","removeItem","clear","keysToRemove","Array","from","keys","forEach","llm","user","getCurrentUser","id","name","email","isAuthenticated"],"mappings":"AAUA,MAAMA,EAAY,IAAIC,IAGhBC,EAAcC,MAAOC,EAAgBC,EAAgBC,KACzDC,QAAQC,IAAI,+BAAgCJ,EAAQ,SAAUC,EAAO,YAAaC,GAC3E,uDAGTJ,EAAYO,KAAON,MAAOO,IACxBH,QAAQC,IAAI,sCAAuCE,GAC5C,CACLC,KAAM,YACNC,QAAS,wDAIbV,EAAYW,SAAWV,MAAOC,IAC5BG,QAAQC,IAAI,wCAAyCJ,GAC9C,yDAGF,MAAMU,EAAe,CAC1BC,GAAI,CACFC,IAAeC,IACb,IACE,MAAMC,EAAQlB,EAAUgB,IAAIC,GAC5B,QAAcE,IAAVD,EACF,OAAOA,EAET,MAAME,EAASC,aAAaC,QAAQL,GACpC,OAAOG,EAASG,KAAKC,MAAMJ,QAAUD,CACvC,CAAE,MAAOM,GAEP,YADAlB,QAAQkB,MAAM,0BAA2BA,EAE3C,GAEFC,IAAK,CAACT,EAAaC,KACjB,IACElB,EAAU0B,IAAIT,EAAKC,GACnBG,aAAaM,QAAQV,EAAKM,KAAKK,UAAUV,GAC3C,CAAE,MAAOO,GACPlB,QAAQkB,MAAM,0BAA2BA,EAC3C,GAEFI,OAASZ,IACP,IACEjB,EAAU6B,OAAOZ,GACjBI,aAAaS,WAAWb,EAC1B,CAAE,MAAOQ,GACPlB,QAAQkB,MAAM,2BAA4BA,EAC5C,GAEFM,MAAO,KACL,IAEE,MAAMC,EAAeC,MAAMC,KAAKlC,EAAUmC,QAC1CnC,EAAU+B,QAEVC,EAAaI,QAAQnB,GAAOI,aAAaS,WAAWb,GACtD,CAAE,MAAOQ,GACPlB,QAAQkB,MAAM,6BAA8BA,EAC9C,GAEFU,KAAM,IAAMF,MAAMC,KAAKlC,EAAUmC,SAGnCE,IAAKnC,EAELoC,KAAM,CACJC,eAAgB,KAAA,CACdC,GAAI,eACJC,KAAM,YACNC,MAAO,qBAETC,gBAAiB,KAAM"} \ No newline at end of file diff --git a/packages/spark-tools/dist/spark.d.ts b/packages/spark-tools/dist/spark.d.ts new file mode 100644 index 0000000..c7f351d --- /dev/null +++ b/packages/spark-tools/dist/spark.d.ts @@ -0,0 +1,8 @@ +/** + * Spark Module - Main export for spark runtime + * + * Re-export spark runtime for '@github/spark/spark' imports + */ +export { default } from './lib/spark'; +export { sparkRuntime } from './lib/spark-runtime'; +//# sourceMappingURL=spark.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/spark.d.ts.map b/packages/spark-tools/dist/spark.d.ts.map new file mode 100644 index 0000000..ce69f60 --- /dev/null +++ b/packages/spark-tools/dist/spark.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"spark.d.ts","sourceRoot":"","sources":["../src/spark.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,EAAE,OAAO,EAAE,MAAM,aAAa,CAAA;AACrC,OAAO,EAAE,YAAY,EAAE,MAAM,qBAAqB,CAAA"} \ No newline at end of file diff --git a/packages/spark-tools/dist/spark.js b/packages/spark-tools/dist/spark.js new file mode 100644 index 0000000..e0e5077 --- /dev/null +++ b/packages/spark-tools/dist/spark.js @@ -0,0 +1,2 @@ +import{s as e}from"./spark-runtime-wNXbhm34.js";"undefined"!=typeof window&&(window.spark=e);export{e as default,e as sparkRuntime}; +//# sourceMappingURL=spark.js.map diff --git a/packages/spark-tools/dist/spark.js.map b/packages/spark-tools/dist/spark.js.map new file mode 100644 index 0000000..d6190bb --- /dev/null +++ b/packages/spark-tools/dist/spark.js.map @@ -0,0 +1 @@ +{"version":3,"file":"spark.js","sources":["../src/lib/spark.ts"],"sourcesContent":["/**\n * Spark Initialization Module\n * \n * This module initializes the Spark runtime and makes it available globally\n * via window.spark. It should be imported early in the application lifecycle.\n */\n\nimport { sparkRuntime } from './spark-runtime'\n\n// Declare global window.spark\ndeclare global {\n interface Window {\n spark: typeof sparkRuntime\n }\n}\n\n// Initialize window.spark\nif (typeof window !== 'undefined') {\n window.spark = sparkRuntime\n}\n\nexport default sparkRuntime\n"],"names":["window","spark","sparkRuntime"],"mappings":"gDAiBsB,oBAAXA,SACTA,OAAOC,MAAQC"} \ No newline at end of file diff --git a/packages/spark-tools/dist/sparkVitePlugin.d.ts b/packages/spark-tools/dist/sparkVitePlugin.d.ts new file mode 100644 index 0000000..3932a91 --- /dev/null +++ b/packages/spark-tools/dist/sparkVitePlugin.d.ts @@ -0,0 +1,12 @@ +/** + * Spark Vite Plugin + * + * This plugin integrates Spark functionality into the Vite build process. + * It handles initialization and configuration of Spark features. + */ +export default function sparkPlugin(): { + name: string; + configResolved(config: any): void; + transformIndexHtml(html: string): string; +}; +//# sourceMappingURL=sparkVitePlugin.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/sparkVitePlugin.d.ts.map b/packages/spark-tools/dist/sparkVitePlugin.d.ts.map new file mode 100644 index 0000000..75ee5ea --- /dev/null +++ b/packages/spark-tools/dist/sparkVitePlugin.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"sparkVitePlugin.d.ts","sourceRoot":"","sources":["../src/sparkVitePlugin.ts"],"names":[],"mappings":"AAAA;;;;;GAKG;AAEH,MAAM,CAAC,OAAO,UAAU,WAAW;;2BAIR,GAAG;6BAID,MAAM;EAKlC"} \ No newline at end of file diff --git a/packages/spark-tools/dist/sparkVitePlugin.js b/packages/spark-tools/dist/sparkVitePlugin.js new file mode 100644 index 0000000..2c0ce9d --- /dev/null +++ b/packages/spark-tools/dist/sparkVitePlugin.js @@ -0,0 +1,2 @@ +function e(){return{name:"spark-vite-plugin",configResolved(e){},transformIndexHtml:e=>e}}export{e as default}; +//# sourceMappingURL=sparkVitePlugin.js.map diff --git a/packages/spark-tools/dist/sparkVitePlugin.js.map b/packages/spark-tools/dist/sparkVitePlugin.js.map new file mode 100644 index 0000000..bf9fcf6 --- /dev/null +++ b/packages/spark-tools/dist/sparkVitePlugin.js.map @@ -0,0 +1 @@ +{"version":3,"file":"sparkVitePlugin.js","sources":["../src/sparkVitePlugin.ts"],"sourcesContent":["/**\n * Spark Vite Plugin\n * \n * This plugin integrates Spark functionality into the Vite build process.\n * It handles initialization and configuration of Spark features.\n */\n\nexport default function sparkPlugin() {\n return {\n name: 'spark-vite-plugin',\n \n configResolved(config: any) {\n // Plugin configuration\n },\n \n transformIndexHtml(html: string) {\n // Inject Spark initialization if needed\n return html\n }\n }\n}\n"],"names":["sparkPlugin","name","configResolved","config","transformIndexHtml","html"],"mappings":"AAOc,SAAUA,IACtB,MAAO,CACLC,KAAM,oBAEN,cAAAC,CAAeC,GAEf,EAEAC,mBAAmBC,GAEVA,EAGb"} \ No newline at end of file diff --git a/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts new file mode 100644 index 0000000..daa89ea --- /dev/null +++ b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts @@ -0,0 +1,12 @@ +/** + * Vite Phosphor Icon Proxy Plugin + * + * This plugin provides a proxy for Phosphor icon imports to improve + * build performance and bundle size. + */ +export default function createIconImportProxy(): { + name: string; + resolveId(id: string): null | undefined; + transform(code: string, id: string): null; +}; +//# sourceMappingURL=vitePhosphorIconProxyPlugin.d.ts.map \ No newline at end of file diff --git a/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts.map b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts.map new file mode 100644 index 0000000..69de9dd --- /dev/null +++ b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.d.ts.map @@ -0,0 +1 @@ +{"version":3,"file":"vitePhosphorIconProxyPlugin.d.ts","sourceRoot":"","sources":["../src/vitePhosphorIconProxyPlugin.ts"],"names":[],"mappings":"AAAA;;;;;GAKG;AAEH,MAAM,CAAC,OAAO,UAAU,qBAAqB;;kBAI3B,MAAM;oBAOJ,MAAM,MAAM,MAAM;EAKrC"} \ No newline at end of file diff --git a/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js new file mode 100644 index 0000000..95cf7f2 --- /dev/null +++ b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js @@ -0,0 +1,2 @@ +function r(){return{name:"vite-phosphor-icon-proxy",resolveId(r){if(r.includes("@phosphor-icons/react"))return null},transform:(r,n)=>null}}export{r as default}; +//# sourceMappingURL=vitePhosphorIconProxyPlugin.js.map diff --git a/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js.map b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js.map new file mode 100644 index 0000000..6bc57bd --- /dev/null +++ b/packages/spark-tools/dist/vitePhosphorIconProxyPlugin.js.map @@ -0,0 +1 @@ +{"version":3,"file":"vitePhosphorIconProxyPlugin.js","sources":["../src/vitePhosphorIconProxyPlugin.ts"],"sourcesContent":["/**\n * Vite Phosphor Icon Proxy Plugin\n * \n * This plugin provides a proxy for Phosphor icon imports to improve\n * build performance and bundle size.\n */\n\nexport default function createIconImportProxy() {\n return {\n name: 'vite-phosphor-icon-proxy',\n \n resolveId(id: string) {\n // Handle icon imports\n if (id.includes('@phosphor-icons/react')) {\n return null // Let Vite handle it normally\n }\n },\n \n transform(code: string, id: string) {\n // Transform icon imports if needed\n return null\n }\n }\n}\n"],"names":["createIconImportProxy","name","resolveId","id","includes","transform","code"],"mappings":"AAOc,SAAUA,IACtB,MAAO,CACLC,KAAM,2BAEN,SAAAC,CAAUC,GAER,GAAIA,EAAGC,SAAS,yBACd,OAAO,IAEX,EAEAC,UAAS,CAACC,EAAcH,IAEf,KAGb"} \ No newline at end of file