|
|
|
@ -1,50 +1,46 @@ |
|
|
|
|
import { ICompletions, IParams, AIRequestType, RemoteBackendOPModel } from "../../types/types"; |
|
|
|
|
import { ICompletions, IParams, AIRequestType, RemoteBackendOPModel, JsonStreamParser } from "../../types/types"; |
|
|
|
|
import { GenerationParams, CompletionParams, InsertionParams } from "../../types/models"; |
|
|
|
|
import { buildSolgptPromt } from "../../prompts/promptBuilder"; |
|
|
|
|
import axios from "axios"; |
|
|
|
|
import EventEmitter from "events"; |
|
|
|
|
import { ChatHistory } from "../../prompts/chat"; |
|
|
|
|
import axios from 'axios'; |
|
|
|
|
|
|
|
|
|
const defaultErrorMessage = `Unable to get a response from AI server` |
|
|
|
|
|
|
|
|
|
export class RemoteInferencer implements ICompletions { |
|
|
|
|
api_url: string |
|
|
|
|
completion_url: string |
|
|
|
|
max_history = 7 |
|
|
|
|
model_op = RemoteBackendOPModel.CODELLAMA // default model operation change this to llama if necessary
|
|
|
|
|
event: EventEmitter |
|
|
|
|
test_env=false |
|
|
|
|
test_url="http://solcodertest.org" |
|
|
|
|
|
|
|
|
|
constructor(apiUrl?:string, completionUrl?:string) { |
|
|
|
|
this.api_url = apiUrl!==undefined ? apiUrl: "https://solcoder.remixproject.org" |
|
|
|
|
this.completion_url = completionUrl!==undefined ? completionUrl : "https://completion.remixproject.org" |
|
|
|
|
this.api_url = apiUrl!==undefined ? apiUrl: this.test_env? this.test_url : "https://solcoder.remixproject.org" |
|
|
|
|
this.completion_url = completionUrl!==undefined ? completionUrl : this.test_env? this.test_url : "https://completion.remixproject.org" |
|
|
|
|
this.event = new EventEmitter() |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private async _makeRequest(data, rType:AIRequestType){ |
|
|
|
|
private async _makeRequest(payload, rType:AIRequestType){ |
|
|
|
|
this.event.emit("onInference") |
|
|
|
|
const requesURL = rType === AIRequestType.COMPLETION ? this.completion_url : this.api_url |
|
|
|
|
const userPrompt = data.data[0] |
|
|
|
|
const requestURL = rType === AIRequestType.COMPLETION ? this.completion_url : this.api_url |
|
|
|
|
|
|
|
|
|
try { |
|
|
|
|
const result = await axios(requesURL, { |
|
|
|
|
method: 'POST', |
|
|
|
|
headers: { |
|
|
|
|
Accept: 'application/json', |
|
|
|
|
'Content-Type': 'application/json', |
|
|
|
|
}, |
|
|
|
|
data: JSON.stringify(data), |
|
|
|
|
}) |
|
|
|
|
const options = { headers: { 'Content-Type': 'application/json', } } |
|
|
|
|
const result = await axios.post(`${requestURL}`, payload, options) |
|
|
|
|
|
|
|
|
|
switch (rType) { |
|
|
|
|
case AIRequestType.COMPLETION: |
|
|
|
|
if (result.statusText === "OK") |
|
|
|
|
return result.data.data[0] |
|
|
|
|
return result.data.generatedText |
|
|
|
|
else { |
|
|
|
|
return defaultErrorMessage |
|
|
|
|
} |
|
|
|
|
case AIRequestType.GENERAL: |
|
|
|
|
if (result.statusText === "OK") { |
|
|
|
|
const resultText = result.data.data[0] |
|
|
|
|
ChatHistory.pushHistory(userPrompt, resultText) |
|
|
|
|
if (result.data?.error) return result.data?.error |
|
|
|
|
const resultText = result.data.generatedText |
|
|
|
|
ChatHistory.pushHistory(payload.prompt, resultText) |
|
|
|
|
return resultText |
|
|
|
|
} else { |
|
|
|
|
return defaultErrorMessage |
|
|
|
@ -54,46 +50,57 @@ export class RemoteInferencer implements ICompletions { |
|
|
|
|
} catch (e) { |
|
|
|
|
ChatHistory.clearHistory() |
|
|
|
|
console.error('Error making request to Inference server:', e.message) |
|
|
|
|
return e |
|
|
|
|
} |
|
|
|
|
finally { |
|
|
|
|
this.event.emit("onInferenceDone") |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private async _streamInferenceRequest(data, rType:AIRequestType){ |
|
|
|
|
private async _streamInferenceRequest(endpoint, payload, rType:AIRequestType){ |
|
|
|
|
let resultText = "" |
|
|
|
|
try { |
|
|
|
|
this.event.emit('onInference') |
|
|
|
|
const requesURL = rType === AIRequestType.COMPLETION ? this.completion_url : this.api_url |
|
|
|
|
const userPrompt = data.data[0] |
|
|
|
|
const response = await axios({ |
|
|
|
|
method: 'post', |
|
|
|
|
url: requesURL, |
|
|
|
|
data: data, |
|
|
|
|
headers: { 'Content-Type': 'application/json', "Accept": "text/event-stream" }, |
|
|
|
|
responseType: 'stream' |
|
|
|
|
const requestURL = rType === AIRequestType.COMPLETION ? this.completion_url : this.api_url |
|
|
|
|
const response = await fetch(requestURL, { |
|
|
|
|
method: 'POST', |
|
|
|
|
headers: { |
|
|
|
|
'Content-Type': 'application/json', |
|
|
|
|
}, |
|
|
|
|
body: JSON.stringify(payload), |
|
|
|
|
}); |
|
|
|
|
|
|
|
|
|
let resultText = "" |
|
|
|
|
response.data.on('data', (chunk: Buffer) => { |
|
|
|
|
if (payload.return_stream_response) { |
|
|
|
|
return response |
|
|
|
|
} |
|
|
|
|
const reader = response.body?.getReader(); |
|
|
|
|
const decoder = new TextDecoder(); |
|
|
|
|
const parser = new JsonStreamParser(); |
|
|
|
|
// eslint-disable-next-line no-constant-condition
|
|
|
|
|
while (true) { |
|
|
|
|
const { done, value } = await reader.read(); |
|
|
|
|
if (done) break; |
|
|
|
|
|
|
|
|
|
try { |
|
|
|
|
const parsedData = JSON.parse(chunk.toString()); |
|
|
|
|
if (parsedData.isGenerating) { |
|
|
|
|
this.event.emit('onStreamResult', parsedData.generatedText); |
|
|
|
|
resultText = resultText + parsedData.generatedText |
|
|
|
|
} else { |
|
|
|
|
// stream generation is complete
|
|
|
|
|
resultText = resultText + parsedData.generatedText |
|
|
|
|
ChatHistory.pushHistory(userPrompt, resultText) |
|
|
|
|
return parsedData.generatedText |
|
|
|
|
console.log("value" + decoder.decode(value)) |
|
|
|
|
const chunk = parser.safeJsonParse<{ generatedText: string; isGenerating: boolean }>(decoder.decode(value, { stream: true })); |
|
|
|
|
for (const parsedData of chunk) { |
|
|
|
|
if (parsedData.isGenerating) { |
|
|
|
|
this.event.emit('onStreamResult', parsedData.generatedText); |
|
|
|
|
resultText = resultText + parsedData.generatedText |
|
|
|
|
} else { |
|
|
|
|
// stream generation is complete
|
|
|
|
|
resultText = resultText + parsedData.generatedText |
|
|
|
|
ChatHistory.pushHistory(payload.prompt, resultText) |
|
|
|
|
return parsedData.generatedText |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} catch (error) { |
|
|
|
|
console.error('Error parsing JSON:', error); |
|
|
|
|
ChatHistory.clearHistory() |
|
|
|
|
} |
|
|
|
|
}); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
return "" // return empty string for now as handled in event
|
|
|
|
|
return resultText |
|
|
|
|
} catch (error) { |
|
|
|
|
ChatHistory.clearHistory() |
|
|
|
|
console.error('Error making stream request to Inference server:', error.message); |
|
|
|
@ -103,39 +110,38 @@ export class RemoteInferencer implements ICompletions { |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async code_completion(prompt, promptAfter, options:IParams=null): Promise<any> { |
|
|
|
|
const payload = !options? |
|
|
|
|
{ "data": [prompt, "code_completion", promptAfter, false, 30, 0.9, 0.90, 50]} : |
|
|
|
|
{ "data": [prompt, "code_completion", promptAfter, options.stream_result, |
|
|
|
|
options.max_new_tokens, options.temperature, options.top_p, options.top_k] |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async code_completion(prompt, promptAfter, options:IParams=CompletionParams): Promise<any> { |
|
|
|
|
const payload = { prompt, 'context':promptAfter, "endpoint":"code_completion", ...options } |
|
|
|
|
return this._makeRequest(payload, AIRequestType.COMPLETION) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async code_insertion(msg_pfx, msg_sfx): Promise<any> { |
|
|
|
|
const payload = { "data":[msg_pfx, "code_insertion", msg_sfx, 1024, 0.5, 0.92, 50]} |
|
|
|
|
async code_insertion(msg_pfx, msg_sfx, options:IParams=InsertionParams): Promise<any> { |
|
|
|
|
const payload = { "endpoint":"code_insertion", msg_pfx, msg_sfx, ...options, prompt: '' } |
|
|
|
|
return this._makeRequest(payload, AIRequestType.COMPLETION) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async code_generation(prompt): Promise<any> { |
|
|
|
|
const payload = { "data":[prompt, "code_completion", "", false,1000,0.9,0.92,50]} |
|
|
|
|
return this._makeRequest(payload, AIRequestType.COMPLETION) |
|
|
|
|
async code_generation(prompt, options:IParams=GenerationParams): Promise<any> { |
|
|
|
|
const payload = { prompt, "endpoint":"code_completion", ...options } |
|
|
|
|
if (options.stream_result) return this._streamInferenceRequest(payload.endpoint, payload, AIRequestType.COMPLETION) |
|
|
|
|
else return this._makeRequest(payload, AIRequestType.COMPLETION) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async solidity_answer(prompt): Promise<any> { |
|
|
|
|
async solidity_answer(prompt, options:IParams=GenerationParams): Promise<any> { |
|
|
|
|
const main_prompt = buildSolgptPromt(prompt, this.model_op) |
|
|
|
|
const payload = { "data":[main_prompt, "solidity_answer", false,2000,0.9,0.8,50]} |
|
|
|
|
return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
const payload = { 'prompt': main_prompt, "endpoint":"solidity_answer", ...options } |
|
|
|
|
if (options.stream_result) return this._streamInferenceRequest(payload.endpoint, payload, AIRequestType.GENERAL) |
|
|
|
|
else return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async code_explaining(prompt, context:string=""): Promise<any> { |
|
|
|
|
const payload = { "data":[prompt, "code_explaining", false,2000,0.9,0.8,50, context]} |
|
|
|
|
return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
async code_explaining(prompt, context:string="", options:IParams=GenerationParams): Promise<any> { |
|
|
|
|
const payload = { prompt, "endpoint":"code_explaining", context, ...options } |
|
|
|
|
if (options.stream_result) return this._streamInferenceRequest(payload.endpoint, payload, AIRequestType.GENERAL) |
|
|
|
|
else return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
async error_explaining(prompt): Promise<any> { |
|
|
|
|
const payload = { "data":[prompt, "error_explaining", false,2000,0.9,0.8,50]} |
|
|
|
|
return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
async error_explaining(prompt, options:IParams=GenerationParams): Promise<any> { |
|
|
|
|
const payload = { prompt, "endpoint":"error_explaining", ...options } |
|
|
|
|
if (options.stream_result) return this._streamInferenceRequest(payload.endpoint, payload, AIRequestType.GENERAL) |
|
|
|
|
else return this._makeRequest(payload, AIRequestType.GENERAL) |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|