Activity
fix stability issue with request already being closed
fix stability issue with request already being closed
revert inference engine changes and restore sibling folder structure …
revert inference engine changes and restore sibling folder structure …
bump ggml (support for redpajama via gpt-neox & starcoder)
bump ggml (support for redpajama via gpt-neox & starcoder)
ggml routes now use inference engine. added scripts for inference eng…
ggml routes now use inference engine. added scripts for inference eng…
fix bug by renaming chatEngine to ChatEngine
fix bug by renaming chatEngine to ChatEngine
remove requirement to pass in ggml, smart detection via model path
remove requirement to pass in ggml, smart detection via model path
clean up completions ggml route
clean up completions ggml route
fix chat routes to use modeltype
fix chat routes to use modeltype
fix ggml route in chatcompletios and trim request to allow for better…
fix ggml route in chatcompletios and trim request to allow for better…
polish chatengine for redpajama
polish chatengine for redpajama
give more leeway for chat interactions
give more leeway for chat interactions
improve filtering initial prompt, actually use model paths from api k…
improve filtering initial prompt, actually use model paths from api k…
init ggml support (PLS DO NOT USE, VERY EXPERIMENTAL AND ROUGH AND NO…
init ggml support (PLS DO NOT USE, VERY EXPERIMENTAL AND ROUGH AND NO…
fix completion bug where streaming always cuts short
fix completion bug where streaming always cuts short
add back context default param. llama has a default of 2048 max conte…
add back context default param. llama has a default of 2048 max conte…
update stderr to only print when readstream closes unexpectely
update stderr to only print when readstream closes unexpectely
update chatroutes to also use prompt length to caluclate response sta…
update chatroutes to also use prompt length to caluclate response sta…
FIX BUG WHERE NULL MAX_TOKENS WILL NOT GENERATE A PROPER RESPONSE!
FIX BUG WHERE NULL MAX_TOKENS WILL NOT GENERATE A PROPER RESPONSE!