4
4
5
5
![ A llama riding a crab, AI-generated] ( ./doc/resources/logo2.png )
6
6
7
- > * Image by [ @darthdeus ] ( https://github.com/darthdeus/ ) , using Stable Diffusion *
7
+ > _ Image by [ @darthdeus ] ( https://github.com/darthdeus/ ) , using Stable Diffusion _
8
8
9
9
[ ![ ko-fi] ( https://ko-fi.com/img/githubbutton_sm.svg )] ( https://ko-fi.com/F1F8DNO5D )
10
10
11
11
[ ![ Latest version] ( https://img.shields.io/crates/v/llama-rs.svg )] ( https://crates.io/crates/llama_rs )
12
12
![ MIT] ( https://img.shields.io/badge/license-MIT-blue.svg )
13
+ [ ![ Discord] ( https://img.shields.io/discord/1085885067601137734 )] ( https://discord.gg/YB9WaXYAWU )
13
14
14
15
![ Gif showcasing language generation using llama-rs] ( ./doc/resources/llama_gif.gif )
15
16
@@ -43,7 +44,7 @@ Some additional things to try:
43
44
44
45
For example, you try the following prompt:
45
46
46
- ``` shell
47
+ ``` shell
47
48
cargo run --release -- -m /data/Llama/LLaMA/7B/ggml-model-q4_0.bin -p " Tell me how cool the Rust programming language is
48
49
` ` `
49
50
@@ -52,14 +53,14 @@ cargo run --release -- -m /data/Llama/LLaMA/7B/ggml-model-q4_0.bin -p "Tell me h
52
53
- ** Q: Why did you do this?**
53
54
- ** A:** It was not my choice. Ferris appeared to me in my dreams and asked me
54
55
to rewrite this in the name of the Holy crab.
55
-
56
+
56
57
- ** Q: Seriously now**
57
58
- ** A:** Come on! I don' t want to get into a flame war. You know how it goes,
58
- *something something* memory *something something* cargo is nice, don' t make
59
+ _something something_ memory _something something_ cargo is nice, don' t make
59
60
me say it, everybody knows this already.
60
61
61
62
- ** Q: I insist.**
62
- - ** A:** * Sheesh ! Okaaay * . After seeing the huge potential for ** llama.cpp** ,
63
+ - ** A:** _Sheesh ! Okaaay_ . After seeing the huge potential for ** llama.cpp** ,
63
64
the first thing I did was to see how hard would it be to turn it into a
64
65
library to embed in my projects. I started digging into the code, and realized
65
66
the heavy lifting is done by ` ggml` (a C library, easy to bind to Rust) and
@@ -69,9 +70,9 @@ cargo run --release -- -m /data/Llama/LLaMA/7B/ggml-model-q4_0.bin -p "Tell me h
69
70
I' m more comfortable.
70
71
71
72
- ** Q: Is this the real reason?**
72
- - ** A:** Haha. Of course * not * . I just like collecting imaginary internet
73
+ - ** A:** Haha. Of course _not_ . I just like collecting imaginary internet
73
74
points, in the form of little stars, that people seem to give to me whenever I
74
- embark on pointless quests for *rewriting X thing, but in Rust * .
75
+ embark on pointless quests for _rewriting X thing, but in Rust_ .
75
76
76
77
# # Known issues / To-dos
77
78
@@ -86,5 +87,5 @@ Contributions welcome! Here's a few pressing issues:
86
87
- [x] The code needs to be "library"-fied. It is nice as a showcase binary, but
87
88
the real potential for this tool is to allow embedding in other services.
88
89
- [x] The code only sets the right CFLAGS on Linux. The `build.rs` script in
89
- `ggml_raw` needs to be fixed, so inference *will be very slow on every
90
- other OS* .
90
+ `ggml_raw` needs to be fixed, so inference _will be very slow on every
91
+ other OS_ .
0 commit comments