This repository was archived by the owner on Jul 4, 2025. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 181
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
bug: cortex run llama3.1 is broken #1273
Copy link
Copy link
Closed
Labels
P1: importantImportant feature / fixImportant feature / fixcategory: model runningInference ux, handling context/parameters, runtimeInference ux, handling context/parameters, runtimetype: bugSomething isn't workingSomething isn't working
Milestone
Description
Cortex version
v70
Problem 1
cortex pull llama3.1-url, download is successfulcortex run llama3.1- Get interactive shell
- Send "hi"
- Get error
❯ cortex-nightly run llama3.1
Inorder to exit, type `exit()`
> hi
libc++abi: terminating due to uncaught exception of type nlohmann::json_abi_v3_11_3::detail::parse_error: [json.exception.parse_error.101] parse error at line 3, column 1: syntax error while parsing value - invalid literal; last read: '"chat.completion.chunk"}<U+000A><U+000A>d'; expected end of input
5![1] 93403 abort cortex-nightly run llama3.1This happened a few times
Problem 2
- Wait a few minutes
- Use the same command, get a different error this time. But model loads!
❯ sudo cortex-nightly run llama3.1
Starting server ...
20240919 15:41:56.580209 UTC 3788319 INFO Host: 127.0.0.1 Port: 3928
- main.cc:32
Error opening file: Permission denied
Server started
Error opening file: Permission denied
Model loaded!
Inorder to exit, type `exit()`
> hi
6!0!47C<%!97++2H5B/F#8(#BA:E+FGD%'E"<9$2-8$==EF>3#B4A5.B!A0>:19)8?<9F'9D8D/<E?
A:)242FB5!$>#42,;#!37EBCH";>E%2.3'0#;"#2#<6AC!0D?19:!:/;8+)9$:7=/7C%8?+7+)/G)E=G&,G'0"!.D?
+6BC&#*!8&$34"4&?99G=!!<;-=&18,<&'1(CG=?7--H4H49;+8<&,&B(??=$/96>(.F8*3.
<!33#3G.CB0"::$HE5,5"8$A(=F;H$A+C#9,2;HH?!:;3,3:0013H))3)//5&7+(36,>AH.!%G2%4+#%:1CD(?B742;<&(0F;2?
>9(F5,H3(4A?D-B#"H18'3F1<:EG.30548.#1A=).D,A*>!:9734FG22A(/(>)57%<E!$7":94#95=(#/0,(%$50?
E*"95"74#/B"076C.E-0<2!'$>%<3.?94A2)!!#372=>H28C2%:23(,(D/0;4H5,B?3B;HE9*Problem 3
- Exit interactive shell
- run
cortex stopjust to be sure - run
cortex run llama3.1again - computer crashes, needs a hard restart
What is your OS?
- MacOS
- Windows
- Linux
What engine are you running?
- cortex.llamacpp (default)
- cortex.tensorrt-llm (Nvidia GPUs)
- cortex.onnx (NPUs, DirectML)
Metadata
Metadata
Assignees
Labels
P1: importantImportant feature / fixImportant feature / fixcategory: model runningInference ux, handling context/parameters, runtimeInference ux, handling context/parameters, runtimetype: bugSomething isn't workingSomething isn't working