-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello author, I am glad that you have made contributions to Jetson users. I have a question I would like to ask #1
Comments
My l4t version is 36.3.0 |
@cplasfwst The container |
@cplasfwst Also for posterity you should run |
When you are building Ollama, the sub-task to build an external library called llama_cpp is unable to properly import libcudart.so from your installed CUDA libraries. Please run the two commands I referenced in the quoted text so we can see if the libcudart.so file is present or not. |
I am not with my computer, I am using my cell phone to type to you, sorry. Please try this command, it is similar but checks a different directory: |
@cplasfwst This command will take longer to run but will scan your entire file system for the library in case I tried searching the wrong one previously.
if that returns nothing, try this (NEW EDIT: CUDA instead of cudart for apt list) if that still finds nothing, please check that you installed Jetpack correctly to your Orin AGX. |
I’ve updated the docker file, it looks like Dusty-NV updated the llama cpp container tag to just be r36.2.0. Can you also run |
Thank you for running the search. I was able to find some suggestions on fixing the problem. It seems like the compiler can’t find Please run |
I have added the path |
The tar file isn’t unzipping correctly, and your image cuts off the error message where the real issue is. I need more of the log. |
Also congratulations on your new child, you should be with your family :) |
The curl command doesn't work for me. Where should I change it to wget? I can get the file using wget |
Since your network is operable now, try the docker build again |
Sorry, I just went out to eat and just came back. I haven't tried this yet. |
I haven't tried this yet, I'm a newbie with a jetson and don't really understand what this is T.T |
A Jetson is just a computer with Linux on it. Its software is sometimes a little different but it’s mostly the same as a regular Linux machine. Here is some recommended setup for Jetsons created by dusty-nv. He is an NVidia engineer who maintains a container repository for Jetson devices. https://github.com/dusty-nv/jetson-containers/blob/master/docs/setup.md |
Thank you for your guidance, looking forward to your good news |
@cplasfwst I originally made this repository to get Ollama working on my Jetson, but when I completed my work I pushed this work to the dusty-nv jetson-containers repository and he integrated it into his project. The Jetson-Containers repository is the most up-to-date repository for Jetson containers and the best place to get updates to running AI on Jetson. Can you explain what your purpose for building the Ollama container is? Do you just want to run Ollama on your Jetson for fun or are you trying to build a custom Ollama binary? I was able to follow the guide from dusty-nv here and run Ollama on my Jetson r36.3.0 with |
After reviewing dusty-nv's container build process, in r36.x.x he employs a multi-stage build process that installs updates from an nvidia repository. I am not currently using this process and it would take quite some time to adopt it for this repo, when he already has the work finished and is able to provide a completed container. Please use dusty-nv's jetson-container repository here |
I let ollama run on my jetson machine so that my machine can use ollama to play. Originally, ollama runs on my laptop, but it is not convenient for my laptop to be turned on 24 hours a day, so I plan to use the jetson machine to run it 24 hours a day. |
I don't quite understand the meaning, does it mean if I use this, I can run ollama normally on the jetson machine? |
Ollama is designed to be both the backend API and a front-end client. For normal operation, someone would run Ollama as a server on the system with the GPU, then you can use Ollama on another terminal to connect to the first process. If you use Jetson-containers command it will run it in a container as a background process and return control to a bash shell running in the container. |
In other words, you run Ollama twice. |
Yes, you can run Ollama on the Jetson with it, it will run on a docker container and will open a listening port that you can connect to with another Ollama instance as a client or with a web client like the Openllm Webui. |
There are 4 cores that are not running |
I use ollama to let One API call it, so that I can use ollama's interface to return information on another machine |
That is normal, it is a mechanism put in place for x64 CPUs since they tend to use SMT/Hyper-Threading. You can technically change the inner workings to adjust that, but it will not be worth the time because most of your processing on the Orin is from the GPU so your performance gain will be minimal. |
It took me a long time to finally get my jetson atx orin 64G to run ollama successfully! Thank you for your patience and help. I will continue to put other dockers on the jetson atx orin 64G device. I plan to put all the original raspberry pi docker images on the jetson atx orin device. |
Stable-Diffusion is really interesting too! |
I can put some Docker images running normally on the server Linux on the Jetson device, right? Because I see that the memory is relatively large, I plan to use Jetson as the main device, put all the Docker images I need on it, and let Jetson proxy the Raspberry Pi 5 to continue running all my services. Is that okay? Running ollama and my other docker images at the same time |
Yes but the problem is AI programs take a lot of memory. Some AI models are huge and require big chunks of memory and you should only run one model at a time. I know you have 64GB but llama 40b will use almost all of that. You can try it and it might work. If you’re trying to use multiple machines like that it can be really complex. It seems like you need clustering with Docker Swarm or Kubernetes, but that’s beyond what I can advise. |
Controlling the video memory to around 32G should not have much impact on the machine, right? |
If it seems okay to you, it should be okay. Have fun! |
grateful
|
@cplasfwst Hello, I am not very familiar with the power profile of the AGX Orin. However, based on nvidia specifications and your screenshot it would appear your current power mode is Please examine this guide and change the power mode to Please also consider that this will increase the heat generated by the device. You might want to check the fan mode and ensure it is set to the highest mode, I think it is |
Thank you very much, I tried turning on the max mode, it really fully utilized the performance of my device! Very good, you solved so many problems for me, thank you very much! You are a very good person。 |
You’re welcome, I hope you have fun with developing on your Jetson! |
Did you change the fan profile? I suggested it to you previously but in your screenshot it doesn’t show your fans at full power. No, 70C does not typically damage processors. |
Hello author, I am glad that you have made contributions to Jetson users. I have a question I would like to ask
Some errors occurred when I used the command to build the docker image. Why?
The text was updated successfully, but these errors were encountered: