{"id":3965,"date":"2025-12-17T17:59:19","date_gmt":"2025-12-17T17:59:19","guid":{"rendered":"https:\/\/in.nau.edu\/arc\/?page_id=3965"},"modified":"2025-12-17T18:35:01","modified_gmt":"2025-12-17T18:35:01","slug":"using-genai-on-monsoon","status":"publish","type":"page","link":"https:\/\/in.nau.edu\/arc\/using-genai-on-monsoon\/","title":{"rendered":"Using GenAI on Monsoon"},"content":{"rendered":"<p>On Monsoon we currently pre-install two key pieces of software that can help you use open-weight GenAI models: 1) <a href=\"https:\/\/github.com\/ollama\/ollama\">Ollama<\/a> and 2) <a href=\"https:\/\/github.com\/open-webui\/open-webui\">Open WebUI<\/a>.\u00a0<em>And if you have any suggestions, please let us know at <a href=\"mailto:ask-arc@nau.edu\">ask-arc@nau.edu<\/a>.<\/em><\/p>\n<p>There are many fully <a href=\"https:\/\/mhit.ai\/docs\/applications-of-ai\/open-weights\">open-weight<\/a> GenAI models that can be run locally. Though these models lack the performance and agentic features (without additional code) of the most popular consumer chatbots such as ChatGPT and Microsoft Copilot, they can be used in Monsoon workloads without data ingress\/egress or compute costs and without entrusting data to other platforms.<\/p>\n<p><strong>Try out the app now<\/strong> by logging into our OpenOndemand instance (<a href=\"https:\/\/ondemand.hpc.nau.edu\">https:\/\/ondemand.hpc.nau.edu<\/a>), clicking the &#8220;Interactive Apps&#8221; tab, and then selecting &#8220;Open WebUI (GenAI)&#8221; in the drop-down menu. Or directly visit the job submission form page for it <a href=\"https:\/\/ondemand.hpc.nau.edu\/pun\/sys\/dashboard\/batch_connect\/sys\/open-webui\/session_contexts\/new\">here<\/a>.<\/p>\n<p>The interface will look like this:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-landscape-image wp-image-3974\" src=\"https:\/\/in.nau.edu\/wp-content\/uploads\/sites\/208\/open-webui-screenshot-cropped-464x348.png\" alt=\"\" width=\"464\" height=\"348\" srcset=\"https:\/\/in.nau.edu\/wp-content\/uploads\/sites\/208\/open-webui-screenshot-cropped-464x348.png 464w, https:\/\/in.nau.edu\/wp-content\/uploads\/sites\/208\/open-webui-screenshot-cropped-800x600.png 800w, https:\/\/in.nau.edu\/wp-content\/uploads\/sites\/208\/open-webui-screenshot-cropped-232x174.png 232w\" sizes=\"auto, (max-width: 464px) 100vw, 464px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p>The OOD app starts a desktop instance with a browser for you. But it will show you in a terminal a command that you can run to create an SSH tunnel so that you can connect to it directly from your local desktop, which will run more smoothly and work better with media features (e.g. uploading an image, audio, etc.).<\/p>\n<p>If you are familiar with Ollama, which serves as the back-end, you may want to run it directly. You can do so from an SSH session on a login node like this:<\/p>\n<pre>srun -G 1 -C a100 --pty \/bin\/bash\r\nmodule load ollama\r\nollama serve &amp;&gt; \/dev\/null &amp;\r\nollama run gpt-oss\r\n<\/pre>\n<p>Once you execute that third command, you will be able to prompt the LLM with minimal delay as it is loaded into the GPU&#8217;s VRAM.<\/p>\n<p>By default, when you load the &#8220;ollama&#8221; module, you&#8217;ll have several models available, which are the ones that we have pre-downloaded for you. But if there is another model that you need, make the ~\/.ollama\/models directory and then reload the module. Like this:<\/p>\n<pre>mkdir -p ~\/.ollama\/models\r\nmodule purge\r\nmodule load ollama\r\n<\/pre>\n<p>Now, you can explore different models available from <a href=\"https:\/\/ollama.com\/search\">the Ollama website<\/a> and once you have picked a model, run ollama and pull the model:<\/p>\n<pre>ollama serve &amp;&gt; \/dev\/null &amp;\r\nollama pull MODEL_NAME\r\n<\/pre>\n<p>But before proceeding with the above, please run it on a data transfer node for any models with a listed size of 10 GB or more. The pull operation is primarily a network download which can bog down the network, and will run faster on a data transfer node (e.g. dtn1 and dtn2).<\/p>\n<p><em>If you have a specific use case and would like some assistance, consider contacting us at <a href=\"mailto:ask-arc@nau.edu\">ask-arc@nau.edu<\/a>.<\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>On Monsoon we currently pre-install two key pieces of software that can help you use open-weight GenAI models: 1) Ollama and 2) Open WebUI.\u00a0And if you have any suggestions, please let us know at ask-arc@nau.edu. There are many fully open-weight GenAI models that can be run locally. Though these models lack the performance and agentic [&hellip;]<\/p>\n","protected":false},"author":544,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_relevanssi_hide_post":"","_relevanssi_hide_content":"","_relevanssi_pin_for_all":"","_relevanssi_pin_keywords":"","_relevanssi_unpin_keywords":"","_relevanssi_related_keywords":"","_relevanssi_related_include_ids":"","_relevanssi_related_exclude_ids":"","_relevanssi_related_no_append":"","_relevanssi_related_not_related":"","_relevanssi_related_posts":"","_relevanssi_noindex_reason":"","ring_central_script_selection":"","footnotes":""},"class_list":["post-3965","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/pages\/3965","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/users\/544"}],"replies":[{"embeddable":true,"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/comments?post=3965"}],"version-history":[{"count":15,"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/pages\/3965\/revisions"}],"predecessor-version":[{"id":4001,"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/pages\/3965\/revisions\/4001"}],"wp:attachment":[{"href":"https:\/\/in.nau.edu\/arc\/wp-json\/wp\/v2\/media?parent=3965"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}