Script 'mail_helper' called by obssrc
Hello community,

here is the log from the commit of package ollama for openSUSE:Factory checked 
in at 2024-11-24 11:04:51
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Comparing /work/SRC/openSUSE:Factory/ollama (Old)
 and      /work/SRC/openSUSE:Factory/.ollama.new.28523 (New)
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Package is "ollama"

Sun Nov 24 11:04:51 2024 rev:25 rq:1225993 version:0.4.2

Changes:
--------
--- /work/SRC/openSUSE:Factory/ollama/ollama.changes    2024-11-07 
18:17:08.252934014 +0100
+++ /work/SRC/openSUSE:Factory/.ollama.new.28523/ollama.changes 2024-11-24 
11:05:07.477151303 +0100
@@ -1,0 +2,46 @@
+Sat Nov 16 16:07:38 UTC 2024 - Eyad Issa <eyadlore...@gmail.com>
+
+- Update to version 0.4.2:
+  * runner.go: Propagate panics back to the user.
+  * runner.go: Increase survivability of main processing loop
+  * build: fix arm container image (#7674)
+  * add line numbers for parser errors (#7326)
+  * chore(deps): bump golang.org/x dependencies (#7655)
+  * runner.go: Don't trim whitespace from inputs
+  * runner.go: Enforce NUM_PARALLEL directly in the runner
+  * cmd: preserve exact bytes when displaying template/system layers (#7586)
+  * fix(mllama): sync backend between batches
+  * runner.go: Fix off-by-one for num predicted
+  * CI: give windows lint more time (#7635)
+  * Jetpack support for Go server (#7217)
+  * doc: capture numeric group requirement (#6941)
+  * docs: Capture docker cgroup workaround (#7519)
+  * runner.go: Make KV entry accounting more robust
+  * readme: add aichat terminal app to community integrations (#7418)
+  * api: fix typos in Go Doc comments (#7620)
+  * readme: add GoLamify to community integrations (#7521)
+  * readme: add browser extension that enables using Ollama for interacting 
with web pages (#5827)
+  * docs: add mentions of Llama 3.2 (#7517)
+  * api: fix typo in python ClientFromEnvironment docs (#7604)
+  * readme: add llama3.2-vision to model list (#7580)
+
+-------------------------------------------------------------------
+Mon Nov 11 13:57:46 UTC 2024 - Eyad Issa <eyadlore...@gmail.com>
+
+- Add patch 01-build-verbose.patch to add the -v option
+  to go build
+
+- Update to version 0.4.1:
+  * runner.go: Check for zero length images
+  * docs: update langchainpy.md with proper model name (#7527)
+  * Set macos min version for all architectures (#7579)
+  * win: remove preview title from installer (#7529)
+  * Workaround buggy P2P ROCm copy on windows (#7466)
+  * Debug logging for nvcuda init (#7532)
+  * Align rocm compiler flags (#7467)
+  * Be explicit for gpu library link dir (#7560)
+  * docs: OLLAMA_NEW_RUNNERS no longer exists
+  * runner.go: Remove unused arguments
+  * sched: Lift parallel restriction for multimodal models except mllama
+
+-------------------------------------------------------------------

Old:
----
  ollama-0.4.0.obscpio

New:
----
  01-build-verbose.patch
  ollama-0.4.2.obscpio

BETA DEBUG BEGIN:
  New:
- Add patch 01-build-verbose.patch to add the -v option
  to go build
BETA DEBUG END:

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Other differences:
------------------
++++++ ollama.spec ++++++
--- /var/tmp/diff_new_pack.KLcDNE/_old  2024-11-24 11:05:08.477192878 +0100
+++ /var/tmp/diff_new_pack.KLcDNE/_new  2024-11-24 11:05:08.481193045 +0100
@@ -17,7 +17,7 @@
 
 
 Name:           ollama
-Version:        0.4.0
+Version:        0.4.2
 Release:        0
 Summary:        Tool for running AI models on-premise
 License:        MIT
@@ -26,6 +26,7 @@
 Source1:        vendor.tar.zstd
 Source2:        ollama.service
 Source3:        %{name}-user.conf
+Patch1:         01-build-verbose.patch
 BuildRequires:  cmake >= 3.24
 BuildRequires:  git
 BuildRequires:  sysuser-tools
@@ -86,7 +87,7 @@
 # pie doesn't work with gcc12 on leap
 export GOFLAGS="-mod=vendor"
 %endif
-go test ./...
+go test -v ./...
 
 %pre -f %{name}.pre
 %service_add_pre %{name}.service

++++++ 01-build-verbose.patch ++++++
diff --git a/llama/make/Makefile.default b/llama/make/Makefile.default
index 95b13a7..6a9b9ef 100644
--- a/llama/make/Makefile.default
+++ b/llama/make/Makefile.default
@@ -24,17 +24,17 @@ all: $(BUILD_RUNNERS) $(DIST_RUNNERS) $(PAYLOAD_RUNNERS)
 $(RUNNERS_BUILD_DIR)/$(DEFAULT_RUNNER)/ollama_llama_server$(EXE_EXT): 
TARGET_CPU_FLAGS=$(CUSTOM_CPU_FLAGS)
 $(RUNNERS_BUILD_DIR)/$(DEFAULT_RUNNER)/ollama_llama_server$(EXE_EXT): *.go 
./runner/*.go $(COMMON_SRCS) $(COMMON_HDRS)
        @-mkdir -p $(dir $@)
-       GOARCH=$(ARCH) go build -buildmode=pie $(CPU_GOFLAGS) -trimpath $(if 
$(CUSTOM_CPU_FLAGS),-tags $(subst $(space),$(comma),$(CUSTOM_CPU_FLAGS)))  -o 
$@ ./runner
+       GOARCH=$(ARCH) go build -buildmode=pie -v $(CPU_GOFLAGS) -trimpath $(if 
$(CUSTOM_CPU_FLAGS),-tags $(subst $(space),$(comma),$(CUSTOM_CPU_FLAGS)))  -o 
$@ ./runner
 
 $(RUNNERS_BUILD_DIR)/cpu_avx/ollama_llama_server$(EXE_EXT): 
TARGET_CPU_FLAGS="avx"
 $(RUNNERS_BUILD_DIR)/cpu_avx/ollama_llama_server$(EXE_EXT): *.go ./runner/*.go 
$(COMMON_SRCS) $(COMMON_HDRS)
        @-mkdir -p $(dir $@)
-       GOARCH=$(ARCH) go build -buildmode=pie $(CPU_GOFLAGS) -trimpath -tags 
$(subst $(space),$(comma),$(TARGET_CPU_FLAGS)) -o $@ ./runner
+       GOARCH=$(ARCH) go build -buildmode=pie -v $(CPU_GOFLAGS) -trimpath 
-tags $(subst $(space),$(comma),$(TARGET_CPU_FLAGS)) -o $@ ./runner
 
 $(RUNNERS_BUILD_DIR)/cpu_avx2/ollama_llama_server$(EXE_EXT): 
TARGET_CPU_FLAGS="avx avx2"
 $(RUNNERS_BUILD_DIR)/cpu_avx2/ollama_llama_server$(EXE_EXT): *.go 
./runner/*.go $(COMMON_SRCS) $(COMMON_HDRS)
        @-mkdir -p $(dir $@)
-       GOARCH=$(ARCH) go build -buildmode=pie $(CPU_GOFLAGS) -trimpath -tags 
$(subst $(space),$(comma),$(TARGET_CPU_FLAGS)) -o $@ ./runner
+       GOARCH=$(ARCH) go build -buildmode=pie -v $(CPU_GOFLAGS) -trimpath 
-tags $(subst $(space),$(comma),$(TARGET_CPU_FLAGS)) -o $@ ./runner
 
 $(RUNNERS_DIST_DIR)/%: $(RUNNERS_BUILD_DIR)/%
        @-mkdir -p $(dir $@)
@@ -44,7 +44,7 @@ $(RUNNERS_PAYLOAD_DIR)/%/ollama_llama_server$(EXE_EXT).gz: 
$(RUNNERS_BUILD_DIR)/
        @-mkdir -p $(dir $@)
        ${GZIP} --best -c $< > $@
 
-clean: 
+clean:
        rm -f $(BUILD_RUNNERS) $(DIST_RUNNERS) $(PAYLOAD_RUNNERS)
 
 .PHONY: clean all

++++++ _service ++++++
--- /var/tmp/diff_new_pack.KLcDNE/_old  2024-11-24 11:05:08.525194874 +0100
+++ /var/tmp/diff_new_pack.KLcDNE/_new  2024-11-24 11:05:08.533195207 +0100
@@ -3,7 +3,7 @@
   <service name="obs_scm" mode="manual">
     <param name="url">https://github.com/ollama/ollama.git</param>
     <param name="scm">git</param>
-    <param name="revision">v0.4.0</param>
+    <param name="revision">v0.4.2</param>
     <param name="versionformat">@PARENT_TAG@</param>
     <param name="versionrewrite-pattern">v(.*)</param>
     <param name="changesgenerate">enable</param>

++++++ _servicedata ++++++
--- /var/tmp/diff_new_pack.KLcDNE/_old  2024-11-24 11:05:08.557196204 +0100
+++ /var/tmp/diff_new_pack.KLcDNE/_new  2024-11-24 11:05:08.561196371 +0100
@@ -1,6 +1,6 @@
 <servicedata>
 <service name="tar_scm">
                 <param name="url">https://github.com/ollama/ollama.git</param>
-              <param 
name="changesrevision">9d71bcc3e2a97c8e62d758450f43aa212346410e</param></service></servicedata>
+              <param 
name="changesrevision">d875e99e4639dc07af90b2e3ea0d175e2e692efb</param></service></servicedata>
 (No newline at EOF)
 

++++++ ollama-0.4.0.obscpio -> ollama-0.4.2.obscpio ++++++
/work/SRC/openSUSE:Factory/ollama/ollama-0.4.0.obscpio 
/work/SRC/openSUSE:Factory/.ollama.new.28523/ollama-0.4.2.obscpio differ: char 
49, line 1

++++++ ollama.obsinfo ++++++
--- /var/tmp/diff_new_pack.KLcDNE/_old  2024-11-24 11:05:08.617198699 +0100
+++ /var/tmp/diff_new_pack.KLcDNE/_new  2024-11-24 11:05:08.617198699 +0100
@@ -1,5 +1,5 @@
 name: ollama
-version: 0.4.0
-mtime: 1730848045
-commit: 9d71bcc3e2a97c8e62d758450f43aa212346410e
+version: 0.4.2
+mtime: 1731700345
+commit: d875e99e4639dc07af90b2e3ea0d175e2e692efb
 

++++++ vendor.tar.zstd ++++++
Binary files /var/tmp/diff_new_pack.KLcDNE/_old and 
/var/tmp/diff_new_pack.KLcDNE/_new differ

Reply via email to