Jared Van Bortel 
							
						 
					 
					
						
						
							
						
						061d1969f8 
					 
					
						
						
							
							expose n_gpu_layers parameter of llama.cpp ( #1890 )  
						
						... 
						
						
						
						Also dynamically limit the GPU layers and context length fields to the maximum supported by the model.
Signed-off-by: Jared Van Bortel <jared@nomic.ai > 
						
						
					 
					
						2024-01-31 14:17:44 -05:00 
						 
				 
			
				
					
						
							
							
								Jared Van Bortel 
							
						 
					 
					
						
						
							
						
						7e9786fccf 
					 
					
						
						
							
							chat: set search path early  
						
						... 
						
						
						
						This fixes the issues with installed versions of v2.6.0. 
						
						
					 
					
						2024-01-11 12:04:18 -05:00 
						 
				 
			
				
					
						
							
							
								ThiloteE 
							
						 
					 
					
						
						
							
						
						2d566710e5 
					 
					
						
						
							
							Address review  
						
						
						
						
					 
					
						2024-01-03 11:13:07 -06:00 
						 
				 
			
				
					
						
							
							
								ThiloteE 
							
						 
					 
					
						
						
							
						
						38d81c14d0 
					 
					
						
						
							
							Fixes  https://github.com/nomic-ai/gpt4all/issues/1760  LLModel ERROR: Could not find CPU LLaMA implementation.  
						
						... 
						
						
						
						Inspired by Microsoft docs for LoadLibraryExA (https://learn.microsoft.com/en-us/windows/win32/api/libloaderapi/nf-libloaderapi-loadlibraryexa ).
When using LOAD_LIBRARY_SEARCH_DLL_LOAD_DIR, the lpFileName parameter must specify a fully qualified path, also it needs to be backslashes (\), not forward slashes (/). 
						
						
					 
					
						2024-01-03 11:13:07 -06:00 
						 
				 
			
				
					
						
							
							
								Jared Van Bortel 
							
						 
					 
					
						
						
							
						
						d1c56b8b28 
					 
					
						
						
							
							Implement configurable context length ( #1749 )  
						
						
						
						
					 
					
						2023-12-16 17:58:15 -05:00 
						 
				 
			
				
					
						
							
							
								Jared Van Bortel 
							
						 
					 
					
						
						
							
						
						3acbef14b7 
					 
					
						
						
							
							fix AVX support by removing direct linking to AVX2 libs ( #1750 )  
						
						
						
						
					 
					
						2023-12-13 12:11:09 -05:00 
						 
				 
			
				
					
						
							
							
								Jared Van Bortel 
							
						 
					 
					
						
						
							
						
						d4ce9f4a7c 
					 
					
						
						
							
							llmodel_c: improve quality of error messages ( #1625 )  
						
						
						
						
					 
					
						2023-11-07 11:20:14 -05:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						f414c28589 
					 
					
						
						
							
							llmodel: whitelist library name patterns  
						
						... 
						
						
						
						this fixes some issues that were being seen on installed windows builds of 2.5.0
only load dlls that actually might be model impl dlls, otherwise we pull all sorts of random junk into the process before it might expect to be
Signed-off-by: Aaron Miller <apage43@ninjawhale.com > 
						
						
					 
					
						2023-10-23 21:40:14 -07:00 
						 
				 
			
				
					
						
							
							
								cebtenzzre 
							
						 
					 
					
						
						
							
						
						7b611b49f2 
					 
					
						
						
							
							llmodel: print an error if the CPU does not support AVX ( #1499 )  
						
						
						
						
					 
					
						2023-10-11 15:09:40 -04:00 
						 
				 
			
				
					
						
							
							
								cebtenzzre 
							
						 
					 
					
						
						
							
						
						7a19047329 
					 
					
						
						
							
							llmodel: do not call magic_match unless build variant is correct ( #1488 )  
						
						
						
						
					 
					
						2023-10-11 11:30:48 -04:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						507753a37c 
					 
					
						
						
							
							macos build fixes  
						
						
						
						
					 
					
						2023-10-05 18:16:19 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						d90d003a1d 
					 
					
						
						
							
							Latest rebase on llama.cpp with gguf support.  
						
						
						
						
					 
					
						2023-10-05 18:16:19 -04:00 
						 
				 
			
				
					
						
							
							
								Cosmic Snow 
							
						 
					 
					
						
						
							
						
						108d950874 
					 
					
						
						
							
							Fix Windows unable to load models on older Windows builds  
						
						... 
						
						
						
						- Replace high-level IsProcessorFeaturePresent
- Reintroduce low-level compiler intrinsics implementation 
						
						
					 
					
						2023-08-09 09:27:43 +02:00 
						 
				 
			
				
					
						
							
							
								cosmic-snow 
							
						 
					 
					
						
						
							
						
						6200900677 
					 
					
						
						
							
							Fix Windows MSVC arch detection ( #1194 )  
						
						... 
						
						
						
						- in llmodel.cpp to fix AVX-only handling
Signed-off-by: cosmic-snow <134004613+cosmic-snow@users.noreply.github.com > 
						
						
					 
					
						2023-07-13 14:44:17 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						315a1f2aa2 
					 
					
						
						
							
							Move it back as internal class.  
						
						
						
						
					 
					
						2023-07-13 14:21:46 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						1f749d7633 
					 
					
						
						
							
							Clean up backend code a bit and hide impl. details.  
						
						
						
						
					 
					
						2023-07-13 14:21:46 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						33557b1f39 
					 
					
						
						
							
							Move the implementation out of llmodel class.  
						
						
						
						
					 
					
						2023-07-13 14:21:46 -04:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						432b7ebbd7 
					 
					
						
						
							
							include windows.h just to be safe  
						
						
						
						
					 
					
						2023-07-12 12:46:46 -04:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						95b8fb312e 
					 
					
						
						
							
							windows/msvc: use high level processor feature detection API  
						
						... 
						
						
						
						see https://learn.microsoft.com/en-us/windows/win32/api/processthreadsapi/nf-processthreadsapi-isprocessorfeaturepresent  
						
						
					 
					
						2023-07-12 12:46:46 -04:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						db34a2f670 
					 
					
						
						
							
							llmodel: skip attempting Metal if model+kvcache > 53% of system ram  
						
						
						
						
					 
					
						2023-06-26 19:46:49 -03:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						d3ba1295a7 
					 
					
						
						
							
							Metal+LLama take two ( #929 )  
						
						... 
						
						
						
						Support latest llama with Metal
---------
Co-authored-by: Adam Treat <adam@nomic.ai >
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de > 
						
						
					 
					
						2023-06-09 16:48:46 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						b162b5c64e 
					 
					
						
						
							
							Revert "llama on Metal ( #885 )"  
						
						... 
						
						
						
						This reverts commit c55f81b860 
						
						
					 
					
						2023-06-09 15:08:46 -04:00 
						 
				 
			
				
					
						
							
							
								Aaron Miller 
							
						 
					 
					
						
						
							
						
						c55f81b860 
					 
					
						
						
							
							llama on Metal ( #885 )  
						
						... 
						
						
						
						Support latest llama with Metal
---------
Co-authored-by: Adam Treat <adam@nomic.ai >
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de > 
						
						
					 
					
						2023-06-09 14:58:12 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						8a9ad258f4 
					 
					
						
						
							
							Fix symbol resolution on windows.  
						
						
						
						
					 
					
						2023-06-05 11:19:02 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						812b2f4b29 
					 
					
						
						
							
							Make installers work with mac/windows for big backend change.  
						
						
						
						
					 
					
						2023-06-05 09:23:17 -04:00 
						 
				 
			
				
					
						
							
							
								AT 
							
						 
					 
					
						
						
							
						
						5f95aa9fc6 
					 
					
						
						
							
							We no longer have an avx_only repository and better error handling for minimum hardware requirements. ( #833 )  
						
						
						
						
					 
					
						2023-06-04 15:28:58 -04:00 
						 
				 
			
				
					
						
							
							
								Richard Guo 
							
						 
					 
					
						
						
							
						
						98420ea6d5 
					 
					
						
						
							
							cleanup  
						
						
						
						
					 
					
						2023-06-02 12:32:26 -04:00 
						 
				 
			
				
					
						
							
							
								Richard Guo 
							
						 
					 
					
						
						
							
						
						c54c42e3fb 
					 
					
						
						
							
							fixed finding model libs  
						
						
						
						
					 
					
						2023-06-02 12:32:26 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						70e3b7e907 
					 
					
						
						
							
							Try and fix build on mac.  
						
						
						
						
					 
					
						2023-06-02 10:47:12 -04:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						a41bd6ac0a 
					 
					
						
						
							
							Trying to shrink the copy+paste code and do more code sharing between backend model impl.  
						
						
						
						
					 
					
						2023-06-02 07:20:59 -04:00 
						 
				 
			
				
					
						
							
							
								niansa/tuxifan 
							
						 
					 
					
						
						
							
						
						27e80e1d10 
					 
					
						
						
							
							Allow user to specify custom search path via $GPT4ALL_IMPLEMENTATIONS_PATH ( #789 )  
						
						
						
						
					 
					
						2023-06-01 17:41:04 +02:00 
						 
				 
			
				
					
						
							
							
								niansa 
							
						 
					 
					
						
						
							
						
						5175db2781 
					 
					
						
						
							
							Fixed double-free in LLModel::Implementation destructor  
						
						
						
						
					 
					
						2023-06-01 11:19:08 -04:00 
						 
				 
			
				
					
						
							
							
								niansa/tuxifan 
							
						 
					 
					
						
						
							
						
						fc60f0c09c 
					 
					
						
						
							
							Cleaned up implementation management ( #787 )  
						
						... 
						
						
						
						* Cleaned up implementation management
* Initialize LLModel::m_implementation to nullptr
* llmodel.h: Moved dlhandle fwd declare above LLModel class 
						
						
					 
					
						2023-06-01 16:51:46 +02:00 
						 
				 
			
				
					
						
							
							
								Adam Treat 
							
						 
					 
					
						
						
							
						
						1eca524171 
					 
					
						
						
							
							Add fixme's and clean up a bit.  
						
						
						
						
					 
					
						2023-06-01 07:57:10 -04:00 
						 
				 
			
				
					
						
							
							
								niansa 
							
						 
					 
					
						
						
							
						
						a3d08cdcd5 
					 
					
						
						
							
							Dlopen better implementation management (Version 2)  
						
						
						
						
					 
					
						2023-06-01 07:44:15 -04:00 
						 
				 
			
				
					
						
							
							
								niansa/tuxifan 
							
						 
					 
					
						
						
							
						
						92407438c8 
					 
					
						
						
							
							Advanced avxonly autodetection ( #744 )  
						
						... 
						
						
						
						* Advanced avxonly requirement detection 
						
						
					 
					
						2023-05-31 21:26:18 -04:00 
						 
				 
			
				
					
						
							
							
								AT 
							
						 
					 
					
						
						
							
						
						48275d0dcc 
					 
					
						
						
							
							Dlopen backend 5 ( #779 )  
						
						... 
						
						
						
						Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved. 
						
						
					 
					
						2023-05-31 17:04:01 -04:00