| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
| |
revert #1198, the block size is limited to 84byte
|
| |
|
| |
|
|
|
|
|
|
|
| |
Co-authored-by: psychocrypt <psychocryptHPC@gmail.com>
Co-authored-by: fireice-uk <fireice-uk@users.noreply.github.com>
Co-authored-by: Lee Clagett <code@leeclagett.com>
Co-authored-by: curie-kief <curie-kief@users.noreply.github.com>
|
|
|
|
|
| |
- reduce startup time for multi gpu systems
- initialize the GPU memory non concurrent
|
|
|
|
|
|
|
| |
- enable L1 cache for Nvidia Volta GPUs and newer
- remove explicit cache controll for Volta GPU and newer
This pull request increases the hash rate for Volta GPUs by ~5%
|
|
|
|
| |
reduce memory usage to 1GiB for NVIDIA devices with <=6 SMX
|
|\
| |
| | |
Updated config.tpl comments to include how to exclude CPU/GPUs
|
| | |
|
|\ \
| | |
| | | |
Rename "MacOS" to "macOS"
|
| |/
| |
| |
| |
| |
| | |
Changed capitalization of "macOS"
Squashed the commit
|
|/ |
|
| |
|
|\
| |
| | |
handle cuda error codes
|
| |
| |
| |
| | |
handle all error codes from the cuda api calls.
|
|/
|
|
|
| |
- fix cuda9.1 compile (remove includ eof device_functions.hpp/ removed with cuda9.1)
- remove NVIDIA Volta gpus for MAC OSX
|
|
|
|
| |
Makes casting more explicit.
|
|
|
|
|
| |
When compiled with VS2017, the negative applied to the uint wait
time is ignored. Fixed by casting first.
|
|
|
|
| |
- fix indention
|
|
|
|
|
|
|
|
| |
Be more conservative with the auto suggestion.
- increase bfactor if `smx <= 6`
- limit memory for pascal < GTX1070 to 2GiB
- limt memory for pascal <= GTX1080 to 4GiB
|
|\
| |
| | |
add message to `CUDA_CHECK...` macros
|
| |
| |
| |
| |
| | |
- add macro `CUDA_CHECK_MSG_KERNEL` and `CUDA_CHECK_MSG`
- add suggestion of typicle errors can be solved
|
|/
|
|
|
| |
- add option `sync_mode`
- update auto suggestion and jconf
|
|\
| |
| | |
fix CUDA launch bounds usage
|
| |
| |
| |
| |
| |
| | |
fix #191
lauch bounds must be placed before the return type but after the template paramater
|
|/
|
|
| |
Increase bfactor for all devices with lesser than 6 multi processors.
|
|\
| |
| | |
fix wrong cuda binary arch detection
|
| |
| |
| |
| | |
fix wrong arch comparsion
|
|/
|
|
|
| |
The lmem is still incalculably and crash the miner very often.
Increase the potential lmem usage to 16kiB to respect lmem alignments, ...
|
| |
|
| |
|
|
|
|
| |
- removed all space indention in `*.hpp` and `*.cpp` files
|
|\
| |
| | |
Network extensions
|
| | |
|
|\ \
| | |
| | | |
check gpu architecture
|
| | |
| | |
| | |
| | |
| | |
| | | |
- check if the gpu architecture is supported by the compiled miner binary
- remove not supported gpus from the auto suggestion
- disallow the selection of a not supported gpu by hand tuning the config
|
|\ \ \
| |/ /
| | | |
fix wrong memory detection
|
| |/
| |
| |
| |
| |
| |
| |
| |
| | |
Free and total memory is only evaluated on the first device.
To detect the gpu memory the gpu must be selected.
- create context on the gpu before the memory is checked
- add smx to the auto detection
- change the result code of `cuda_get_deviceinfo()`
|
|/
|
|
|
|
| |
* CONF_NO_TLS fix
* gpu error fix + always return to sched after pin
|
|
|
|
|
| |
- add type of the backend to each backend-plugin
- add `gteName` to `iBackend` to get the name of the backend
|
|
|
|
|
| |
- avoid creation of a config with zero threads or blocks
- WINDOWS: reduce the used memory for the auto suggestion by the amount of already used memory
|
|
|
|
|
| |
- fix windows linker error during compile
- fix wrong parameter to call aeon (nvidia-backend)
|
|
|
|
|
| |
- increase safty memory from 64 to 128 MiB
- NVIDIA: increase lmem reserve per thread to 1kiB
|
|
|
|
|
|
| |
- rename all `xmr` to `monero`
- be insensitive while check for set currency
- add function to compate two strings insensitive
|
|
|
|
|
| |
- add template parameter to kernel to support aeon and xmr
- update auto suggestion
|
|
|
|
| |
- check if affinity vale before calling `thd_setaffinity`
|
|
|
|
|
|
|
|
| |
This is a follow up of #43 and use the some mechanism to set the thread affinity for
non cpu backends correct.
- use cpu affinity workflow for nvidia and amd
- cpu: move messages of thread spawning before thread creation
|
|\
| |
| | |
Implement pool-controlled nonce allocation
|
| | |
|
| | |
|