-
Updated
Dec 9, 2021 - Makefile
CUDA
CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
Here are 3,263 public repositories matching this topic...
-
Updated
Dec 23, 2021 - Shell
I am working on creating a WandbCallback for Weights and Biases. I am glad that CatBoost has a callback system in place but it would be great if we can extend the interface.
The current callback only supports after_iteration that takes info. Taking inspiration from XGBoost callback system it would be great if we can have before iteration that takes info, before_training, and `after
-
Updated
Dec 29, 2021 - C++
-
Updated
Dec 29, 2021 - Python
Description
Change the signature of cupy.{percentile,quantile} to provide exactly the same API as NumPy.
I think it's ok to implement overwrite_input as nop (just ignore the option).
Additional Information
No response
-
Updated
Dec 22, 2021 - Go
Based on @karthikeyann's work on this PR rapidsai/cudf#9767 I'm wondering if it makes sense to consider removing the defaults for the stream parameters in various detail functions. It is pretty surprising how often these are getting missed.
The most common case seems to be in factory functions and various ::create functions. Maybe just do it for those?
请问可以直接training tmfile出来吗? 因为tengine-convert-tool covert 会有error
tengine-lite library version: 1.4-dev
Get input tensor failed

或是有例子能training出下面tmfile 呢?


I see comments suggesting adding this to understand how loops are being handled by numba, and in the their own FAQ (https://numba.pydata.org/numba-doc/latest/user/faq.html)
You would then create your njit function and run it, and I believe the idea is that it prints debug information about whether