Comments (6)
For the question about the number provided to TchDevice::Cuda, I was curious as well and looked through the documentation. According to the backend, it is the device index.
from burn.
I think I saw that same source code which prompted me to try 0 and 1. It's still a bit unclear from the libraries what the value should be, is there a way to list valid device indices?
from burn.
I made a pull request that should fix the problem. The reason I did not see this before is that I always test using the --release
flag, since it speeds up quite a bit the training without noticeable compilation time issue. I discovered that there was an underflow with the file checkpointer on debug builds.
I also added documentation on the TchDevice struct, the usize
parameter is indeed the device index, which starts at zero. If you only have 1 GPU, it should just be zero.
I will work soon on error handling and proper logging to help understand those issues more clearly.
from burn.
Have you tried this trick in your toml to get the speed optimizations without needing --release
and losing debug info:
[profile.dev]
opt-level = 0
[profile.dev.package."*"]
opt-level = 3
from burn.
I didn't think about this but yeah this is probably a good default and something to include in the example!
from burn.
In turns out that you can't set the optimization profile in packages under a workspace, I'm not sure how to make debug builds use a different optimization level (only for examples).
from burn.
Related Issues (20)
- error when import onnx of yolo8 HOT 3
- Panic w/ backwards pass when combining gather and max_dim HOT 2
- Convert ONNX graph inputs of 0-dim tensors to scalars HOT 4
- Support for 0-Dimensional Tensors in Burn HOT 5
- No adapter found for graphics API AutoGraphicsApi HOT 1
- [Fusion] Support multi-precision fusion
- burn and drug discovery
- loss.backward() hangs after burn update 0.12 -> 0.13 HOT 3
- Help Wanted: Implementing ONNX Ops
- Implement multi-dimensional repeat operation and rename existing repeat method HOT 2
- [Tensor] Add `cumsum` operation HOT 1
- .select_assign does not work with Autodiff<NdArray> backend
- Add indentation in contributing book
- Text classification example gives "Shader validation error" when run on multiple GPUs HOT 5
- Upgrade all dependencies
- Better memory management in Burn Compute
- Config Derive: Generic Types? HOT 1
- Optimizer / Visitor / Mapper confusion, no documentation HOT 4
- clamp_min does not handle -inf correctly on Autodiff<NdArray> backend
- Update tch to 0.16+
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from burn.