Comments (4)
Hi @rose-jinyang, it seems that the issue is related to this question on stackoverflow, which may be caused by corrupted faiss_k_160.npz
. Could you please check the file to see whether the size is normal or whether it can be read by np.load
?
For the kNN setting, we found two empirical rules in our experiments:
- The k for GCN-V and the k for GCN-E are not necessarily to be the same. As GCN-E is capable of predicting robust linkage, increasing the k for GCN-E usually brings performance gain, especially when the neighborhood structure is complex. But as you said, they can share the same k value. In the first version of our arxiv paper, we adopt k=80 for both GCN-V and GCN-E. In current version, we adopt k=80 for GCN-V and k=160 for GCN-E.
- The k for GCN-E training and the k for GCN-E testing are not necessarily to be the same. We can choose to train GCN-E with complex graphs with large k value, and test it with easy local graph with small k value. However, there is a trade-off that if we use a small k value, the candidate set will be small and may not contain the correct one for connection. For simplicity, we usually keep them the same.
from learn-to-cluster.
Thanks for your reply.
I trained with k=80 for GCN-V and k=160 for GCN-E.
I tried with k=80 for GCN-V and k=160 for GCN-E even when testing.
GCN-V test was successful but GCN-E test was not.
So I changed k for GCN-V test to 160 and the GCN-E test was successful.
But I am strange that the class_num after GCN-E test is greater that one after GCN-V test.
I tested with 8000 classes.
The number of the predicted classes after GCN-V test was 9383.
But the number of the predicted classes after GCN-E test was 9416.
How can I understand this?
and I am going to use k=160 for testing both GCN-V and GCN-E.
Please let me know if it is okay.
Thanks.
from learn-to-cluster.
(1) As the error is raised when loading knn, I think it may not be caused by the inconsistent k value between GCN-V and GCN-E. But I am not very sure why this happened according to current message. If you can provide a minimum example to reproduce the error, I am happy to help diagnose the problem.
(2) There is no guarantee between the number of predicted classes of GCN-V and GCN-E. If the number of predicted classes of GCN-E increases, it means GCN-E cuts down some linkages according to predictions, partitioning a class into several sub-classes, and verse vice.
(3) Sure. It is okay to use k=160 for both GCN-V and GCN-E, but it will increase the testing time as the built affinity graph is more dense.
from learn-to-cluster.
Thanks for your reply.
I trained with k=80 for GCN-V and k=160 for GCN-E.
I tried with k=80 for GCN-V and k=160 for GCN-E even when testing.
GCN-V test was successful but GCN-E test was not.
So I changed k for GCN-V test to 160 and the GCN-E test was successful.
But I am strange that the class_num after GCN-E test is greater that one after GCN-V test.
I tested with 8000 classes.
The number of the predicted classes after GCN-V test was 9383.
But the number of the predicted classes after GCN-E test was 9416.
How can I understand this?
and I am going to use k=160 for testing both GCN-V and GCN-E.
Please let me know if it is okay.
Thanks.
您好,请问你聚类的结果如何呢,我使用自己的数据进行在vgcn上训练,效果不是很好,可以和您交流一下吗
from learn-to-cluster.
Related Issues (20)
- output error HOT 1
- Dataset issue HOT 2
- Issue while training LGCN HOT 1
- 您好,想请教一下训练GCN-D时报错run out of input HOT 1
- 没有修改任何参数跑作者提供的pretrained model for ms1m Fscore 只有0.32
- Clustering 512D features using pretrained model HOT 3
- building symmetric adjacency matrix in utils HOT 1
- why using output of relu for rebuilding knn graph HOT 1
- is not a checkpoint filemodels/pretrained_gcn_e_ms1m.pth HOT 3
- about extracting features HOT 3
- About how to make a .meta file HOT 1
- Can we manually set the number of clusters during the test HOT 1
- 大佬,请教下train_cluster_seg_iop_vox.sh 的test阶段报错的问题 HOT 2
- 两个face recognition模型,推荐使用哪个,hfsoftmax还要用到python 2 ? HOT 1
- About pretrained_model to get the the extracted feature HOT 2
- GCN-D MMdistributedDataparallel train HOT 2
- 关于性能 HOT 1
- How to apply lgcn for clustering on custom embeddings of collected face images? HOT 1
- On custom dataset preparation
- On the numpy version which may cause exception HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from learn-to-cluster.