Comments (3)
-
Because some data are more important than others. You may have heard about publications talking about how to maintain a consistent personality of a chatbot, this is a practical way to do it. You manual create a small data set, in which, you answer questions about the name, age, birthday, gender, place of living, basic personal interest of the chatbot. You can also try to get rid of the possible conflicting pairs in the base dataset, and repeating your important data, therefore, your model can remember what you want it do remember. You can think that, the base dataset is to train the model to learn grammar, but the dataset you are repeating is to train the model for knowledge. That's same way as a human to learn: you spend more time on more important/useful knowledge, while you also need to read a lot of books, such as novels.
-
It depends on your model and your data. Normally, I suggest the epoch number be in 50 to 100. Generally, a small model with more epochs may get a slightly better result, but may take you much longer time to train.
Hope these make sense to you.
from chatlearner.
@bshao001 thank you for your response. I have doubts about the small model. Right now I'm using embedding_size = 1024 and num_units = 1024, as used in your code. I don't have a sense of a smaller model. So, do these values come in a range of small model values, large model values or moderate models?
from chatlearner.
There is no absolute large model or small model. It is described based on your data set. You have to experience quite several rounds of training and testing (involve some manual review and check) in order to get a feeling what is a good size of the model for your data set. When you reach that step, you may fine tune your model, in which case, you will understand what I have originally said above.
from chatlearner.
Related Issues (20)
- botui not picking trained files HOT 2
- SessionID max length (int) HOT 4
- session Id HOT 1
- How do we continue training from checkpoint? HOT 1
- Allow Data to be gzipped and predictor to read gzipped data HOT 2
- settings.py
- chat service not available at this moment. Please try again later HOT 8
- get wrong result when use trained model to infer HOT 4
- when copy the project to a new folder and try training, change do not work HOT 1
- Training killed HOT 1
- how do we utilize multiple GPUs? HOT 2
- cannot train when new vocab file is generated . HOT 1
- Reddit data cleaner
- stacked lstm HOT 1
- Does the project use HRED for Multi-round dialogue
- Tensorflow 2.0 HOT 2
- Does this NMT model have attention mech in there? HOT 1
- How to Adapt the bot for commercial use?
- Can someone please send me the already train file
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from chatlearner.