Comments (2)
This is kind of a tricky case. It is somehow possible already by first defining the sub-website, crawling it, which will generate a second so-called "parent" webentity for this one which won't be crawled, and afterwards merge the parent into the sub-one making them only one (while only the sub one will have been crawled). Features to redefine and merge webentities aren't completely offered yet on the web interface but these are possible functionnalities.
from hyphe.
I think that there is a non-technical discussion here. I will reopen this issue so that we have this discussion if needed.
The coincidence of a web entity and the limits of a crawl is intentional. We want to crawl the web and we need to define the limits of a crawl. We tried to stick to user's needs and users think in terms of websites (most of the time). In order to fit to that need we implemented web entities. They are what you have crawled. Of course you can edit a web entity and then reach a state where a web entity is only partially crawled. But this is a side effect and we want users to fix that situation so that every web entity is crawled. In other terms, web entities deserve the purpose of helping users to manage their crawl.
Web entities are good because they are a simple way to cope with a difficult problem. This problem is to define the limits of a crawl so that we have meaningful entities even if the web is large, heterogeneous, and full of singularities (such as redirects). Web entities are the incarnation of a design strategy. We aim at presenting features in terms of results for the user. The user comes with a need: "I want to have a website in my data." We would rather say "Let's define this website (we call that a web entity) and then harvest it, knowing that it requires several steps" than "We have a harvesting feature requiring several steps, starting with the definition of what you call a website". The user searches for a way to achieve goals, and features must appear as answers to these goals. Web entities are our concept for leading the user to cope with the issues of crawling. As a design trick I find it quite efficient, since users seem to quickly understand the concept, while we are able to use it as a solution to different hard-to-design features. We just ask the user to keep believing that web entities are the result of the crawl, and then we lead the user to the different methodological questions of the crawl.
How is it that some users like you want to separate the crawl from a web entity? Maybe the concept of web entity is so transparent that people see different things in it. This is somehow a design success, since you accept the concept of web entity while discussing the issues of crawling. But you probably understand now that if we separate crawl settings from web entities, it leads us to a bigger issue about how to explain the issue of crawling to users. We can nevertheless explore this design space if you have ideas. Feel free to detail the system you would like to use!
from hyphe.
Related Issues (20)
- Renew code to load recent user agents HOT 1
- Add a tool to cleanup/merge entities
- Allow to set and filter tags from listwebentities
- Rename a corpus
- Enable/Disable multivalued tags
- [network] allow to view selected tag when hovering in legend
- [crawls] Give exact datetimes on hover on human ones (such as 4 months ago)
- Force specific User Agent per crawl
- "Node has no left sibling" when calling `paginate_webentity_pagelinks` HOT 6
- Use ural's shorteners list
- Various Ideas from RESPADON Sprint
- Actions pending alert remains active after changing page
- add actions pending warner on other pages such as StartCrawls
- Bugs with advanced crawl options
- Disable Suspicious crawl status on webentity crawled at the page level
- Link to web.archive url of a crawled page within a webentity page is not clickable
- Bug with network when tag category named type ?
- detection permalink web.archive not working with http version of web.archive
- Add ability to remove a web entity HOT 2
- [Text indexation] Unicode errors sometimes when indexing into ES
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from hyphe.