Qortal Project

The future of blockchain platforms

User Tools

Site Tools


qort_data_hosting_updates

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
qort_data_hosting_updates [11/17/2021 17:29] gfactorqort_data_hosting_updates [07/21/2022 05:29] (current) gfactor
Line 1: Line 1:
-====== QORT Data Hosting Updates ====== +This page has moved[[QDN Updates]]
-{{:qortal_official_logo_transparent_.png?400|}} +
- +
-**From CalDescent 8/18/22:** +
- +
-For anyone interested in checking out the progress, you can do so here: https://github.com/CalDescent14/qortal-data/commits/master +
- +
-You will see that we already have a fully operational decentralized storage system. +
- +
-The most recent feature introduces layering - similar to Docker - so that on-chain updates can be published to existing sites/resources without having to upload the entire file structure again. We can also now access sites via registered names (will demo this shortly). +
- +
-**From CalDescent 10/26/21:** +
- +
-The files are "grouped" by the transaction that they are contained  in. So in the case of a website, you would have to download all files relating to a transaction (and potentially past transactions) in order to access any file contained within it. For example: +
- +
-A user wants to view an image at /site/QortalDemo/wp-content/uploads/2021/02/qortal-logo.png (real image is here: http://node1.qortal.uk:12393/site/QortalDemo/wp-content/uploads/2021/02/qortal-logo.png) +
- +
-In order to do this, they need to build the QortalDemo website resource. +
- +
-The build system in the node gets notified that they have requested the file, and it goes and looks up arbitrary transactions for name QortalDemo with the service WEBSITE. +
- +
-It may find just a recent PUT transaction, which contains the entire site, or it may find a PUT followed by multiple PATCH transactions. +
- +
-Once it knows what layers are needed, it puts out requests to the network to retrieve the encrypted data files associated with these layers. For each file it checks against the hashes on chain to ensure they are correct. +
- +
-Once it has all the files, it builds the website resource by decrypting+extracting each layer, applying the patches, and validating the hashes of the end result to ensure they match what was uploaded. +
- +
-So the first time a website is accessed, it can take some time to build it depending on the complexity of the site. During this time a loading screen is shown which we can brand with the Qortal logo. Once built, the site is cached on the node so that subsequent accesses are instant (as seen in the QortalDemo and QortalWikiDemo). +
- +
-From that point onwards, the node will be hosting a copy of all the files so that other nodes can request them and repeat the above process. They could optionally delete the files for some of the layers or even just a few chunks for a single layer, and they would still be helping the network by serving fragments of the complete site, in the same way it works with torrents. +
- +
-I’m planning on having a “storage policy” setting with various different possible values such as following only / viewed only / following and viewed / all / none. Then we can default to following and viewed, and people have the option of increasing or reducing the scope of their storage. I’ll be working on this part in the next few weeks. +
- +
-**From CalDescent 11/17/21:**  +
- +
-Each website is broken up into chunks (currently 1MB each but will probably increase to 2MB+ at some point), and these chunks can be distributed in any way we choose. You could have one chunk per node, or all chunks on all nodes, or anything in between. The system pieces them together from all over the world when it needs to build the website back into its original state. The actual logic of the distribution of chunks is yet to be implemented; it currently attempts to store all chunks on each node that follows the name or views the content. But soon this will become more intelligent. I imagine we will have a total storage limit per name that is being followed. Once we reach the limit, we'll start dropping chunks evenly so that no-one is holding the complete set. The node owner will have full oversight. By default, no content will be hosted by each node. They will only store the hashes of the data which are on the chain that everyone has a copy of. Once a node owner either views some content, or follows a content creator, they will then start hosting a copy of the content that is viewed or followed. The idea behind following is that you can support creators on the network that you trust, without having to view every single piece of their content. +
- +
-With concern to transferring illegal content across nodes, this was probably the #1 consideration of mine when I started working on the data hosting. I've given it more attention than any other aspect of the project. As it's very important to me that I don't have all sorts of unknown/unvetted stuff going through my nodes and internet connection. So the whole thing will be opt in. The default settings will be: +
- +
-- You host individual data resources that you view +
- +
-- You also host complete datasets from creators that you follow +
- +
-- Your node won't store or relay data unless it meets the above criteria +
- +
-We may have an opt-in setting to allow you to become a "relay" - this allows data outside of the criteria to be relayed via your node but not stored. That will be disabled by default so it is entirely an opt-in feature. Relay nodes will help the network function more smoothly, so anyone that does it will be helping content to flow more easily. The whole thing is going to be a bit of an experiment. But I think we've covered enough of the basics for it to succeed, after a few rounds of fixes of course.+
qort_data_hosting_updates.1637188162.txt.gz · Last modified: 11/17/2021 17:29 by gfactor